Since you stumbled upon this article, you might be wondering what the heck genetic algorithms are. To put it simply: genetic algorithms employ the same tactics used in natural selection to find an optimal solution to an optimization problem. Genetic algorithms are often used in high dimensional problems where the optimal solutions are not apparent. Genetic algorithms are commonly used to tune the hyper-parameters of a program. However, this algorithm can be used in any scenario where you have a function which defines how well a solution is. Many people have used genetic algorithms in video games to auto learn the weaknesses of players.
Shortly after working on my Steam Friends Graph ,I had the idea of extending the project to include the GitHub network. I used BrickHack V as the opportunity to work on this project with my friends. Rather than simply use the code that was used in the Steam friends graph, the architecture was completely revamped to reflect both the differences between the Steam and GitHub networks and my improved web development skills.
We created an interactive website which allows you to make graphs based on the Github network. Currently the site generates three types of graphs-- the most popular and entertaining of which is the friends graph. The friends graph helps you visualize clusters of friends/collaborators on GitHub. Similar to the Steam Friends Project, I hope that this project will make people more interested in learning about big data. The visual aspect of this website makes learning about topics such as clustering and graph databases more intuitive.
Suppose that you have a large quantity of files that you want your program to ingest. Is it faster to sequentially read each file or, read the files in parallel using multiple threads? Browsing this question online will give varied answers from confused people on discussion threads. Reading resent research papers on multi threaded IO did not quite provide a clear answer my question.
A ton of people argue that multiple threads don't increase file throughput because at the end of the day a HHD can only read one file at a time. Adding more CPU cores into the mix would actually slow down the file ingest because the HHD would have to take turns between reading fragments of different files. The seek speed of the HHD would heavily degrade the performance.
I occasionally cooked during High School, however, college was my first real experience cooking on a daily basis. Since I have lived to tell the tail, I want to share some of the recipes that I created/modified which got me through my first two years of college.
Being stuck in a dorm room you are limited to only being able to use the microwave. Desperate times calls for depraved cooking solutions.
The concept of lazy loading is a design pattern where you only download the required scripts/content when the client needs it rather than when the page is loaded. The most iconic example is the "infinite" feed that social media sites like Instagram use. Rather than dump a ton of content to the client at once, Instagram dynamically loads more content once you reach the end of the page. This allows websites to save bandwidth by only sending the client what they need when they need it.
When a youtube video is embedded into a website, it pulls a ton of scripts which are required to play the video. However, when you embed a youtube video on a blog post, most people are not going to watch that video. Even if the visitor wanted to watch that video, there is no reason why all the scripts should be loaded when you initially land on the page. Rather than embed the video normally, you can instead simply have a the thumbnail of the video with a play button on it. When the client clicks the play button, the youtube scripts are then loaded.