How many make threads to use?
When I (re-)build large systems on a desktop/laptop computer, I tell make
to use more than one thread to speed up the compilation speed, like this:
Starting on a large project, should I split it into more concise parts?
Background: Currently an intern at a company and working on repurposing older programs (still in use, but for different departments) for use in testing new hardware. The new hardware is different in terms of register layout and firmware interface from the older iterations of hardware. I have been given free reign on how to setup this project as it is to be used for the new hardware exclusively.
Using replacement to get possible outcomes to then search through HUGE amount of data
I have a database table holding 40 million records (table A).
Amortized Analysis? (Worst-case Performance Guarantees)
What is Amortized Analysis? And how can it help me achieve worst-case performance guarantees in my programs?
Is it ever too early to learn a new language? [closed]
As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for […]
Combining multiple events into one action/ Defer refreshing
So in a GUI program I have multiple events that trigger a update of a workspace. This update call is costly so I would want it to not happen very often.
When should code favour optimization over readability and ease-of-use?
I am in the process of designing a small library, where one of my design goals is that the API should be as close to the domain language as possible. While working on the design, I’ve noticed that there are some cases in the code where a more intuitive, readable attribute/method call requires some functionally unnecessary encapsulation. Since the final product will not necessarily require high performance, I am unconcerned about making the decision to favour ease-of-use in my current project over the most efficient implementation of the code in question.
At what point is asynchronous reading of disk I/O more efficient than synchronous?
Assuming there is some bit of code that reads files for multiple consumers, and the files are of any arbitrary size: At what size does it become more efficient to read the file asynchronously? Or to put it another way, how small must a file be for it to be faster just to read it synchronously?
What is an efficient algorithm for randomly assigning a pool of objects to a parent using specific rules
I need some expert answers to help me determine the most efficient algorithm in this scenario.
Effective handling of variables in non-object oriented programming
What is the best method to use and share variables between functions in non object-oriented program languages?