Suppose a transistor on an integrated circuit chip were 2 microns in size. According to Moore’s Law, how large would that transistor be in 2 years? How is Moore’s law relevant to programmers?
In 1965, Intel founder Gordon Moore stated that “The density of transistors in an integrated circuit will double every year.”
The current version of this prediction is usually conveyed as “the density of silicon chips doubles every 18 months” means Moore’s Law states that every 18 months a CPU’s transistors will shrink in size by a factor of two.This is called as Moore’s Law.
The size of a transistor on an integrated circuit chip =2 microns
The size of the transistor be in 2 years = 18months/2 years
= 18months/24 months
= 18/24
= 0.75microns.
The popular programming languages like C, C++, C#, Objective-C, Java, Python, Ruby are ill-equipped to stand up to the challenge of multicore programming. While developers can still use these languages for single CPU programming, they face the drawback of leaving the speed of the program.