Answering The Nagging Apache Hadoop/Spark Question
- Details
- Written by Douglas Eadline
- Hits: 8733
(or How to Avoid the Trough of Disillusionment)
A recent blog post, Why not so Hadoop?, is worth reading if you are interested in big data analytics, Hadoop, Spark, and all that. The article contains the 2015 Gartner Hype Cycle. The 2016 version is worth examining as well. Some points similar to the blog can be made here:
- Big data was at the "Trough of Disillusionment" stage in 2014, but is not seen in the 2015/16 Hype cycle.
- The "Internet of Things" (a technology that is expected to fill the big data pipeline) was on the peak for two years and now has been given "platform status."
Return of the Free Lunch (sort of)
- Details
- Written by Douglas Eadline
- Hits: 5201
From the do-you-want-fries-with-that department
As processor frequencies continue to level off and mainstream processors keep sprouting cores, the end of the "frequency free lunch" has finally arrived. That is, in the good old days each new generation of processor would bring with it a faster system clock that would result in a "free" performance bump for many software applications -- no reprogramming needed. Can we ever get back to the good old days?
You (Still) Can't Always Get What You Want
- Details
- Written by Douglas Eadline
- Hits: 4028
You can't always get what you want. But that doesn't stop me from asking. Besides, Buddha's got my back. Here's my HPC wish list.
[Note: A recent article by Al Geist, How To Kill A Supercomputer: Dirty Power, Cosmic Rays, and Bad Solder reminds us that statistics at scale represents a big challenge for HPC. The following updated article was originally published in Linux Magazine in May 2006 and offers some further thoughts on this important issue. ]
Twenty five years ago I wrote a short article in a now defunct parallel computing magazine (Parallelogram) entitled "How Will You Program 1000 Processors?" Back then, it was a good question that had no easy answer. Today, it's still a good question with no easy answer, except now it seems a bit more urgent as we step into the "multi-core" era. Indeed, when I originally wrote the article, using 1,000 processors was a far off, but real possibility. Today, 1,000 processors are a reality for many practitioners of high-performance computing (HPC). And as dual-cores (and now 18-core processors) hit the server room, effectively doubling processor counts, many more people will be joining the 1,000P club very soon.
So let's get adventurous and ask, "How will you program 10,000 processors?" As I realized twenty five years ago, such a question may never really have a complete answer. In the history of computers, no one has ever answered such a question to my liking — even when considering ten processors. Of course, there are plenty of methods and ideas like threads, messages, barrier synchronization, and so on, but when I have to think more about the computer than about my problem, something is wrong.
The Ignorance is Bliss Approach To Parallel Computing
- Details
- Written by Douglas Eadline
- Hits: 4312
from the random thoughts department
[Note: The following updated article was originally published in Linux Magazine in June 2006 and offers some further thoughts on the concept of dynamic execution.]
In a previous article, I talked about programming large numbers of cluster nodes. By large, I mean somewhere around 10,000. To recap quickly, I pointed out that dependence on large numbers of things increase the chance that one of them will fail. I then proposed that it would be cheaper to develop software that can live with failure than try to engineer hardware redundancy. Finally, I concluded that adapting to failure requires dynamic software. As opposed to statically scheduled programs, dynamic software adapts at run-time. The ultimate goal is to make cluster programming easier: focus more on the problem and less on the minutiae of message passing. (Not that there is anything wrong with message passing or MPI. At some level messages (memory) needs to be transferred between cluster nodes.)
Search
Login And Newsletter
Feedburner
Who's Online
We have 50 guests and no members online
Latest Stories/News
Popular
HPCWire
-
NSF Awards $6 Million Per Year to Trusted CI, Names Berkeley Lab as Steward
NSF Awards $6 Million Per Year to Trusted CI, Names Berkeley Lab as Steward
Oct. 3, 2024 — The U.S. National Science Foundation has awarded Trusted CI, the NSF Cybersecurity Center of Excellence, a five-year, $6-million per-year award to run through September 2029. Lawrence Berkeley National Laboratory (Berkeley Lab) will now serve as Trusted CI’s[…]
Source: HPCwire
Created on: Oct 3, 2024 | 16:25 pm
HPCwire | Oct 3, 2024 | 16:25 pm -
DOE Announces Up to $500M for Basic Research to Advance the Frontiers of Science
DOE Announces Up to $500M for Basic Research to Advance the Frontiers of Science
WASHINGTON, Oct. 3, 2024 — The U.S. Department of Energy (DOE) has announced up to $500 million in funding for basic research in support of DOE’s clean energy, economic, and national security goals. The funding will advance the priorities of[…]
Source: HPCwire
Created on: Oct 3, 2024 | 16:12 pm
HPCwire | Oct 3, 2024 | 16:12 pm -
GIGABYTE Announces General Availability of AmpereOne Servers for Cloud-Native Workloads
GIGABYTE Announces General Availability of AmpereOne Servers for Cloud-Native Workloads
October 3, 2024 – Giga Computing, a subsidiary of GIGABYTE and an industry leader in servers for x86 and ARM platforms as well as advanced cooling technologies, today announced its initial wave of GIGABYTE servers that support the full stack[…]
Source: HPCwire
Created on: Oct 3, 2024 | 15:43 pm
HPCwire | Oct 3, 2024 | 15:43 pm
InsideHPC
-
HPC-AI Tech Secures $50M Series A Funding
Oct 2, 2024 | 19:58 pm
-
SimOps Launches Simulation Operations Automation Initiative
Oct 2, 2024 | 19:26 pm
-
DDN Expands AI Research Partnership with YTexas
Oct 2, 2024 | 18:36 pm