
Sign up to save your podcasts
Or


When computers became commodity hardware and storage became incredibly cheap, we entered the era of so-call "big" data. Most definitions of big data will include something about not being able to process all the data on a single machine. Distributed computing is required for such large datasets.
Getting an algorithm to run on data spread out over a variety of different machines introduced new challenges for designing large-scale systems. First, there are concerns about the best strategy for spreading that data over many machines in an orderly fashion. Resolving ambiguity or disagreements across sources is sometimes required.
This episode discusses how such algorithms related to the complexity class NC.
By Kyle Polich4.4
475475 ratings
When computers became commodity hardware and storage became incredibly cheap, we entered the era of so-call "big" data. Most definitions of big data will include something about not being able to process all the data on a single machine. Distributed computing is required for such large datasets.
Getting an algorithm to run on data spread out over a variety of different machines introduced new challenges for designing large-scale systems. First, there are concerns about the best strategy for spreading that data over many machines in an orderly fashion. Resolving ambiguity or disagreements across sources is sometimes required.
This episode discusses how such algorithms related to the complexity class NC.

32,243 Listeners

30,635 Listeners

288 Listeners

1,107 Listeners

629 Listeners

583 Listeners

305 Listeners

345 Listeners

209 Listeners

205 Listeners

313 Listeners

100 Listeners

554 Listeners

102 Listeners

229 Listeners