data-algorithms-book
data-algorithms-book copied to clipboard
MapReduce, Spark, Java, and Scala for Data Algorithms Book
Bumps hadoop-common from 2.10.1 to 3.2.3. [](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a...
I am wondering to know when the second edition will be out? Thanks,
Your project uses some dependencies with CVEs. I found that the buggy methods of the CVEs are in the program execution path of your project, which makes your project at...
Hi, I have a question about the FastaRecordReader class data-algorithms-book/src/main/java/org/dataalgorithms/chap24/mapreduce/FastaRecordReader.java I have been trying to use it for large genomes (fasta files much larger than a HDFS block, ie: ftp://ftp.ncbi.nlm.nih.gov/genomes/all/GCF/000/001/405/GCF_000001405.38_GRCh38.p12/GCF_000001405.38_GRCh38.p12_genomic.fna.gz)...
I want to run these source code ,but these source code need special development environment . so can you provide a sanbox like cdh-sanbox that can source code can run...
my teacher ask me to finish unit 21,as 50% for my final exam,,,i want to allele code...
ClientArguments only takes on argument in current version. Can you update this example? I'm not sure it's possible to do it this way (anymore), ack or nack? Thanks
Hi when am invoking https://github.com/mahmoudparsian/data-algorithms-book/blob/master/misc/how-to-submit-spark-job-to-yarn-from-java-code.md this class from a running spark-submit from local it is getting invoked and able to submit the spark-submit to the yarn cluster. But when am...
This pull request is focused on resolving occurrences of Sonar rule squid:S2097 - “ "equals(Object obj)" should test argument type ”. You can find more information about the issue here:...
Dear mahmoudparsian, Sorry to bother you. Actually, it is known that two methods can be used in the propose of saving output when one scala-spark program finishes. As you do...