简体   繁体   中英

Will reducer out of java heap space

I am implementing a program using Hadoop. My question is how to deal with java out of space problem, I added some property configuration into xml file, but it didn't work. Increasing number of reducers doesn't work for me either. Because in my program every reducer needs large sparse whole matrix, and I am not allowed to change this logic. Yet every reducer will receive an entry with column id as key, and column vector as value.Is there any way I can get out of this dilemma?

PS:I was first thinking that reducer will take column one by one, and that won't cause out of memory issue. But it seems not this way. Actually, I am trying to implement an algorithm described in a paper. The algorithm for reducer is listed below. Line 11 is the column I mentioned, and each reducer has every column of the matrix. 在此处输入图片说明

The reason why you use something like hadoop is because you cant fit the entire data set into memory. Either you don't change the logic and try to find a computer that's big enough or you parallel-ize the algorithm and exploit hadoop.

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM