简体   繁体   English

Hadoop Facebook使用Mapreduce的共同朋友

[英]Hadoop facebook mutual friends using mapreduce

I was trying a mapreduce program in hadoop (Java version) , to find the mutual friends list from a json file . 我正在尝试使用hadoop(Java版)中的mapreduce程序,以从json文件中找到共同的朋友列表。 The json file content has the following pattern : json文件内容具有以下模式:

{"name":"abc","id":123} [{"name":"xyz","id":124},{"name":"def","id":125},{"name":"cxf","id":155}]
{"name":"cxf","id":155} [{"name":"xyz","id":124},{"name":"abc","id":123},{"name":"yyy","id":129}]

Pattern to be interpreted as follows : 模式解释如下:

friend json tab separated by array of related friends json's 朋友json标签,由相关朋友json的数组分隔

Hence abc has xyz , def and cxf as friends cxf has xyz abc and yyy as friends . 因此abc拥有xyz,def和cxf作为朋友cxf拥有xyz abc和yyy作为朋友。

Given the above the mutual friends between abc and cxf is xyz . 鉴于以上所述,abc和cxf之间的共同朋友是xyz。

tried to implement the same using mapreduce by creating custom writables , with the mapper emitting following key values , key being pair of friends and value being related friends of the first friend in the key (ie , pair of friends) 尝试通过创建自定义可写对象使用mapreduce来实现相同的功能,而映射器发出以下键值,键是一对朋友,而值是该键中第一个朋友(即一对朋友)的相关朋友

K->V
(abc,xyz) -> [xyz,def,cxf]
(abc,def) -> [xyz,def,cxf]
(abc,cxf) -> [xyz,def,cxf]
(cxf,xyz) -> [xyz,abc,yyy]
(cxf,abc) -> [xyz,abc,yyy]
(cxf,yyy) -> [xyz,abc,yyy]

The key here is actually a Custom writable , created a class which extends WritableComparable and i have overridden the compareTo method so that both these pairs (a,b) and (b,a) are same . 这里的键实际上是Custom可写的,创建了一个扩展WritableComparable的类,并且我重写了compareTo方法,以使这两个对(a,b)和(b,a)相同。 But the problem i am facing is that the compareTo method is not invoked for all combinations of pairs and hence the reducer logic is failing. 但是我面临的问题是,并非对所有组合都调用compareTo方法,因此reducer逻辑失败了。

Based on the above example , there are 6 K, V pairs emitted by the mapper . 根据以上示例,映射器发出了6 K,V对。 But compareTo is invoked only 5 times key1.compareTo(key2) , key2.compareTo(key3), key3.compareTo(key4),key4.compareTo(key5),,key5.compareTo(key6) . 但是compareTo仅被调用5次key1.compareTo(key2),key2.compareTo(key3),key3.compareTo(key4),key4.compareTo(key5),key5.compareTo(key6)。

Any idea why this is happening ? 知道为什么会这样吗?

Below is the code as per the logic suggested by f11ler 以下是根据f11ler建议的逻辑的代码

Driver class : 驾驶员等级:

package com.facebook.updated;

import java.io.IOException;

import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
import org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;
import org.apache.log4j.Logger;

public class FacebookMain extends Configured implements Tool

{

    Logger logger = Logger.getLogger(FacebookMain.class);

    public static void main(String[] args) throws Exception {
        System.exit(ToolRunner.run(new FacebookMain(), args));

    }

    @Override
    public int run(String[] args) throws IOException, ClassNotFoundException, InterruptedException {
        logger.info("Running======>");
        Job job = Job.getInstance();

        job.setJarByClass(FacebookMain.class);
        job.setJobName("FBApp");

        job.setMapOutputKeyClass(Friend.class);
        job.setMapOutputValueClass(Friend.class);

        job.setOutputKeyClass(FriendPair.class);
        job.setOutputValueClass(Friend.class);

        job.setMapperClass(FacebookMapper.class);
        job.setReducerClass(FacebookReducer.class);

        job.setInputFormatClass(org.apache.hadoop.mapreduce.lib.input.TextInputFormat.class);
        job.setOutputFormatClass(SequenceFileOutputFormat.class);

        FileInputFormat.setInputPaths(job, new Path(args[0]));
        FileOutputFormat.setOutputPath(job, new Path(args[1]));

        boolean val = job.waitForCompletion(true);

        return val ? 0 : 1;

    }

}

The customWritables (used to represent a friend and friendpair) customWritables(用于表示朋友和对子)

package com.facebook.updated;

import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;

import lombok.Getter;
import lombok.Setter;

import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.io.WritableComparable;
import org.apache.log4j.Logger;

@Getter
@Setter
public class Friend implements WritableComparable<Friend> {

    Logger logger = Logger.getLogger(Friend.class);

    private IntWritable id;
    private Text name;

    public Friend() {
        this.id = new IntWritable();
        this.name = new Text();
    }

    @Override
    public int compareTo(Friend arg0) {
        int val = getId().compareTo(arg0.getId());
        logger.info("compareTo Friend ======> " + arg0 + " and " + this + " compare is " + val);
        return val;
    }

    @Override
    public void readFields(DataInput in) throws IOException {
        id.readFields(in);
        name.readFields(in);
    }

    @Override
    public void write(DataOutput out) throws IOException {
        id.write(out);
        name.write(out);
    }

    @Override
    public boolean equals(Object obj) {
        Friend f2 = (Friend) obj;
        boolean val = this.getId().equals(f2.getId());
        //logger.info("equals Friend ======> " + obj + " and " + this);
        return val;
    }

    @Override
    public String toString() {
        return id + ":" + name + " ";
    }
}

package com.facebook.updated;

import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;

import lombok.Getter;
import lombok.Setter;

import org.apache.hadoop.io.WritableComparable;
import org.apache.log4j.Logger;

@Getter
@Setter
public class FriendPair implements WritableComparable<FriendPair> {

    Logger logger = Logger.getLogger(FriendPair.class);

    private Friend first;
    private Friend second;

    public FriendPair() {
        this.first = new Friend();
        this.second = new Friend();
    }

    public FriendPair(Friend f1, Friend f2) {
        this.first = f1;
        this.second = f2;
    }

    @Override
    public int compareTo(FriendPair o) {

        logger.info("compareTo FriendPair ======> " + o + " and " + this);
        FriendPair pair2 = o;
        int cmp = -1;

        if (getFirst().compareTo(pair2.getFirst()) == 0 || getFirst().compareTo(pair2.getSecond()) == 0) {
            cmp = 0;
        }
        if (cmp != 0) {
            // logger.info("compareTo FriendPair ======> " + o + " and " + this
            // + " comparison is " + cmp);
            return cmp;
        }
        cmp = -1;
        if (getSecond().compareTo(pair2.getFirst()) == 0 || getSecond().compareTo(pair2.getSecond()) == 0) {
            cmp = 0;
        }

        // logger.info("compareTo FriendPair ======> " + o + " and " + this +
        // " comparison is " + cmp);

        // logger.info("getFirst() " + getFirst());
        // logger.info("pair2.getFirst() " + pair2.getFirst());
        // logger.info("getFirst().compareTo(pair2.getFirst()) " +
        // getFirst().compareTo(pair2.getFirst()));
        // logger.info("getFirst().compareTo(pair2.getSecond()) " +
        // getFirst().compareTo(pair2.getSecond()));
        // logger.info("getSecond().compareTo(pair2.getFirst()) " +
        // getSecond().compareTo(pair2.getFirst()));
        // logger.info("getSecond().compareTo(pair2.getSecond()) " +
        // getSecond().compareTo(pair2.getSecond()));
        // logger.info("pair2.getSecond() " + pair2.getSecond());
        // logger.info("getSecond() " + getSecond());
        // logger.info("pair2.getFirst() " + pair2.getFirst());
        // logger.info("pair2.getSecond() " + pair2.getSecond());

        return cmp;
    }

    @Override
    public boolean equals(Object obj) {

        FriendPair pair1 = this;
        FriendPair pair2 = (FriendPair) obj;

        boolean eq = false;

        logger.info("equals FriendPair ======> " + obj + " and " + this);

        if (pair1.getFirst().equals(pair2.getFirst()) || pair1.getFirst().equals(pair2.getSecond()))
            eq = true;

        if (!eq) {
            // logger.info("equals FriendPair ======> " + obj + " and " + this +
            // " equality is " + eq);
            return false;
        }
        if (pair1.getSecond().equals(pair2.getFirst()) || pair1.getSecond().equals(pair2.getSecond()))
            eq = true;

        // logger.info("equals FriendPair ======> " + obj + " and " + this +
        // " equality is " + eq);

        return eq;
    }

    @Override
    public void readFields(DataInput in) throws IOException {
        first.readFields(in);
        second.readFields(in);
    }

    @Override
    public void write(DataOutput out) throws IOException {
        first.write(out);
        second.write(out);
    }

    @Override
    public String toString() {
        return "[" + first + ";" + second + "]";
    }

    @Override
    public int hashCode() {
        logger.info("hashCode FriendPair ======> " + this);
        return first.getId().hashCode() + second.getId().hashCode();
    }
}

Mapper and Reducer 映射器和减速器

package com.facebook.updated;

import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
import java.util.StringTokenizer;

import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.log4j.Logger;

import com.mongodb.BasicDBList;
import com.mongodb.BasicDBObject;
import com.mongodb.util.JSON;

public class FacebookMapper extends Mapper<LongWritable, Text, Friend, Friend> {

    Logger log = Logger.getLogger(FacebookMapper.class);

    @Override
    protected void map(LongWritable key, Text value, Mapper<LongWritable, Text, Friend, Friend>.Context context)
            throws IOException, InterruptedException {

        String line = value.toString();
        StringTokenizer st = new StringTokenizer(line, "\t");
        String person = st.nextToken();
        String friends = st.nextToken();

        BasicDBObject personObj = (BasicDBObject) JSON.parse(person);
        BasicDBList friendsList = (BasicDBList) JSON.parse(friends);

        List<Friend> frndJavaList = new ArrayList<>();

        for (Object frndObj : friendsList) {
            frndJavaList.add(getFriend((BasicDBObject) frndObj));
        }

        Friend frnd = getFriend(personObj);
        Friend[] array = frndJavaList.toArray(new Friend[frndJavaList.size()]);
        for (Friend f : array) {
            log.info("Map output is " + f + " and " + frnd);
            context.write(f, frnd);
        }
    }

    private static Friend getFriend(BasicDBObject personObj) {
        Friend frnd = new Friend();
        frnd.setId(new IntWritable(personObj.getInt("id")));
        frnd.setName(new Text(personObj.getString("name")));
        frnd.setHomeTown(new Text(personObj.getString("homeTown")));
        return frnd;
    }
}

package com.facebook.updated;

import java.io.IOException;
import java.util.ArrayList;
import java.util.List;

import org.apache.hadoop.mapreduce.Reducer;
import org.apache.log4j.Logger;

public class FacebookReducer extends Reducer<Friend, Friend, FriendPair, Friend> {

    Logger log = Logger.getLogger(FacebookReducer.class);

    @Override
    protected void reduce(Friend friend, Iterable<Friend> vals,
            Reducer<Friend, Friend, FriendPair, Friend>.Context context) throws IOException, InterruptedException {
        List<Friend> friends = new ArrayList<>();
        for (Friend frnd : vals) {
            friends.add(frnd);
        }
        log.info("Reducer output is " + friend + " and values are " + friends);
        if (friends.size() == 2) {
            FriendPair key = new FriendPair(friends.get(0), friends.get(1));
            context.write(key, friend);
        } else {
            //log.info("Size of friends is not 2 key is " + friend + " and values are " + friends);
        }

    }
}

Input json file containing 2 lines 输入包含2行的json文件

{"name":"abc","id":123} [{"name":"xyz","id":124},{"name":"def","id":125},{"name":"cxf","id":155}]
{"name":"cxf","id":155} [{"name":"xyz","id":124},{"name":"abc","id":123},{"name":"yyy","id":129}]

Output of reducer (abc,abc)->xyz 减速器的输出 (abc,abc)-> xyz

compareTo method is required for sorting, this relation should be transitive. compareTo方法是排序所必需的,此关系应该是可传递的。 This mean that if a > b and b > c then a > c. 这意味着如果a> b且b> c,则a> c。 Probably this is not true for your implementation. 可能这对您的实现不正确。

Why you generate this kind of records in mapper? 为什么在mapper中生成此类记录? If "being a friend" is a symmetric relation you can simply do a mapper-only job with this logic (pseudo-code): 如果“成为朋友”是对称关系,则可以使用此逻辑(伪代码)简单地执行仅映射器的工作:

for(int i = 0; i < values.length; ++i)
    for(int j = 0; j < values.length; ++j)
        if (i ==j)
            continue
        emmit (values[i], values[j]), key

Update: If this is not symmetric (which means that "xyz has friend abc" not follows from "abc has friend xyz") then we need reverse records: 更新:如果这不是对称的(这意味着“ xyz有朋友abc”不跟在“ abc有朋友xyz”之后),那么我们需要反向记录:

Mapper: 映射器:

for(int i = 0; i < values.length; ++i)
    emmit values[i], key

Reducer (same as mapper before): 减速器(与之前的映射器相同):

for(int i = 0; i < values.length; ++i)
    for(int j = 0; j < values.length; ++j)
        if (i ==j)
            continue
        emmit (values[i], values[j]), key

Update2: 更新2:

Lets see how this algorithm works with your example: 让我们看看该算法如何与您的示例配合使用:

The result of mapper: 映射器的结果:

xyz -> abc
def -> abc
cxf -> abc
xyz -> cxf
abc -> cxf
yyy -> cxf

Mapreduce wiil group this values by key, so the input of reducer: Mapreduce将按键将此值分组,所以化简器的输入为:

xyz -> [abc,cxf]
def -> [abc]
cxf -> [abc]
abc -> [cxf]
yyy -> [cxf]

In reducer we do a nested loop by values, but skip comparing with self. 在reducer中,我们按值进行嵌套循环,但是跳过与self的比较。 Result: 结果:

(abc, cxf) -> xyz

This is what we want to get. 这就是我们想要得到的。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM