Mapreducetrabajos

Filtro

Mis búsquedas recientes
Filtrar por:
Presupuesto
a
a
a
Tipo
Habilidades
Idiomas
    Estado del trabajo
    976 mapreduce trabajados encontrados, precios en USD

    The project I want to be implemented can be downloaded from here: ~kpm/TopGC/ The above project is written in Java and I need someone to convert it so it would be able to run in Apache Hadoop using the MapReduce framework. I want to make a project in Apache Hadoop (with MapReduce) written in Java, which does the same things as TopGC. I'd like also to take a look at this pdf (~kpm/TopGC/), which describes the functionality of the system. In the last page you can find the pseudocode. The things I want to recieve (when the project is finished) are the source codes with comments and a small tutorial with instructions to be able to run it on my own PC. I need it completed till 15/6/2012.

    $227 (Avg Bid)
    $227 Oferta promedio
    7 ofertas
    Web Demo of MapReduce Finalizado left

    The goal of this project is to create a demo project that will use Amazon EC2 (or another service if you have a good reason) to implement a basic login/account functionality as well as a MapReduce backend to do a search on the site. The basic demo will work as follows: After logging in the user can create a new search. They will enter places they've been in one column and places they want to go in the other. The service will then search the database and return the matches of people who have the opposite criteria (i.e. someone who has been to New york city and wants to go to paris would be matched with someone who wants to go to new york city but has already been to paris). Results will be sorted based on number of relevant matches. There will also need to be s...

    $307 (Avg Bid)
    $307 Oferta promedio
    4 ofertas
    Simple MapReduce Finalizado left

    I have a simple database structure for tracking a users time on a website. The session_time entry is accumulative. i.e The first session_time entry when you search by session_id with always be 0 last entry will always be the total session time I require a MongoDB mapreduce function/query that can collect the total, hourly, daily, monthly etc session time for a user_id within a date range using created_at. The structure of the data is like: { _id "4f2d1122ae06ff0003000002" user_id "d77c8001-adc4-49c6-9e83-d58208e78ebd" session_id "1e49ab89c2f634b4" session_time 0 created_at 2012-02-04 11:06:10 UTC updated_at 2012-02-04 11:06:10 UTC }, { _id "4f2d1141ae06ff0003000004" user_id "d77c8001-adc4-49c6-...

    $30 - $250
    Sellado
    $30 - $250
    2 ofertas

    You will create a massive information aggregation and retrieval system. You should know the entire SMAQ stack (Storage, MapReduce and Query) and be ready to start immediately. You will be working with other engineers at our company. You absolutely must be able to demonstrate other large projects you've finished, and we will need to talk to those who you worked for. This is a huge project, and needs to be started and finished this month. I'm sure you are good, but this project will be challenging and we need the absolute best. Don't apply if you are still learning the SMAQ stack and have never done a large project in this space. The bid will be ignored. More details will be given if you fit the qualifications and message us.

    $4187 (Avg Bid)
    $4187 Oferta promedio
    8 ofertas
    Hadoop Project Finalizado left

    All the examples in should be implemented in Java and they should use Hadoop MapReduce (programming model/software framework). is an implemented example in Java using Hadoop MapReduce framework. You can also see the source code of in the file. All the examples in should be implemented as the example in the Example.doc.

    $139 (Avg Bid)
    $139 Oferta promedio
    10 ofertas

    I'd like to understand how to solve a specific problem with map reduce/hadoop whereby I have to do a "fuzzy" join on large data sets. Please see the details. ## Deliverables Let's say I have two fairly large data sets - the first is called "Base" and it contains 200 million tab delimited rows and the second is call "MatchSet" which has 10 million tab delimited rows of similar data. Let's say I then also have an arbitrary function called Match(row1, row2) and Match() essentially contains some heuristics for looking at row1 (from MatchSet) and comparing it to row2 (from Base) and determining if they are similar in some way. Let's say the rules implemented in Match() are custom and complex rules, aka not a simple string match, i...

    $160 (Avg Bid)
    $160 Oferta promedio
    2 ofertas

    ...robust, user friendly architectures that? In Phase 1: (2 weeks from start) i. Host on Amazon's EC2 and optimization to reduce costs (i.e. using a virtual instance on EC2); ii. Link multiple databases together - the core Hbase (to store time series data) and MYSQL (which keys the data to tags and keeps data statistics); iii. AS A BONUS -- Simply demonstrate an ability to use Mapreduce for function calls across the Hbase database; In Phase 2: (6 weeks from start) iv. Build relatively rich GUIs using JQuery and other graphic libraries to graph the data quickly (using <> or another solution where it is graphed on the server then pushed to the client); v. Build the system from extensible program strucutres with libraries and APIs to allow

    $6500 (Avg Bid)
    $6500 Oferta promedio
    4 ofertas

    Need help, with: ================== 1. Setting up HBase on Amazon EC2 2. Design of table on HBase 3. Using Thrift through PHP to: insert/update/delete/query data from HBase 4. Doing a simple MapReduce Bid ONLY, if: ============== 1. You've had experience with all of the above, nothing is new to you and you can easily explain everything. 2. You know best practices regarding all of the above. If you'll be able to deliver in time, quality and according to the above requirements, full time position will be offered. Any question, please submit privately, Thank you

    $250 - $750
    Destacado Sellado
    $250 - $750
    6 ofertas

    It's about a very simple MapReduce program running on Hadoop and written in Java, which gets information from a specific website and makes some statistic calculations. Then it returns the results into a file. For someone with experience in MapReduce programming it will be an easy task to complete.

    $209 (Avg Bid)
    $209 Oferta promedio
    9 ofertas

    It's about a very simple MapReduce program running on Hadoop and written in Java, which gets information from a specific website and makes some statistic calculations. Then it returns the results into a file. For someone with experience in MapReduce programming it will be an easy task to complete.

    N/A
    N/A
    0 ofertas

    The Disco Project. Disco is a distributed computing framework based on the MapReduce paradigm. Please do not bid on this project unless you have experience working with Disco, Erlang and Python. I have a project that is powered by PHP. I need it to be converted to run in a distributed Disco environment, using The Disco Project, Erlang and Python. CouchDB should be used for the database. Please provide proof of experience with these technologies. More information and links will be provided to qualified bidders.

    $750 - $1500
    $750 - $1500
    0 ofertas

    DELIVERABLE Cloud infrastructure and web UI using R (e.g. RStudio), distributed file systems (e.g. HDFS, S3) and distributed computing (e.g. MapReduce). USERS AND USER GROUPS Users log in using username and password or credential providers (Google, facebook, etc) into a secure work environment. Each work environment may be accessed by more than one user, and thus the set of users with access to a work environment constitute a user group. Work environment access rights are of three kinds: author, reviewer and admin. Authors may read and write, Reviewers only read, and Admins read, write are able to invite and exclude users from the group. Work environment content includes data, R scripts and messages between users. ENVIRONMENT FUNCTIONALITY Authors and Admins are ...

    $30 - $5000
    $30 - $5000
    0 ofertas

    Cloud computing analytics UI for analyzing data through Amazon Web Services. The UI will allow users to: 1. Connect their data to the VPC 2. Upload data to S3 3. Manage their data through SimpleDB and RDS 4. Create and maintain R scripts for analyzing the data in the VPC or S3 through Elastic MapReduce in an environment such as Elastic-R 5. Compare analytics results and select best alternatives 6. Communicate with other users via chat and discussion boards 7. Run results reports through CloudFront 8. Pay subscription fees through FWS

    $2068 (Avg Bid)
    $2068 Oferta promedio
    3 ofertas

    ...to mitigate and stop the attack.**************************** * ******************************Disaster Recovery -Helps find the cause of the Disaster, Recover and rebuild from your backup.****************************** The server administrator needs to have high skills in """Server **Optimize and Tweak High-Traffic Servers tools"""" ** *MapReduce * Hadoop * Apache Cassandra, * Memcached *Load balancer in front of webserver nodes, *Load Balancer Node *Web Servers Node *Database Server Node *CISCO Compression: <> fastbid guide [][7][~kewu/fastbit/doc/][8] and other

    $1 / hr (Avg Bid)
    $1 / hr Oferta promedio
    1 ofertas

    I have html files I need to parse. I want to use MapReduce to do the work, but don't know how. I wrote the java program to loop through a local directory with the files and parse them into a CSV file. I want to upload all my files to Amazon S3 and then use a mapreduce job to parse the files into 1 or more CSV files. For this project, I would like someone to take the attached java program and convert it into a java mapreduce job that I can run on Amazon. I would also like detailed, step-by-step instructions on how to initiate the job and get the results. Also, sample html files are in the attached zip file. ## Deliverables 0a) Convert attached java program into a java mapreduce job that I can run on Amazon. 0b) Detailed, step-by-step instructions on how t...

    $48 (Avg Bid)
    $48 Oferta promedio
    2 ofertas
    Hadoop jobserver Finalizado left

    We need a jobserver where we can dispatch jobs to a hadoop cluster. It is approximately the same we try to achieve that is discussed at: @ We would like to target amazon mapreduce, see and the project should include integration with this service. The jobserver should be a standard war file that is deployable in any standard servlet container. Preferable it should be written using struts/hibernate/mysql/jquery/ext/guice or other similar open source technologies We should be able to administrate jobs through an ui of the server that is accessed through a browser. The first job that should be implemented on the server is for processing html pages fetched from real estate

    $959 (Avg Bid)
    $959 Oferta promedio
    7 ofertas
    Web Crawler! Finalizado left

    Hello, everyone, we are seeking a well tailored web crawler for our need of vertical search market in china, here is the details : 1. able to run multiple instance simultaneously for multiple download of web pages 2. able to set up index in form of word count statistic of each page, also the hyperlink structure of the page h...they shall be delivered through web page view such as asp or php, but asp is our currently preferred, 9. remote access is highly preferred 10. good security in design and coding, particularly if you use language like c or c++, and other security best practices industrial wide shall be used, issues such as data privacy and integrity and authorization is essential 11. integration with google mapreduce or bigtable is highly preferred but not essential b...

    $525 (Avg Bid)
    $525 Oferta promedio
    2 ofertas

    ...experience in programming, but wishes to produce an application based on artificial neural networks, programmed in Java, and run on the Elastic MapReduce model. Additionally, the manager has produced an incomplete process flow, describing the application. The application itself has a limited scope and relatively simple structure, so creation of this SRS is not expected take a great deal of time. This SRS will be used to estimate cost, to seek bids for the development of the application, and to guide the development itself. Application components: * Browser-based GUI * Java-based neural network training * Java-based report generation * Java-based MapReduce programs Process: 1. 05/25/09: Interview with manager, to be scheduled between 3:00 p.m. and 12:00 a.m. G...

    $181 (Avg Bid)
    $181 Oferta promedio
    4 ofertas

    Enable a Python application to run in [][1][Amazon Elastic MapReduce][1] environment by modifying well-documented and well-structured source-code. The original application was developed to retrieve Wikimapia information and designed to enable proto-parallel processing: it can subdivide one task in order to run it in parallel in multiple computers and then collates the results. However, it was not developed to take advantage of [][2][Hadoop][2]. ## Deliverables # Preliminary Analysis: Amazon provides an extended example on how to distribute Python processes ??" check [][3][Finding Similar Items with Amazon Elastic MapReduce, Python, and Hadoop Streaming][4]? to get an idea of the desired result. The application to adapt has fewer than 900 lines. See attached [] for...

    $446 (Avg Bid)
    $446 Oferta promedio
    2 ofertas

    ... There are several in interpreted languages, which are not my first choice, but I will consider if the coder feels that they can provide the required performance. Hadoop pipes Ruby starfish Ruby skynet If possible, I am looking for a coder who has created a similar application and is familiar with the suggested solution. Please feel free to ask questions or make suggestions that are outside of the parameters that have been outlined. I have included sample files with a small number of chemical structures. Feel free to duplicate these to create larger input files or to request a larger

    $85 (Avg Bid)
    $85 Oferta promedio
    1 ofertas
    Hadoop Genius Wanted Finalizado left

    We are a web property, shortly to launch, that has a requireme...requirement for a very very very large and elegantly designed database architecture capable of handling a data set of up to 20 million "records" with 50 points of relevance on day one. This base will grow daily based on the input and interraction of our registered users, and the entire database needs to be quickly ported to requests from our corporate customers. We are interested in any hadoop or mapreduce experts opinion on how you would construct such an architecture from the code to the systems, storage arrays, etc on a PAID CONSULTANT basis, or, full time employment as this will be an ongoing entity begginning in NOV 2008. Email us for more info, or with relevant work you have done on large, rapidly scali...

    PHP
    $5000 - $20000
    $5000 - $20000
    0 ofertas

    ...hadoop distribute filesystem. See <> . The input for this program will be an xml file, so what we need is we can index, search, add and delete in the hadoop distribute filesystem, and this filesystem will be distribute on multiple servers, which means when 1 or 2 servers went down, the whole system should still be functional. It should be Mapreduce programming and XML-RPC API or Soap API. Please test it in your own servers first, and we will eventually set up three or more servers for developing and testing purpose. _To be qualified for this project, you must know Java, Lucens, and Hadoop well, and also have knowledge of how to manage Linux servers._ **There is a high possibility that a nice completion of this project will lead to a

    $212 (Avg Bid)
    $212 Oferta promedio
    1 ofertas

    ...hadoop distribute filesystem. See <> . The input for this program will be an xml file, so what we need is we can index, search, add and delete in the hadoop distribute filesystem, and this filesystem will be distribute on multiple servers, which means when 1 or 2 servers went down, the whole system should still be functional. It should se Mapreduce programming and XML-RPC API or Soap API. We will set up three or more servers for you. To be qualified for this project, you must know Java, Lucens, and Hadoop extremely well, and have rich experience in this area. ## Deliverables 1) Complete and fully-functional working program(s) in executable form as well as complete source code of all work done. 2) Deliverables must be in

    $361 (Avg Bid)
    $361 Oferta promedio
    2 ofertas

    Principales artículos de la comunidad mapreduce