Pig Latin
Fermé
Chercheur000
Messages postés
5
Date d'inscription
dimanche 18 décembre 2016
Statut
Membre
Dernière intervention
20 décembre 2016
-
Modifié par Chercheur000 le 18/12/2016 à 22:30
mamiemando Messages postés 33446 Date d'inscription jeudi 12 mai 2005 Statut Modérateur Dernière intervention 20 décembre 2024 - 21 déc. 2016 à 12:01
mamiemando Messages postés 33446 Date d'inscription jeudi 12 mai 2005 Statut Modérateur Dernière intervention 20 décembre 2024 - 21 déc. 2016 à 12:01
2 réponses
mamiemando
Messages postés
33446
Date d'inscription
jeudi 12 mai 2005
Statut
Modérateur
Dernière intervention
20 décembre 2024
7 812
19 déc. 2016 à 10:27
19 déc. 2016 à 10:27
Bonjour,
Si j'en crois ton message d'erreur :
... j'en déduis que parmi les explications possibles, il y a :
- le chemin vers le fichier d'entrée est faux
- le fichier n'a pas des droits suffisants pour être lu par le profil utilisateur qui lance l'application que tu mentionnes
- le fichier est corrompu ou n'a pas le format attendu par ton logiciel.
Bonne chance
Si j'en crois ton message d'erreur :
Failed to read data from "file:///home/soukaina/Bureau/hdfile"
... j'en déduis que parmi les explications possibles, il y a :
- le chemin vers le fichier d'entrée est faux
- le fichier n'a pas des droits suffisants pour être lu par le profil utilisateur qui lance l'application que tu mentionnes
- le fichier est corrompu ou n'a pas le format attendu par ton logiciel.
Bonne chance
mamiemando
Messages postés
33446
Date d'inscription
jeudi 12 mai 2005
Statut
Modérateur
Dernière intervention
20 décembre 2024
7 812
20 déc. 2016 à 00:05
20 déc. 2016 à 00:05
Ce n'est pas vraiment mon domaine, mais l'erreur est :
Du coup peut être que parmi A,B,C,D,E l'un des jobs a planté, ou en peut être que le résultats ne sont pas homogène et ne permettent pas de faire le "GROUP BY". As-tu regardé si ces 5 variables contenaient le résultat attendu ?
Bonne chance
Failed Jobs:
JobId Alias Feature Message Outputs
job_local1472906397_0003 A,B,C,D,E GROUP_BY,COMBINER Message: Job failed! file:///home/soukaina/Bureau/results,
Du coup peut être que parmi A,B,C,D,E l'un des jobs a planté, ou en peut être que le résultats ne sont pas homogène et ne permettent pas de faire le "GROUP BY". As-tu regardé si ces 5 variables contenaient le résultat attendu ?
Bonne chance
Chercheur000
Messages postés
5
Date d'inscription
dimanche 18 décembre 2016
Statut
Membre
Dernière intervention
20 décembre 2016
Modifié par Chercheur000 le 20/12/2016 à 12:33
Modifié par Chercheur000 le 20/12/2016 à 12:33
Oui ces 5 variables contiennent le résultat attendu.
en fait j'ai exécuté un programme pareil en mode mapreduce et tout a bien marché même la commande STORE
le problème reste au niveau de la commande STORE en mode local.
merci.
en fait j'ai exécuté un programme pareil en mode mapreduce et tout a bien marché même la commande STORE
le problème reste au niveau de la commande STORE en mode local.
merci.
mamiemando
Messages postés
33446
Date d'inscription
jeudi 12 mai 2005
Statut
Modérateur
Dernière intervention
20 décembre 2024
7 812
21 déc. 2016 à 12:01
21 déc. 2016 à 12:01
Alors, ce n'est pas du tout ma spécialité, mais je vois le mot clé GROUP BY, COMBINER.
- Est-ce que cette primitive a du sens par rapport à chacuns des résultats observés dans A,B,C,D,E ?
- As-tu essayé de l'appliquer progressivement (à A, puis A,B, etc... pour localiser la cause du problème) ?
- Est-ce que cette primitive a du sens par rapport à chacuns des résultats observés dans A,B,C,D,E ?
- As-tu essayé de l'appliquer progressivement (à A, puis A,B, etc... pour localiser la cause du problème) ?
19 déc. 2016 à 21:43
19 déc. 2016 à 21:57
2016-12-19 20:50:36,657 [main] INFO org.apache.pig.tools.pigstats.mapreduce.SimplePigStats - Script Statistics:
HadoopVersion PigVersion UserId StartedAt FinishedAt Features
2.6.0 0.15.0 hduser 2016-12-19 20:50:24 2016-12-19 20:50:36 GROUP_BY
Success!
Job Stats (time in seconds):
JobId Maps Reduces MaxMapTime MinMapTime AvgMapTime MedianMapTime MaxReduceTime MinReduceTime AvgReduceTime MedianReducetime Alias Feature Outputs
job_local2023858608_0002 1 1 n/a n/a n/a n/a n/a n/a n/a n/a A,B,C,D GROUP_BY file:/tmp/temp766825122/tmp1271703645,
Input(s):
Successfully read 4 records from: "file:///home/soukaina/Bureau/hdfile"
Output(s):
Successfully stored 7 records in: "file:/tmp/temp766825122/tmp1271703645"
Counters:
Total records written : 7
Total bytes written : 0
Spillable Memory Manager spill count : 0
Total bags proactively spilled: 0
Total records proactively spilled: 0
Job DAG:
job_local2023858608_0002
2016-12-19 20:50:36,702 [main] INFO org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics with processName=JobTracker, sessionId= - already initialized
2016-12-19 20:50:36,707 [main] INFO org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics with processName=JobTracker, sessionId= - already initialized
2016-12-19 20:50:36,726 [main] INFO org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics with processName=JobTracker, sessionId= - already initialized
2016-12-19 20:50:36,822 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Success!
2016-12-19 20:50:36,857 [main] INFO org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS
2016-12-19 20:50:36,859 [main] INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2016-12-19 20:50:36,860 [main] WARN org.apache.pig.data.SchemaTupleBackend - SchemaTupleBackend has already been initialized
2016-12-19 20:50:37,123 [main] INFO org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths to process : 1
2016-12-19 20:50:37,124 [main] INFO org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil - Total input paths to process : 1
(je,{(je),(je),(je)})
(à,{(à)})
(fes,{(fes)})
(suis,{(suis),(suis),(suis)})
(contente,{(contente)})
(j'habite,{(j'habite)})
(soukaina,{(soukaina),(soukaina)})
D: {group: chararray,C: {(mot: chararray)}}
E: {mot: chararray,occurrences: long}
2016-12-19 20:50:38,122 [main] INFO org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS
2016-12-19 20:50:38,134 [main] INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2016-12-19 20:50:38,174 [main] INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.textoutputformat.separator is deprecated. Instead, use mapreduce.output.textoutputformat.separator
2016-12-19 20:50:38,704 [main] INFO org.apache.pig.tools.pigstats.ScriptState - Pig features used in the script: GROUP_BY
2016-12-19 20:50:39,074 [main] INFO org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS
2016-12-19 20:50:39,075 [main] INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2016-12-19 20:50:39,075 [main] INFO org.apache.pig.data.SchemaTupleBackend - Key [pig.schematuple] was not set... will not generate code.
2016-12-19 20:50:39,076 [main] INFO org.apache.pig.newplan.logical.optimizer.LogicalPlanOptimizer - {RULES_ENABLED=[AddForEach, ColumnMapKeyPrune, ConstantCalculator, GroupByConstParallelSetter, LimitOptimizer, LoadTypeCastInserter, MergeFilter, MergeForEach, PartitionFilterOptimizer, PredicatePushdownOptimizer, PushDownForEachFlatten, PushUpFilter, SplitFilter, StreamTypeCastInserter]}
2016-12-19 20:50:39,099 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler - File concatenation threshold: 100 optimistic? false
2016-12-19 20:50:39,124 [main] INFO org.apache.pig.backend.hadoop.executionengine.util.CombinerOptimizerUtil - Choosing to move algebraic foreach to combiner
2016-12-19 20:50:39,273 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer - MR plan size before optimization: 1
2016-12-19 20:50:39,274 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer - MR plan size after optimization: 1
2016-12-19 20:50:39,329 [main] INFO org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS
2016-12-19 20:50:39,331 [main] INFO org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics with processName=JobTracker, sessionId= - already initialized
2016-12-19 20:50:39,336 [main] INFO org.apache.pig.tools.pigstats.mapreduce.MRScriptState - Pig script settings are added to the job
2016-12-19 20:50:39,339 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - mapred.job.reduce.markreset.buffer.percent is not set, set to default 0.3
2016-12-19 20:50:39,340 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Reduce phase detected, estimating # of required reducers.
2016-12-19 20:50:39,341 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Using reducer estimator: org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.InputSizeReducerEstimator
2016-12-19 20:50:39,374 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.InputSizeReducerEstimator - BytesPerReducer=1000000000 maxReducers=999 totalInputFileSize=67
2016-12-19 20:50:39,375 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Setting Parallelism to 1
2016-12-19 20:50:39,431 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Setting up single store job
2016-12-19 20:50:39,440 [main] INFO org.apache.pig.data.SchemaTupleFrontend - Key [pig.schematuple] is false, will not generate code.
2016-12-19 20:50:39,441 [main] INFO org.apache.pig.data.SchemaTupleFrontend - Starting process to move generated code to distributed cacche
2016-12-19 20:50:39,441 [main] INFO org.apache.pig.data.SchemaTupleFrontend - Distributed cache not supported or needed in local mode. Setting key [pig.schematuple.local.dir] with code temp directory: /tmp/1482180639439-0
2016-12-19 20:50:39,738 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 1 map-reduce job(s) waiting for submission.
2016-12-19 20:50:39,762 [JobControl] INFO org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics with processName=JobTracker, sessionId= - already initialized
2016-12-19 20:50:40,148 [JobControl] WARN org.apache.hadoop.mapreduce.JobSubmitter - No job jar file set. User classes may not be found. See Job or Job#setJar(String).
2016-12-19 20:50:40,225 [JobControl] INFO org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths to process : 1
2016-12-19 20:50:40,231 [JobControl] INFO org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil - Total input paths to process : 1
2016-12-19 20:50:40,234 [JobControl] INFO org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil - Total input paths (combined) to process : 1
2016-12-19 20:50:40,680 [JobControl] INFO org.apache.hadoop.mapreduce.JobSubmitter - number of splits:1
2016-12-19 20:50:41,022 [JobControl] INFO org.apache.hadoop.mapreduce.JobSubmitter - Submitting tokens for job: job_local1354877940_0003
2016-12-19 20:50:41,896 [JobControl] INFO org.apache.hadoop.mapreduce.Job - The url to track the job: http://localhost:8080/
2016-12-19 20:50:41,901 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - HadoopJobId: job_local1354877940_0003
2016-12-19 20:50:41,902 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Processing aliases A,B,C,D,E
2016-12-19 20:50:41,902 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - detailed locations: M: A[1,4],B[-1,-1],C[5,4],E[10,4],D[7,4] C: E[10,4],D[7,4] R: E[10,4]
2016-12-19 20:50:41,944 [Thread-50] INFO org.apache.hadoop.mapred.LocalJobRunner - OutputCommitter set in config null
2016-12-19 20:50:41,949 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 0% complete
2016-12-19 20:50:41,950 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Running jobs are [job_local1354877940_0003]
2016-12-19 20:50:41,996 [Thread-50] INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.textoutputformat.separator is deprecated. Instead, use mapreduce.output.textoutputformat.separator
2016-12-19 20:50:42,006 [Thread-50] INFO org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS
2016-12-19 20:50:42,006 [Thread-50] INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.reduce.markreset.buffer.percent is deprecated. Instead, use mapreduce.reduce.markreset.buffer.percent
2016-12-19 20:50:42,007 [Thread-50] INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.reduce.tasks is deprecated. Instead, use mapreduce.job.reduces
2016-12-19 20:50:42,009 [Thread-50] INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2016-12-19 20:50:42,011 [Thread-50] INFO org.apache.hadoop.mapred.LocalJobRunner - OutputCommitter is org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputCommitter
2016-12-19 20:50:42,037 [Thread-50] ERROR org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter - Mkdirs failed to create file:/home/soukaina/Bureau/results/_temporary/0
2016-12-19 20:50:42,047 [Thread-50] INFO org.apache.hadoop.mapred.LocalJobRunner - Waiting for map tasks
2016-12-19 20:50:42,051 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.LocalJobRunner - Starting task: attempt_local1354877940_0003_m_000000_0
2016-12-19 20:50:42,094 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.Task - Using ResourceCalculatorProcessTree : [ ]
2016-12-19 20:50:42,116 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - Processing split: Number of splits :1
Total Length = 67
Input split[0]:
Length = 67
ClassName: org.apache.hadoop.mapreduce.lib.input.FileSplit
Locations:
-----------------------
2016-12-19 20:50:42,154 [LocalJobRunner Map Task Executor #0] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader - Current split being processed file:/home/soukaina/Bureau/hdfile:0+67
2016-12-19 20:50:42,786 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - (EQUATOR) 0 kvi 26214396(104857584)
2016-12-19 20:50:42,786 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - mapreduce.task.io.sort.mb: 100
2016-12-19 20:50:42,793 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - soft limit at 83886080
2016-12-19 20:50:42,793 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - bufstart = 0; bufvoid = 104857600
2016-12-19 20:50:42,794 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - kvstart = 26214396; length = 6553600
2016-12-19 20:50:42,818 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
2016-12-19 20:50:42,873 [LocalJobRunner Map Task Executor #0] INFO org.apache.pig.data.SchemaTupleBackend - Key [pig.schematuple] was not set... will not generate code.
2016-12-19 20:50:42,950 [LocalJobRunner Map Task Executor #0] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Map - Aliases being processed per job phase (AliasName[line,offset]): M: A[1,4],B[-1,-1],C[5,4],E[10,4],D[7,4] C: E[10,4],D[7,4] R: E[10,4]
2016-12-19 20:50:42,983 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.LocalJobRunner -
2016-12-19 20:50:42,985 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - Starting flush of map output
2016-12-19 20:50:42,989 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - Spilling map output
2016-12-19 20:50:42,990 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - bufstart = 0; bufend = 151; bufvoid = 104857600
2016-12-19 20:50:42,990 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - kvstart = 26214396(104857584); kvend = 26214352(104857408); length = 45/6553600
2016-12-19 20:50:43,098 [LocalJobRunner Map Task Executor #0] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigCombiner$Combine - Aliases being processed per job phase (AliasName[line,offset]): M: A[1,4],B[-1,-1],C[5,4],E[10,4],D[7,4] C: E[10,4],D[7,4] R: E[10,4]
2016-12-19 20:50:43,118 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.MapTask - Finished spill 0
2016-12-19 20:50:43,135 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.Task - Task:attempt_local1354877940_0003_m_000000_0 is done. And is in the process of committing
2016-12-19 20:50:43,147 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.LocalJobRunner - map
2016-12-19 20:50:43,150 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.Task - Task 'attempt_local1354877940_0003_m_000000_0' done.
2016-12-19 20:50:43,152 [LocalJobRunner Map Task Executor #0] INFO org.apache.hadoop.mapred.LocalJobRunner - Finishing task: attempt_local1354877940_0003_m_000000_0
2016-12-19 20:50:43,156 [Thread-50] INFO org.apache.hadoop.mapred.LocalJobRunner - map task executor complete.
2016-12-19 20:50:43,159 [Thread-50] INFO org.apache.hadoop.mapred.LocalJobRunner - Waiting for reduce tasks
2016-12-19 20:50:43,164 [pool-8-thread-1] INFO org.apache.hadoop.mapred.LocalJobRunner - Starting task: attempt_local1354877940_0003_r_000000_0
2016-12-19 20:50:43,287 [pool-8-thread-1] INFO org.apache.hadoop.mapred.Task - Using ResourceCalculatorProcessTree : [ ]
2016-12-19 20:50:43,298 [pool-8-thread-1] INFO org.apache.hadoop.mapred.ReduceTask - Using ShuffleConsumerPlugin: org.apache.hadoop.mapreduce.task.reduce.Shuffle@15517d3
2016-12-19 20:50:43,333 [pool-8-thread-1] INFO org.apache.hadoop.mapreduce.task.reduce.MergeManagerImpl - MergerManager: memoryLimit=709551680, maxSingleShuffleLimit=177387920, mergeThreshold=468304128, ioSortFactor=10, memToMemMergeOutputsThreshold=10
2016-12-19 20:50:43,355 [EventFetcher for fetching Map Completion Events] INFO org.apache.hadoop.mapreduce.task.reduce.EventFetcher - attempt_local1354877940_0003_r_000000_0 Thread started: EventFetcher for fetching Map Completion Events
2016-12-19 20:50:43,390 [localfetcher#2] INFO org.apache.hadoop.mapreduce.task.reduce.LocalFetcher - localfetcher#2 about to shuffle output of map attempt_local1354877940_0003_m_000000_0 decomp: 110 len: 114 to MEMORY
2016-12-19 20:50:43,415 [localfetcher#2] INFO org.apache.hadoop.mapreduce.task.reduce.InMemoryMapOutput - Read 110 bytes from map-output for attempt_local1354877940_0003_m_000000_0
2016-12-19 20:50:43,422 [localfetcher#2] INFO org.apache.hadoop.mapreduce.task.reduce.MergeManagerImpl - closeInMemoryFile -> map-output of size: 110, inMemoryMapOutputs.size() -> 1, commitMemory -> 0, usedMemory ->110
2016-12-19 20:50:43,476 [EventFetcher for fetching Map Completion Events] INFO org.apache.hadoop.mapreduce.task.reduce.EventFetcher - EventFetcher is interrupted.. Returning
2016-12-19 20:50:43,508 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 50% complete
2016-12-19 20:50:43,514 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Running jobs are [job_local1354877940_0003]
2016-12-19 20:50:43,516 [pool-8-thread-1] INFO org.apache.hadoop.mapred.LocalJobRunner - 1 / 1 copied.
2016-12-19 20:50:43,535 [pool-8-thread-1] INFO org.apache.hadoop.mapreduce.task.reduce.MergeManagerImpl - finalMerge called with 1 in-memory map-outputs and 0 on-disk map-outputs
2016-12-19 20:50:43,575 [pool-8-thread-1] INFO org.apache.hadoop.mapred.Merger - Merging 1 sorted segments
2016-12-19 20:50:43,598 [pool-8-thread-1] INFO org.apache.hadoop.mapred.Merger - Down to the last merge-pass, with 1 segments left of total size: 103 bytes
2016-12-19 20:50:43,614 [pool-8-thread-1] INFO org.apache.hadoop.mapreduce.task.reduce.MergeManagerImpl - Merged 1 segments, 110 bytes to disk to satisfy reduce memory limit
2016-12-19 20:50:43,615 [pool-8-thread-1] INFO org.apache.hadoop.mapreduce.task.reduce.MergeManagerImpl - Merging 1 files, 114 bytes from disk
2016-12-19 20:50:43,630 [pool-8-thread-1] INFO org.apache.hadoop.mapreduce.task.reduce.MergeManagerImpl - Merging 0 segments, 0 bytes from memory into reduce
2016-12-19 20:50:43,630 [pool-8-thread-1] INFO org.apache.hadoop.mapred.Merger - Merging 1 sorted segments
2016-12-19 20:50:43,631 [pool-8-thread-1] INFO org.apache.hadoop.mapred.Merger - Down to the last merge-pass, with 1 segments left of total size: 103 bytes
2016-12-19 20:50:43,635 [pool-8-thread-1] INFO org.apache.hadoop.mapred.LocalJobRunner - 1 / 1 copied.
2016-12-19 20:50:43,719 [Thread-50] INFO org.apache.hadoop.mapred.LocalJobRunner - reduce task executor complete.
2016-12-19 20:50:43,759 [Thread-50] WARN org.apache.hadoop.mapred.LocalJobRunner - job_local1354877940_0003
java.lang.Exception: java.io.IOException: Mkdirs failed to create file:/home/soukaina/Bureau/results/_temporary/0/_temporary/attempt_local1354877940_0003_r_000000_0 (exists=false, cwd=file:/home/soukaina/Bureau)
at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462)
at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529)
Caused by: java.io.IOException: Mkdirs failed to create file:/home/soukaina/Bureau/results/_temporary/0/_temporary/attempt_local1354877940_0003_r_000000_0 (exists=false, cwd=file:/home/soukaina/Bureau)
at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:442)
at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:428)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:908)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:889)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:786)
at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigTextOutputFormat.getRecordWriter(PigTextOutputFormat.java:98)
at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.getRecordWriter(PigOutputFormat.java:81)
at org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540)
at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
at org.apache.hadoop.mapred.LocalJobRunner$Job$ReduceTaskRunnable.run(LocalJobRunner.java:319)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
2016-12-19 20:50:43,922 [main] WARN org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Ooops! Some job has failed! Specify -stop_on_failure if you want Pig to stop immediately on failure.
2016-12-19 20:50:43,986 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - job job_local1354877940_0003 has failed! Stop running all dependent jobs
2016-12-19 20:50:43,988 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 100% complete
2016-12-19 20:50:44,081 [main] INFO org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics with processName=JobTracker, sessionId= - already initialized
2016-12-19 20:50:44,096 [main] INFO org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics with processName=JobTracker, sessionId= - already initialized
2016-12-19 20:50:44,102 [main] ERROR org.apache.pig.tools.pigstats.mapreduce.MRPigStatsUtil - 1 map reduce job(s) failed!
2016-12-19 20:50:44,104 [main] INFO org.apache.pig.tools.pigstats.mapreduce.SimplePigStats - Script Statistics:
HadoopVersion PigVersion UserId StartedAt FinishedAt Features
2.6.0 0.15.0 hduser 2016-12-19 20:50:39 2016-12-19 20:50:44 GROUP_BY
Failed!
Failed Jobs:
JobId Alias Feature Message Outputs
job_local1354877940_0003 A,B,C,D,E GROUP_BY,COMBINER Message: Job failed! file:///home/soukaina/Bureau/results,
Input(s):
Failed to read data from "file:///home/soukaina/Bureau/hdfile"
Output(s):
Failed to produce result in "file:///home/soukaina/Bureau/results"
Counters:
Total records written : 0
Total bytes written : 0
Spillable Memory Manager spill count : 0
Total bags proactively spilled: 0
Total records proactively spilled: 0
Job DAG:
job_local1354877940_0003
2016-12-19 20:50:44,132 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Failed!
2016-12-19 20:50:44,252 [main] ERROR org.apache.pig.tools.grunt.GruntParser - ERROR 2244: Job job_local1354877940_0003 failed, hadoop does not return any error message
2016-12-19 20:50:44,269 [main] WARN org.apache.pig.tools.grunt.GruntParser - There is no log file to write to.
2016-12-19 20:50:44,270 [main] ERROR org.apache.pig.tools.grunt.GruntParser - org.apache.pig.backend.executionengine.ExecException: ERROR 2244: Job job_local1354877940_0003 failed, hadoop does not return any error message
at org.apache.pig.tools.grunt.GruntParser.executeBatch(GruntParser.java:179)
at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:234)
at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:205)
at org.apache.pig.tools.grunt.Grunt.exec(Grunt.java:81)
at org.apache.pig.Main.run(Main.java:631)
at org.apache.pig.Main.main(Main.java:177)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
2016-12-19 20:50:44,604 [main] INFO org.apache.pig.Main - Pig script completed in 1 minute, 1 second and 532 milliseconds (61532 ms)
2016-12-19 20:50:49,372 [communication thread] INFO org.apache.hadoop.mapred.LocalJobRunner - reduce > reduce
19 déc. 2016 à 21:58