Package org.apache.hadoop.hive.ql.metadata

Examples of org.apache.hadoop.hive.ql.metadata.HiveException


        } else {
          fs.delete(tmpPath);
        }
      }
    } catch (IOException e) {
      throw new HiveException (e);
    }
    super.jobClose(hconf, success);
  }
View Full Code Here


        outStream.write(terminator);
      }
      ((FSDataOutputStream)outStream).close();
    } catch (FileNotFoundException e) {
      LOG.info("show partitions: " + StringUtils.stringifyException(e));
      throw new HiveException(e.toString());
    } catch (IOException e) {
      LOG.info("show partitions: " + StringUtils.stringifyException(e));
      throw new HiveException(e.toString());
    } catch (Exception e) {
      throw new HiveException(e.toString());
    }

    return 0;
  }
View Full Code Here

      return 1;
    } catch (IOException e) {
      LOG.info("show table: " + StringUtils.stringifyException(e));
      return 1;
    } catch (Exception e) {
      throw new HiveException(e.toString());
    }
    return 0;
  }
View Full Code Here

      return 1;
    } catch (IOException e) {
      LOG.info("describe table: " + StringUtils.stringifyException(e));
      return 1;
    } catch (Exception e) {
      throw new HiveException(e.toString());
    }

    return 0;
  }
View Full Code Here

        reducer.startGroup();
      }
      try {
        keyObject = inputKeyDeserializer.deserialize(keyWritable);
      } catch (SerDeException e) {
        throw new HiveException(e);
      }
      // System.err.print(keyObject.toString());
      while (values.hasNext()) {
        Writable valueWritable = (Writable) values.next();
        //System.err.print(who.getHo().toString());
        try {
          valueObject[tag] = inputValueDeserializer[tag].deserialize(valueWritable);
        } catch (SerDeException e) {
          throw new HiveException(e);
        }
        row.clear();
        row.add(keyObject);
        row.add(valueObject[tag]);
        row.add(tag);
View Full Code Here

      String auxJars = HiveConf.getVar(conf, HiveConf.ConfVars.HIVEAUXJARS);
      if (StringUtils.isNotBlank(auxJars)) {
        try {
          addToClassPath(StringUtils.split(auxJars, ","));
        } catch (Exception e) {
          throw new HiveException(e.getMessage(), e);
        }
      }
    }

    mapredWork plan = Utilities.deserializeMapRedWork(pathData);
View Full Code Here

        }
        argumentString.append("} of size " + arguments.length);
      }

      e.printStackTrace();
      throw new HiveException("Unable to execute method " + m + " "
          + " on object " + thisObjectString
          + " with arguments " + argumentString.toString()
          + ":" + e.getMessage());
    }
    return o;
View Full Code Here

          aggregationsAggregateMethods[i] = m;
          break;
        }
      }
      if (null == aggregationsAggregateMethods[i]) {
        throw new HiveException("Cannot find " + aggregateMethodName + " method of UDAF class "
                                 + aggregationClasses[i].getName() + " that accepts "
                                 + aggregationParameterFields[i].length + " parameters!");
      }
      // aggregationsEvaluateMethods
      try {
        aggregationsEvaluateMethods[i] = aggregationClasses[i].getMethod(evaluateMethodName);
      } catch (Exception e) {
        throw new HiveException("Unable to get the method named " + evaluateMethodName + " from "
            + aggregationClasses[i] + ": " + e.getMessage());
      }

      if (null == aggregationsEvaluateMethods[i]) {
        throw new HiveException("Cannot find " + evaluateMethodName + " method of UDAF class "
                                 + aggregationClasses[i].getName() + "!");
      }
      assert(aggregationsEvaluateMethods[i] != null);
    }
View Full Code Here

    for(int i=0; i<aggregationClasses.length; i++) {
      try {
        aggs[i] = aggregationClasses[i].newInstance();
      } catch (Exception e) {
        e.printStackTrace();
        throw new HiveException("Unable to create an instance of class " + aggregationClasses[i] + ": " + e.getMessage());
      }
      aggs[i].init();
    }
    return aggs;
  }
View Full Code Here

      else
        processAggr(row, rowInspector, newKeys);
    } catch (HiveException e) {
      throw e;
    } catch (Exception e) {
      throw new HiveException(e);
    }
  }
View Full Code Here

TOP

Related Classes of org.apache.hadoop.hive.ql.metadata.HiveException

Copyright © 2018 www.massapicom. All rights reserved.
All source code are property of their respective owners. Java is a trademark of Sun Microsystems, Inc and owned by ORACLE Inc. Contact coftware#gmail.com.