carbondata-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From chenerlu <...@git.apache.org>
Subject [GitHub] carbondata pull request #877: [CARBONDATA-951] create table ddl can specify ...
Date Thu, 11 May 2017 07:01:49 GMT
Github user chenerlu commented on a diff in the pull request:

    https://github.com/apache/carbondata/pull/877#discussion_r115916800
  
    --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/chunk/impl/BitMapDimensionDataChunk.java
---
    @@ -0,0 +1,171 @@
    +/*
    + * Licensed to the Apache Software Foundation (ASF) under one or more
    + * contributor license agreements.  See the NOTICE file distributed with
    + * this work for additional information regarding copyright ownership.
    + * The ASF licenses this file to You under the Apache License, Version 2.0
    + * (the "License"); you may not use this file except in compliance with
    + * the License.  You may obtain a copy of the License at
    + *
    + *    http://www.apache.org/licenses/LICENSE-2.0
    + *
    + * Unless required by applicable law or agreed to in writing, software
    + * distributed under the License is distributed on an "AS IS" BASIS,
    + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    + * See the License for the specific language governing permissions and
    + * limitations under the License.
    + */
    +package org.apache.carbondata.core.datastore.chunk.impl;
    +
    +import java.util.BitSet;
    +import java.util.List;
    +
    +import org.apache.carbondata.core.datastore.chunk.store.DimensionChunkStoreFactory;
    +import org.apache.carbondata.core.datastore.chunk.store.DimensionChunkStoreFactory.DimensionStoreType;
    +import org.apache.carbondata.core.datastore.chunk.store.impl.safe.SafeBitMapDimensionDataChunkStore;
    +import org.apache.carbondata.core.scan.executor.infos.KeyStructureInfo;
    +import org.apache.carbondata.core.scan.filter.executer.AbstractFilterExecuter.FilterOperator;
    +import org.apache.carbondata.core.scan.result.vector.CarbonColumnVector;
    +import org.apache.carbondata.core.scan.result.vector.ColumnVectorInfo;
    +
    +/**
    + * This class is gives access to fixed length dimension data chunk store
    + */
    +public class BitMapDimensionDataChunk extends AbstractDimensionDataChunk {
    +
    +  // private List<Integer> bitmap_encoded_dictionaries;
    +  // private List<Integer> bitmap_data_pages_length;
    +  /**
    +   * Constructor
    +   *
    +   * @param dataChunk            data chunk
    +   * @param invertedIndex        inverted index
    +   * @param invertedIndexReverse reverse inverted index
    +   * @param numberOfRows         number of rows
    +   * @param columnValueSize      size of each column value
    +   */
    +  public BitMapDimensionDataChunk(byte[] dataChunk, List<Integer> bitmap_encoded_dictionaries,
    +      List<Integer> bitmap_data_pages_length, int numberOfRows, int columnValueSize)
{
    +    // this.bitmap_encoded_dictionaries = bitmap_encoded_dictionaries;
    +    // this.bitmap_data_pages_length = bitmap_data_pages_length;
    +    long totalSize = dataChunk.length;
    +    dataChunkStore = DimensionChunkStoreFactory.INSTANCE.getDimensionChunkStore(columnValueSize,
    +        false, bitmap_encoded_dictionaries.size(), totalSize, DimensionStoreType.BITMAP,
    +        bitmap_encoded_dictionaries, bitmap_data_pages_length);
    +    dataChunkStore.putArray(null, null, dataChunk);
    +  }
    +
    +  /**
    +   * Below method will be used to fill the data based on offset and row id
    +   *
    +   * @param data             data to filed
    +   * @param offset           offset from which data need to be filed
    +   * @param index            row id of the chunk
    +   * @param keyStructureInfo define the structure of the key
    +   * @return how many bytes was copied
    +   */
    +  @Override public int fillChunkData(byte[] data, int offset, int index,
    +      KeyStructureInfo keyStructureInfo) {
    +    dataChunkStore.fillRow(index, data, offset);
    +    return dataChunkStore.getColumnValueSize();
    +  }
    +
    +  /**
    +   * Converts to column dictionary integer value
    +   *
    +   * @param rowId
    +   * @param columnIndex
    +   * @param row
    +   * @param restructuringInfo
    +   * @return
    +   */
    +  @Override public int fillConvertedChunkData(int rowId, int columnIndex, int[] row,
    +      KeyStructureInfo restructuringInfo) {
    +    row[columnIndex] = dataChunkStore.getSurrogate(rowId);
    +    return columnIndex + 1;
    +  }
    +
    +  /**
    +   * Fill the data to vector
    +   *
    +   * @param vectorInfo
    +   * @param column
    +   * @param restructuringInfo
    +   * @return next column index
    +   */
    +  @Override public int fillConvertedChunkData(ColumnVectorInfo[] vectorInfo, int column,
    +      KeyStructureInfo restructuringInfo) {
    +    ColumnVectorInfo columnVectorInfo = vectorInfo[column];
    +    int offset = columnVectorInfo.offset;
    +    int vectorOffset = columnVectorInfo.vectorOffset;
    +    int len = columnVectorInfo.size + offset;
    +    CarbonColumnVector vector = columnVectorInfo.vector;
    +    for (int j = offset; j < len; j++) {
    +      int dict = dataChunkStore.getSurrogate(j);
    +      if (columnVectorInfo.directDictionaryGenerator == null) {
    +        vector.putInt(vectorOffset++, dict);
    +      } else {
    +        Object valueFromSurrogate =
    +            columnVectorInfo.directDictionaryGenerator.getValueFromSurrogate(dict);
    +        if (valueFromSurrogate == null) {
    +          vector.putNull(vectorOffset++);
    +        } else {
    +          switch (columnVectorInfo.directDictionaryGenerator.getReturnType()) {
    +            case INT:
    +              vector.putInt(vectorOffset++, (int) valueFromSurrogate);
    +              break;
    +            case LONG:
    +              vector.putLong(vectorOffset++, (long) valueFromSurrogate);
    +              break;
    +          }
    +        }
    +      }
    +    }
    +    return column + 1;
    +  }
    +
    +  /**
    +   * Fill the data to vector
    +   *
    +   * @param rowMapping
    +   * @param vectorInfo
    +   * @param column
    +   * @param restructuringInfo
    +   * @return next column index
    +   */
    +  @Override public int fillConvertedChunkData(int[] rowMapping, ColumnVectorInfo[] vectorInfo,
    +      int column, KeyStructureInfo restructuringInfo) {
    +    ColumnVectorInfo columnVectorInfo = vectorInfo[column];
    +    int offset = columnVectorInfo.offset;
    +    int vectorOffset = columnVectorInfo.vectorOffset;
    +    int len = columnVectorInfo.size + offset;
    +    CarbonColumnVector vector = columnVectorInfo.vector;
    +    for (int j = offset; j < len; j++) {
    +      int dict = dataChunkStore.getSurrogate(rowMapping[j]);
    +      if (columnVectorInfo.directDictionaryGenerator == null) {
    --- End diff --
    
    This part of code is same with last fillConvertedChunkData, the different is the dict.
So I wonder if it is better to create a new method for this same part ? What do you think
?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

Mime
View raw message