• 设为首页
  • 点击收藏
  • 手机版
    手机扫一扫访问
    迪恩网络手机版
  • 关注官方公众号
    微信扫一扫关注
    迪恩网络公众号

Java StreamingHistogram类代码示例

原作者: [db:作者] 来自: [db:来源] 收藏 邀请

本文整理汇总了Java中org.apache.cassandra.utils.StreamingHistogram的典型用法代码示例。如果您正苦于以下问题:Java StreamingHistogram类的具体用法?Java StreamingHistogram怎么用?Java StreamingHistogram使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。



StreamingHistogram类属于org.apache.cassandra.utils包,在下文中一共展示了StreamingHistogram类的20个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Java代码示例。

示例1: ColumnStats

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public ColumnStats(int columnCount,
                   long minTimestamp,
                   long maxTimestamp,
                   int maxLocalDeletionTime,
                   StreamingHistogram tombstoneHistogram,
                   List<ByteBuffer> minColumnNames,
                   List<ByteBuffer> maxColumnNames,
                   boolean hasLegacyCounterShards)
{
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.maxLocalDeletionTime = maxLocalDeletionTime;
    this.columnCount = columnCount;
    this.tombstoneHistogram = tombstoneHistogram;
    this.minColumnNames = minColumnNames;
    this.maxColumnNames = maxColumnNames;
    this.hasLegacyCounterShards = hasLegacyCounterShards;
}
 
开发者ID:vcostet,项目名称:cassandra-kmean,代码行数:19,代码来源:ColumnStats.java


示例2: legacySerialize

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
/**
 * Used to serialize to an old version - needed to be able to update sstable level without a full compaction.
 *
 * @deprecated will be removed when it is assumed that the minimum upgrade-from-version is the version that this
 * patch made it into
 *
 * @param sstableStats
 * @param legacyDesc
 * @param out
 * @throws IOException
 */
@Deprecated
public void legacySerialize(SSTableMetadata sstableStats, Set<Integer> ancestors, Descriptor legacyDesc, DataOutput out) throws IOException
{
    EstimatedHistogram.serializer.serialize(sstableStats.estimatedRowSize, out);
    EstimatedHistogram.serializer.serialize(sstableStats.estimatedColumnCount, out);
    ReplayPosition.serializer.serialize(sstableStats.replayPosition, out);
    out.writeLong(sstableStats.minTimestamp);
    out.writeLong(sstableStats.maxTimestamp);
    if (legacyDesc.version.tracksMaxLocalDeletionTime)
        out.writeInt(sstableStats.maxLocalDeletionTime);
    if (legacyDesc.version.hasBloomFilterFPChance)
        out.writeDouble(sstableStats.bloomFilterFPChance);
    out.writeDouble(sstableStats.compressionRatio);
    out.writeUTF(sstableStats.partitioner);
    out.writeInt(ancestors.size());
    for (Integer g : ancestors)
        out.writeInt(g);
    StreamingHistogram.serializer.serialize(sstableStats.estimatedTombstoneDropTime, out);
    out.writeInt(sstableStats.sstableLevel);
    if (legacyDesc.version.tracksMaxMinColumnNames)
        serializeMinMaxColumnNames(sstableStats.minColumnNames, sstableStats.maxColumnNames, out);
}
 
开发者ID:pgaref,项目名称:ACaZoo,代码行数:34,代码来源:SSTableMetadata.java


示例3: SSTableMetadata

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
private SSTableMetadata(EstimatedHistogram rowSizes,
                        EstimatedHistogram columnCounts,
                        ReplayPosition replayPosition,
                        long minTimestamp,
                        long maxTimestamp,
                        double cr,
                        String partitioner,
                        StreamingHistogram estimatedTombstoneDropTime)
{
    this.estimatedRowSize = rowSizes;
    this.estimatedColumnCount = columnCounts;
    this.replayPosition = replayPosition;
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.compressionRatio = cr;
    this.partitioner = partitioner;
    this.estimatedTombstoneDropTime = estimatedTombstoneDropTime;
}
 
开发者ID:dprguiuc,项目名称:Cassandra-Wasef,代码行数:19,代码来源:SSTableMetadata.java


示例4: StatsMetadata

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public StatsMetadata(EstimatedHistogram estimatedRowSize,
                     EstimatedHistogram estimatedColumnCount,
                     ReplayPosition replayPosition,
                     long minTimestamp,
                     long maxTimestamp,
                     int maxLocalDeletionTime,
                     double compressionRatio,
                     StreamingHistogram estimatedTombstoneDropTime,
                     int sstableLevel,
                     List<ByteBuffer> minColumnNames,
                     List<ByteBuffer> maxColumnNames)
{
    this.estimatedRowSize = estimatedRowSize;
    this.estimatedColumnCount = estimatedColumnCount;
    this.replayPosition = replayPosition;
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.maxLocalDeletionTime = maxLocalDeletionTime;
    this.compressionRatio = compressionRatio;
    this.estimatedTombstoneDropTime = estimatedTombstoneDropTime;
    this.sstableLevel = sstableLevel;
    this.minColumnNames = minColumnNames;
    this.maxColumnNames = maxColumnNames;
}
 
开发者ID:mafernandez-stratio,项目名称:cassandra-cqlMod,代码行数:25,代码来源:StatsMetadata.java


示例5: StatsMetadata

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public StatsMetadata(EstimatedHistogram estimatedPartitionSize,
                     EstimatedHistogram estimatedColumnCount,
                     IntervalSet<CommitLogPosition> commitLogIntervals,
                     long minTimestamp,
                     long maxTimestamp,
                     int minLocalDeletionTime,
                     int maxLocalDeletionTime,
                     int minTTL,
                     int maxTTL,
                     double compressionRatio,
                     StreamingHistogram estimatedTombstoneDropTime,
                     int sstableLevel,
                     List<ByteBuffer> minClusteringValues,
                     List<ByteBuffer> maxClusteringValues,
                     boolean hasLegacyCounterShards,
                     long repairedAt,
                     long totalColumnsSet,
                     long totalRows)
{
    this.estimatedPartitionSize = estimatedPartitionSize;
    this.estimatedColumnCount = estimatedColumnCount;
    this.commitLogIntervals = commitLogIntervals;
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.minLocalDeletionTime = minLocalDeletionTime;
    this.maxLocalDeletionTime = maxLocalDeletionTime;
    this.minTTL = minTTL;
    this.maxTTL = maxTTL;
    this.compressionRatio = compressionRatio;
    this.estimatedTombstoneDropTime = estimatedTombstoneDropTime;
    this.sstableLevel = sstableLevel;
    this.minClusteringValues = minClusteringValues;
    this.maxClusteringValues = maxClusteringValues;
    this.hasLegacyCounterShards = hasLegacyCounterShards;
    this.repairedAt = repairedAt;
    this.totalColumnsSet = totalColumnsSet;
    this.totalRows = totalRows;
}
 
开发者ID:Netflix,项目名称:sstable-adaptor,代码行数:39,代码来源:StatsMetadata.java


示例6: StatsMetadata

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public StatsMetadata(EstimatedHistogram estimatedRowSize,
                     EstimatedHistogram estimatedColumnCount,
                     ReplayPosition replayPosition,
                     long minTimestamp,
                     long maxTimestamp,
                     int maxLocalDeletionTime,
                     double compressionRatio,
                     StreamingHistogram estimatedTombstoneDropTime,
                     int sstableLevel,
                     List<ByteBuffer> minColumnNames,
                     List<ByteBuffer> maxColumnNames,
                     boolean hasLegacyCounterShards,
                     long repairedAt)
{
    this.estimatedRowSize = estimatedRowSize;
    this.estimatedColumnCount = estimatedColumnCount;
    this.replayPosition = replayPosition;
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.maxLocalDeletionTime = maxLocalDeletionTime;
    this.compressionRatio = compressionRatio;
    this.estimatedTombstoneDropTime = estimatedTombstoneDropTime;
    this.sstableLevel = sstableLevel;
    this.minColumnNames = minColumnNames;
    this.maxColumnNames = maxColumnNames;
    this.hasLegacyCounterShards = hasLegacyCounterShards;
    this.repairedAt = repairedAt;
}
 
开发者ID:vcostet,项目名称:cassandra-kmean,代码行数:29,代码来源:StatsMetadata.java


示例7: SSTableMetadata

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
private SSTableMetadata(EstimatedHistogram rowSizes,
                        EstimatedHistogram columnCounts,
                        ReplayPosition replayPosition,
                        long minTimestamp,
                        long maxTimestamp,
                        int maxLocalDeletionTime,
                        double bloomFilterFPChance,
                        double compressionRatio,
                        String partitioner,
                        StreamingHistogram estimatedTombstoneDropTime,
                        int sstableLevel,
                        List<ByteBuffer> minColumnNames,
                        List<ByteBuffer> maxColumnNames)
{
    this.estimatedRowSize = rowSizes;
    this.estimatedColumnCount = columnCounts;
    this.replayPosition = replayPosition;
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.maxLocalDeletionTime = maxLocalDeletionTime;
    this.bloomFilterFPChance = bloomFilterFPChance;
    this.compressionRatio = compressionRatio;
    this.partitioner = partitioner;
    this.estimatedTombstoneDropTime = estimatedTombstoneDropTime;
    this.sstableLevel = sstableLevel;
    this.minColumnNames = minColumnNames;
    this.maxColumnNames = maxColumnNames;
}
 
开发者ID:pgaref,项目名称:ACaZoo,代码行数:29,代码来源:SSTableMetadata.java


示例8: ColumnStats

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public ColumnStats(int columnCount, long minTimestamp, long maxTimestamp, int maxLocalDeletionTime, StreamingHistogram tombstoneHistogram, List<ByteBuffer> minColumnNames, List<ByteBuffer> maxColumnNames)
{
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.maxLocalDeletionTime = maxLocalDeletionTime;
    this.columnCount = columnCount;
    this.tombstoneHistogram = tombstoneHistogram;
    this.minColumnNames = minColumnNames;
    this.maxColumnNames = maxColumnNames;
}
 
开发者ID:pgaref,项目名称:ACaZoo,代码行数:11,代码来源:ColumnStats.java


示例9: StatsMetadata

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public StatsMetadata(EstimatedHistogram estimatedPartitionSize,
                     EstimatedHistogram estimatedColumnCount,
                     IntervalSet<ReplayPosition> commitLogIntervals,
                     long minTimestamp,
                     long maxTimestamp,
                     int minLocalDeletionTime,
                     int maxLocalDeletionTime,
                     int minTTL,
                     int maxTTL,
                     double compressionRatio,
                     StreamingHistogram estimatedTombstoneDropTime,
                     int sstableLevel,
                     List<ByteBuffer> minClusteringValues,
                     List<ByteBuffer> maxClusteringValues,
                     boolean hasLegacyCounterShards,
                     long repairedAt,
                     long totalColumnsSet,
                     long totalRows)
{
    this.estimatedPartitionSize = estimatedPartitionSize;
    this.estimatedColumnCount = estimatedColumnCount;
    this.commitLogIntervals = commitLogIntervals;
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.minLocalDeletionTime = minLocalDeletionTime;
    this.maxLocalDeletionTime = maxLocalDeletionTime;
    this.minTTL = minTTL;
    this.maxTTL = maxTTL;
    this.compressionRatio = compressionRatio;
    this.estimatedTombstoneDropTime = estimatedTombstoneDropTime;
    this.sstableLevel = sstableLevel;
    this.minClusteringValues = minClusteringValues;
    this.maxClusteringValues = maxClusteringValues;
    this.hasLegacyCounterShards = hasLegacyCounterShards;
    this.repairedAt = repairedAt;
    this.totalColumnsSet = totalColumnsSet;
    this.totalRows = totalRows;
}
 
开发者ID:scylladb,项目名称:scylla-tools-java,代码行数:39,代码来源:StatsMetadata.java


示例10: LazilyCompactedRow

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public LazilyCompactedRow(CompactionController controller, List<? extends ICountableColumnIterator> rows)
{
    super(rows.get(0).getKey());
    this.rows = rows;
    this.controller = controller;
    indexer = controller.cfs.indexManager.updaterFor(key);

    long maxDelTimestamp = Long.MIN_VALUE;
    for (OnDiskAtomIterator row : rows)
    {
        ColumnFamily cf = row.getColumnFamily();
        maxDelTimestamp = Math.max(maxDelTimestamp, cf.deletionInfo().maxTimestamp());

        if (emptyColumnFamily == null)
            emptyColumnFamily = cf;
        else
            emptyColumnFamily.delete(cf);
    }
    this.shouldPurge = controller.shouldPurge(key, maxDelTimestamp);

    try
    {
        indexAndWrite(null);
    }
    catch (IOException e)
    {
        throw new RuntimeException(e);
    }
    // reach into the reducer used during iteration to get column count, size, max column timestamp
    // (however, if there are zero columns, iterator() will not be called by ColumnIndexer and reducer will be null)
    columnStats = new ColumnStats(reducer == null ? 0 : reducer.columns, 
                                  reducer == null ? Long.MAX_VALUE : reducer.minTimestampSeen, 
                                  reducer == null ? maxDelTimestamp : Math.max(maxDelTimestamp, reducer.maxTimestampSeen),
                                  reducer == null ? new StreamingHistogram(SSTable.TOMBSTONE_HISTOGRAM_BIN_SIZE) : reducer.tombstones
    );
    columnSerializedSize = reducer == null ? 0 : reducer.serializedSize;
    reducer = null;
}
 
开发者ID:dprguiuc,项目名称:Cassandra-Wasef,代码行数:39,代码来源:LazilyCompactedRow.java


示例11: deserialize

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public Pair<SSTableMetadata, Set<Integer>> deserialize(DataInputStream dis, Descriptor desc) throws IOException
{
    EstimatedHistogram rowSizes = EstimatedHistogram.serializer.deserialize(dis);
    EstimatedHistogram columnCounts = EstimatedHistogram.serializer.deserialize(dis);
    ReplayPosition replayPosition = desc.version.metadataIncludesReplayPosition
                                  ? ReplayPosition.serializer.deserialize(dis)
                                  : ReplayPosition.NONE;
    if (!desc.version.metadataIncludesModernReplayPosition)
    {
        // replay position may be "from the future" thanks to older versions generating them with nanotime.
        // make sure we don't omit replaying something that we should.  see CASSANDRA-4782
        replayPosition = ReplayPosition.NONE;
    }
    long minTimestamp = desc.version.tracksMinTimestamp ? dis.readLong() : Long.MIN_VALUE;
    long maxTimestamp = desc.version.containsTimestamp() ? dis.readLong() : Long.MAX_VALUE;
    if (!desc.version.tracksMaxTimestamp) // see javadoc to Descriptor.containsTimestamp
        maxTimestamp = Long.MAX_VALUE;
    double compressionRatio = desc.version.hasCompressionRatio
                            ? dis.readDouble()
                            : NO_COMPRESSION_RATIO;
    String partitioner = desc.version.hasPartitioner ? dis.readUTF() : null;
    int nbAncestors = desc.version.hasAncestors ? dis.readInt() : 0;
    Set<Integer> ancestors = new HashSet<Integer>(nbAncestors);
    for (int i = 0; i < nbAncestors; i++)
        ancestors.add(dis.readInt());
    StreamingHistogram tombstoneHistogram = desc.version.tracksTombstones
                                           ? StreamingHistogram.serializer.deserialize(dis)
                                           : defaultTombstoneDropTimeHistogram();
    return Pair.create(new SSTableMetadata(rowSizes,
                                           columnCounts,
                                           replayPosition,
                                           minTimestamp,
                                           maxTimestamp,
                                           compressionRatio,
                                           partitioner,
                                           tombstoneHistogram),
                                           ancestors);
}
 
开发者ID:dprguiuc,项目名称:Cassandra-Wasef,代码行数:39,代码来源:SSTableMetadata.java


示例12: ColumnStats

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public ColumnStats(int columnCount, long minTimestamp, long maxTimestamp, StreamingHistogram tombstoneHistogram)
{
    this.minTimestamp = minTimestamp;
    this.maxTimestamp = maxTimestamp;
    this.columnCount = columnCount;
    this.tombstoneHistogram = tombstoneHistogram;
}
 
开发者ID:dprguiuc,项目名称:Cassandra-Wasef,代码行数:8,代码来源:ColumnStats.java


示例13: defaultTombstoneDropTimeHistogram

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
static StreamingHistogram defaultTombstoneDropTimeHistogram()
{
    return new StreamingHistogram(SSTable.TOMBSTONE_HISTOGRAM_BIN_SIZE);
}
 
开发者ID:Netflix,项目名称:sstable-adaptor,代码行数:5,代码来源:MetadataCollector.java


示例14: mergeTombstoneHistogram

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public MetadataCollector mergeTombstoneHistogram(StreamingHistogram histogram)
{
    estimatedTombstoneDropTime.merge(histogram);
    return this;
}
 
开发者ID:Netflix,项目名称:sstable-adaptor,代码行数:6,代码来源:MetadataCollector.java


示例15: deserialize

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public StatsMetadata deserialize(Version version, DataInputPlus in) throws IOException
{
    EstimatedHistogram partitionSizes = EstimatedHistogram.serializer.deserialize(in);
    EstimatedHistogram columnCounts = EstimatedHistogram.serializer.deserialize(in);
    CommitLogPosition commitLogLowerBound = CommitLogPosition.NONE, commitLogUpperBound;
    commitLogUpperBound = CommitLogPosition.serializer.deserialize(in);
    long minTimestamp = in.readLong();
    long maxTimestamp = in.readLong();
    // We use MAX_VALUE as that's the default value for "no deletion time"
    int minLocalDeletionTime = version.storeRows() ? in.readInt() : Integer.MAX_VALUE;
    int maxLocalDeletionTime = in.readInt();
    int minTTL = version.storeRows() ? in.readInt() : 0;
    int maxTTL = version.storeRows() ? in.readInt() : Integer.MAX_VALUE;
    double compressionRatio = in.readDouble();
    StreamingHistogram tombstoneHistogram = StreamingHistogram.serializer.deserialize(in);
    int sstableLevel = in.readInt();
    long repairedAt = 0;
    if (version.hasRepairedAt())
        repairedAt = in.readLong();

    int colCount = in.readInt();
    List<ByteBuffer> minClusteringValues = new ArrayList<>(colCount);
    for (int i = 0; i < colCount; i++)
        minClusteringValues.add(ByteBufferUtil.readWithShortLength(in));

    colCount = in.readInt();
    List<ByteBuffer> maxClusteringValues = new ArrayList<>(colCount);
    for (int i = 0; i < colCount; i++)
        maxClusteringValues.add(ByteBufferUtil.readWithShortLength(in));

    boolean hasLegacyCounterShards = true;
    if (version.tracksLegacyCounterShards())
        hasLegacyCounterShards = in.readBoolean();

    long totalColumnsSet = version.storeRows() ? in.readLong() : -1L;
    long totalRows = version.storeRows() ? in.readLong() : -1L;

    if (version.hasCommitLogLowerBound())
        commitLogLowerBound = CommitLogPosition.serializer.deserialize(in);
    IntervalSet<CommitLogPosition> commitLogIntervals;
    if (version.hasCommitLogIntervals())
        commitLogIntervals = commitLogPositionSetSerializer.deserialize(in);
    else
        commitLogIntervals = new IntervalSet<CommitLogPosition>(commitLogLowerBound, commitLogUpperBound);

    return new StatsMetadata(partitionSizes,
                             columnCounts,
                             commitLogIntervals,
                             minTimestamp,
                             maxTimestamp,
                             minLocalDeletionTime,
                             maxLocalDeletionTime,
                             minTTL,
                             maxTTL,
                             compressionRatio,
                             tombstoneHistogram,
                             sstableLevel,
                             minClusteringValues,
                             maxClusteringValues,
                             hasLegacyCounterShards,
                             repairedAt,
                             totalColumnsSet,
                             totalRows);
}
 
开发者ID:Netflix,项目名称:sstable-adaptor,代码行数:65,代码来源:StatsMetadata.java


示例16: deserialize

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
/**
 * Legacy serializer deserialize all components no matter what types are specified.
 */
@Override
public Map<MetadataType, MetadataComponent> deserialize(Descriptor descriptor, EnumSet<MetadataType> types) throws IOException
{
    Map<MetadataType, MetadataComponent> components = Maps.newHashMap();

    File statsFile = new File(descriptor.filenameFor(Component.STATS));
    if (!statsFile.exists() && types.contains(MetadataType.STATS))
    {
        components.put(MetadataType.STATS, MetadataCollector.defaultStatsMetadata());
    }
    else
    {
        try (DataInputStream in = new DataInputStream(new BufferedInputStream(new FileInputStream(statsFile))))
        {
            EstimatedHistogram rowSizes = EstimatedHistogram.serializer.deserialize(in);
            EstimatedHistogram columnCounts = EstimatedHistogram.serializer.deserialize(in);
            ReplayPosition replayPosition = ReplayPosition.serializer.deserialize(in);
            long minTimestamp = in.readLong();
            long maxTimestamp = in.readLong();
            int maxLocalDeletionTime = in.readInt();
            double bloomFilterFPChance = in.readDouble();
            double compressionRatio = in.readDouble();
            String partitioner = in.readUTF();
            int nbAncestors = in.readInt();
            Set<Integer> ancestors = new HashSet<>(nbAncestors);
            for (int i = 0; i < nbAncestors; i++)
                ancestors.add(in.readInt());
            StreamingHistogram tombstoneHistogram = StreamingHistogram.serializer.deserialize(in);
            int sstableLevel = 0;
            if (in.available() > 0)
                sstableLevel = in.readInt();

            int colCount = in.readInt();
            List<ByteBuffer> minColumnNames = new ArrayList<>(colCount);
            for (int i = 0; i < colCount; i++)
                minColumnNames.add(ByteBufferUtil.readWithShortLength(in));

            colCount = in.readInt();
            List<ByteBuffer> maxColumnNames = new ArrayList<>(colCount);
            for (int i = 0; i < colCount; i++)
                maxColumnNames.add(ByteBufferUtil.readWithShortLength(in));

            if (types.contains(MetadataType.VALIDATION))
                components.put(MetadataType.VALIDATION,
                               new ValidationMetadata(partitioner, bloomFilterFPChance));
            if (types.contains(MetadataType.STATS))
                components.put(MetadataType.STATS,
                               new StatsMetadata(rowSizes,
                                                 columnCounts,
                                                 replayPosition,
                                                 minTimestamp,
                                                 maxTimestamp,
                                                 maxLocalDeletionTime,
                                                 compressionRatio,
                                                 tombstoneHistogram,
                                                 sstableLevel,
                                                 minColumnNames,
                                                 maxColumnNames,
                                                 true,
                                                 ActiveRepairService.UNREPAIRED_SSTABLE));
            if (types.contains(MetadataType.COMPACTION))
                components.put(MetadataType.COMPACTION,
                               new CompactionMetadata(ancestors, null));
        }
    }
    return components;
}
 
开发者ID:vcostet,项目名称:cassandra-kmean,代码行数:71,代码来源:LegacyMetadataSerializer.java


示例17: deserialize

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public StatsMetadata deserialize(Descriptor.Version version, DataInput in) throws IOException
{
    EstimatedHistogram rowSizes = EstimatedHistogram.serializer.deserialize(in);
    EstimatedHistogram columnCounts = EstimatedHistogram.serializer.deserialize(in);
    ReplayPosition replayPosition = ReplayPosition.serializer.deserialize(in);
    long minTimestamp = in.readLong();
    long maxTimestamp = in.readLong();
    int maxLocalDeletionTime = in.readInt();
    double compressionRatio = in.readDouble();
    StreamingHistogram tombstoneHistogram = StreamingHistogram.serializer.deserialize(in);
    int sstableLevel = in.readInt();
    long repairedAt = 0;
    if (version.hasRepairedAt)
        repairedAt = in.readLong();

    int colCount = in.readInt();
    List<ByteBuffer> minColumnNames = new ArrayList<>(colCount);
    for (int i = 0; i < colCount; i++)
        minColumnNames.add(ByteBufferUtil.readWithShortLength(in));

    colCount = in.readInt();
    List<ByteBuffer> maxColumnNames = new ArrayList<>(colCount);
    for (int i = 0; i < colCount; i++)
        maxColumnNames.add(ByteBufferUtil.readWithShortLength(in));

    boolean hasLegacyCounterShards = true;
    if (version.tracksLegacyCounterShards)
        hasLegacyCounterShards = in.readBoolean();

    return new StatsMetadata(rowSizes,
                             columnCounts,
                             replayPosition,
                             minTimestamp,
                             maxTimestamp,
                             maxLocalDeletionTime,
                             compressionRatio,
                             tombstoneHistogram,
                             sstableLevel,
                             minColumnNames,
                             maxColumnNames,
                             hasLegacyCounterShards,
                             repairedAt);
}
 
开发者ID:vcostet,项目名称:cassandra-kmean,代码行数:44,代码来源:StatsMetadata.java


示例18: appendFromStream

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
/**
 * @throws IOException if a read from the DataInput fails
 * @throws FSWriteError if a write to the dataFile fails
 */
public long appendFromStream(DecoratedKey key, CFMetaData metadata, DataInput in, Descriptor.Version version) throws IOException
{
    long currentPosition = beforeAppend(key);

    // deserialize each column to obtain maxTimestamp and immediately serialize it.
    long minTimestamp = Long.MAX_VALUE;
    long maxTimestamp = Long.MIN_VALUE;
    int maxLocalDeletionTime = Integer.MIN_VALUE;
    List<ByteBuffer> minColumnNames = Collections.emptyList();
    List<ByteBuffer> maxColumnNames = Collections.emptyList();
    StreamingHistogram tombstones = new StreamingHistogram(TOMBSTONE_HISTOGRAM_BIN_SIZE);
    ColumnFamily cf = ArrayBackedSortedColumns.factory.create(metadata);

    cf.delete(DeletionTime.serializer.deserialize(in));

    ColumnIndex.Builder columnIndexer = new ColumnIndex.Builder(cf, key.key, dataFile.stream);
    int columnCount = Integer.MAX_VALUE;
    if (version.hasRowSizeAndColumnCount)
    {
        // skip row size
        FileUtils.skipBytesFully(in, 8);
        columnCount = in.readInt();
    }
    Iterator<OnDiskAtom> iter = metadata.getOnDiskIterator(in, columnCount, ColumnSerializer.Flag.PRESERVE_SIZE, Integer.MIN_VALUE, version);
    try
    {
        while (iter.hasNext())
        {
            // deserialize column with PRESERVE_SIZE because we've written the dataSize based on the
            // data size received, so we must reserialize the exact same data
            OnDiskAtom atom = iter.next();
            if (atom == null)
                break;
            if (atom instanceof CounterColumn)
                atom = ((CounterColumn) atom).markDeltaToBeCleared();

            int deletionTime = atom.getLocalDeletionTime();
            if (deletionTime < Integer.MAX_VALUE)
            {
                tombstones.update(deletionTime);
            }
            minTimestamp = Math.min(minTimestamp, atom.minTimestamp());
            maxTimestamp = Math.max(maxTimestamp, atom.maxTimestamp());
            minColumnNames = ColumnNameHelper.minComponents(minColumnNames, atom.name(), metadata.comparator);
            maxColumnNames = ColumnNameHelper.maxComponents(maxColumnNames, atom.name(), metadata.comparator);
            maxLocalDeletionTime = Math.max(maxLocalDeletionTime, atom.getLocalDeletionTime());

            columnIndexer.add(atom); // This write the atom on disk too
        }

        columnIndexer.finish();
        dataFile.stream.writeShort(END_OF_ROW);
    }
    catch (IOException e)
    {
        throw new FSWriteError(e, dataFile.getPath());
    }

    sstableMetadataCollector.updateMinTimestamp(minTimestamp);
    sstableMetadataCollector.updateMaxTimestamp(maxTimestamp);
    sstableMetadataCollector.updateMaxLocalDeletionTime(maxLocalDeletionTime);
    sstableMetadataCollector.addRowSize(dataFile.getFilePointer() - currentPosition);
    sstableMetadataCollector.addColumnCount(columnIndexer.writtenAtomCount());
    sstableMetadataCollector.mergeTombstoneHistogram(tombstones);
    sstableMetadataCollector.updateMinColumnNames(minColumnNames);
    sstableMetadataCollector.updateMaxColumnNames(maxColumnNames);
    afterAppend(key, currentPosition, RowIndexEntry.create(currentPosition, cf.deletionInfo().getTopLevelDeletion(), columnIndexer.build()));
    return currentPosition;
}
 
开发者ID:pgaref,项目名称:ACaZoo,代码行数:74,代码来源:SSTableWriter.java


示例19: mergeTombstoneHistogram

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public void mergeTombstoneHistogram(StreamingHistogram histogram)
{
    estimatedTombstoneDropTime.merge(histogram);
}
 
开发者ID:pgaref,项目名称:ACaZoo,代码行数:5,代码来源:SSTableMetadata.java


示例20: deserialize

import org.apache.cassandra.utils.StreamingHistogram; //导入依赖的package包/类
public Pair<SSTableMetadata, Set<Integer>> deserialize(DataInputStream in, Descriptor desc, boolean loadSSTableLevel) throws IOException
{
    EstimatedHistogram rowSizes = EstimatedHistogram.serializer.deserialize(in);
    EstimatedHistogram columnCounts = EstimatedHistogram.serializer.deserialize(in);
    ReplayPosition replayPosition = ReplayPosition.serializer.deserialize(in);
    long minTimestamp = in.readLong();
    long maxTimestamp = in.readLong();
    int maxLocalDeletionTime = desc.version.tracksMaxLocalDeletionTime ? in.readInt() : Integer.MAX_VALUE;
    double bloomFilterFPChance = desc.version.hasBloomFilterFPChance ? in.readDouble() : NO_BLOOM_FLITER_FP_CHANCE;
    double compressionRatio = in.readDouble();
    String partitioner = in.readUTF();
    int nbAncestors = in.readInt();
    Set<Integer> ancestors = new HashSet<Integer>(nbAncestors);
    for (int i = 0; i < nbAncestors; i++)
        ancestors.add(in.readInt());
    StreamingHistogram tombstoneHistogram = StreamingHistogram.serializer.deserialize(in);
    int sstableLevel = 0;

    if (loadSSTableLevel && in.available() > 0)
        sstableLevel = in.readInt();

    List<ByteBuffer> minColumnNames;
    List<ByteBuffer> maxColumnNames;
    if (desc.version.tracksMaxMinColumnNames)
    {
        int colCount = in.readInt();
        minColumnNames = new ArrayList<ByteBuffer>(colCount);
        for (int i = 0; i < colCount; i++)
        {
            minColumnNames.add(ByteBufferUtil.readWithShortLength(in));
        }
        colCount = in.readInt();
        maxColumnNames = new ArrayList<ByteBuffer>(colCount);
        for (int i = 0; i < colCount; i++)
        {
            maxColumnNames.add(ByteBufferUtil.readWithShortLength(in));
        }
    }
    else
    {
        minColumnNames = Collections.emptyList();
        maxColumnNames = Collections.emptyList();
    }
    return Pair.create(new SSTableMetadata(rowSizes,
                               columnCounts,
                               replayPosition,
                               minTimestamp,
                               maxTimestamp,
                               maxLocalDeletionTime,
                               bloomFilterFPChance,
                               compressionRatio,
                               partitioner,
                               tombstoneHistogram,
                               sstableLevel,
                               minColumnNames,
                               maxColumnNames), ancestors);
}
 
开发者ID:pgaref,项目名称:ACaZoo,代码行数:58,代码来源:SSTableMetadata.java



注:本文中的org.apache.cassandra.utils.StreamingHistogram类示例整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。


鲜花

握手

雷人

路过

鸡蛋
该文章已有0人参与评论

请发表评论

全部评论

专题导读
上一篇:
Java TaskAttemptReportProtoOrBuilder类代码示例发布时间:2022-05-22
下一篇:
Java SnmpEngineFactory类代码示例发布时间:2022-05-22
热门推荐
阅读排行榜

扫描微信二维码

查看手机版网站

随时了解更新最新资讯

139-2527-9053

在线客服(服务时间 9:00~18:00)

在线QQ客服
地址:深圳市南山区西丽大学城创智工业园
电邮:jeky_zhao#qq.com
移动电话:139-2527-9053

Powered by 互联科技 X3.4© 2001-2213 极客世界.|Sitemap