2019-05-19 18:01:58 +08:00
|
|
|
package erasure_coding
|
|
|
|
|
2020-08-19 08:04:28 +08:00
|
|
|
import (
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
|
|
|
)
|
|
|
|
|
2019-05-19 18:01:58 +08:00
|
|
|
type Interval struct {
|
2019-05-27 16:29:46 +08:00
|
|
|
BlockIndex int
|
|
|
|
InnerBlockOffset int64
|
2020-08-19 08:04:28 +08:00
|
|
|
Size types.Size
|
2019-05-27 16:29:46 +08:00
|
|
|
IsLargeBlock bool
|
|
|
|
LargeBlockRowsCount int
|
2019-05-19 18:01:58 +08:00
|
|
|
}
|
|
|
|
|
2020-08-19 08:04:28 +08:00
|
|
|
func LocateData(largeBlockLength, smallBlockLength int64, datSize int64, offset int64, size types.Size) (intervals []Interval) {
|
2019-05-19 18:01:58 +08:00
|
|
|
blockIndex, isLargeBlock, innerBlockOffset := locateOffset(largeBlockLength, smallBlockLength, datSize, offset)
|
|
|
|
|
2019-05-27 16:29:46 +08:00
|
|
|
// adding DataShardsCount*smallBlockLength to ensure we can derive the number of large block size from a shard size
|
|
|
|
nLargeBlockRows := int((datSize + DataShardsCount*smallBlockLength) / (largeBlockLength * DataShardsCount))
|
2019-05-19 18:01:58 +08:00
|
|
|
|
|
|
|
for size > 0 {
|
|
|
|
interval := Interval{
|
2019-05-27 16:29:46 +08:00
|
|
|
BlockIndex: blockIndex,
|
|
|
|
InnerBlockOffset: innerBlockOffset,
|
|
|
|
IsLargeBlock: isLargeBlock,
|
|
|
|
LargeBlockRowsCount: nLargeBlockRows,
|
2019-05-19 18:01:58 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
blockRemaining := largeBlockLength - innerBlockOffset
|
|
|
|
if !isLargeBlock {
|
|
|
|
blockRemaining = smallBlockLength - innerBlockOffset
|
|
|
|
}
|
|
|
|
|
|
|
|
if int64(size) <= blockRemaining {
|
2019-05-27 16:29:46 +08:00
|
|
|
interval.Size = size
|
2019-05-19 18:01:58 +08:00
|
|
|
intervals = append(intervals, interval)
|
|
|
|
return
|
|
|
|
}
|
2020-08-19 08:04:28 +08:00
|
|
|
interval.Size = types.Size(blockRemaining)
|
2019-05-19 18:01:58 +08:00
|
|
|
intervals = append(intervals, interval)
|
|
|
|
|
2019-05-27 16:29:46 +08:00
|
|
|
size -= interval.Size
|
2019-05-19 18:01:58 +08:00
|
|
|
blockIndex += 1
|
|
|
|
if isLargeBlock && blockIndex == nLargeBlockRows*DataShardsCount {
|
|
|
|
isLargeBlock = false
|
|
|
|
blockIndex = 0
|
|
|
|
}
|
|
|
|
innerBlockOffset = 0
|
|
|
|
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
func locateOffset(largeBlockLength, smallBlockLength int64, datSize int64, offset int64) (blockIndex int, isLargeBlock bool, innerBlockOffset int64) {
|
|
|
|
largeRowSize := largeBlockLength * DataShardsCount
|
|
|
|
nLargeBlockRows := datSize / (largeBlockLength * DataShardsCount)
|
|
|
|
|
|
|
|
// if offset is within the large block area
|
|
|
|
if offset < nLargeBlockRows*largeRowSize {
|
|
|
|
isLargeBlock = true
|
|
|
|
blockIndex, innerBlockOffset = locateOffsetWithinBlocks(largeBlockLength, offset)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
isLargeBlock = false
|
|
|
|
offset -= nLargeBlockRows * largeRowSize
|
|
|
|
blockIndex, innerBlockOffset = locateOffsetWithinBlocks(smallBlockLength, offset)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
func locateOffsetWithinBlocks(blockLength int64, offset int64) (blockIndex int, innerBlockOffset int64) {
|
|
|
|
blockIndex = int(offset / blockLength)
|
|
|
|
innerBlockOffset = offset % blockLength
|
|
|
|
return
|
|
|
|
}
|
2019-05-28 02:59:03 +08:00
|
|
|
|
|
|
|
func (interval Interval) ToShardIdAndOffset(largeBlockSize, smallBlockSize int64) (ShardId, int64) {
|
|
|
|
ecFileOffset := interval.InnerBlockOffset
|
|
|
|
rowIndex := interval.BlockIndex / DataShardsCount
|
|
|
|
if interval.IsLargeBlock {
|
|
|
|
ecFileOffset += int64(rowIndex) * largeBlockSize
|
|
|
|
} else {
|
|
|
|
ecFileOffset += int64(interval.LargeBlockRowsCount)*largeBlockSize + int64(rowIndex)*smallBlockSize
|
|
|
|
}
|
|
|
|
ecFileIndex := interval.BlockIndex % DataShardsCount
|
|
|
|
return ShardId(ecFileIndex), ecFileOffset
|
|
|
|
}
|