本文整理匯總了Golang中github.com/gholt/brimtime.TimeToUnixMicro函數的典型用法代碼示例。如果您正苦於以下問題:Golang TimeToUnixMicro函數的具體用法?Golang TimeToUnixMicro怎麽用?Golang TimeToUnixMicro使用的例子?那麽, 這裏精選的函數代碼示例或許可以為您提供幫助。
在下文中一共展示了TimeToUnixMicro函數的15個代碼示例,這些例子默認根據受歡迎程度排序。您可以為喜歡或者感覺有用的代碼點讚,您的評價將有助於係統推薦出更棒的Golang代碼示例。
示例1: GroupWrite
func GroupWrite(c *GroupClientConfig) {
defer c.wg.Done()
var err error
var opts []grpc.DialOption
var creds credentials.TransportAuthenticator
creds = credentials.NewTLS(&tls.Config{
InsecureSkipVerify: true,
})
opts = append(opts, grpc.WithTransportCredentials(creds))
conn, err := grpc.Dial(c.addr, opts...)
if err != nil {
log.Fatalln(fmt.Sprintf("Failed to dial server: %s", err))
}
defer conn.Close()
client := gp.NewGroupStoreClient(conn)
w := &gp.WriteRequest{
Value: *c.value,
}
empty := []byte("")
for i, _ := range c.wm {
ctx, _ := context.WithTimeout(context.Background(), 10*time.Second)
c.wm[i].Value = *c.value
w.TimestampMicro = brimtime.TimeToUnixMicro(time.Now())
res, err := client.Write(ctx, c.wm[i])
if err != nil {
log.Println("Client", c.id, ":", err)
}
if res.TimestampMicro > w.TimestampMicro {
log.Printf("TSM is newer than attempted, Key %d-%d Got %s, Sent: %s", c.id, i, brimtime.UnixMicroToTime(res.TimestampMicro), brimtime.UnixMicroToTime(w.TimestampMicro))
}
c.wm[i].Value = empty
}
}
示例2: RevokeAddrFS
// RevokeAddrFS ...
func (s *FileSystemAPIServer) RevokeAddrFS(ctx context.Context, r *pb.RevokeAddrFSRequest) (*pb.RevokeAddrFSResponse, error) {
var err error
srcAddr := ""
// Get incomming ip
pr, ok := peer.FromContext(ctx)
if ok {
srcAddr = pr.Addr.String()
}
// Validate Token
_, err = s.validateToken(r.Token)
if err != nil {
log.Printf("%s REVOKE FAILED %s\n", srcAddr, "PermissionDenied")
return nil, errf(codes.PermissionDenied, "%v", "Invalid Token")
}
// REVOKE an file system entry for the addr
// delete /fs/FSID/addr addr AddrRef
pKey := fmt.Sprintf("/fs/%s/addr", r.FSid)
pKeyA, pKeyB := murmur3.Sum128([]byte(pKey))
cKeyA, cKeyB := murmur3.Sum128([]byte(r.Addr))
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
_, err = s.gstore.Delete(context.Background(), pKeyA, pKeyB, cKeyA, cKeyB, timestampMicro)
if store.IsNotFound(err) {
log.Printf("%s REVOKE FAILED %s %s\n", srcAddr, r.FSid, r.Addr)
return nil, errf(codes.NotFound, "%v", "Not Found")
}
// return Addr was revoked
// Log Operation
log.Printf("%s REVOKE SUCCESS %s %s\n", srcAddr, r.FSid, r.Addr)
return &pb.RevokeAddrFSResponse{Data: r.FSid}, nil
}
示例3: RevokeAddrFS
// RevokeAddrFS ...
func (s *FileSystemAPIServer) RevokeAddrFS(ctx context.Context, r *pb.RevokeAddrFSRequest) (*pb.RevokeAddrFSResponse, error) {
var err error
var acctID string
var value []byte
var fsRef FileSysRef
srcAddr := ""
// Get incomming ip
pr, ok := peer.FromContext(ctx)
if ok {
srcAddr = pr.Addr.String()
}
// Validate Token
acctID, err = s.validateToken(r.Token)
if err != nil {
log.Printf("%s REVOKE FAILED %s\n", srcAddr, "PermissionDenied")
return nil, errf(codes.PermissionDenied, "%v", "Invalid Token")
}
// Validate Token/Account owns this file system
// Read FileSysRef entry to determine if it exists
pKey := fmt.Sprintf("/fs")
pKeyA, pKeyB := murmur3.Sum128([]byte(pKey))
cKeyA, cKeyB := murmur3.Sum128([]byte(r.FSid))
_, value, err = s.gstore.Read(context.Background(), pKeyA, pKeyB, cKeyA, cKeyB, nil)
if store.IsNotFound(err) {
log.Printf("%s REVOKE FAILED %s NOTFOUND", srcAddr, r.FSid)
return nil, errf(codes.NotFound, "%v", "Not Found")
}
if err != nil {
log.Printf("%s REVOKE FAILED %v\n", srcAddr, err)
return nil, errf(codes.Internal, "%v", err)
}
err = json.Unmarshal(value, &fsRef)
if err != nil {
log.Printf("%s REVOKE FAILED %v\n", srcAddr, err)
return nil, errf(codes.Internal, "%v", err)
}
if fsRef.AcctID != acctID {
log.Printf("$s REVOKE FAILED %v ACCOUNT MISMATCH", r.FSid)
return nil, errf(codes.FailedPrecondition, "%v", "Account Mismatch")
}
// REVOKE an file system entry for the addr
// delete /fs/FSID/addr addr AddrRef
pKey = fmt.Sprintf("/fs/%s/addr", r.FSid)
pKeyA, pKeyB = murmur3.Sum128([]byte(pKey))
cKeyA, cKeyB = murmur3.Sum128([]byte(r.Addr))
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
_, err = s.gstore.Delete(context.Background(), pKeyA, pKeyB, cKeyA, cKeyB, timestampMicro)
if store.IsNotFound(err) {
log.Printf("%s REVOKE FAILED %s %s\n", srcAddr, r.FSid, r.Addr)
return nil, errf(codes.NotFound, "%v", "Not Found")
}
// return Addr was revoked
// Log Operation
log.Printf("%s REVOKE SUCCESS %s %s\n", srcAddr, r.FSid, r.Addr)
return &pb.RevokeAddrFSResponse{Data: r.FSid}, nil
}
示例4: VSTests
func VSTests() {
vsconfigs := make([]ValueClientConfig, *clients)
var wg sync.WaitGroup
for w := 0; w < *clients; w++ {
vsconfigs[w].addr = *vsServer
vsconfigs[w].id = w
vsconfigs[w].count = perClient
vsconfigs[w].value = &value
vsconfigs[w].wg = &wg
vsconfigs[w].wm = make([]*vp.WriteRequest, perClient)
vsconfigs[w].rm = make([]*vp.ReadRequest, perClient)
for k := 0; k < perClient; k++ {
vsconfigs[w].wm[k] = &vp.WriteRequest{}
vsconfigs[w].rm[k] = &vp.ReadRequest{}
vsconfigs[w].wm[k].KeyA, vsconfigs[w].wm[k].KeyB = murmur3.Sum128([]byte(fmt.Sprintf("somethingtestkey%d-%d", vsconfigs[w].id, k)))
vsconfigs[w].wm[k].TimestampMicro = brimtime.TimeToUnixMicro(time.Now())
vsconfigs[w].rm[k].KeyA = vsconfigs[w].wm[k].KeyA
vsconfigs[w].rm[k].KeyB = vsconfigs[w].wm[k].KeyB
}
}
log.Println("ValueStore Key/hash generation complete. Spawning tests.")
// ValueStore Tests
if *vsWriteTest {
t := time.Now()
for w := 0; w < *clients; w++ {
wg.Add(1)
if *streamTest {
go ValueStreamWrite(&vsconfigs[w])
} else {
go ValueWrite(&vsconfigs[w])
}
}
wg.Wait()
log.Println("Issued", *clients*perClient, "VS WRITES")
ts := time.Since(t).Seconds()
log.Println("Total run time was:", ts, "seconds")
log.Printf("Per second: %.2f\n", float64(*clients*perClient)/ts)
}
if *vsReadTest {
t := time.Now()
for w := 0; w < *clients; w++ {
wg.Add(1)
if *streamTest {
go ValueStreamRead(&vsconfigs[w])
} else {
go ValueRead(&vsconfigs[w])
}
}
wg.Wait()
log.Println("Issued", *clients*perClient, "VS READS")
ts := time.Since(t).Seconds()
log.Println("Total run time was:", ts, "seconds")
log.Printf("Per second: %.2f\n", float64(*clients*perClient)/ts)
}
}
示例5: GrantAddrFS
// GrantAddrFS ...
func (s *FileSystemAPIServer) GrantAddrFS(ctx context.Context, r *fb.GrantAddrFSRequest) (*fb.GrantAddrFSResponse, error) {
var status string
var err error
var acctData AcctPayLoad
var fsData FileSysPayLoad
var addrData AddrPayLoad
var dataB []byte
// Get incomming ip
pr, ok := peer.FromContext(ctx)
if ok {
fmt.Println(pr.Addr)
}
// getAcct data
acctData, err = s.getAcct("/acct", r.Acctnum)
if err != nil {
log.Printf("Error %v on lookup for account %s", err, r.Acctnum)
return nil, err
}
// validate token
if acctData.Token != r.Token {
return nil, errf(codes.PermissionDenied, "%s", "Invalid Token")
}
// getFS data
fs := fmt.Sprintf("/acct/%s/fs", r.Acctnum)
fsData, err = s.getFS(fs, r.FSid)
if err != nil {
log.Printf("Error %v on lookup for File system %s", err, r.Acctnum)
return nil, err
}
if fsData.Status == "active" {
log.Println("FileSystem is active")
}
// write out the ip address
parentKey := fmt.Sprintf("/fs/%s/addr", r.FSid)
childKey := r.Addr
parentKeyA, parentKeyB := murmur3.Sum128([]byte(parentKey))
childKeyA, childKeyB := murmur3.Sum128([]byte(childKey))
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
addrData.Addr = r.Addr
dataB, err = json.Marshal(addrData)
if err != nil {
log.Printf("Marshal Error: %v\n...", err)
return nil, errf(codes.Internal, "%v", err)
}
_, err = s.fsws.gstore.Write(context.Background(), parentKeyA, parentKeyB, childKeyA, childKeyB, timestampMicro, dataB)
if err != nil {
log.Printf("Write Error: %v", err)
return nil, errf(codes.Internal, "%v", err)
}
// DO stuff
status = fmt.Sprintf("addr %s for filesystem %s with account id %s was granted", r.Addr, r.FSid, r.Acctnum)
return &fb.GrantAddrFSResponse{Status: status}, nil
}
示例6: WriteValue
func (o *StoreComms) WriteValue(ctx context.Context, id, data []byte) error {
keyA, keyB := murmur3.Sum128(id)
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
oldTimestampMicro, err := o.vstore.Write(ctx, keyA, keyB, timestampMicro, data)
if err != nil {
return err
}
if oldTimestampMicro >= timestampMicro {
return ErrStoreHasNewerValue
}
return nil
}
示例7: writeGStore
// writeGStore ...
func (fsws *FileSystemWS) writeGStore(g string, m string, p []byte) (string, error) {
// prepare groupVal and memberVal
log.Println("Starting a Write to the Group Store")
keyA, keyB := murmur3.Sum128([]byte(g))
childKeyA, childKeyB := murmur3.Sum128([]byte(m))
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
newTimestampMicro, err := fsws.gstore.Write(context.Background(), keyA, keyB, childKeyA, childKeyB, timestampMicro, p)
if err != nil {
return "", err
}
log.Println("Successfully wrote something to the Group Store")
return fmt.Sprintf("TSM: %d", newTimestampMicro), nil
}
示例8: Remove
func (o *OortFS) Remove(ctx context.Context, parent []byte, name string) (int32, error) {
v, err := o.validateIP(ctx)
if err != nil {
return 1, err
}
if !v {
return 1, errors.New("Unknown or unauthorized FS use")
}
// Get the ID from the group list
b, err := o.comms.ReadGroupItem(ctx, parent, []byte(name))
if store.IsNotFound(err) {
return 1, nil
} else if err != nil {
return 1, err
}
d := &pb.DirEntry{}
err = proto.Unmarshal(b, d)
if err != nil {
return 1, err
}
// TODO: More error handling needed
// TODO: Handle possible race conditions where user writes and deletes the same file over and over
// Mark the item deleted in the group
t := &pb.Tombstone{}
tsm := brimtime.TimeToUnixMicro(time.Now())
t.Dtime = tsm
t.Qtime = tsm
t.FsId = []byte("1") // TODO: Make sure this gets set when we are tracking fsids
inode, err := o.GetInode(ctx, d.Id)
if err != nil {
return 1, err
}
t.Blocks = inode.Blocks
t.Inode = inode.Inode
d.Tombstone = t
b, err = proto.Marshal(d)
if err != nil {
return 1, err
}
// NOTE: The tsm-1 is kind of a hack because the timestamp needs to be updated on this write, but if we choose tsm, once the actual delete comes through, it will not work because it is going to try to delete with a timestamp of tsm.
err = o.comms.WriteGroupTS(ctx, parent, []byte(name), b, tsm-1)
if err != nil {
return 1, err // Not really sure what should be done here to try to recover from err
}
o.deleteChan <- &DeleteItem{
parent: parent,
name: name,
}
return 0, nil
}
示例9: GrantAddrFS
// GrantAddrFS ...
func (s *FileSystemAPIServer) GrantAddrFS(ctx context.Context, r *pb.GrantAddrFSRequest) (*pb.GrantAddrFSResponse, error) {
var err error
var addrData AddrRef
var addrByte []byte
srcAddr := ""
// Get incomming ip
pr, ok := peer.FromContext(ctx)
if ok {
srcAddr = pr.Addr.String()
}
// validate token
_, err = s.validateToken(r.Token)
if err != nil {
log.Printf("%s GRANT FAILED %s\n", srcAddr, "PermissionDenied")
return nil, errf(codes.PermissionDenied, "%v", "Invalid Token")
}
// GRANT an file system entry for the addr
// write /fs/FSID/addr addr AddrRef
pKey := fmt.Sprintf("/fs/%s/addr", r.FSid)
pKeyA, pKeyB := murmur3.Sum128([]byte(pKey))
cKeyA, cKeyB := murmur3.Sum128([]byte(r.Addr))
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
addrData.Addr = r.Addr
addrData.FSID = r.FSid
addrByte, err = json.Marshal(addrData)
if err != nil {
log.Printf("%s GRANT FAILED %v\n", srcAddr, err)
return nil, errf(codes.Internal, "%v", err)
}
_, err = s.gstore.Write(context.Background(), pKeyA, pKeyB, cKeyA, cKeyB, timestampMicro, addrByte)
if err != nil {
log.Printf("%s GRANT FAILED %v\n", srcAddr, err)
return nil, errf(codes.Internal, "%v", err)
}
// return Addr was Granted
// Log Operation
log.Printf("%s GRANT SUCCESS %s %s\n", srcAddr, r.FSid, r.Addr)
return &pb.GrantAddrFSResponse{Data: r.FSid}, nil
}
示例10: RevokeAddrFS
// RevokeAddrFS ...
func (s *FileSystemAPIServer) RevokeAddrFS(ctx context.Context, r *fb.RevokeAddrFSRequest) (*fb.RevokeAddrFSResponse, error) {
var status string
var err error
var acctData AcctPayLoad
// Get incomming ip
pr, ok := peer.FromContext(ctx)
if ok {
fmt.Println(pr.Addr)
}
// getAcct data
acctData, err = s.getAcct("/acct", r.Acctnum)
if err != nil {
log.Printf("Error %v on lookup for account %s", err, r.Acctnum)
return nil, errf(codes.NotFound, "%v", err)
}
// validate token
if acctData.Token != r.Token {
return nil, errf(codes.PermissionDenied, "%s", "Invalid Token")
}
parentKey := fmt.Sprintf("/fs/%s/addr", r.FSid)
childKey := r.Addr
parentKeyA, parentKeyB := murmur3.Sum128([]byte(parentKey))
childKeyA, childKeyB := murmur3.Sum128([]byte(childKey))
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
// Delete addr
_, err = s.fsws.gstore.Delete(context.Background(), parentKeyA, parentKeyB, childKeyA, childKeyB, timestampMicro)
if store.IsNotFound(err) {
log.Printf("/fs/%s/addr/%s did not exist to delete", r.FSid, r.Addr)
return nil, errf(codes.NotFound, "%s", "Addr not found")
} else if err != nil {
return nil, errf(codes.Internal, "%s", err)
}
// DO stuff
status = fmt.Sprintf("addr %s for filesystem %s with account id %s was revoked", r.Addr, r.FSid, r.Acctnum)
return &fb.RevokeAddrFSResponse{Status: status}, nil
}
示例11: outPullReplicationPass
func (store *defaultValueStore) outPullReplicationPass(notifyChan chan *bgNotification) *bgNotification {
if store.msgRing == nil {
return nil
}
ring := store.msgRing.Ring()
if ring == nil || ring.ReplicaCount() < 2 || ring.NodeCount() < 2 {
return nil
}
begin := time.Now()
defer func() {
elapsed := time.Now().Sub(begin)
store.logDebug("outPullReplication: pass took %s", elapsed)
atomic.StoreInt64(&store.outPullReplicationNanoseconds, elapsed.Nanoseconds())
}()
rightwardPartitionShift := 64 - ring.PartitionBitCount()
partitionCount := uint64(1) << ring.PartitionBitCount()
if store.pullReplicationState.outIteration == math.MaxUint16 {
store.pullReplicationState.outIteration = 0
} else {
store.pullReplicationState.outIteration++
}
ringVersion := ring.Version()
ws := store.pullReplicationState.outWorkers
for uint64(len(store.pullReplicationState.outKTBFs)) < ws {
store.pullReplicationState.outKTBFs = append(store.pullReplicationState.outKTBFs, newValueKTBloomFilter(store.pullReplicationState.outBloomN, store.pullReplicationState.outBloomP, 0))
}
var abort uint32
f := func(p uint64, w uint64, ktbf *valueKTBloomFilter) {
pb := p << rightwardPartitionShift
rb := pb + ((uint64(1) << rightwardPartitionShift) / ws * w)
var re uint64
if w+1 == ws {
if p+1 == partitionCount {
re = math.MaxUint64
} else {
re = ((p + 1) << rightwardPartitionShift) - 1
}
} else {
re = pb + ((uint64(1) << rightwardPartitionShift) / ws * (w + 1)) - 1
}
timestampbitsnow := uint64(brimtime.TimeToUnixMicro(time.Now())) << _TSB_UTIL_BITS
cutoff := timestampbitsnow - store.replicationIgnoreRecent
var more bool
for atomic.LoadUint32(&abort) == 0 {
rbThis := rb
ktbf.reset(store.pullReplicationState.outIteration)
rb, more = store.locmap.ScanCallback(rb, re, 0, _TSB_LOCAL_REMOVAL, cutoff, store.pullReplicationState.outBloomN, func(keyA uint64, keyB uint64, timestampbits uint64, length uint32) bool {
ktbf.add(keyA, keyB, timestampbits)
return true
})
ring2 := store.msgRing.Ring()
if ring2 == nil || ring2.Version() != ringVersion {
break
}
reThis := re
if more {
reThis = rb - 1
}
prm := store.newOutPullReplicationMsg(ringVersion, uint32(p), cutoff, rbThis, reThis, ktbf)
atomic.AddInt32(&store.outPullReplications, 1)
store.msgRing.MsgToOtherReplicas(prm, uint32(p), store.pullReplicationState.outMsgTimeout)
if !more {
break
}
}
}
wg := &sync.WaitGroup{}
wg.Add(int(ws))
for w := uint64(0); w < ws; w++ {
go func(w uint64) {
ktbf := store.pullReplicationState.outKTBFs[w]
pb := partitionCount / ws * w
for p := pb; p < partitionCount; p++ {
if atomic.LoadUint32(&abort) != 0 {
break
}
ring2 := store.msgRing.Ring()
if ring2 == nil || ring2.Version() != ringVersion {
break
}
if ring.Responsible(uint32(p)) {
f(p, w, ktbf)
}
}
for p := uint64(0); p < pb; p++ {
if atomic.LoadUint32(&abort) != 0 {
break
}
ring2 := store.msgRing.Ring()
if ring2 == nil || ring2.Version() != ringVersion {
break
}
if ring.Responsible(uint32(p)) {
f(p, w, ktbf)
}
}
wg.Done()
}(w)
}
waitChan := make(chan struct{}, 1)
//.........這裏部分代碼省略.........
示例12: inPullReplication
// inPullReplication actually processes incoming pull-replication messages;
// there may be more than one of these workers.
func (store *defaultValueStore) inPullReplication(wg *sync.WaitGroup) {
k := make([]uint64, store.bulkSetState.msgCap/_VALUE_BULK_SET_MSG_MIN_ENTRY_LENGTH*2)
v := make([]byte, store.valueCap)
for {
prm := <-store.pullReplicationState.inMsgChan
if prm == nil {
break
}
if store.msgRing == nil {
store.pullReplicationState.inFreeMsgChan <- prm
continue
}
ring := store.msgRing.Ring()
if ring == nil {
store.pullReplicationState.inFreeMsgChan <- prm
continue
}
k = k[:0]
// This is what the remote system used when making its bloom filter,
// computed via its config.ReplicationIgnoreRecent setting. We want to
// use the exact same cutoff in our checks and possible response.
cutoff := prm.cutoff()
tombstoneCutoff := (uint64(brimtime.TimeToUnixMicro(time.Now())) << _TSB_UTIL_BITS) - store.tombstoneDiscardState.age
ktbf := prm.ktBloomFilter()
l := int64(store.bulkSetState.msgCap)
callback := func(keyA uint64, keyB uint64, timestampbits uint64, length uint32) bool {
if timestampbits&_TSB_DELETION == 0 || timestampbits >= tombstoneCutoff {
if !ktbf.mayHave(keyA, keyB, timestampbits) {
k = append(k, keyA, keyB)
l -= _VALUE_BULK_SET_MSG_ENTRY_HEADER_LENGTH + int64(length)
if l <= 0 {
return false
}
}
}
return true
}
// Based on the replica index for the local node, start the scan at
// different points. For example, in a three replica system the first
// replica would start scanning at the start, the second a third
// through, the last would start two thirds through. This is so that
// pull-replication messages, which are sent concurrently to all other
// replicas, will get different responses back instead of duplicate
// items if there is a lot of data to be sent.
responsibleReplica := ring.ResponsibleReplica(uint32(prm.rangeStart() >> (64 - ring.PartitionBitCount())))
if responsibleReplica < 0 {
responsibleReplica = 0
}
scanStart := prm.rangeStart() + (prm.rangeStop()-prm.rangeStart())/uint64(ring.ReplicaCount())*uint64(responsibleReplica)
scanStop := prm.rangeStop()
store.locmap.ScanCallback(scanStart, scanStop, 0, _TSB_LOCAL_REMOVAL, cutoff, math.MaxUint64, callback)
if scanStart != prm.rangeStart() && l > 0 {
scanStop = scanStart - 1
scanStart = prm.rangeStart()
store.locmap.ScanCallback(scanStart, scanStop, 0, _TSB_LOCAL_REMOVAL, cutoff, math.MaxUint64, callback)
}
nodeID := prm.nodeID()
store.pullReplicationState.inFreeMsgChan <- prm
if len(k) > 0 {
bsm := store.newOutBulkSetMsg()
// Indicate that a response to this bulk-set message is not
// necessary. If the message fails to reach its destination, that
// destination will simply resend another pull replication message
// on its next pass.
binary.BigEndian.PutUint64(bsm.header, 0)
var t uint64
var err error
for i := 0; i < len(k); i += 2 {
t, v, err = store.read(k[i], k[i+1], v[:0])
if IsNotFound(err) {
if t == 0 {
continue
}
} else if err != nil {
continue
}
if t&_TSB_LOCAL_REMOVAL == 0 {
if !bsm.add(k[i], k[i+1], t, v) {
break
}
atomic.AddInt32(&store.outBulkSetValues, 1)
}
}
if len(bsm.body) > 0 {
atomic.AddInt32(&store.outBulkSets, 1)
store.msgRing.MsgToNode(bsm, nodeID, store.pullReplicationState.inResponseMsgTimeout)
}
}
}
wg.Done()
}
示例13: DeleteValue
func (o *StoreComms) DeleteValue(ctx context.Context, id []byte) error {
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
return o.DeleteValueTS(ctx, id, timestampMicro)
}
示例14: GSTests
func GSTests() {
gsconfigs := make([]GroupClientConfig, *clients)
var wg sync.WaitGroup
for w := 0; w < *clients; w++ {
gsconfigs[w].addr = *gsServer
gsconfigs[w].id = w
gsconfigs[w].count = perClient
gsconfigs[w].value = &value
gsconfigs[w].wg = &wg
perGroup := perClient / *groups
for g := 0; g < *groups; g++ {
grpA, grpB := murmur3.Sum128([]byte(fmt.Sprintf("group%d-%d", gsconfigs[w].id, g)))
for k := 0; k < perGroup; k++ {
tsm := brimtime.TimeToUnixMicro(time.Now())
wr := &gp.WriteRequest{
KeyA: grpA,
KeyB: grpB,
TimestampMicro: tsm,
}
wr.ChildKeyA, wr.ChildKeyB = murmur3.Sum128([]byte(fmt.Sprintf("somethingtestkey%d-%d", gsconfigs[w].id, k)))
rr := &gp.ReadRequest{
KeyA: grpA,
KeyB: grpB,
ChildKeyA: wr.ChildKeyA,
ChildKeyB: wr.ChildKeyB,
}
gsconfigs[w].wm = append(gsconfigs[w].wm, wr)
gsconfigs[w].rm = append(gsconfigs[w].rm, rr)
}
}
}
log.Println("GroupStore Key/hash generation complete. Spawning tests.")
if *gsWriteTest {
t := time.Now()
for w := 0; w < *clients; w++ {
wg.Add(1)
if *streamTest {
go GroupStreamWrite(&gsconfigs[w])
} else {
go GroupWrite(&gsconfigs[w])
}
}
wg.Wait()
log.Println("Issued", *clients*perClient, "GS WRITES")
ts := time.Since(t).Seconds()
log.Println("Total run time was:", ts, "seconds")
log.Printf("Per second: %.2f\n", float64(*clients*perClient)/ts)
}
if *gsReadTest {
t := time.Now()
for w := 0; w < *clients; w++ {
wg.Add(1)
if *streamTest {
go GroupStreamRead(&gsconfigs[w])
} else {
go GroupRead(&gsconfigs[w])
}
}
wg.Wait()
log.Println("Issued", *clients*perClient, "GS READS")
ts := time.Since(t).Seconds()
log.Println("Total run time was:", ts, "seconds")
log.Printf("Per second: %.2f\n", float64(*clients*perClient)/ts)
}
}
示例15: DeleteGroupItem
func (o *StoreComms) DeleteGroupItem(ctx context.Context, key, childKey []byte) error {
timestampMicro := brimtime.TimeToUnixMicro(time.Now())
return o.DeleteGroupItemTS(ctx, key, childKey, timestampMicro)
}