query
stringlengths
8
6.75k
document
stringlengths
9
1.89M
negatives
listlengths
19
19
metadata
dict
Count returns the number of nodes and leaves in the part of tree having the given prefix, or the whole tree if searchPrefix is nil.
func (idx *Tree) Count(searchPrefix []byte) (nodes, leaves int) { return NewCounter(idx).Count(searchPrefix) }
[ "func (c *Counter) Count(searchPrefix []byte) (nodes, leaves int) {\n\tstart := time.Now()\n\tc.reset()\n\talloc := c.idx.GetAllocator()\n\tdefer c.idx.ReleaseAllocator(alloc)\n\talloc.startOp()\n\tdefer alloc.endOp()\n\tvar p counterPosition\n\n\tp.raw, _ = c.idx.partialSearch(searchPrefix)\n\tif p.raw == 0 {\n\t\treturn\n\t}\n\tif isLeaf(p.raw) {\n\t\tleaves++\n\t\treturn\n\t}\n\tnodes++\n\nsearchLoop:\n\tfor {\n\t\t_, node, count, prefixLen := explodeNode(p.raw)\n\t\tdata := c.idx.getNodeData(node)\n\t\tif prefixLen > 0 {\n\t\t\tprefixSlots := ((prefixLen + 7) >> 3)\n\t\t\tif prefixLen == 255 {\n\t\t\t\tc.LargePrefixes++\n\t\t\t\tprefixLen = int(data[0])\n\t\t\t\tprefixSlots = ((prefixLen + 15) >> 3)\n\t\t\t}\n\t\t\tc.NodesWithPrefix++\n\t\t\tc.TotalPrefixBytes += prefixSlots * 8\n\t\t\tc.TotalPrefixLength += prefixLen\n\t\t\tdata = data[prefixSlots:]\n\t\t}\n\n\t\tc.NodeSizes[count]++\n\n\t\tfor k := p.nextChild; k < count; k++ {\n\t\t\ta := atomic.LoadUint64(&data[k])\n\t\t\tif a != 0 {\n\t\t\t\tif isLeaf(a) {\n\t\t\t\t\tleaves++\n\t\t\t\t} else {\n\t\t\t\t\tnodes++\n\t\t\t\t\tnext := k + 1\n\t\t\t\t\tif count >= fullAllocFrom {\n\t\t\t\t\t\t// since it's a node-256 we need to check if there are any more non-zero children in this node\n\t\t\t\t\t\tfor ; next < count; next++ {\n\t\t\t\t\t\t\tif atomic.LoadUint64(&data[next]) != 0 {\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tif next < count {\n\t\t\t\t\t\tc.stack = append(c.stack, counterPosition{p.raw, next})\n\t\t\t\t\t}\n\t\t\t\t\tp.raw = a\n\t\t\t\t\tp.nextChild = 0\n\t\t\t\t\tcontinue searchLoop\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif len(c.stack) == 0 {\n\t\t\tbreak\n\t\t}\n\t\tp, c.stack = c.stack[len(c.stack)-1], c.stack[:len(c.stack)-1]\n\t}\n\n\tc.Elapsed = time.Since(start)\n\treturn nodes, leaves\n}", "func (db *DB) CountPrefix(prefix interface{}, value interface{}, count *int) error {\n\treturn db.bolt.View(func(tx *bolt.Tx) error {\n\t\treturn db.CountPrefixTx(tx, prefix, value, count)\n\t})\n}", "func (db *DB) CountPrefixTx(tx *bolt.Tx, prefix interface{}, value interface{}, count *int) error {\n\t*count = 0\n\tbb := db.bucket(value)\n\tb := tx.Bucket(bb)\n\tif b == nil {\n\t\treturn nil\n\t}\n\n\tpb, err := db.encode(prefix)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tc := b.Cursor()\n\tfor k, _ := c.Seek(pb); k != nil && bytes.HasPrefix(k, pb); k, _ = c.Next() {\n\t\t*count++\n\t}\n\treturn nil\n}", "func (n *Node) HasPrefix(prefix string) (*Node, error) {\n\tletters := []rune(prefix)\n\tcurrent := n\n\tfor i := 0; current != nil && i < len(letters); i++ {\n\t\tcurrent = current.GetChild(letters[i])\n\t}\n\tif current == nil {\n\t\terr := fmt.Errorf(\"%s not found\", prefix)\n\t\treturn nil, err\n\t}\n\treturn current, nil\n}", "func (pT *PrefixTree) Search(prefix []interface{}) [][]interface{} {\n\tvalues := make([][]interface{}, 0)\n\tlastNode, to := pT.searchLastNode(prefix)\n\tif to+1 != len(prefix) {\n\t\treturn values\n\t}\n\tvalues = *lastNode.getFinalChainValues()\n\tprefixInTrie := prefix[:to]\n\taddPrefixToValues(prefixInTrie, &values)\n\treturn values\n}", "func (n *Node) WalkPrefix(prefix []byte, fn WalkFn) {\n\tsearch := prefix\n\tfor {\n\t\t// Check for key exhaution\n\t\tif len(search) == 0 {\n\t\t\trecursiveWalk(n, fn)\n\t\t\treturn\n\t\t}\n\n\t\t// Look for an edge\n\t\t_, n = n.getEdge(search[0])\n\t\tif n == nil {\n\t\t\tbreak\n\t\t}\n\n\t\t// Consume the search prefix\n\t\tif bytes.HasPrefix(search, n.prefix) {\n\t\t\tsearch = search[len(n.prefix):]\n\n\t\t} else if bytes.HasPrefix(n.prefix, search) {\n\t\t\t// Child may be under our search prefix\n\t\t\trecursiveWalk(n, fn)\n\t\t\treturn\n\t\t} else {\n\t\t\tbreak\n\t\t}\n\t}\n}", "func prefixCount(words []string, pref string) int {\n\tr := 0\n\tfor _, w := range words {\n\t\tif len(pref) <= len(words) && strings.HasPrefix(w, pref) {\n\t\t\tr++\n\t\t}\n\t}\n\n\treturn r\n}", "func (t *trie) PrefixMatch(prefix string, total int) []string {\n\tvar result []string\n\tfor i := 0; i < len(prefix); i++ {\n\t\tindex := prefix[i]\n\t\tif t.ChildrenNode[index] == nil {\n\t\t\treturn result\n\t\t}\n\t\tt = t.ChildrenNode[index]\n\t}\n\tif t.isTerminal && t.isLastNode() {\n\t\tresult = append(result, prefix)\n\t\treturn result\n\t}\n\tkeys := []string{}\n\tif t.isTerminal {\n\t\tkeys = append(keys, prefix)\n\t\tif total != -1 {\n\t\t\ttotal--\n\t\t}\n\t}\n\tif !t.isLastNode() {\n\t\t_, result = t.find(prefix, keys, total)\n\t}\n\treturn result\n}", "func (t *BPTree) PrefixScan(prefix []byte, offsetNum int, limitNum int) (records Records, off int, err error) {\n\tvar (\n\t\tn *Node\n\t\tscanFlag bool\n\t\tkeys [][]byte\n\t\tpointers []interface{}\n\t\ti, j, numFound int\n\t)\n\n\tn = t.FindLeaf(prefix)\n\n\tif n == nil {\n\t\treturn nil, off, ErrPrefixScansNoResult\n\t}\n\n\tfor j = 0; j < n.KeysNum && compare(n.Keys[j], prefix) < 0; {\n\t\tj++\n\t}\n\n\tscanFlag = true\n\tnumFound = 0\n\n\tcoff := 0\n\n\tfor n != nil && scanFlag {\n\t\tfor i = j; i < n.KeysNum; i++ {\n\n\t\t\tif !bytes.HasPrefix(n.Keys[i], prefix) {\n\t\t\t\tscanFlag = false\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif coff < offsetNum {\n\t\t\t\tcoff++\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tkeys = append(keys, n.Keys[i])\n\t\t\tpointers = append(pointers, n.pointers[i])\n\t\t\tnumFound++\n\n\t\t\tif limitNum > 0 && numFound == limitNum {\n\t\t\t\tscanFlag = false\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tn, _ = n.pointers[order-1].(*Node)\n\t\tj = 0\n\t}\n\n\toff = coff\n\n\tesr, err := getRecordWrapper(numFound, keys, pointers)\n\treturn esr, off, err\n}", "func (t *BPTree) PrefixSearchScan(prefix []byte, reg string, offsetNum int, limitNum int) (records Records, off int, err error) {\n\tvar (\n\t\tn *Node\n\t\tscanFlag bool\n\t\tkeys [][]byte\n\t\tpointers []interface{}\n\t\ti, j, numFound int\n\t)\n\n\trgx, err := regexp.Compile(reg)\n\tif err != nil {\n\t\treturn nil, off, ErrBadRegexp\n\t}\n\n\tn = t.FindLeaf(prefix)\n\n\tif n == nil {\n\t\treturn nil, off, ErrPrefixSearchScansNoResult\n\t}\n\n\tfor j = 0; j < n.KeysNum && compare(n.Keys[j], prefix) < 0; {\n\t\tj++\n\t}\n\n\tscanFlag = true\n\tnumFound = 0\n\n\tcoff := 0\n\n\tfor n != nil && scanFlag {\n\t\tfor i = j; i < n.KeysNum; i++ {\n\n\t\t\tif !bytes.HasPrefix(n.Keys[i], prefix) {\n\t\t\t\tscanFlag = false\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif coff < offsetNum {\n\t\t\t\tcoff++\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif !rgx.Match(bytes.TrimPrefix(n.Keys[i], prefix)) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tkeys = append(keys, n.Keys[i])\n\t\t\tpointers = append(pointers, n.pointers[i])\n\t\t\tnumFound++\n\n\t\t\tif limitNum > 0 && numFound == limitNum {\n\t\t\t\tscanFlag = false\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tn, _ = n.pointers[order-1].(*Node)\n\t\tj = 0\n\t}\n\n\toff = coff\n\n\tesr, err := getRecordWrapper(numFound, keys, pointers)\n\treturn esr, off, err\n}", "func (d Digest) NPrefix() int {\n\tfor i := d.h.Size() - 1; i >= 0; i-- {\n\t\tif d.b[i] != 0 {\n\t\t\treturn i + 1\n\t\t}\n\t}\n\treturn 0\n}", "func (t *PrefixStoreByteTrie) PrefixSearch(prefix []byte) *list.List {\n\tif len(prefix) > t.maxKeySizeInBytes {\n\t\treturn list.New()\n\t}\n\tsubTrie := t.get(prefix)\n\tif subTrie == nil {\n\t\treturn list.New()\n\t}\n\n\tentries := subTrie.list(prefix)\n\tif subTrie.isLast {\n\t\tentries.PushFront(prefix)\n\t}\n\treturn entries\n}", "func PrefixLength(fulltext, prefix []byte) int {\n\ti := 0\n\n\tlimit := len(prefix)\n\tif len(fulltext) < len(prefix) {\n\t\tlimit = len(fulltext)\n\t}\n\n\tfor {\n\t\tif i >= limit {\n\t\t\tbreak\n\t\t}\n\n\t\tif prefix[i] != fulltext[i] {\n\t\t\tbreak\n\t\t}\n\n\t\ti++\n\t}\n\n\treturn i\n}", "func (pT *PrefixTree) Length() int {\n\treturn pT.length\n}", "func (t *ArtTree) PrefixSearch(key []byte) []interface{} {\n\tret := make([]interface{}, 0)\n\tfor r := range t.PrefixSearchChan(key) {\n\t\tret = append(ret, r.Value)\n\t}\n\treturn ret\n}", "func (d *Driver) Count(ctx context.Context, prefix string) (revRet int64, count int64, err error) {\n\tstart := time.Now()\n\tdefer func() {\n\t\tdur := time.Since(start)\n\t\tfStr := \"COUNT %s => rev=%d, count=%d, err=%v, duration=%s\"\n\t\td.logMethod(dur, fStr, prefix, revRet, count, err, dur)\n\t}()\n\n\tentries, err := d.getKeys(ctx, prefix, false)\n\tif err != nil {\n\t\treturn 0, 0, err\n\t}\n\t// current revision\n\tcurrentRev, err := d.currentRevision()\n\tif err != nil {\n\t\treturn 0, 0, err\n\t}\n\treturn currentRev, int64(len(entries)), nil\n}", "func (n IPv4Network) GetPrefixLength() int {\n\tslice := strings.Split(n.CIDR, \"/\")\n\tlen, _ := strconv.Atoi(slice[1])\n\treturn len\n}", "func IsPrefixOfWord(sentence string, searchWord string) int {\n\tvar l1 int = len(sentence)\n\tvar l2 int = len(searchWord)\n\tvar left int = 0\n\tvar res int= -1\n\tvar idx int = 1\n\tfor left < l1{\n\t\tvar visit int = 0\n\t\tfor left < l1 && visit < l2 && sentence[left] == searchWord[visit]{\n\t\t\tleft++\n\t\t\tvisit++\n\t\t}\n\t\tif visit == l2{\n\t\t\treturn idx\n\t\t}\n\t\tfor left < l1 && sentence[left] != ' '{\n\t\t\tleft++\n\t\t}\n\t\tidx++\n\t\tleft++\n\t}\n\treturn res\n}", "func (this *Trie) StartsWith(prefix string) bool {\n node := this.SearchWord(prefix, 0)\n if node == nil {\n return false\n } \n return true\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
LookupCalled looks up ssa.Instruction that call the `fn` func in the given instr
func LookupCalled(instr ssa.Instruction, fn *types.Func) ([]ssa.Instruction, bool) { instrs := []ssa.Instruction{} call, ok := instr.(ssa.CallInstruction) if !ok { return instrs, false } ssaCall := call.Value() if ssaCall == nil { return instrs, false } common := ssaCall.Common() if common == nil { return instrs, false } val := common.Value called := false switch fnval := val.(type) { case *ssa.Function: for _, block := range fnval.Blocks { for _, instr := range block.Instrs { if analysisutil.Called(instr, nil, fn) { called = true instrs = append(instrs, instr) } } } } return instrs, called }
[ "func (m *Module) instCall(old *ast.InstCall, resolved, unresolved map[ast.NamedValue]value.Named) bool {\n\tif isUnresolved(unresolved, old.Callee) || isUnresolved(unresolved, old.Args...) {\n\t\treturn false\n\t}\n\tv := m.getLocal(old.Name)\n\tinst, ok := v.(*ir.InstCall)\n\tif !ok {\n\t\tpanic(fmt.Errorf(\"invalid instruction type for instruction %s; expected *ir.InstCall, got %T\", enc.Local(old.Name), v))\n\t}\n\tvar (\n\t\ttyp *types.PointerType\n\t\tsig *types.FuncType\n\t)\n\tvar args []value.Value\n\tfor _, oldArg := range old.Args {\n\t\targ := m.irValue(oldArg)\n\t\targs = append(args, arg)\n\t}\n\tcallee := m.irValue(old.Callee)\n\tif c, ok := callee.(*ir.InlineAsm); ok {\n\t\tswitch t := c.Typ.(type) {\n\t\tcase *types.FuncType:\n\t\t\tsig = t\n\t\tdefault:\n\t\t\t// Result type stored in t, get parameters for function signature from\n\t\t\t// arguments. Not perfect, but the best we can do. In particular, this\n\t\t\t// approach is known to fail for variadic parameters.\n\t\t\tvar params []*types.Param\n\t\t\tfor _, arg := range args {\n\t\t\t\tparam := types.NewParam(\"\", arg.Type())\n\t\t\t\tparams = append(params, param)\n\t\t\t}\n\t\t\tsig = types.NewFunc(t, params...)\n\t\t}\n\t\ttyp = types.NewPointer(sig)\n\t} else {\n\t\tvar ok bool\n\t\ttyp, ok = callee.Type().(*types.PointerType)\n\t\tif !ok {\n\t\t\tpanic(fmt.Errorf(\"invalid callee type, expected *types.PointerType, got %T\", callee.Type()))\n\t\t}\n\t\tsig, ok = typ.Elem.(*types.FuncType)\n\t\tif !ok {\n\t\t\tpanic(fmt.Errorf(\"invalid callee signature type, expected *types.FuncType, got %T\", typ.Elem))\n\t\t}\n\t}\n\tinst.Callee = callee\n\tinst.Sig = sig\n\t// TODO: Validate old.Type against inst.Sig.\n\tinst.Args = args\n\tinst.CallConv = ir.CallConv(old.CallConv)\n\tinst.Metadata = m.irMetadata(old.Metadata)\n\treturn true\n}", "func (r *eventCBRegister) lookup(i int) EventCallbackFn {\n\tr.Lock()\n\tdefer r.Unlock()\n\treturn r.fns[i]\n}", "func FnCall() bool {\n\treturn fnCall\n}", "func (th *Thread) CallLookup(this Value, method string, args ...Value) Value {\n\treturn th.CallThis(th.Lookup(this, method), this, args...)\n}", "func (l *Loader) SetLookupFn(fn func(string) (string, bool)) {\n\tl.lock.Lock()\n\tdefer l.lock.Unlock()\n\n\tl.lookUp = fn\n}", "func (l *InfosT) ByRuntimeFunc(argFunc *runtime.Func) Info {\n\n\tfor i := 0; i < len(*l); i++ {\n\t\tlpFunc := (*l)[i].Handler\n\t\tlpFuncType := reflect.ValueOf(lpFunc)\n\t\tif lpFuncType.Pointer() == argFunc.Entry() {\n\t\t\treturn (*l)[i]\n\t\t}\n\t}\n\tlog.Panicf(\"unknown runtime func %+v\", argFunc)\n\treturn Info{}\n}", "func (_f6 *FakeResolver) ResolveCalledWith(name string) (found bool) {\n\tfor _, call := range _f6.ResolveCalls {\n\t\tif reflect.DeepEqual(call.Parameters.Name, name) {\n\t\t\tfound = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn\n}", "func callFn(name string, unit reflect.Value, args []reflect.Value) error {\n\tif fn := unit.MethodByName(name); fn.IsValid() == false {\n\t\treturn nil\n\t} else if ret := fn.Call(args); len(ret) != 1 {\n\t\treturn nil\n\t} else if err, ok := ret[0].Interface().(error); ok {\n\t\treturn err\n\t} else if err == nil {\n\t\treturn nil\n\t} else {\n\t\treturn gopi.ErrBadParameter.WithPrefix(name)\n\t}\n}", "func Lookup(key string) RunnerFunc {\n\tv, ok := runners.Load(key)\n\tif !ok {\n\t\treturn nil\n\t}\n\treturn v.(RunnerFunc)\n}", "func (_f2 *FakeInterfacer) InterfaceCalledWith(ident1 interface{}) (found bool) {\n\tfor _, call := range _f2.InterfaceCalls {\n\t\tif reflect.DeepEqual(call.Parameters.Ident1, ident1) {\n\t\t\tfound = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn\n}", "func (_f6 *FakeconsulRegistry) ServiceRegisterCalledWith(ident1 *consul.AgentServiceRegistration) (found bool) {\n\tfor _, call := range _f6.ServiceRegisterCalls {\n\t\tif reflect.DeepEqual(call.Parameters.Ident1, ident1) {\n\t\t\tfound = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn\n}", "func (bi *BinaryInfo) LookupGenericFunc() map[string][]*Function {\n\tif bi.lookupGenericFunc == nil {\n\t\tbi.lookupGenericFunc = make(map[string][]*Function)\n\t\tfor i := range bi.Functions {\n\t\t\tdn := bi.Functions[i].NameWithoutTypeParams()\n\t\t\tif dn != bi.Functions[i].Name {\n\t\t\t\tbi.lookupGenericFunc[dn] = append(bi.lookupGenericFunc[dn], &bi.Functions[i])\n\t\t\t}\n\t\t}\n\t}\n\treturn bi.lookupGenericFunc\n}", "func (b *busMock) CallFunction(\n\t_ context.Context,\n\tdest string,\n\tobjPath dbus.ObjectPath,\n\tfuncName string) error {\n\tif strings.HasSuffix(funcName, \"StartDiscovery\") {\n\t\treturn nil\n\t}\n\tif strings.HasSuffix(funcName, \"StopDiscovery\") {\n\t\treturn nil\n\t}\n\treturn fmt.Errorf(\"CallFunction(%s) not yet mocked\", funcName)\n}", "func (a *ActorInfo) LookupMethod(num uint64) (MethodInfo, bool) {\n\tif idx, ok := a.methodMap[num]; ok {\n\t\treturn a.Methods[idx], true\n\t}\n\n\treturn MethodInfo{}, false\n}", "func (mr *MockNoSQLEngineMockRecorder) Lookup(arg0, arg1, arg2, arg3 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Lookup\", reflect.TypeOf((*MockNoSQLEngine)(nil).Lookup), arg0, arg1, arg2, arg3)\n}", "func (frame *StackFrame) Func() *runtime.Func {\n\tif frame.ProgramCounter == 0 {\n\t\treturn nil\n\t}\n\treturn runtime.FuncForPC(frame.ProgramCounter)\n}", "func (mr *MockKBFSOpsMockRecorder) Lookup(ctx, dir, name interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Lookup\", reflect.TypeOf((*MockKBFSOps)(nil).Lookup), ctx, dir, name)\n}", "func (bi *BinaryInfo) symLookup(addr uint64) (string, uint64) {\n\tfn := bi.PCToFunc(addr)\n\tif fn != nil {\n\t\tif fn.Entry == addr {\n\t\t\t// only report the function name if it's the exact address because it's\n\t\t\t// easier to read the absolute address than function_name+offset.\n\t\t\treturn fn.Name, fn.Entry\n\t\t}\n\t\treturn \"\", 0\n\t}\n\tif sym, ok := bi.SymNames[addr]; ok {\n\t\treturn sym.Name, addr\n\t}\n\ti := sort.Search(len(bi.packageVars), func(i int) bool {\n\t\treturn bi.packageVars[i].addr >= addr\n\t})\n\tif i >= len(bi.packageVars) {\n\t\treturn \"\", 0\n\t}\n\tif bi.packageVars[i].addr > addr {\n\t\t// report previous variable + offset if i-th variable starts after addr\n\t\ti--\n\t}\n\tif i >= 0 && bi.packageVars[i].addr != 0 {\n\t\treturn bi.packageVars[i].name, bi.packageVars[i].addr\n\t}\n\treturn \"\", 0\n}", "func (frame *stackFrame) Func() *runtime.Func {\n\tif frame.ProgramCounter == 0 {\n\t\treturn nil\n\t}\n\treturn runtime.FuncForPC(frame.ProgramCounter)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
HasArgs returns whether the given ssa.Instruction has `typ` type args
func HasArgs(instr ssa.Instruction, typ types.Type) bool { call, ok := instr.(ssa.CallInstruction) if !ok { return false } ssaCall := call.Value() if ssaCall == nil { return false } for _, arg := range ssaCall.Call.Args { if types.Identical(arg.Type(), typ) { return true } } return false }
[ "func (i Instruction) IsVariadic() bool {\n\treturn len(i.Arities()) > 1\n}", "func hasArgs(args []Argument) bool {\n\treturn len(args) > 0\n}", "func (TypesObject) IsVariadicParam() bool { return boolResult }", "func (f flagString) HasArg() bool {\n\treturn true\n}", "func (t *itype) isVariadic() bool {\n\tswitch t.cat {\n\tcase funcT:\n\t\treturn len(t.arg) > 0 && t.arg[len(t.arg)-1].cat == variadicT\n\tcase valueT:\n\t\tif t.rtype.Kind() == reflect.Func {\n\t\t\treturn t.rtype.IsVariadic()\n\t\t}\n\t}\n\treturn false\n}", "func (f flagBool) HasArg() bool {\n\treturn false\n}", "func (Reg) isArg() {}", "func isTypeParam(_ *ast.Field, _ []*ast.FuncDecl, _ []*ast.FuncLit) bool {\n\treturn false\n}", "func (p *FuncInfo) IsVariadic() bool {\n\tif p.nVariadic == 0 {\n\t\tlog.Panicln(\"FuncInfo is unintialized.\")\n\t}\n\treturn p.nVariadic == nVariadicVariadicArgs\n}", "func IsValidArgs(args []string) bool {\n\treturn len(args) > 1\n}", "func (t *Type) IsFuncArgStruct() bool", "func hasContextParam(handlerType reflect.Type) bool {\n\t//if the handler doesn't take arguments, false\n\tif handlerType.NumIn() == 0 {\n\t\treturn false\n\t}\n\n\t//if the first argument is not a pointer, false\n\tp1 := handlerType.In(0)\n\tif p1.Kind() != reflect.Ptr {\n\t\treturn false\n\t}\n\t//but if the first argument is a context, true\n\tif p1.Elem() == contextType {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (t Type) HasAttr(name string) bool {\n\treturn t.AttrType(name) != NilType\n}", "func (m *Method) HasParameters() bool {\n\treturn len(m.Parameters) != 0\n}", "func (args *Args) Present() bool {\n\treturn args.Len() != 0\n}", "func IsHasMethod(st interface{}, methodName string) bool {\n\treturn HasMethod(st, methodName)\n}", "func (info *ConceptInfo) HasArgument(argument Concept) bool {\n\tswitch argument {\n\tcase \"beer\":\n\t\treturn true\n\t}\n\tfor _, concept := range info.ValidArguments {\n\t\tif concept == argument {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (tr *transform) isType(t dataType, name string) bool {\n\tif name == \"\" {\n\t\treturn false\n\t}\n\n\tname = strings.SplitN(name, \"<<\", 2)[0] // could have a tag\n\n\tfor funcId := tr.funcId; funcId >= 0; funcId-- {\n\t\tfor blockId := tr.blockId; blockId >= 0; blockId-- {\n\t\t\tif _, ok := tr.vars[funcId][blockId][name]; ok { // variable found\n\t\t\t\tswitch t {\n\t\t\t\tcase sliceType:\n\t\t\t\t\tif _, ok := tr.slices[funcId][blockId][name]; ok {\n\t\t\t\t\t\treturn true\n\t\t\t\t\t}\n\t\t\t\tcase mapType:\n\t\t\t\t\tif _, ok := tr.maps[funcId][blockId][name]; ok {\n\t\t\t\t\t\treturn true\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func (n Name) HasType() bool {\n\t_, s := n.GetLookupAndType()\n\treturn s != \"\"\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
withClerk invokes the specified callback with a clerk that manages the specified queue. If there is no such clerk, withClerk first creates one. The clerk's reference count is incremented for the duration of the callback. withClerk returns whatever the callback returns.
func (registry *Registry) withClerk(queue string, callback func(*clerk) error) error { registry.mutex.Lock() entry, found := registry.queues[queue] if !found { entry = &refCountedClerk{ Clerk: clerk{ // TODO: context Enqueue: make(chan messageSend), Dequeue: make(chan messageReceive), Registry: registry, Queue: queue, }, RefCount: 0, } go entry.Clerk.Run() registry.queues[queue] = entry } entry.RefCount++ registry.mutex.Unlock() result := callback(&entry.Clerk) registry.mutex.Lock() entry.RefCount-- registry.mutex.Unlock() return result }
[ "func (this *ModuleManager) CallWithCallback(topic string, f, cb interface{}, cbParams, params []interface{}) (err error) {\n\tif m := this.GetModule(topic); m != nil {\n\t\terr = m.CallWithCallback(f, cb, cbParams, params)\n\t} else {\n\t\t// fmt.Println(this)\n\t\terr = Post.PutQueueWithCallback(f, cb, cbParams, params...)\n\t}\n\treturn\n}", "func (client *Client) GetClusterQueueInfoWithCallback(request *GetClusterQueueInfoRequest, callback func(response *GetClusterQueueInfoResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetClusterQueueInfoResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetClusterQueueInfo(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func GoCallbackWrapper(ptr_q *unsafe.Pointer, ptr_nfad *unsafe.Pointer) int {\n q := (*Queue)(unsafe.Pointer(ptr_q))\n payload := build_payload(q.c_gh, ptr_nfad)\n return q.cb(payload)\n}", "func (q *Queue) SetCallback(cb Callback) error {\n\tq.cb = cb\n\treturn nil\n}", "func (x *MQQueueManager) Ctl(goOperation int32, goctlo *MQCTLO) error {\n\tvar mqrc C.MQLONG\n\tvar mqcc C.MQLONG\n\tvar mqOperation C.MQLONG\n\tvar mqctlo C.MQCTLO\n\n\tmqOperation = C.MQLONG(goOperation)\n\tcopyCTLOtoC(&mqctlo, goctlo)\n\n\t// Need to make sure control information is available before the callback\n\t// is enabled. So this gets setup even if the MQCTL fails.\n\tkey := makePartialKey(x.hConn)\n\tmapLock()\n\tfor k, info := range cbMap {\n\t\tif strings.HasPrefix(k, key) {\n\t\t\tinfo.connectionArea = goctlo.ConnectionArea\n\t\t}\n\t}\n\tmapUnlock()\n\n\tC.MQCTL(x.hConn, mqOperation, (C.PMQVOID)(unsafe.Pointer(&mqctlo)), &mqcc, &mqrc)\n\n\tmqreturn := MQReturn{MQCC: int32(mqcc),\n\t\tMQRC: int32(mqrc),\n\t\tverb: \"MQCTL\",\n\t}\n\n\tif mqcc != C.MQCC_OK {\n\t\treturn &mqreturn\n\t}\n\n\treturn nil\n}", "func (client *Client) RecognizeFlowerWithCallback(request *RecognizeFlowerRequest, callback func(response *RecognizeFlowerResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *RecognizeFlowerResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.RecognizeFlower(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetClusterQueueInfoWithChan(request *GetClusterQueueInfoRequest) (<-chan *GetClusterQueueInfoResponse, <-chan error) {\n\tresponseChan := make(chan *GetClusterQueueInfoResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetClusterQueueInfo(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) QueryCubePerformanceWithCallback(request *QueryCubePerformanceRequest, callback func(response *QueryCubePerformanceResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *QueryCubePerformanceResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.QueryCubePerformance(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (p *asyncPipeline) DoCmdWithCallback(cmder redis.Cmder, callback func(redis.Cmder) error) error {\n\n\tif cmder == nil {\n\t\treturn fmt.Errorf(\"Cmder passing in is nil\")\n\t}\n\n\tcmders := make([]redis.Cmder, 0)\n\tcmders = append(cmders, cmder)\n\tp.chQueue <- &asyncPipelineCmd{\n\t\tcmders: cmders,\n\t\tcallback: func(cmders []redis.Cmder) error {\n\t\t\tfor _, cmder := range cmders {\n\t\t\t\tif callback != nil {\n\t\t\t\t\tif err := callback(cmder); err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t}\n\treturn nil\n}", "func WithConcurrency(concurrency int) Option {\n\treturn func(c *queue) {\n\t\tc.concurrency = concurrency\n\t}\n}", "func (c *chain) consumeDeferred() {\n\tfor kevt := range c.deferredKevts {\n\t\tif c.cb != nil {\n\t\t\terr := c.cb(kevt)\n\t\t\tif err == nil {\n\t\t\t\tdeferredEnqueued.Add(1)\n\t\t\t}\n\t\t}\n\t}\n}", "func (client *Client) PutConfigurationRecorderWithCallback(request *PutConfigurationRecorderRequest, callback func(response *PutConfigurationRecorderResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *PutConfigurationRecorderResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.PutConfigurationRecorder(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetTotalQueueReportWithCallback(request *GetTotalQueueReportRequest, callback func(response *GetTotalQueueReportResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetTotalQueueReportResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetTotalQueueReport(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func NewQueue(cli *clientv3.Client) (Queue, error) {\n\t// issue linearized read to ensure leader election\n\tglog.Infof(\"GET request to endpoint %v\", cli.Endpoints())\n\tctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)\n\t_, err := cli.Get(ctx, \"foo\")\n\tcancel()\n\tglog.Infof(\"GET request succeeded on endpoint %v\", cli.Endpoints())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx, cancel = context.WithCancel(context.Background())\n\treturn &queue{\n\t\tcli: cli,\n\t\trootCtx: ctx,\n\t\trootCancel: cancel,\n\t}, nil\n}", "func (k Keeper) CandQueueIterate(ctx sdk.Context, handler func(uint64) (stop bool)) {\n\tstore := ctx.KVStore(k.key)\n\titer := sdk.KVStorePrefixIterator(store, prefixCandQueue)\n\tdefer iter.Close()\n\tfor ; iter.Valid(); iter.Next() {\n\t\tvar programID uint64\n\t\tk.cdc.MustUnmarshalBinaryLengthPrefixed(iter.Value(), &programID)\n\n\t\tif handler(programID) {\n\t\t\tbreak\n\t\t}\n\t}\n}", "func (client *Client) CreateFabricChannelWithCallback(request *CreateFabricChannelRequest, callback func(response *CreateFabricChannelResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *CreateFabricChannelResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.CreateFabricChannel(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func consumer(queueChannel chan int, doneChannel chan bool) {\n\tdefer reset.Done()\n\n\tvar counter = 0\n\tfor value := range queueChannel {\n\t\tprintln(\"Consuming value from queue: \" + strconv.Itoa(value))\n\t\tcounter++\n\t\tif counter == 23 {\n\t\t\tprintln(\"Consumed 23 values from queue Channel, stop consuming.\")\n\t\t\tclose(doneChannel)\n\t\t\treturn\n\t\t}\n\t}\n}", "func (c *Controller) enqueue(desc *appsapi.Description) {\n\tkey, err := cache.MetaNamespaceKeyFunc(desc)\n\tif err != nil {\n\t\tutilruntime.HandleError(err)\n\t\treturn\n\t}\n\tc.workqueue.Add(key)\n}", "func (ec *EstablishingController) QueueCRD(key string, timeout time.Duration) {\r\n\tec.queue.AddAfter(key, timeout)\r\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
optionalTimeout returns a channel on which the current time will be sent when the specified deadline arrives. Or, if deadline is nil, optionalTimeout returns a nil channel (receives will block forever, i.e. no timeout).
func optionalTimeout(deadline *time.Time) <-chan time.Time { var timeout <-chan time.Time if deadline != nil { timeout = time.NewTimer(time.Until(*deadline)).C } return timeout }
[ "func (s *StreamSession) deadline(ctx context.Context, now time.Time) (earliest time.Time) {\n\tif s.Timeout != 0 { // including negative, for historical reasons\n\t\tearliest = now.Add(s.Timeout)\n\t}\n\tif d, ok := ctx.Deadline(); ok {\n\t\tearliest = minNonzeroTime(earliest, d)\n\t}\n\treturn minNonzeroTime(earliest, s.Deadline)\n}", "func getTimeout(ctx context.Context, defaultTimeout time.Duration) time.Duration {\n\tdeadline, ok := ctx.Deadline()\n\tif !ok {\n\t\treturn defaultTimeout\n\t}\n\tresult := time.Until(deadline)\n\tif result < 0 {\n\t\treturn 0\n\t}\n\treturn result\n}", "func createTimeoutChannel(t time.Duration) (timeout chan bool) {\n\ttimeout = make(chan bool, 1)\n\tgo func() {\n\t\ttime.Sleep(t * time.Second)\n\t\ttimeout <- true\n\t}()\n\treturn\n}", "func (f *Sink) TimeoutChan() <-chan time.Time {\n\treturn f.timeoutTimer.C\n}", "func TimeRemaining() Option {\n\treturn funcOption(func(o *options) { o.timeRemaining = true })\n}", "func RequireContextHasDefaultTimeout(t *testing.T, ctx context.Context, timeout time.Duration) {\n\ttm, hasDeadline := ctx.Deadline()\n\n\trequire.True(t, hasDeadline, \"deadline must exist, we always set an operation timeout\")\n\tduration := time.Until(tm)\n\n\trequire.Greater(t, duration, time.Duration(0))\n\trequire.LessOrEqual(t, duration, timeout)\n}", "func (w *ChannelWriter) SetDeadline(t time.Time) {\n\tw.deadline = t\n}", "func WithDeadline(d time.Duration) RingOption {\n\treturn func(r *Ring) error {\n\t\tr.deadline = d\n\t\ts := newRingSubmitter(r, d)\n\t\t// This is an ugly hack....\n\t\tgo s.run()\n\t\tr.submitter = s\n\t\treturn nil\n\t}\n}", "func (r *Raft) timeoutNow(rpc RPC, req *pb.TimeoutNowRequest) {\n\n}", "func Timeout() <-chan time.Time {\n\treturn Timeouts(GetNonSubscribeTimeout())\n}", "func NewTimeoutChan(ctx context.Context, resolution time.Duration, limit int) *TimeoutChan {\n\tsize := limit\n\tif limit == 0 {\n\t\tsize = 1024\n\t}\n\tin := make(chan Deadliner)\n\tout := make(chan Deadliner)\n\ttc := &TimeoutChan{\n\t\tIn: in,\n\t\tOut: out,\n\n\t\tctx: ctx,\n\t\tpushCtrl: NewController(ctx, \"TimeoutChan Push\"),\n\t\tpopCtrl: NewController(ctx, \"TimeoutChan Pop\"),\n\t\tresolution: resolution,\n\t\tlimit: limit,\n\t\tin: in,\n\t\tout: out,\n\t\tresumePush: make(chan interface{}),\n\t\tresumePop: make(chan interface{}),\n\t\treschedule: make(chan interface{}),\n\t\tclosePush: make(chan interface{}),\n\n\t\tmu: &sync.RWMutex{},\n\t\tpq: NewPriorityQueue(false, size),\n\t\tpushed: 0,\n\t\tpopped: 0,\n\t\tcleared: 0,\n\t}\n\ttc.popCtrl.Go(tc.popProcess)\n\ttc.pushCtrl.Go(tc.pushProcess)\n\treturn tc\n}", "func TestDeadline(t *testing.T) {\n\tmsgChan, _, wg := initTest()\n\teb := eventbus.New()\n\tsupervisor, err := monitor.Launch(eb, unixSoc)\n\tassert.NoError(t, err)\n\n\tlog.AddHook(supervisor)\n\n\t// Send an error entry, to trigger Send\n\tlog.Errorln(\"pippo\")\n\n\tmsg := <-msgChan\n\tassert.Equal(t, \"error\", msg[\"level\"])\n\tassert.Equal(t, \"pippo\", msg[\"msg\"])\n\n\t// The write deadline is 3 seconds, so let's wait for that to expire\n\ttime.Sleep(3 * time.Second)\n\n\tblk := helper.RandomBlock(t, 23, 4)\n\tmsgBlk := message.New(topics.AcceptedBlock, *blk)\n\teb.Publish(topics.AcceptedBlock, msgBlk)\n\n\t// Should get the accepted block message on the msgchan\n\tfor {\n\t\tmsg = <-msgChan\n\t\t// We should discard any other messages\n\t\tif msg[\"code\"] == \"round\" {\n\t\t\t// Success\n\t\t\tbreak\n\t\t}\n\t}\n\n\t_ = supervisor.Stop()\n\twg.Wait()\n}", "func timeoutDialer(secs int) func(net, addr string) (c net.Conn, err error) {\n\treturn func(netw, addr string) (net.Conn, error) {\n\t\tc, err := net.Dial(netw, addr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tc.SetDeadline(time.Now().Add(time.Duration(secs) * time.Second))\n\t\treturn c, nil\n\t}\n}", "func NewOptionalTicker(d time.Duration) *OptionalTicker {\n\tvar ticker OptionalTicker\n\tif d != 0 {\n\t\tticker.t = time.NewTicker(d)\n\t\tticker.C = ticker.t.C\n\t}\n\n\treturn &ticker\n}", "func ReceiveWait(t time.Duration) ReceiveOpt {\n\treturn func(m ReceiveMatcher) ReceiveMatcher {\n\t\tm.timeout = t\n\t\treturn m\n\t}\n}", "func ContextWithOptionalTimeout(parent context.Context, timeout time.Duration) (context.Context, context.CancelFunc) {\n\tif timeout < 0 {\n\t\t// This should be handled in validation\n\t\tklog.Errorf(\"Timeout for context shall not be negative!\")\n\t\ttimeout = 0\n\t}\n\n\tif timeout == 0 {\n\t\treturn context.WithCancel(parent)\n\t}\n\n\treturn context.WithTimeout(parent, timeout)\n}", "func (jc JobConfig) TimeoutOrDefault() time.Duration {\n\treturn jc.Timeout\n}", "func (w *realTimeoutFactory) TimeoutCh() (<-chan time.Time, func() bool) {\n\tif w.timeout == 0 {\n\t\treturn neverExitWatch, func() bool { return false }\n\t}\n\tt := time.NewTimer(w.timeout)\n\treturn t.C, t.Stop\n}", "func TestMock_WithDeadlineCancel(t *testing.T) {\n\tm := NewMock()\n\tctx, cancel := m.WithDeadline(context.Background(), m.Now().Add(time.Second))\n\tcancel()\n\tselect {\n\tcase <-ctx.Done():\n\t\tif !errors.Is(ctx.Err(), context.Canceled) {\n\t\t\tt.Error(\"invalid type of error returned after cancellation\")\n\t\t}\n\tcase <-time.After(time.Second):\n\t\tt.Error(\"context is not cancelled after cancel was called\")\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
copyHeader adds the HTTP header whose name is the specified `which` from the specified headers `from` to the specified headers `to`.
func copyHeader(which string, from http.Header, to http.Header) { for _, value := range from.Values(which) { to.Add(which, value) } }
[ "func copyHeader(target, source http.Header) {\n\tfor k, vs := range source {\n\t\ttarget[k] = vs\n\t}\n}", "func CopyRequestHeaders(from, to *http.Request, headers []string) {\n\tfor _, header := range headers {\n\t\tvalue := from.Header.Get(header)\n\t\tif value != \"\" {\n\t\t\tto.Header.Set(header, value)\n\t\t}\n\t}\n}", "func (c *Client) CopyHeader(req *http.Request) {\n\tfor hn, ha := range req.Header {\n\t\tswitch hn {\n\t\tcase \"Cookie\", \"Date\":\n\t\t\t//ingore cookies\n\t\tdefault:\n\t\t\tfor _, hv := range ha {\n\t\t\t\tc.Header.Set(hn, hv)\n\t\t\t}\n\t\t}\n\t}\n}", "func copyHeaders(dst, src http.Header) {\n\tfor key, vals := range src {\n\t\tfor _, val := range vals {\n\t\t\tdst.Add(key, val)\n\t\t}\n\t}\n}", "func copySourceHeaders(sh http.Header) (th http.Header) {\n\tth = make(http.Header)\n\n\tif sh == nil {\n\t\treturn nil\n\t}\n\n\tfor key, values := range sh {\n\t\tif dhHeadersRe.MatchString(key) {\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, val := range values {\n\t\t\tth.Add(key, val)\n\t\t}\n\t}\n\n\treturn th\n}", "func copyHeaders(rOrig, rNew *http.Request) {\n\tif rOrig == nil || rNew == nil {\n\t\treturn\n\t}\n\n\tfor key, value := range rOrig.Header {\n\t\trNew.Header[key] = value\n\t}\n}", "func CopyHeaders(dst http.Header, src http.Header) {\n\tfor k, vv := range src {\n\t\tdst[k] = append(dst[k], vv...)\n\t}\n}", "func copyHeaders(headerNames []string, srcRequest *http.Request, destRequest *http.Request) {\n for _, headerName := range headerNames {\n // TODO: make sure headerName exists in srcRequest.Header\n if headerValue := srcRequest.Header.Get(headerName); headerValue != \"\" {\n destRequest.Header.Set(headerName, headerValue)\n }\n }\n}", "func CopyHeadersRedirect(req *http.Request, via []*http.Request) error {\n\tif len(via) >= 10 {\n\t\treturn errors.New(\"stopped after 10 redirects\")\n\t}\n\tlastRequest := via[len(via)-1]\n\n\tfor attr, val := range lastRequest.Header {\n\t\t// if hosts do not match do not copy Authorization header\n\t\tif attr == \"Authorization\" && req.Host != lastRequest.Host {\n\t\t\tcontinue\n\t\t}\n\t\tif _, ok := req.Header[attr]; !ok {\n\t\t\treq.Header[attr] = val\n\t\t}\n\t}\n\treturn nil\n}", "func ToHeader(hs []*httpgrpc.Header, header http.Header) {\n\tfor _, h := range hs {\n\t\tif skipHopHeader[h.Key] {\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, v := range h.Values {\n\t\t\theader.Add(h.Key, v)\n\t\t}\n\t}\n}", "func CopyHeader(h *Header) *Header {\n\tcpy := *h\n\tif cpy.Time = new(big.Int); h.Time != nil {\n\t\tcpy.Time.Set(h.Time)\n\t}\n\tif cpy.SnailNumber = new(big.Int); h.SnailNumber != nil {\n\t\tcpy.SnailNumber.Set(h.SnailNumber)\n\t}\n\tif cpy.Number = new(big.Int); h.Number != nil {\n\t\tcpy.Number.Set(h.Number)\n\t}\n\tif len(h.Extra) > 0 {\n\t\tcpy.Extra = make([]byte, len(h.Extra))\n\t\tcopy(cpy.Extra, h.Extra)\n\t}\n\treturn &cpy\n}", "func CopyHeader(dst *LogBuffer, src *LogBuffer) {\n\tsrc.headerMU.Lock()\n\tdup, err := copystructure.Copy(src.header)\n\tdupBanner := src.AddBanner\n\tsrc.headerMU.Unlock()\n\n\tdst.headerMU.Lock()\n\tif err != nil {\n\t\tdst.header = map[string]interface{}{}\n\t} else {\n\t\tdst.header = dup.(map[string]interface{})\n\t}\n\tdst.AddBanner = dupBanner\n\tdst.headerMU.Unlock()\n}", "func copyHeaders(resp *http.Response, w http.ResponseWriter) {\n\tfor key, values := range resp.Header {\n\t\tfor _, value := range values {\n\t\t\tw.Header().Add(key, value)\n\t\t}\n\t}\n}", "func copyResponseHeader(w http.ResponseWriter, resp *http.Response) {\n\tlog.Println(\"start copyResponseHeader\")\n\tif resp == nil {\n\t\tlog.Println(\"null resp in copyResponseHeader\")\n\t\treturn\n\t}\n\tnewHeader := w.Header()\n\tfor key, values := range resp.Header {\n\t\tfor _, v := range values {\n\t\t\tnewHeader.Add(key, v)\n\t\t}\n\t}\n\n\tw.WriteHeader(resp.StatusCode)\n\tlog.Println(\"finished copyResponseHeader\")\n\t\n}", "func (br *ResponseBuffer) Copy(w http.ResponseWriter) error {\n\tfor k, v := range br.sent {\n\t\tw.Header()[k] = v\n\t}\n\tw.WriteHeader(br.code)\n\t_, err := io.Copy(w, br)\n\treturn err\n}", "func FormatHeader(dest , src http.Header ){\r\n\r\n\tfor k, vs := range src {\r\n\t\tfor _, v := range vs {\r\n\t\t\tdest.Add(k, v)\r\n\t\t}\r\n\t}\r\n\tdest.Del(\"Proxy-Connection\")\r\n\tdest.Del(\"Proxy-Authenticate\")\r\n\tdest.Del(\"Proxy-Authorization\")\r\n\tdest.Del(\"Connection\")\r\n}", "func CopyHeaders(dst map[string][]string, src map[string][]string) {\n\tif dst == nil || src == nil {\n\t\treturn\n\t}\n\n\tfor k, vv := range src {\n\t\tv := make([]string, len(vv))\n\t\tcopy(v, vv)\n\t\tdst[k] = v\n\t}\n}", "func FromHeader(hs http.Header) []*httpgrpc.Header {\n\tresult := make([]*httpgrpc.Header, 0, len(hs))\n\tfor k, vs := range hs {\n\t\tif skipHopHeader[k] {\n\t\t\tcontinue\n\t\t}\n\n\t\tresult = append(result, &httpgrpc.Header{\n\t\t\tKey: k,\n\t\t\tValues: vs,\n\t\t})\n\t}\n\n\treturn result\n}", "func (w *responseWrapper) copy(rw http.ResponseWriter) {\n\trw.WriteHeader(w.status)\n\n\tfor k, v := range w.header {\n\t\tfor _, vv := range v {\n\t\t\trw.Header().Add(k, vv)\n\t\t}\n\t}\n\tio.Copy(rw, w.buffer)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
RemoveIfUnused deletes the entry for the specified queue in this registry if its reference count is zero. Return whether the entry has been removed, and any error that occurred.
func (registry *Registry) RemoveIfUnused(queue string) (bool, error) { registry.mutex.Lock() defer registry.mutex.Unlock() entry, found := registry.queues[queue] if !found { // "Not found" is unexpected, since I expect only the clerk itself // would be trying to remove its queue (so why, then, is it already // missing?). return true, fmt.Errorf("no entry for queue %q", queue) } if entry.RefCount > 0 { // Someone is using the queue. Return `false`, meaning "not removed." return false, nil } // Nobody is using the queue. Remove it. delete(registry.queues, queue) // Return `true`, meaning "removed." return true, nil }
[ "func (o *consumer) removeFromRedeliverQueue(seq uint64) bool {\n\tif !o.onRedeliverQueue(seq) {\n\t\treturn false\n\t}\n\tfor i, rseq := range o.rdq {\n\t\tif rseq == seq {\n\t\t\tif len(o.rdq) == 1 {\n\t\t\t\to.rdq, o.rdqi = nil, nil\n\t\t\t} else {\n\t\t\t\to.rdq = append(o.rdq[:i], o.rdq[i+1:]...)\n\t\t\t\tdelete(o.rdqi, seq)\n\t\t\t}\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (recv *Queue) Remove(data uintptr) bool {\n\tc_data := (C.gconstpointer)(data)\n\n\tretC := C.g_queue_remove((*C.GQueue)(recv.native), c_data)\n\tretGo := retC == C.TRUE\n\n\treturn retGo\n}", "func (q *Queue) MaybeRemoveMissing(names []string) int {\n\tq.mu.Lock()\n\tsameSize := len(q.items) == len(names)\n\tq.mu.Unlock()\n\n\t// heuristically skip expensive work\n\tif sameSize {\n\t\treturn -1\n\t}\n\n\tset := make(map[string]struct{}, len(names))\n\tfor _, name := range names {\n\t\tset[name] = struct{}{}\n\t}\n\n\tq.mu.Lock()\n\tdefer q.mu.Unlock()\n\n\tcount := 0\n\tfor name, item := range q.items {\n\t\tif _, ok := set[name]; ok {\n\t\t\tcontinue\n\t\t}\n\n\t\tif item.heapIdx >= 0 {\n\t\t\theap.Remove(&q.pq, item.heapIdx)\n\t\t}\n\t\titem.setIndexState(\"\")\n\t\tdelete(q.items, name)\n\t\tcount++\n\t}\n\n\tmetricQueueLen.Set(float64(len(q.pq)))\n\tmetricQueueCap.Set(float64(len(q.items)))\n\n\treturn count\n}", "func TestRemoveFromEmptyQueue(t *testing.T) {\n\ttarget := teaser.New()\n\tdeleted := target.Delete(\"6e0c9774-2674-4c6f-906e-6ccaebad3772\")\n\tif deleted {\n\t\tt.Fatal(\"deleting from an empty queue should not be possbile!\")\n\t}\n}", "func (q *SliceQueue) DeQueue() bool {\n\tq.lock.Lock()\n\tdefer q.lock.Unlock()\n\n\tif q.IsEmpty() {\n\t\treturn false\n\t}\n\n\tq.Data = q.Data[1:]\n\treturn true\n}", "func (cuo *CityUpdateOne) RemoveQueue(q ...*QueueItem) *CityUpdateOne {\n\tids := make([]int, len(q))\n\tfor i := range q {\n\t\tids[i] = q[i].ID\n\t}\n\treturn cuo.RemoveQueueIDs(ids...)\n}", "func (q *Queue) Remove() (interface{}, bool) {\n\tq.mu.Lock()\n\tdefer q.mu.Unlock()\n\n\tif q.queueLen == 0 {\n\t\treturn nil, false\n\t}\n\treturn q.popFront(), true\n}", "func (h *UintHeap) Remove(tm UintValue) (bool, error) {\n\tif tm.Index() < 3 || tm.Index() >= h.size {\n\t\treturn false, ErrRemove\n\t}\n\n\ti := tm.Index()\n\ttm.SetIndex(0)\n\th.size--\n\tl := h.size\n\tif i != l {\n\t\th.move(l, i)\n\t\th.down(i)\n\t\th.up(i)\n\t}\n\th.chomp()\n\treturn i == 3, nil\n}", "func (cu *CityUpdate) RemoveQueue(q ...*QueueItem) *CityUpdate {\n\tids := make([]int, len(q))\n\tfor i := range q {\n\t\tids[i] = q[i].ID\n\t}\n\treturn cu.RemoveQueueIDs(ids...)\n}", "func (q *Queue) Dequeue() (int, bool) {\n\tif q.length == 0 {\n\t\treturn 0, false\n\t}\n\tn := q.start\n\tif q.length > 1 {\n\t\t// queue has more than one item, proceed as normal\n\t\tq.start = q.start.next\n\t} else {\n\t\t// queue has only one item and needs special case\n\t\tq.start = nil\n\t\tq.end = nil\n\t}\n\t// decrement length\n\tq.length--\n\treturn n.value, true\n}", "func (c *caches) remove(qname string) (ok bool) {\n\tc.Lock()\n\tfor e := c.lru.Front(); e != nil; e = e.Next() {\n\t\tanswer := e.Value.(*answer)\n\t\tif answer.qname != qname {\n\t\t\tcontinue\n\t\t}\n\n\t\tc.lru.Remove(e)\n\t\tdelete(c.v, qname)\n\t\tanswer.clear()\n\t\tok = true\n\t\tbreak\n\t}\n\tif !ok {\n\t\t_, ok = c.v[qname]\n\t\tif ok {\n\t\t\t// If the qname is not found in non-local caches, it\n\t\t\t// may exist as local answer.\n\t\t\tdelete(c.v, qname)\n\t\t}\n\t}\n\tc.Unlock()\n\treturn ok\n}", "func (block *SimpleQueueBlock) clearOldStorage(ctx context.Context, q *SimpleQueue) (notNeed bool, err *mft.Error) {\n\n\tif !block.mxFileSave.TryLock(ctx) {\n\t\treturn false, GenerateError(10017000)\n\t}\n\tdefer block.mxFileSave.Unlock()\n\n\tif len(block.RemoveMarks) == 0 {\n\t\treturn true, nil\n\t}\n\n\tfileName := block.blockFileName()\n\n\tfor _, stName := range block.RemoveMarks {\n\t\tif block.Mark == stName {\n\t\t\tcontinue\n\t\t}\n\t\tst, err := q.getStorageLock(ctx, stName)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\terr = storage.DeleteIfExists(ctx, st, fileName)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t}\n\n\tif !q.mx.TryLock(ctx) {\n\t\treturn false, GenerateError(10017001)\n\t}\n\n\tblock.RemoveMarks = make([]string, 0)\n\n\tq.ChangesRv = q.IDGenerator.RvGetPart()\n\n\tq.mx.Unlock()\n\n\treturn false, nil\n}", "func (b *backend) removeUnsafe(id *entroq.TaskID) {\n\tif !b.existsIDVersionUnsafe(id) {\n\t\tlog.Panicf(\"Item not found for removal: %v\", id)\n\t}\n\titem := b.byID[id.ID]\n\tif item.task.Version != id.Version {\n\t\tlog.Panicf(\"Item removal version mismatch: wanted %q, got %q\", id, item.task.IDVersion())\n\t}\n\th, ok := b.heaps[item.task.Queue]\n\tif !ok {\n\t\tlog.Panicf(\"Queues not in sync; found item to remove in queues but not index: %v\", id)\n\t}\n\n\tdelete(b.byID, id.ID)\n\theap.Remove(h, item.idx)\n\tif h.Len() == 0 {\n\t\tdelete(b.heaps, item.task.Queue)\n\t}\n}", "func (lru *LRUMap) removeFromQueue(node *keyValNode) {\n\tif node.prev != nil {\n\t\tnode.prev.next = node.next\n\t} else {\n\t\tlru.rear = node.next\n\t}\n\n\tif node.next != nil {\n\t\tnode.next.prev = node.prev\n\t} else {\n\t\tlru.front = node.prev\n\t}\n}", "func (k Keeper) CandQueueRemove(ctx sdk.Context, endBlock int64, programID uint64) {\n\tstore := ctx.KVStore(k.key)\n\tstore.Delete(keyCandidate(endBlock, programID))\n}", "func (q *Queue) Remove(v interface{}) bool {\r\n\tq.mu.Lock()\r\n\tdefer q.mu.Unlock()\r\n\r\n\tel, ok := q.m[v]\r\n\tif !ok {\r\n\t\treturn false\r\n\t}\r\n\r\n\tfirst := q.first\r\n\tdelete(q.m, v)\r\n\tq.remove(el)\r\n\r\n\t// If the element was first, we need to start a new timer\r\n\tif first == el && q.first != nil {\r\n\t\tgo q.timer(q.first, q.first.time)\r\n\t}\r\n\treturn true\r\n}", "func (gc *GceCache) UnregisterMig(toBeRemoved Mig) bool {\n\tgc.cacheMutex.Lock()\n\tdefer gc.cacheMutex.Unlock()\n\n\t_, found := gc.migs[toBeRemoved.GceRef()]\n\tif found {\n\t\tklog.V(1).Infof(\"Unregistered Mig %s\", toBeRemoved.GceRef().String())\n\t\tdelete(gc.migs, toBeRemoved.GceRef())\n\t\tgc.removeMigInstances(toBeRemoved.GceRef())\n\t\treturn true\n\t}\n\treturn false\n}", "func (q *Queue) DeQueue() error {\r\n\tif len(q.QueueList) > 0 {\r\n\t\tq.QueueList = q.QueueList[1:]\r\n\t\treturn nil\r\n\t}\r\n\treturn errors.New(\"Queue is empty\")\r\n}", "func (ch *Channel) QueueDelete(name string, ifUnused, ifEmpty, noWait bool) (int, error) {\n\treq := &queueDelete{\n\t\tQueue: name,\n\t\tIfUnused: ifUnused,\n\t\tIfEmpty: ifEmpty,\n\t\tNoWait: noWait,\n\t}\n\tres := &queueDeleteOk{}\n\n\terr := ch.call(req, res)\n\n\treturn int(res.MessageCount), err\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetSecret returns the value of the NovaScheduler.Spec.Secret
func (n NovaScheduler) GetSecret() string { return n.Spec.Secret }
[ "func (c *constantSecret) GetSecret() (string, error) {\n\treturn c.secret, nil\n}", "func (s *Secret) GetSecret() []byte {\n\treturn s.s\n}", "func GetSecret(tid string) ([]byte, error) {\n\tvar token string\n\n\tif err := db.QueryRow(\"SELECT token FROM event_tasks WHERE id = $1\", tid).\n\t\tScan(&token); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn []byte(token), nil\n}", "func (k *Instance) GetSecret(sel *core_api.SecretKeySelector, namespace string) ([]byte, error) {\n\tif sel == nil {\n\t\treturn nil, nil\n\t}\n\tsec, err := k.k8sOps.GetSecret(sel.Name, namespace)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn sec.Data[sel.Key], nil\n}", "func GetSecret() (string, string, error) {\n\tsecretMap := make(map[string]map[string]string)\n\tsecretCfg := os.Getenv(\"SQLFLOW_WORKFLOW_SECRET\")\n\tif secretCfg == \"\" {\n\t\treturn \"\", \"\", nil\n\t}\n\tif e := json.Unmarshal([]byte(secretCfg), &secretMap); e != nil {\n\t\treturn \"\", \"\", e\n\t}\n\tif len(secretMap) != 1 {\n\t\treturn \"\", \"\", fmt.Errorf(`SQLFLOW_WORKFLOW_SECRET should be a json string, e.g. {name: {key: value, ...}}`)\n\t}\n\tname := reflect.ValueOf(secretMap).MapKeys()[0].String()\n\tvalue, e := json.Marshal(secretMap[name])\n\tif e != nil {\n\t\treturn \"\", \"\", e\n\t}\n\treturn name, string(value), nil\n}", "func (r *Robot) GetSecret(name string) string {\n\tcryptKey.RLock()\n\tinitialized := cryptKey.initialized\n\tkey := cryptKey.key\n\tcryptKey.RUnlock()\n\tif !initialized {\n\t\tr.Log(Warn, \"GetSecret called but encryption not initialized\")\n\t\treturn \"\"\n\t}\n\n\tvar secret []byte\n\tvar exists bool\n\tvar ret RetVal\n\n\tc := r.getContext()\n\tif !c.secrets.retrieved {\n\t\t// if it fails, there's little point in multiple lookups in a single\n\t\t// pipeline\n\t\tc.secrets.retrieved = true\n\t\t_, exists, ret = checkoutDatum(secretKey, &c.secrets, false)\n\t\tif ret != Ok {\n\t\t\tr.Log(Error, \"Error retrieving secrets in GetSecret: %s\", ret)\n\t\t\treturn \"\"\n\t\t}\n\t\tif !exists {\n\t\t\tr.Log(Warn, \"GetSecret called for '%s', but no secrets stored\", name)\n\t\t\treturn \"\"\n\t\t}\n\t}\n\ttask, _, _ := getTask(c.currentTask)\n\tsecfound := false\n\tif len(c.nsExtension) > 0 {\n\t\tvar nsMap map[string][]byte\n\t\tfound := false\n\t\tnsMap, exists = c.secrets.RepositoryParams[c.nsExtension]\n\t\tif exists {\n\t\t\tfound = true\n\t\t\tif secret, exists = nsMap[name]; exists {\n\t\t\t\tsecfound = true\n\t\t\t}\n\t\t}\n\t\tif !secfound {\n\t\t\tcmp := strings.Split(c.nsExtension, \"/\")\n\t\t\trepo := strings.Join(cmp[0:len(cmp)-1], \"/\")\n\t\t\tnsMap, exists = c.secrets.RepositoryParams[repo]\n\t\t\tif exists {\n\t\t\t\tfound = true\n\t\t\t\tif secret, exists = nsMap[name]; exists {\n\t\t\t\t\tsecfound = true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tr.Log(Debug, \"Secrets not found for extended namespace '%s'\", c.nsExtension)\n\t\t} else if !secfound {\n\t\t\tr.Log(Debug, \"Secret '%s' not found for extended namespace '%s'\", name, c.nsExtension)\n\t\t}\n\t}\n\t// Fall back to task secrets if namespace secret not found\n\tif !secfound {\n\t\tvar tMap map[string][]byte\n\t\ttMap, exists = c.secrets.TaskParams[task.NameSpace]\n\t\tif !exists {\n\t\t\tr.Log(Debug, \"Secrets not found for task/namespace '%s'\", task.NameSpace)\n\t\t} else if secret, exists = tMap[name]; !exists {\n\t\t\tr.Log(Debug, \"Secret '%s' not found for task/namespace '%s'\", name, task.NameSpace)\n\t\t} else {\n\t\t\tsecfound = true\n\t\t}\n\t}\n\tif !secfound {\n\t\tr.Log(Warn, \"Secret '%s' not found for extended namespace '%s' or task/namespace '%s'\", name, c.nsExtension, task.NameSpace)\n\t\treturn \"\"\n\t}\n\tvar value []byte\n\tvar err error\n\tif value, err = decrypt(secret, key); err != nil {\n\t\tr.Log(Error, \"Error decrypting secret '%s': %v\", name, err)\n\t\treturn \"\"\n\t}\n\treturn string(value)\n}", "func GetSecret(c *gin.Context) {\n\tvar (\n\t\trepo = session.Repo(c)\n\t\tname = c.Param(\"secret\")\n\t)\n\tsecret, err := Config.Services.Secrets.SecretFind(repo, name)\n\tif err != nil {\n\t\tc.String(404, \"Error getting secret %q. %s\", name, err)\n\t\treturn\n\t}\n\tc.JSON(200, secret.Copy())\n}", "func (m *Manager) GetSecret() string {\n\treturn m.user.Secret\n}", "func (a *Authentication) GetSecret() string {\n\treturn a.Secret\n}", "func (rm *redisResourceManager) GetSecret() *corev1.Secret {\n\tname := rm.GetSecretName()\n\n\tconst SecretLen = 8\n\tpassStr := common.RandomString(SecretLen, \"a\")\n\n\treturn &corev1.Secret{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: name,\n\t\t\tNamespace: rm.cluster.Namespace,\n\t\t\tLabels: rm.GetLabels(),\n\t\t},\n\t\tStringData: map[string]string{\n\t\t\t\"redis-password\": passStr,\n\t\t\t\"password\": passStr,\n\t\t},\n\t}\n}", "func (system *System) GetSecret(ctx context.Context) (Secret, error) {\n\tsecret := Secret{}\n\terr := Connection.WithContext(ctx).Where(\"system_id = ?\", system.ID).First(&secret).Error\n\tif err != nil {\n\t\treturn secret, fmt.Errorf(\"unable to get hashed secret for clientID %s: %s\", system.ClientID, err.Error())\n\t}\n\n\tif strings.TrimSpace(secret.Hash) == \"\" {\n\t\treturn secret, fmt.Errorf(\"stored hash of secret for clientID %s is blank\", system.ClientID)\n\t}\n\n\treturn secret, nil\n}", "func (system *System) GetSecret() (string, error) {\n\tdb := GetGORMDbConnection()\n\tdefer Close(db)\n\n\tsecret := Secret{}\n\n\terr := db.Where(\"system_id = ?\", system.ID).First(&secret).Error\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"unable to get hashed secret for clientID %s: %s\", system.ClientID, err.Error())\n\t}\n\n\tif secret.Hash == \"\" {\n\t\treturn \"\", fmt.Errorf(\"stored hash of secret for clientID %s is blank\", system.ClientID)\n\t}\n\n\treturn secret.Hash, nil\n}", "func getSecret() (error, string) {\n\n\t//Create a Secrets Manager client\n\tsvc := secretsmanager.New(session.New(),\n aws.NewConfig().WithRegion(*region))\n\tinput := &secretsmanager.GetSecretValueInput{\n\t\tSecretId: aws.String(*secretName),\n\t\tVersionStage: aws.String(\"AWSCURRENT\"), // VersionStage defaults to AWSCURRENT if unspecified\n\t}\n\n\t// In this sample we only handle the specific exceptions for the 'GetSecretValue' API.\n\t// See https://docs.aws.amazon.com/secretsmanager/latest/apireference/API_GetSecretValue.html\n\n\tresult, err := svc.GetSecretValue(input)\n\tif err != nil {\n\t\treturn err, \"\"\n\t}\n\n\t// Decrypts secret using the associated KMS CMK.\n\t// Depending on whether the secret is a string or binary, one of these fields will be populated.\n\tvar secretString, decodedBinarySecret string\n\tvar keyValueMap map[string]interface{}\n\n\tif result.SecretString != nil {\n\t\tsecretString = *result.SecretString\n\n\t\tif *keyValue == \"\" {\n\t\t\treturn nil, secretString\n\t\t}\n\n\t\t// b, _ := json.Marshal(secretString)\n\t\tjson.Unmarshal([]byte(secretString), &keyValueMap)\n\t\tif keyValueMap[*keyValue] == nil {\n\t\t\treturn nil, \"\"\n\t\t} else {\n\t\t\treturn nil, keyValueMap[*keyValue].(string)\n\t\t}\n\t} else {\n\t\tdecodedBinarySecretBytes := make([]byte, base64.StdEncoding.DecodedLen(len(result.SecretBinary)))\n\t\tlen, err := base64.StdEncoding.Decode(decodedBinarySecretBytes, result.SecretBinary)\n\t\tif err != nil {\n\t\t\t// fmt.Println(\"Base64 Decode Error:\", err)\n\t\t\treturn err, \"\"\n\t\t}\n\t\tdecodedBinarySecret = string(decodedBinarySecretBytes[:len])\n\n\t\tif *keyValue == \"\" {\n\t\t\treturn nil, decodedBinarySecret\n\t\t}\n\n\t\t// b, _ := json.Marshal(decodedBinarySecret)\n\t\tjson.Unmarshal([]byte(decodedBinarySecret), &keyValueMap)\n\t\tif keyValueMap[*keyValue] == nil {\n\t\t\treturn nil, \"\"\n\t\t} else {\n\t\t\treturn nil, keyValueMap[*keyValue].(string)\n\t\t}\t\n\t}\n}", "func (p Prompt) GetSecret(message, help string) (string, error) {\n\tprompt := &survey.Password{\n\t\tMessage: message,\n\t\tHelp: help,\n\t}\n\n\tvar result string\n\n\terr := p(prompt, &result, stdio(), icons())\n\n\treturn result, err\n}", "func getSecret() (string, error) {\n\tlog.Println(prefixLog, os.Args)\n\tif len(os.Args) == 1 || len(os.Args[1]) < 10 {\n\t\treturn \"\", errors.New(\"worker secret key invalid\")\n\t}\n\treturn os.Args[1], nil\n}", "func (s *SecretAPI) GetSecret(secretID, versionID, versionStage string) (*Secret, error) {\n\n\tif s.err != nil {\n\t\treturn nil, s.err\n\t}\n\n\tcs := C.CString(secretID)\n\tcv := C.CString(versionID)\n\tct := C.CString(versionStage)\n\n\tdefer func() {\n\t\ts.close()\n\t\tC.free(unsafe.Pointer(cs))\n\t\tC.free(unsafe.Pointer(cv))\n\t\tC.free(unsafe.Pointer(ct))\n\t}()\n\n\tres := C.gg_request_result{}\n\n\ts.APIRequest.initialize()\n\n\te := GreenGrassCode(C.gg_get_secret_value(s.request, cs, cv, ct, &res))\n\n\ts.response = s.handleRequestResponse(\n\t\t\"Failed to get secret\",\n\t\te, RequestStatus(res.request_status),\n\t)\n\n\tif s.err != nil {\n\t\treturn nil, s.err\n\t}\n\n\tLog(LogLevelInfo, \"Got Secret: %s\", s.response)\n\n\tvar secret Secret\n\tif err := json.Unmarshal([]byte(s.response), &secret); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &secret, nil\n\n}", "func getSecret() (error, map[string]interface{}) {\n\n\t//Create a Secrets Manager client\n\tsvc := secretsmanager.New(session.New(),\n\t\taws.NewConfig().WithRegion(secretConfig.Region))\n\tinput := &secretsmanager.GetSecretValueInput{\n\t\tSecretId: aws.String(secretConfig.Secrets),\n\t\tVersionStage: aws.String(\"AWSCURRENT\"), // VersionStage defaults to AWSCURRENT if unspecified\n\t}\n\n\t// In this sample we only handle the specific exceptions for the 'GetSecretValue' API.\n\t// See https://docs.aws.amazon.com/secretsmanager/latest/apireference/API_GetSecretValue.html\n\n\tresult, err := svc.GetSecretValue(input)\n\tif err != nil {\n\t\treturn err, nil\n\t}\n\n\t// Decrypts secret using the associated KMS CMK.\n\t// Depending on whether the secret is a string or binary, one of these fields will be populated.\n\tvar secretString, decodedBinarySecret string\n\tvar keyValueMap map[string]interface{}\n\n\tif result.SecretString != nil {\n\t\tsecretString = *result.SecretString\n\t\tjson.Unmarshal([]byte(secretString), &keyValueMap)\n\n\t\treturn nil, keyValueMap\n\t} else {\n\t\tdecodedBinarySecretBytes := make([]byte, base64.StdEncoding.DecodedLen(len(result.SecretBinary)))\n\t\tlen, err := base64.StdEncoding.Decode(decodedBinarySecretBytes, result.SecretBinary)\n\t\tif err != nil {\n\t\t\t// fmt.Println(\"Base64 Decode Error:\", err)\n\t\t\treturn err, nil\n\t\t}\n\t\tdecodedBinarySecret = string(decodedBinarySecretBytes[:len])\n\n\t\t// b, _ := json.Marshal(decodedBinarySecret)\n\t\tjson.Unmarshal([]byte(decodedBinarySecret), &keyValueMap)\n\t\treturn nil, keyValueMap\n\t}\n}", "func (s Secret) Get() (*corev1.Secret, error) {\n\tsec := &corev1.Secret{}\n\terr := s.client.Get(context.TODO(), s.NamespacedName, sec)\n\treturn sec, err\n}", "func (v *Vault) getSecret(ginCtx *gin.Context) {\n\tappID := ginCtx.Params.ByName(\"appID\")\n\tglog.Infof(\"App %s waiting to read cubbyhole token.\\n\", appID)\n\tcreateChannelIfNotExistent(appID)\n\tvalue := <-sharedSecret[appID]\n\tglog.Infof(\"Token for app %s will be sent\\n\", appID)\n\tginCtx.JSON(http.StatusOK, gin.H{\"secret\": value})\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
NewTravisBuildListCommand will add a `travis build list` command which is responsible for showing a list of build
func NewTravisBuildListCommand(client *travis.Client) *cobra.Command { cmd := &cobra.Command{ Use: "list", Short: "List all the builds", RunE: func(cmd *cobra.Command, args []string) error { return ListBuilds(client, os.Stdout) }, } return cmd }
[ "func NewListCommand() cli.Command {\n\treturn NewListCommandWithEnv(commoncli.DefaultEnv)\n}", "func NewTravisCommand() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"travis\",\n\t\tShort: \"TravisCI related commands\",\n\t}\n\n\tclient := travis.NewClient(\n\t\ttravis.TRAVIS_API_DEFAULT_URL,\n\t\tviper.GetString(\"TRAVIS_CI_TOKEN\"),\n\t)\n\n\tcmd.AddCommand(\n\t\tbuild.NewTravisBuildListCommand(client),\n\t)\n\n\treturn cmd\n}", "func buildList(c *cli.Context) error {\n\trepo := c.Args().First()\n\towner, name, err := parseRepo(repo)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclient, err := newClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbuilds, err := client.BuildList(owner, name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttmpl, err := template.New(\"_\").Parse(c.String(\"format\") + \"\\n\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbranch := c.String(\"branch\")\n\tevent := c.String(\"event\")\n\tstatus := c.String(\"status\")\n\tlimit := c.Int(\"limit\")\n\n\tvar count int\n\tfor _, build := range builds {\n\t\tif count >= limit {\n\t\t\tbreak\n\t\t}\n\t\tif branch != \"\" && build.Branch != branch {\n\t\t\tcontinue\n\t\t}\n\t\tif event != \"\" && build.Event != event {\n\t\t\tcontinue\n\t\t}\n\t\tif status != \"\" && build.Status != status {\n\t\t\tcontinue\n\t\t}\n\t\ttmpl.Execute(os.Stdout, build)\n\t\tcount++\n\t}\n\treturn nil\n}", "func NewListCommand() cli.Command {\n\treturn newListCommand(defaultEnv, newClients)\n}", "func NewAdoBuildListCommand(client *azuredevops.Client) *cobra.Command {\n\tvar opts ListOptions\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"List all the builds\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\topts.flags = args\n\t\t\treturn ListBuilds(client, opts, os.Stdout)\n\t\t},\n\t}\n\n\tflags := cmd.Flags()\n\tflags.StringVar(&opts.Branches, \"branches\", \"master\", \"Which branches should be displayed\")\n\n\treturn cmd\n}", "func buildPipelineListCmd() *cobra.Command {\n\tvars := listPipelineVars{}\n\tcmd := &cobra.Command{\n\t\tUse: \"ls\",\n\t\tShort: \"Lists all the deployed pipelines in an application.\",\n\t\tExample: `\n Lists all the pipelines for the frontend application.\n /code $ copilot pipeline ls -a frontend`,\n\t\tRunE: runCmdE(func(cmd *cobra.Command, args []string) error {\n\t\t\topts, err := newListPipelinesOpts(vars)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif err := opts.Ask(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn opts.Execute()\n\t\t}),\n\t}\n\n\tcmd.Flags().StringVarP(&vars.appName, appFlag, appFlagShort, tryReadingAppName(), appFlagDescription)\n\tcmd.Flags().BoolVar(&vars.shouldOutputJSON, jsonFlag, false, jsonFlagDescription)\n\treturn cmd\n}", "func NewListCommand(c Command, run RunListFunc, subCommands SubCommands, mod ...CommandModifier) *cobra.Command {\n\treturn newCommand(c, run, subCommands, mod...)\n}", "func newListCommand(cmd *cliCommand) cli.Command {\n\treturn cli.Command{\n\t\tName: \"list\",\n\t\tAliases: []string{\"ls\"},\n\t\tUsage: \"providing a file listing of the files currently in there\",\n\t\tFlags: []cli.Flag{\n\t\t\tcli.BoolFlag{\n\t\t\t\tName: \"l, long\",\n\t\t\t\tUsage: \"provide a detailed / long listing of the files in the bucket\",\n\t\t\t},\n\t\t\tcli.StringFlag{\n\t\t\t\tName: \"b, bucket\",\n\t\t\t\tUsage: \"the name of the s3 bucket containing the encrypted files\",\n\t\t\t\tEnvVar: \"AWS_S3_BUCKET\",\n\t\t\t},\n\t\t\tcli.BoolTFlag{\n\t\t\t\tName: \"r, recursive\",\n\t\t\t\tUsage: \"enable recursive option and transverse all subdirectories\",\n\t\t\t},\n\t\t},\n\t\tAction: func(cx *cli.Context) error {\n\t\t\treturn handleCommand(cx, []string{\"l:bucket:s\"}, cmd, listFiles)\n\t\t},\n\t}\n}", "func (b Build) List(c *gin.Context) {\n\tproject := c.DefaultQuery(\"project\", \"\")\n\tpublic := c.DefaultQuery(\"public\", \"\")\n\tbuilds := []models.Build{}\n\tvar err error\n\n\tif public == \"true\" {\n\t\tbuilds, err = b.publicBuilds()\n\t} else {\n\t\tbuilds, err = b.userBuilds(c, project)\n\t}\n\tif err != nil && err != gorm.ErrRecordNotFound {\n\t\tsugar.InternalError(c, err)\n\t\treturn\n\t}\n\n\tsugar.SuccessResponse(c, 200, builds)\n}", "func (c OSClientBuildClient) List(namespace string, opts kapi.ListOptions) (*buildapi.BuildList, error) {\n\treturn c.Client.Builds(namespace).List(opts)\n}", "func (c *BuildClient) List(ctx context.Context, opts *BuildListOptions) ([]*resource.Build, *Pager, error) {\n\tif opts == nil {\n\t\topts = NewBuildListOptions()\n\t}\n\tvar res resource.BuildList\n\terr := c.client.get(ctx, path.Format(\"/v3/builds?%s\", opts.ToQueryString()), &res)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tpager := NewPager(res.Pagination)\n\treturn res.Resources, pager, nil\n}", "func NewBuildLogsCommand(p *config.KfParams, client builds.Client) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"build-logs BUILD_NAME\",\n\t\tShort: \"Get the logs of the given Build.\",\n\t\tExample: \"kf build-logs build-12345\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tValidArgsFunction: completion.BuildCompletionFn(p),\n\t\tSilenceUsage: true,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tctx := cmd.Context()\n\t\t\tif err := p.ValidateSpaceTargeted(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuildName := args[0]\n\n\t\t\treturn client.Tail(ctx, p.Space, buildName, cmd.OutOrStdout())\n\t\t},\n\t}\n\n\treturn cmd\n}", "func NewListCmd(f *cmdutil.Factory) *cobra.Command {\n\tvar output string\n\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tAliases: []string{\"ls\"},\n\t\tShort: \"List available gitignores\",\n\t\tLong: cmdutil.LongDesc(`\n\t\t\tLists all gitignore templates available via the GitHub Gitignores API (https://docs.github.com/en/rest/reference/gitignore#get-all-gitignore-templates).`),\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclient := gitignore.NewClient(f.HTTPClient())\n\n\t\t\tgitignores, err := client.ListTemplates(context.Background())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tswitch output {\n\t\t\tcase \"name\":\n\t\t\t\tfor _, gitignore := range gitignores {\n\t\t\t\t\tfmt.Fprintln(f.IOStreams.Out, gitignore)\n\t\t\t\t}\n\t\t\tdefault:\n\t\t\t\ttw := cli.NewTableWriter(f.IOStreams.Out)\n\t\t\t\ttw.SetHeader(\"Name\")\n\n\t\t\t\tfor _, gitignore := range gitignores {\n\t\t\t\t\ttw.Append(gitignore)\n\t\t\t\t}\n\n\t\t\t\ttw.Render()\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\tcmdutil.AddOutputFlag(cmd, &output, \"table\", \"name\")\n\n\treturn cmd\n}", "func BuildList(builds ...buildapi.Build) buildapi.BuildList {\n\treturn buildapi.BuildList{\n\t\tItems: builds,\n\t}\n}", "func NewCmdProjectList(name, fullName string) *cobra.Command {\n\to := NewProjectListOptions()\n\tprojectListCmd := &cobra.Command{\n\t\tUse: name,\n\t\tShort: listLongDesc,\n\t\tLong: listLongDesc,\n\t\tExample: fmt.Sprintf(listExample, fullName),\n\t\tArgs: cobra.ExactArgs(0),\n\t\tAnnotations: map[string]string{\"machineoutput\": \"json\"},\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tgenericclioptions.GenericRun(o, cmd, args)\n\t\t},\n\t}\n\treturn projectListCmd\n}", "func NewListCommand(activityRepo core.ActivityRepository) *cobra.Command {\n\tlistCmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"Lists all activities\",\n\t\tLong: \"Lists all the current registered activities in the system\",\n\t\tArgs: cobra.NoArgs,\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tExitIfAppNotConfigured()\n\t\t\tactivities, err := activityRepo.List()\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err.Error())\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\n\t\t\tfor _, act := range activities {\n\t\t\t\tfmt.Println(act.ToPrintableString())\n\t\t\t}\n\t\t},\n\t}\n\treturn listCmd\n}", "func newBuildCmd() *cobra.Command {\n\tvar (\n\t\toptions core.BuildOptions\n\t)\n\n\tbuildCmd := cobra.Command{\n\t\tUse: \"build PROJECT\",\n\t\tShort: `Build your verless project`,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tvar path = \".\"\n\t\t\tif len(args) == 1 {\n\t\t\t\tpath = args[0]\n\t\t\t}\n\t\t\ttargetFs := afero.NewOsFs()\n\n\t\t\tbuild, err := core.NewBuild(targetFs, path, options)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn build.Run()\n\t\t},\n\t}\n\n\taddBuildOptions(&buildCmd, &options, true)\n\n\treturn &buildCmd\n}", "func ListCommand(cli *cli.SensuCli) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"list extensions\",\n\t\tRunE: runList(cli.Config.Format(), cli.Client, cli.Config.Namespace(), cli.Config.Format()),\n\t}\n\n\thelpers.AddAllNamespace(cmd.Flags())\n\thelpers.AddFormatFlag(cmd.Flags())\n\thelpers.AddFieldSelectorFlag(cmd.Flags())\n\thelpers.AddLabelSelectorFlag(cmd.Flags())\n\thelpers.AddChunkSizeFlag(cmd.Flags())\n\n\treturn cmd\n}", "func (p *TestProviderRunner) BuildList(app string, limit int64) (structs.Builds, error) {\n\tp.Called(app, limit)\n\treturn p.Builds, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Bind a `RunnableWithContext` to this runner. WARNING: invoking it when state is StateRunning may cause blocked.
func (s *DeterminationWithContext) Bind(ctx context.Context, runnable RunnableWithContext) *DeterminationWithContext { defer s.lock.Unlock() /*_*/ s.lock.Lock() s.rctx = ctx s.runn = runnable return s }
[ "func (m *Mrb) RunWithContext(v Value, self Value, stackKeep int) (int, *MrbValue, error) {\n\tif self == nil {\n\t\tself = m.TopSelf()\n\t}\n\n\tmrbV := v.MrbValue(m)\n\tmrbSelf := self.MrbValue(m)\n\tproc := C._go_mrb_proc_ptr(mrbV.value)\n\n\ti := C.int(stackKeep)\n\n\tvalue := C._go_mrb_context_run(m.state, proc, mrbSelf.value, &i)\n\n\tif exc := checkException(m.state); exc != nil {\n\t\treturn stackKeep, nil, exc\n\t}\n\n\treturn int(i), newValue(m.state, value), nil\n}", "func RunnerWithContext(ctx context.Context) *SingleRunner {\n\tgrp, _ := errgroup.WithContext(ctx)\n\treturn &SingleRunner{\n\t\tgrp: grp,\n\t\tm: make(map[string]context.CancelFunc),\n\t}\n}", "func (tx *WriteTx) RunWithContext(ctx context.Context) error {\n\tif tx.err != nil {\n\t\treturn tx.err\n\t}\n\tinput, err := tx.input()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = retry(ctx, func() error {\n\t\tout, err := tx.db.client.TransactWriteItemsWithContext(ctx, input)\n\t\tif tx.cc != nil && out != nil {\n\t\t\tfor _, cc := range out.ConsumedCapacity {\n\t\t\t\taddConsumedCapacity(tx.cc, cc)\n\t\t\t}\n\t\t}\n\t\treturn err\n\t})\n\treturn err\n}", "func (r *Retry) RunWithContext(\n\tctx context.Context, funcToRetry func(context.Context) (interface{}, error),\n) (interface{}, error) {\n\t// create a random source which is used in setting the jitter\n\n\tattempts := 0\n\tfor {\n\t\t// run the function\n\t\tresult, err := funcToRetry(ctx)\n\t\t// no error, then we are done!\n\t\tif err == nil {\n\t\t\treturn result, nil\n\t\t}\n\t\t// max retries is reached return the error\n\t\tattempts++\n\t\tif attempts == r.maxTries {\n\t\t\treturn nil, err\n\t\t}\n\t\t// wait for the next duration or context canceled|time out, whichever comes first\n\t\tt := time.NewTimer(getNextBackoff(attempts, r.initialDelay, r.maxDelay))\n\t\tselect {\n\t\tcase <-t.C:\n\t\t\t// nothing to be done as the timer is killed\n\t\tcase <-ctx.Done():\n\t\t\t// context cancelled, kill the timer if it is not killed, and return the last error\n\t\t\tif !t.Stop() {\n\t\t\t\t<-t.C\n\t\t\t}\n\t\t\treturn nil, err\n\t\t}\n\t}\n}", "func (e *executor) RunWithContext(ctx context.Context, cmd *exec.Cmd) (string, error) {\n\tvar out bytes.Buffer\n\n\tprefix := color.BlueString(cmd.Args[0])\n\tlogger := e.logger.WithContext(log.ContextWithPrefix(prefix))\n\n\tcmd.Stdout = io.MultiWriter(&out, log.LineWriter(logger.Info))\n\tcmd.Stderr = io.MultiWriter(&out, log.LineWriter(logger.Error))\n\n\treturn e.run(ctx, &out, cmd)\n}", "func runWithContext(fun func(ctx context.Context) error) (context.CancelFunc, chan error) {\n\tctx, cancel := context.WithCancel(context.Background())\n\tdone := make(chan error, 1)\n\tgo func() {\n\t\tdefer close(done)\n\t\tdone <- fun(ctx)\n\t}()\n\n\treturn cancel, done\n}", "func RunWithContext(ctx context.Context, cmd *exec.Cmd) (string, error) {\n\treturn DefaultExecutor.RunWithContext(ctx, cmd)\n}", "func BenchWithContext(ctx context.Context, r ep.Runner, datasets ...ep.Dataset) (err error) {\n\tout := make(chan ep.Dataset)\n\tinp := make(chan ep.Dataset, len(datasets))\n\tfor _, data := range datasets {\n\t\tinp <- data\n\t}\n\tclose(inp)\n\n\tgo ep.Run(ctx, r, inp, out, nil, &err)\n\n\tfor range out {\n\t}\n\n\treturn err\n}", "func (m *MinikubeRunner) RunWithContext(ctx context.Context, cmdStr string, wait ...bool) (string, string, error) {\n\tprofileArg := fmt.Sprintf(\"-p=%s \", m.Profile)\n\tcmdStr = profileArg + cmdStr\n\tcmdArgs := strings.Split(cmdStr, \" \")\n\tpath, _ := filepath.Abs(m.BinaryPath)\n\n\tcmd := exec.CommandContext(ctx, path, cmdArgs...)\n\tLogf(\"RunWithContext: %s\", cmd.Args)\n\treturn m.teeRun(cmd, wait...)\n}", "func (suite *DispatcherTestSuite) TestRunWithContext() {\n\t// Make some fake SQS messages.\n\tmsg1 := makeMsg(\"msg1\")\n\tmsg2 := makeMsg(\"msg2\")\n\tmsg3 := makeMsg(\"msg3\")\n\tsuite.SUT.MaxConcurrent = 2\n\n\t// Create a mock handler that will be call once for each message.\n\thandler := &mocks.Handler{}\n\thandler.Test(suite.T())\n\n\t// Generate the sequence off message visibility timeouts we expect.\n\tboff := suite.SUT.backoff()\n\ttimeout1 := boff()\n\ttimeout2 := boff()\n\ttimeout3 := boff()\n\n\ttotalExpectedRunTime := 0 * time.Nanosecond // Keep track of how long we expect the test to take.\n\n\t// Describe sequence of expected mock calls, then call RunWithContext().\n\t//\n\t// Step 1: Dispatcher will try to receives messages, gets none within PollTimeout.\n\texpectedReceiveInput := &sqs.ReceiveMessageInput{\n\t\tQueueUrl: aws.String(testQueueURL),\n\t\tMaxNumberOfMessages: aws.Int64(int64(suite.SUT.MaxConcurrent)),\n\t\tVisibilityTimeout: aws.Int64(numSeconds(ceilSeconds(suite.SUT.InitialVisibilityTimeout))),\n\t\tWaitTimeSeconds: aws.Int64(numSeconds(suite.SUT.PollTime)),\n\t}\n\tsuite.MockSQS.On(\"ReceiveMessageWithContext\", AnyContext, expectedReceiveInput).\n\t\tAfter(suite.SUT.PollTime).\n\t\tReturn(&sqs.ReceiveMessageOutput{Messages: []*sqs.Message{}}, error(nil)).\n\t\tOnce()\n\ttotalExpectedRunTime += suite.SUT.PollTime\n\n\t// Step 2: Dispatcher will try to receives messages again, this time gets the first two.\n\tsuite.MockSQS.On(\"ReceiveMessageWithContext\", AnyContext, expectedReceiveInput).\n\t\tAfter(suite.SUT.PollTime/2).\n\t\tReturn(&sqs.ReceiveMessageOutput{Messages: []*sqs.Message{msg1, msg2}}, error(nil)).\n\t\tOnce()\n\ttotalExpectedRunTime += suite.SUT.PollTime / 2\n\n\t// Step 3a: handler is called for msg1. It succeeds immediately and gets deleted.\n\thandler.On(\"Handle\", AnyContext, msg1).\n\t\tReturn(error(nil)).\n\t\tOnce()\n\tsuite.MockSQS.On(\"DeleteMessageWithContext\", AnyContext, &sqs.DeleteMessageInput{\n\t\tQueueUrl: aws.String(testQueueURL),\n\t\tReceiptHandle: msg1.ReceiptHandle,\n\t}).Return(&sqs.DeleteMessageOutput{}, error(nil)).Once()\n\n\t// Step 3b: handler is called for msg2. It succeeds but only after\n\t// the visibility timeout needs to be updated, then gets deleted.\n\thandler.On(\"Handle\", AnyContext, msg2).\n\t\tAfter(timeout1 + timeout2/2). // Return half way through the second timeout\n\t\tReturn(error(nil)).\n\t\tOnce()\n\tsuite.MockSQS.On(\"ChangeMessageVisibilityWithContext\", AnyContext, &sqs.ChangeMessageVisibilityInput{\n\t\tQueueUrl: aws.String(testQueueURL),\n\t\tReceiptHandle: msg2.ReceiptHandle,\n\t\tVisibilityTimeout: aws.Int64(numSeconds(ceilSeconds(timeout2))),\n\t}).Return(&sqs.ChangeMessageVisibilityOutput{}, error(nil)).Once()\n\tsuite.MockSQS.On(\"DeleteMessageWithContext\", AnyContext, &sqs.DeleteMessageInput{\n\t\tQueueUrl: aws.String(testQueueURL),\n\t\tReceiptHandle: msg2.ReceiptHandle,\n\t}).Return(&sqs.DeleteMessageOutput{}, error(nil)).Once()\n\n\t// Step 4: Dispatcher will try to receives messages again, this time gets the last one.\n\texpectedReceiveInput = &sqs.ReceiveMessageInput{\n\t\tQueueUrl: aws.String(testQueueURL),\n\t\tMaxNumberOfMessages: aws.Int64(1), // 1 because msg2 should still be in progress.\n\t\tVisibilityTimeout: aws.Int64(numSeconds(ceilSeconds(suite.SUT.InitialVisibilityTimeout))),\n\t\tWaitTimeSeconds: aws.Int64(numSeconds(suite.SUT.PollTime)),\n\t}\n\tsuite.MockSQS.On(\"ReceiveMessageWithContext\", AnyContext, expectedReceiveInput).\n\t\tAfter(suite.SUT.PollTime/2).\n\t\tReturn(&sqs.ReceiveMessageOutput{Messages: []*sqs.Message{msg3}}, error(nil)).\n\t\tOnce()\n\t// Thereafter, calls to ReceiveMessageWithContext return no messages.\n\tsuite.MockSQS.On(\"ReceiveMessageWithContext\", AnyContext, expectedReceiveInput).\n\t\tAfter(suite.SUT.PollTime).\n\t\tReturn(&sqs.ReceiveMessageOutput{Messages: []*sqs.Message{}}, error(nil)).\n\t\tMaybe()\n\ttotalExpectedRunTime += suite.SUT.PollTime / 2\n\n\t// Step 5: handler is called for msg3. It fails on the third timeout update.\n\tmsg3Err := errors.New(\"msg3Err\")\n\tfailAfter := timeout1 + timeout2 + timeout3/2 // Return half way through the third timeout.\n\thandler.On(\"Handle\", AnyContext, msg3).\n\t\tAfter(failAfter).\n\t\tReturn(msg3Err).\n\t\tOnce()\n\tsuite.MockSQS.On(\"ChangeMessageVisibilityWithContext\", AnyContext, &sqs.ChangeMessageVisibilityInput{\n\t\tQueueUrl: aws.String(testQueueURL),\n\t\tReceiptHandle: msg3.ReceiptHandle,\n\t\tVisibilityTimeout: aws.Int64(numSeconds(ceilSeconds(timeout2))),\n\t}).Return(&sqs.ChangeMessageVisibilityOutput{}, error(nil)).Once()\n\tsuite.MockSQS.On(\"ChangeMessageVisibilityWithContext\", AnyContext, &sqs.ChangeMessageVisibilityInput{\n\t\tQueueUrl: aws.String(testQueueURL),\n\t\tReceiptHandle: msg3.ReceiptHandle,\n\t\tVisibilityTimeout: aws.Int64(numSeconds(ceilSeconds(timeout3))),\n\t}).Return(&sqs.ChangeMessageVisibilityOutput{}, error(nil)).Once()\n\ttotalExpectedRunTime += failAfter\n\n\t// Run the Dispatcher with a timeout.\n\tctx, cancel := context.WithCancel(context.Background())\n\tshouldFinishWithin := time.Duration(totalExpectedRunTime * 2) // Add a little buffer time.\n\tdoCancel := time.AfterFunc(shouldFinishWithin, cancel)\n\terr := suite.SUT.RunWithContext(ctx, handler)\n\n\tsuite.True(doCancel.Stop(), \"timed out\")\n\tsuite.Equal(msg3Err, err)\n\thandler.AssertCalled(suite.T(), \"Handle\", AnyContext, msg1)\n\thandler.AssertCalled(suite.T(), \"Handle\", AnyContext, msg2)\n\thandler.AssertCalled(suite.T(), \"Handle\", AnyContext, msg3)\n\thandler.AssertExpectations(suite.T())\n}", "func (NoopRunner) Run(context.Context) {}", "func (g *Workgroup) AddWithContext(fn func(context.Context)) {\n\tg.fn = append(g.fn, func(stop <-chan struct{}) error {\n\t\tfnctx, fncancel := context.WithCancel(context.Background())\n\t\tdone := make(chan int)\n\n\t\tgo func() {\n\t\t\tdefer close(done)\n\t\t\tfn(fnctx)\n\t\t}()\n\n\t\t// wait for stop signal from the workgroup\n\t\t<-stop\n\n\t\t// cancel the context passed to the function\n\t\tfncancel()\n\n\t\t// wait for function to exit\n\t\t<-done\n\n\t\treturn nil\n\t})\n}", "func (p *Pipeline) RunWithContext(ctx context.Context, cancel context.CancelFunc) {\n\tif p.err != nil {\n\t\treturn\n\t}\n\tp.ctx, p.cancel = ctx, cancel\n\tdataSize := p.source.Prepare(p.ctx)\n\tfilteredSize := dataSize\n\tfor index := 0; index < len(p.nodes); {\n\t\tif p.nodes[index].Begin(p, index, &filteredSize) {\n\t\t\tindex++\n\t\t} else {\n\t\t\tp.nodes = append(p.nodes[:index], p.nodes[index+1:]...)\n\t\t}\n\t}\n\tif p.err != nil {\n\t\treturn\n\t}\n\tif len(p.nodes) > 0 {\n\t\tfor index := 0; index < len(p.nodes)-1; {\n\t\t\tif p.nodes[index].TryMerge(p.nodes[index+1]) {\n\t\t\t\tp.nodes = append(p.nodes[:index+1], p.nodes[index+2:]...)\n\t\t\t} else {\n\t\t\t\tindex++\n\t\t\t}\n\t\t}\n\t\tfor index := len(p.nodes) - 1; index >= 0; index-- {\n\t\t\tif _, ok := p.nodes[index].(*strictordnode); ok {\n\t\t\t\tfor index = index - 1; index >= 0; index-- {\n\t\t\t\t\tswitch node := p.nodes[index].(type) {\n\t\t\t\t\tcase *seqnode:\n\t\t\t\t\t\tnode.kind = Ordered\n\t\t\t\t\tcase *lparnode:\n\t\t\t\t\t\tnode.makeOrdered()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif dataSize < 0 {\n\t\t\tp.finalizeVariableBatchSize()\n\t\t\tfor seqNo, batchSize := 0, p.batchInc; p.source.Fetch(batchSize) > 0; seqNo, batchSize = seqNo+1, p.nextBatchSize(batchSize) {\n\t\t\t\tp.nodes[0].Feed(p, 0, seqNo, p.source.Data())\n\t\t\t\tif err := p.source.Err(); err != nil {\n\t\t\t\t\tp.SetErr(err)\n\t\t\t\t\treturn\n\t\t\t\t} else if p.Err() != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tbatchSize := ((dataSize - 1) / p.NofBatches(0)) + 1\n\t\t\tif batchSize == 0 {\n\t\t\t\tbatchSize = 1\n\t\t\t}\n\t\t\tfor seqNo := 0; p.source.Fetch(batchSize) > 0; seqNo++ {\n\t\t\t\tp.nodes[0].Feed(p, 0, seqNo, p.source.Data())\n\t\t\t\tif err := p.source.Err(); err != nil {\n\t\t\t\t\tp.SetErr(err)\n\t\t\t\t\treturn\n\t\t\t\t} else if p.Err() != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tfor _, node := range p.nodes {\n\t\tnode.End()\n\t}\n\tif p.err == nil {\n\t\tp.err = p.source.Err()\n\t}\n}", "func withContext(borrower ContextBorrower, worker Worker) Worker {\n\n\treturn func(t *T, _ Context) {\n\n\t\tif t.Failed() {\n\t\t\treturn\n\t\t}\n\n\t\tctx, release, err := borrower.Borrow()\n\t\tif err != nil {\n\t\t\tt.Errorf(\"%s\", err)\n\t\t\tt.FailNow()\n\t\t}\n\n\t\tdefer release()\n\t\tworkerRunner(nil, worker, t, ctx)\n\t}\n}", "func (h *Handler) BindContext(ctx context.Context, server core.Server) {\n\tswitch s := server.(type) {\n\tcase *http.Server:\n\t\ts.Handler = h\n\t\ts.BaseContext = func(l net.Listener) context.Context {\n\t\t\treturn ctx\n\t\t}\n\tcase *fasthttp.Server:\n\t\ts.Handler = h.ServeFastHTTP\n\t}\n}", "func (c *Consumer) ListenWithContext(ctx context.Context, fn MessageProcessor) {\n\tc.consume(ctx)\n\tvar wg sync.WaitGroup\n\tdefer wg.Wait()\n\n\tfor {\n\t\tmsg, ok := <-c.messages\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tc.Stats.AddDelivered(1)\n\t\t// For simplicity, did not do the pipe of death here. If POD is received, we may deliver a\n\t\t// couple more messages (especially since select is random in which channel is read from).\n\t\tc.concurrencySem <- empty{}\n\t\twg.Add(1)\n\t\tgo func(msg *Message) {\n\t\t\tdefer func() {\n\t\t\t\t<-c.concurrencySem\n\t\t\t}()\n\t\t\tstart := time.Now()\n\t\t\tfn(msg)\n\t\t\tc.Stats.UpdateProcessedDuration(time.Since(start))\n\t\t\tc.Stats.AddProcessed(1)\n\t\t\twg.Done()\n\t\t}(msg)\n\t}\n}", "func AddServantWithContext(v dispatch, f interface{}, obj string) {\n\taddServantCommon(v, f, obj, true)\n}", "func SleepWithContext(ctx Context, dur time.Duration) error {\n\tt := time.NewTimer(dur)\n\tdefer t.Stop()\n\n\tselect {\n\tcase <-t.C:\n\t\tbreak\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\t}\n\n\treturn nil\n}", "func (p *Pipeline) RunWithContext(ctx context.Context, cancel context.CancelFunc) {\n\tif p.err != nil {\n\t\treturn\n\t}\n\tp.ctx, p.cancel = ctx, cancel\n\tdataSize := p.source.Prepare(p.ctx)\n\tfilteredSize := dataSize\n\tfor index := 0; index < len(p.nodes); {\n\t\tif p.nodes[index].Begin(p, index, &filteredSize) {\n\t\t\tindex++\n\t\t} else {\n\t\t\tp.nodes = append(p.nodes[:index], p.nodes[index+1:]...)\n\t\t}\n\t}\n\tif p.err != nil {\n\t\treturn\n\t}\n\tif len(p.nodes) > 0 {\n\t\tfor index := 0; index < len(p.nodes)-1; {\n\t\t\tif p.nodes[index].TryMerge(p.nodes[index+1]) {\n\t\t\t\tp.nodes = append(p.nodes[:index+1], p.nodes[index+2:]...)\n\t\t\t} else {\n\t\t\t\tindex++\n\t\t\t}\n\t\t}\n\t\tif dataSize < 0 {\n\t\t\tfor seqNo, batchSize := 0, batchInc; p.source.Fetch(batchSize) > 0; seqNo, batchSize = seqNo+1, nextBatchSize(batchSize) {\n\t\t\t\tp.nodes[0].Feed(p, 0, seqNo, p.source.Data())\n\t\t\t\tif err := p.source.Err(); err != nil {\n\t\t\t\t\tp.SetErr(err)\n\t\t\t\t\treturn\n\t\t\t\t} else if p.Err() != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tbatchSize := ((dataSize - 1) / p.NofBatches(0)) + 1\n\t\t\tif batchSize == 0 {\n\t\t\t\tbatchSize = 1\n\t\t\t}\n\t\t\tfor seqNo := 0; p.source.Fetch(batchSize) > 0; seqNo++ {\n\t\t\t\tp.nodes[0].Feed(p, 0, seqNo, p.source.Data())\n\t\t\t\tif err := p.source.Err(); err != nil {\n\t\t\t\t\tp.SetErr(err)\n\t\t\t\t\treturn\n\t\t\t\t} else if p.Err() != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tfor _, node := range p.nodes {\n\t\tnode.End()\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
WhileRunning do something if it is running. It provides a context o check if it stops.
func (s *DeterminationWithContext) WhileRunning(do func(context.Context) error) error { return s.whilerunning(func() error { select { case <-s.sctx.Done(): return ErrRunnerIsClosing default: return do(s.sctx) } }) }
[ "func (s *Determination) WhileRunning(do func(<-chan struct{}) error) error {\n\treturn s.whilerunning(func() error {\n\t\tselect {\n\t\tcase <-s.exit:\n\t\t\treturn ErrRunnerIsClosing\n\t\tdefault:\n\t\t\treturn do(s.exit)\n\t\t}\n\t})\n}", "func (m *Manager) WaitUntilRunning() {\n\tfor {\n\t\tm.mu.Lock()\n\n\t\tif m.ctx != nil {\n\t\t\tm.mu.Unlock()\n\t\t\treturn\n\t\t}\n\n\t\tstarted := m.started\n\t\tm.mu.Unlock()\n\n\t\t// Block until we have been started.\n\t\t<-started\n\t}\n}", "func (s *Stopwatch) isRunning() bool {\n\treturn !s.refTime.IsZero()\n}", "func (w *watcher) checkLoop(ctx context.Context) {\n\tfor atomic.LoadInt32(&w.state) == isRunning {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\tw.Close()\n\t\t\tw.dispose()\n\t\t\treturn\n\t\tdefault:\n\t\t\tw.check(ctx)\n\t\t\ttime.Sleep(w.interval)\n\t\t}\n\t}\n}", "func (ff *FFmpeg) IsRunning() bool {\n\treturn !ff.closed\n}", "func (w *worker) isRunning() bool {\n\treturn atomic.LoadInt32(&w.running) == 1\n}", "func (p *TaskRunner) isRunning() bool {\n\treturn atomic.LoadUint32(&p.state) == startedState\n}", "func isRunning(ctx context.Context) (context.Context, bool) {\n\tif val := ctx.Value(backoffRunningContextKey); val != nil {\n\t\tif v, ok := val.(string); ok && v == backoffRunningContextValue {\n\t\t\treturn ctx, true\n\t\t}\n\t}\n\treturn context.WithValue(ctx, backoffRunningContextKey, backoffRunningContextValue), false\n}", "func (f *flushDaemon) isRunning() bool {\n\tf.mu.Lock()\n\tdefer f.mu.Unlock()\n\treturn f.stopC != nil\n}", "func runWhileFilesystemLives(f func() error, label string, filesystemId string, errorBackoff, successBackoff time.Duration) {\n\tdeathChan := make(chan interface{})\n\tdeathObserver.Subscribe(filesystemId, deathChan)\n\tstillAlive := true\n\tfor stillAlive {\n\t\tselect {\n\t\tcase _ = <-deathChan:\n\t\t\tstillAlive = false\n\t\tdefault:\n\t\t\terr := f()\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\n\t\t\t\t\t\"Error in runWhileFilesystemLives(%s@%s), retrying in %s: %s\",\n\t\t\t\t\tlabel, filesystemId, errorBackoff, err)\n\t\t\t\ttime.Sleep(errorBackoff)\n\t\t\t} else {\n\t\t\t\ttime.Sleep(successBackoff)\n\t\t\t}\n\t\t}\n\t}\n\tdeathObserver.Unsubscribe(filesystemId, deathChan)\n}", "func (run *RunContext) Running() <-chan struct{} {\n\treturn run.running\n}", "func (p *adapter) Running() bool {\n\tif p.cmd == nil || p.cmd.Process == nil || p.cmd.Process.Pid == 0 || p.state() != nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func running() bool {\n\treturn runCalled.Load() != 0\n}", "func (i *info) IsRunning() bool {\n\t_, ok := i.driver.activeContainers[i.ID]\n\treturn ok\n}", "func (o *runner) IsRunning() bool {\n\treturn o.running\n}", "func (b *Backtest) IsRunning() (ret bool) {\n\treturn b.running\n}", "func (b *base) IsRunning() bool {\n\treturn b.isRunning\n}", "func (h *Module) IsRunning()bool{\n\treturn h.running\n}", "func (p *AbstractRunProvider) IsRunning() bool {\n\treturn p.running\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Close this DeterminationWithContext and wait until stop running. Using it in `WhileRunning` will cause deadlock. Please use `CloseAsync()` instead.
func (s *DeterminationWithContext) Close() error { return s.close(s.trigger) }
[ "func CloseWaitContext(ctx context.Context, lc StartedLifecycle) error {\n\tlc.Close()\n\treturn WaitContext(ctx, lc)\n}", "func (s *DeterminationWithContext) WhileRunning(do func(context.Context) error) error {\n\treturn s.whilerunning(func() error {\n\t\tselect {\n\t\tcase <-s.sctx.Done():\n\t\t\treturn ErrRunnerIsClosing\n\t\tdefault:\n\t\t\treturn do(s.sctx)\n\t\t}\n\t})\n}", "func (p *Poller) Stop(ctx context.Context) error {\n\tc := make(chan struct{})\n\tgo func() {\n\t\tdefer close(c)\n\t\tp.wg.Wait()\n\t}()\n\n\tclose(p.quit)\n\tfor {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn ctx.Err()\n\t\tcase <-c:\n\t\t\treturn nil\n\t\t}\n\t}\n}", "func (d *Drummer) Stop() {\n\tplog.Infof(\"drummer's server ctx is going to be stopped\")\n\td.cancel()\n\td.electionManager.stop()\n\td.stopper.Stop()\n\taddr := d.nh.RaftAddress()\n\tplog.Debugf(\"going to stop the election manager, %s\", addr)\n\tplog.Debugf(\"election manager stopped on %s\", addr)\n\tif d.grpcServer != nil {\n\t\tplog.Debugf(\"grpc server will be stopped on %s\", addr)\n\t\td.grpcServer.Stop()\n\t\tplog.Debugf(\"grpc server stopped on %s\", addr)\n\t}\n}", "func (i *FanInReflect) CloseAsync() {\n\tfor _, input := range i.inputs {\n\t\tif closable, ok := input.(types.Closable); ok {\n\t\t\tclosable.CloseAsync()\n\t\t}\n\t}\n\tif atomic.CompareAndSwapInt32(&i.running, 1, 0) {\n\t\tclose(i.closeChan)\n\t}\n}", "func (node *DataNode) Stop() error {\n\tnode.stopOnce.Do(func() {\n\t\tnode.cancel()\n\t\t// https://github.com/milvus-io/milvus/issues/12282\n\t\tnode.UpdateStateCode(commonpb.StateCode_Abnormal)\n\t\tnode.flowgraphManager.close()\n\n\t\tnode.eventManagerMap.Range(func(_ string, m *channelEventManager) bool {\n\t\t\tm.Close()\n\t\t\treturn true\n\t\t})\n\n\t\tif node.allocator != nil {\n\t\t\tlog.Info(\"close id allocator\", zap.String(\"role\", typeutil.DataNodeRole))\n\t\t\tnode.allocator.Close()\n\t\t}\n\n\t\tif node.closer != nil {\n\t\t\tnode.closer.Close()\n\t\t}\n\n\t\tif node.session != nil {\n\t\t\tnode.session.Stop()\n\t\t}\n\n\t\tnode.wg.Wait()\n\t})\n\treturn nil\n}", "func (o *Switch) CloseAsync() {\n\to.close()\n}", "func (d *Dispatcher) Close() error {\n\td.inflightWg.Wait() // Wait for all inflight operations to complete\n\td.mu.Lock()\n\tif d.started {\n\t\tclose(d.results)\n\t\tclose(d.msgs)\n\t\td.started = false\n\t}\n\td.mu.Unlock()\n\td.wg.Wait() // Wait for all channel readers to return\n\treturn nil\n}", "func (sm *stateMachine) Close() error {\n\tif sm.running.CAS(true, false) {\n\t\tdefer func() {\n\t\t\tsm.cancel()\n\t\t}()\n\n\t\tsm.discovery.Close()\n\n\t\tsm.logger.Info(\"state machine stop successfully\",\n\t\t\tlogger.String(\"type\", sm.stateMachineType.String()))\n\t}\n\treturn nil\n}", "func runWithContext(fun func(ctx context.Context) error) (context.CancelFunc, chan error) {\n\tctx, cancel := context.WithCancel(context.Background())\n\tdone := make(chan error, 1)\n\tgo func() {\n\t\tdefer close(done)\n\t\tdone <- fun(ctx)\n\t}()\n\n\treturn cancel, done\n}", "func (svc *Service) Close(ctx context.Context) error {\n\ttimeout := svc.Cfg.ShutdownTimeout\n\tlog.Info(ctx, \"commencing graceful shutdown\", log.Data{\"graceful_shutdown_timeout\": timeout})\n\tctx, cancel := context.WithTimeout(ctx, timeout)\n\thasShutdownError := false\n\n\t// Gracefully shutdown the application closing any open resources.\n\tgo func() {\n\t\tdefer cancel()\n\n\t\t// stop healthcheck, as it depends on everything else\n\t\tif svc.HealthCheck != nil {\n\t\t\tsvc.HealthCheck.Stop()\n\t\t}\n\n\t\t// stop any incoming requests\n\t\tif svc.Server != nil {\n\t\t\tif err := svc.Server.Shutdown(ctx); err != nil {\n\t\t\t\tlog.Error(ctx, \"failed to shutdown http server\", err)\n\t\t\t\thasShutdownError = true\n\t\t\t}\n\t\t}\n\n\t\t// Close MongoDB (if it exists)\n\t\tif svc.FilterStore != nil {\n\t\t\tlog.Info(ctx, \"closing mongoDB filter data store\")\n\t\t\tif err := svc.FilterStore.Close(ctx); err != nil {\n\t\t\t\t// if err := mongolib.Close(ctx, svc.filterStore.Session); err != nil {\n\t\t\t\tlog.Error(ctx, \"unable to close mongo filter data store\", err)\n\t\t\t\thasShutdownError = true\n\t\t\t}\n\t\t}\n\n\t\t// Close Kafka Producer (it if exists)\n\t\tif svc.FilterOutputSubmittedProducer != nil {\n\t\t\tlog.Info(ctx, \"closing filter output submitted producer\")\n\t\t\tif err := svc.FilterOutputSubmittedProducer.Close(ctx); err != nil {\n\t\t\t\tlog.Error(ctx, \"unable to close filter output submitted producer\", err)\n\t\t\t\thasShutdownError = true\n\t\t\t}\n\t\t}\n\t}()\n\n\t// wait for shutdown success (via cancel) or failure (timeout)\n\t<-ctx.Done()\n\n\t// timeout expired\n\tif ctx.Err() == context.DeadlineExceeded {\n\t\tlog.Error(ctx, \"shutdown timed out\", ctx.Err())\n\t\treturn ctx.Err()\n\t}\n\n\t// other error\n\tif hasShutdownError {\n\t\terr := errors.New(\"failed to shutdown gracefully\")\n\t\tlog.Error(ctx, \"failed to shutdown gracefully \", err)\n\t\treturn err\n\t}\n\n\tlog.Info(ctx, \"graceful shutdown was successful\")\n\treturn nil\n}", "func (c *context) Close() (err error) {\n\tif c.child == nil {\n\t\treturn nil\n\t}\n\tdefer func() {\n\t\tc.child = nil\n\t\terr1 := os.RemoveAll(c.appDir)\n\t\tif err == nil {\n\t\t\terr = err1\n\t\t}\n\t}()\n\n\tif p := c.child.Process; p != nil {\n\t\terrc := make(chan error, 1)\n\t\tgo func() {\n\t\t\terrc <- c.child.Wait()\n\t\t}()\n\n\t\t// Call the quit handler on the admin server.\n\t\tres, err := http.Get(c.adminURL + \"/quit\")\n\t\tif err != nil {\n\t\t\tp.Kill()\n\t\t\treturn fmt.Errorf(\"unable to call /quit handler: %v\", err)\n\t\t}\n\t\tres.Body.Close()\n\n\t\tselect {\n\t\tcase <-time.After(15 * time.Second):\n\t\t\tp.Kill()\n\t\t\treturn errors.New(\"timeout killing child process\")\n\t\tcase err = <-errc:\n\t\t\t// Do nothing.\n\t\t}\n\t}\n\treturn\n}", "func (context Context) Close() {\n\tC.gfal2_context_free(context.cContext)\n\tcontext.cContext = nil\n}", "func UntilContext(ctx context.Context, d time.Duration) bool {\n\treturn Until(ctx.Done(), d)\n}", "func (l *Locker) Stop(_ context.Context) error {\n\tif err := l.ss.Close(); err != nil {\n\t\treturn err\n\t}\n\tif err := l.cli.Close(); err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(\"Stopped etcd locker\")\n\treturn nil\n}", "func (p *AbstractRunProvider) Close() error {\n\tp.SetRunning(false)\n\treturn nil\n}", "func (c *stoppableContext) stop() {\n\tc.stopOnce.Do(func() {\n\t\tclose(c.stopped)\n\t})\n\n\tc.stopWg.Wait()\n}", "func (f *Filler) Terminate(err error) {\n\tf.closeo.Do(func() {\n\t\tf.drop(err)\n\t\tclose(f.closeq)\n\t})\n}", "func ShutdownContext() {\n\tif aptly.EnableDebug {\n\t\tif context.fileMemProfile != nil {\n\t\t\tpprof.WriteHeapProfile(context.fileMemProfile)\n\t\t\tcontext.fileMemProfile.Close()\n\t\t\tcontext.fileMemProfile = nil\n\t\t}\n\t\tif context.fileCPUProfile != nil {\n\t\t\tpprof.StopCPUProfile()\n\t\t\tcontext.fileCPUProfile.Close()\n\t\t\tcontext.fileCPUProfile = nil\n\t\t}\n\t\tif context.fileMemProfile != nil {\n\t\t\tcontext.fileMemProfile.Close()\n\t\t\tcontext.fileMemProfile = nil\n\t\t}\n\t}\n\tif context.database != nil {\n\t\tcontext.database.Close()\n\t}\n\tif context.downloader != nil {\n\t\tcontext.downloader.Shutdown()\n\t}\n\tif context.progress != nil {\n\t\tcontext.progress.Shutdown()\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
RecordReqDuration records the duration of given operation in metrics system
func (pr PrometheusRecorder) RecordReqDuration(jenkinsService, operation string, code int, elapsedTime float64) { reportRequestDuration(jenkinsService, operation, code, elapsedTime) }
[ "func recordDuration(name string, hostname string, path string, method string, duration float64) {\n\trequestDurations.WithLabelValues(name, hostname, path, method).Observe(duration)\n}", "func (me *Metrics) RecordRequestTime(labels Labels, length time.Duration) {\n\t// Only record times for successful requests, as we don't have labels to screen out bad requests.\n\tif labels.RequestStatus == RequestStatusOK {\n\t\tme.RequestTimer.Update(length)\n\t}\n}", "func PromAddRequestDuration(sc int, m string, d time.Duration) {\n\tRequestDuration.With(prometheus.Labels{\"code\": fmt.Sprintf(\"%v\", sc), \"method\": m}).Set(d.Seconds())\n}", "func recordTiming(operation string, err *error) func() {\n\tstartTime := time.Now()\n\treturn func() {\n\t\tendTime := time.Since(startTime)\n\t\tendTimeMS := float64(endTime.Milliseconds())\n\t\tsuccess := \"true\"\n\t\tif *err != nil {\n\t\t\tsuccess = \"false\"\n\t\t}\n\t\toperationDurationInMsSummary.WithLabelValues(operation, success).Observe(endTimeMS)\n\t}\n}", "func recordOperation(operation string, start time.Time) {\n\tmetrics.RuntimeOperations.WithLabelValues(operation).Inc()\n\tmetrics.RuntimeOperationsDuration.WithLabelValues(operation).Observe(metrics.SinceInSeconds(start))\n}", "func TraceRequestTime(method, action, status string, elapsedTime float64) {\n\thisto.WithLabelValues(method, action, status).Observe(elapsedTime)\n}", "func (me *Metrics) RecordRequest(labels Labels) {\n\tme.RequestStatuses[labels.RType][labels.RequestStatus].Mark(1)\n\tif labels.Source == DemandApp {\n\t\tme.AppRequestMeter.Mark(1)\n\t} else {\n\t\tif labels.CookieFlag == CookieFlagNo {\n\t\t\t// NOTE: Old behavior was log me.AMPNoCookieMeter here for AMP requests.\n\t\t\t// AMP is still new and OpenRTB does not do this, so changing to match\n\t\t\t// OpenRTB endpoint\n\t\t\tme.NoCookieMeter.Mark(1)\n\t\t}\n\t}\n\n\t// Handle the account metrics now.\n\tam := me.getAccountMetrics(labels.PubID)\n\tam.requestMeter.Mark(1)\n}", "func (_m *Recorder) ObserveHTTPRequestDuration(id string, duration time.Duration, method string, code string) {\n\t_m.Called(id, duration, method, code)\n}", "func recordRequest(name string, hostname string, path string, method string) {\n\trequestsProcessed.WithLabelValues(name, hostname, path, method).Inc()\n}", "func observeRequestSimDuration(jobName string, extJobID uuid.UUID, vrfVersion vrfcommon.Version, pendingReqs []pendingRequest) {\n\tnow := time.Now().UTC()\n\tfor _, request := range pendingReqs {\n\t\t// First time around lastTry will be zero because the request has not been\n\t\t// simulated yet. It will be updated every time the request is simulated (in the event\n\t\t// the request is simulated multiple times, due to it being underfunded).\n\t\tif request.lastTry.IsZero() {\n\t\t\tvrfcommon.MetricTimeUntilInitialSim.\n\t\t\t\tWithLabelValues(jobName, extJobID.String(), string(vrfVersion)).\n\t\t\t\tObserve(float64(now.Sub(request.utcTimestamp)))\n\t\t} else {\n\t\t\tvrfcommon.MetricTimeBetweenSims.\n\t\t\t\tWithLabelValues(jobName, extJobID.String(), string(vrfVersion)).\n\t\t\t\tObserve(float64(now.Sub(request.lastTry)))\n\t\t}\n\t}\n}", "func (m *metricsReporter) ReportRequest(_ context.Context, startTime time.Time, action string, err error) {\n\tm.requestDurationMetric.With(kitmetrics.Field{Key: \"action\", Value: action}).Observe(time.Since(startTime))\n\tm.requestCounterMetric.With(kitmetrics.Field{Key: \"action\", Value: action}).Add(1)\n\tif err != nil {\n\t\tm.errorCounterMetric.With(kitmetrics.Field{Key: \"action\", Value: action}).Add(1)\n\t}\n}", "func (m *MetricsProvider) AddOperationTime(value time.Duration) {\n}", "func Request(requestID string, resource string, method string, statusCode int) {\n\tfmt.Println(\"METRIC\", request, requestID, resource, method, statusCode)\n}", "func (s *Stats) RequestDuration() (time.Duration, error) {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\tif s.start.IsZero() || s.end.IsZero() {\n\t\treturn time.Duration(0), fmt.Errorf(\"request is not completed\")\n\t}\n\n\treturn s.end.Sub(s.start), nil\n}", "func ReportLibRequestMetric(system, handler, method, status string, started time.Time) {\n\trequestsTotalLib.WithLabelValues(system, handler, method, status).Inc()\n\trequestLatencyLib.WithLabelValues(system, handler, method, status).Observe(time.Since(started).Seconds())\n}", "func RecordKMSOperationLatency(providerName, methodName string, duration time.Duration, err error) {\n\tKMSOperationsLatencyMetric.WithLabelValues(providerName, methodName, getErrorCode(err)).Observe(duration.Seconds())\n}", "func BenchmarkRecordReqCommand(b *testing.B) {\n\tw := newWorker()\n\n\tregister := &registerViewReq{views: []*View{view}, err: make(chan error, 1)}\n\tregister.handleCommand(w)\n\tif err := <-register.err; err != nil {\n\t\tb.Fatal(err)\n\t}\n\n\tconst tagCount = 10\n\tctxs := make([]context.Context, 0, tagCount)\n\tfor i := 0; i < tagCount; i++ {\n\t\tctx, _ := tag.New(context.Background(),\n\t\t\ttag.Upsert(k1, fmt.Sprintf(\"v%d\", i)),\n\t\t\ttag.Upsert(k2, fmt.Sprintf(\"v%d\", i)),\n\t\t\ttag.Upsert(k3, fmt.Sprintf(\"v%d\", i)),\n\t\t\ttag.Upsert(k4, fmt.Sprintf(\"v%d\", i)),\n\t\t\ttag.Upsert(k5, fmt.Sprintf(\"v%d\", i)),\n\t\t\ttag.Upsert(k6, fmt.Sprintf(\"v%d\", i)),\n\t\t\ttag.Upsert(k7, fmt.Sprintf(\"v%d\", i)),\n\t\t\ttag.Upsert(k8, fmt.Sprintf(\"v%d\", i)),\n\t\t)\n\t\tctxs = append(ctxs, ctx)\n\t}\n\n\tb.ReportAllocs()\n\tb.ResetTimer()\n\n\tfor i := 0; i < b.N; i++ {\n\t\trecord := &recordReq{\n\t\t\tms: []stats.Measurement{\n\t\t\t\tm.M(1),\n\t\t\t\tm.M(1),\n\t\t\t\tm.M(1),\n\t\t\t\tm.M(1),\n\t\t\t\tm.M(1),\n\t\t\t\tm.M(1),\n\t\t\t\tm.M(1),\n\t\t\t\tm.M(1),\n\t\t\t},\n\t\t\ttm: tag.FromContext(ctxs[i%len(ctxs)]),\n\t\t}\n\t\trecord.handleCommand(w)\n\t}\n}", "func RecordLatency(seconds float64) RecordOption {\n\treturn func(r *RecordStream) {\n\t\tr.createRequest.BufferFragSize = uint32(seconds*float64(r.createRequest.Rate)) * uint32(r.createRequest.Channels) * uint32(r.bytesPerSample)\n\t\tr.createRequest.BufferMaxLength = 2 * r.createRequest.BufferFragSize\n\t\tr.createRequest.AdjustLatency = true\n\t}\n}", "func ReportAPIRequestMetric(handler, method, status string, started time.Time) {\n\trequestsTotalAPI.WithLabelValues(handler, method, status).Inc()\n\trequestLatencyAPI.WithLabelValues(handler, method, status).Observe(time.Since(started).Seconds())\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Generates a random, 64bit token, encoded as base62.
func GenToken() (string, error) { var data [64]byte if _, err := io.ReadFull(rand.Reader, data[:]); err != nil { return "", err } return basex.Base62StdEncoding.EncodeToString(data[:]), nil }
[ "func TokenB64(n int) string {\n\tif n < 1 {\n\t\treturn \"\"\n\t}\n\trand.Seed(time.Now().UTC().UnixNano())\n\tbts := make([]byte, n)\n\trand.Read(bts)\n\treturn base64.StdEncoding.EncodeToString(bts)\n}", "func randToken() string {\n\tb := make([]byte, 32)\n\trand.Read(b)\n\treturn base64.StdEncoding.EncodeToString(b)\n}", "func generateToken() (string, time.Time) {\n now := time.Now();\n\n randData := make([]byte, TOKEN_RANDOM_BYTE_LEN);\n rand.Read(randData);\n\n timeBinary := make([]byte, 8);\n binary.LittleEndian.PutUint64(timeBinary, uint64(now.UnixNano() / 1000));\n\n tokenData := bytes.Join([][]byte{timeBinary, randData}, []byte{});\n\n return base64.URLEncoding.EncodeToString(tokenData), now;\n}", "func generateToken() string {\n\tcharset := \"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\"\n\trand.Seed(time.Now().UnixNano())\n\n\tchars := make([]byte, tokenLength)\n for i := range chars {\n chars[i] = charset[rand.Intn(len(charset))]\n }\n\n\tmsg := string(chars)\n\tfmt.Println(msg)\n\treturn msg\n}", "func generateRandomToken() string {\n\tsleep()\n\trandomBytes := generateRandomBytes()\n\treturn base64.StdEncoding.EncodeToString(randomBytes)\n}", "func NewUuid62() string {\n return ConvertUp(NewUuid(), base62alphabet)\n}", "func randToken(i int) string {\n\tb := make([]byte, i)\n\trand.Read(b)\n\treturn base64.StdEncoding.EncodeToString(b)\n}", "func generateSessionToken() string {\n\treturn strconv.FormatInt(rand.Int63(), 16)\n}", "func Token() string {\n bytes := make([]byte, 10)\n _, err := rand.Read(bytes)\n if err != nil {\n panic(err)\n }\n return hex.EncodeToString(bytes)\n}", "func randomBase64(n int) (string, error) {\n\tb, err := randomBytes(n)\n\treturn base64.URLEncoding.EncodeToString(b), err\n}", "func Base62(n int) string { return String(n, Base62Chars) }", "func GenerateToken() string {\n\trand.Seed(time.Now().UnixNano())\n\tletters := []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\")\n\ttokenBytes := make([]rune, 16)\n\tfor i := range tokenBytes {\n\t\ttokenBytes[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(tokenBytes)\n}", "func randomToken(n int) string {\n\t// buffer to store n bytes\n\tb := make([]byte, n)\n\n\t// read random bytes based on size of b\n\t_, err := rand.Read(b)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\t// convert buffer to URL friendly string\n\treturn base64.URLEncoding.EncodeToString(b)\n}", "func generateSessionToken() string {\n\t// DO NOT USE THIS IN PRODUCTION\n\treturn strconv.FormatInt(rand.Int63(), 16)\n}", "func createToken() (string, error) {\n\tc := 30\n\tb := make([]byte, c)\n\t_, err := rand.Read(b)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"could not create random bytes, %s\", err)\n\t}\n\treturn base64.StdEncoding.EncodeToString(b), nil\n}", "func GenerateBase64(length int) string {\n\tbase := new(big.Int)\n\tbase.SetString(\"64\", 10)\n\n\tbase64 := \"0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ-_\"\n\ttempKey := \"\"\n\tfor i := 0; i < length; i++ {\n\t\tindex, _ := rand.Int(rand.Reader, base)\n\t\ttempKey += string(base64[int(index.Int64())])\n\t}\n\treturn tempKey\n}", "func (t *Token) gen(tl TokenLifetime) (string, error) {\n\tif timeutil.Now().Before(t.NextAt.Time) {\n\t\treturn \"\", ErrTooManyTokens\n\t}\n\n\tv := uniuri.NewLenChars(uniuri.StdLen, _tokenChars)\n\n\th, err := bcrypt.GenerateFromPassword([]byte(v), bcrypt.DefaultCost)\n\tif err != nil {\n\t\t// unlikely to happen\n\t\treturn \"\", err\n\t}\n\n\tt.ExpiresAt = null.TimeFrom(timeutil.Now().Add(tl.Interval))\n\tt.NextAt = null.TimeFrom(timeutil.Now().Add(tl.Cooldown))\n\tt.Hash = h\n\n\treturn v, nil\n}", "func Base62String(length int) string {\n\treturn generateString(length, base62Alphabet)\n}", "func Uint64() uint64 { return globalRand.Uint64() }" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
CreateListFromDiscriminatorValue creates a new instance of the appropriate class based on discriminator value
func CreateListFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) { return NewList(), nil }
[ "func CreateChecklistItemFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewChecklistItem(), nil\n}", "func CreateBrowserSiteListFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewBrowserSiteList(), nil\n}", "func CreateAppListItemFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewAppListItem(), nil\n}", "func CreateSharepointIdsFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewSharepointIds(), nil\n}", "func CreateVulnerabilityFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewVulnerability(), nil\n}", "func CreateHashesFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewHashes(), nil\n}", "func CreateDetectionRuleFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewDetectionRule(), nil\n}", "func CreateDeviceCategoryFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewDeviceCategory(), nil\n}", "func CreateBundleFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewBundle(), nil\n}", "func CreateJsonFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewJson(), nil\n}", "func CreateIosDeviceFeaturesConfigurationFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewIosDeviceFeaturesConfiguration(), nil\n}", "func CreateCloudCommunicationsFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewCloudCommunications(), nil\n}", "func CreateSiteCollectionFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewSiteCollection(), nil\n}", "func CreateWin32LobAppRegistryRuleFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewWin32LobAppRegistryRule(), nil\n}", "func CreateSetFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewSet(), nil\n}", "func CreateSubCategoryFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewSubCategory(), nil\n}", "func CreateAppVulnerabilityManagedDeviceFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewAppVulnerabilityManagedDevice(), nil\n}", "func CreateWin32LobAppRegistryDetectionFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewWin32LobAppRegistryDetection(), nil\n}", "func CreateDeviceCompliancePolicyPolicySetItemFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewDeviceCompliancePolicyPolicySetItem(), nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetColumns gets the columns property value. The collection of field definitions for this list.
func (m *List) GetColumns()([]ColumnDefinitionable) { return m.columns }
[ "func (m *Site) GetColumns()([]ColumnDefinitionable) {\n val, err := m.GetBackingStore().Get(\"columns\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]ColumnDefinitionable)\n }\n return nil\n}", "func (m *Model) GetColumns() []Column {\n\treturn m.Columns\n}", "func (fmd *FakeMysqlDaemon) GetColumns(ctx context.Context, dbName, table string) ([]*querypb.Field, []string, error) {\n\treturn []*querypb.Field{}, []string{}, nil\n}", "func (c *Client) GetColumns(sheetID string) (cols []Column, err error) {\n\tpath := fmt.Sprintf(\"sheets/%v/columns\", sheetID)\n\n\tbody, err := c.Get(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer body.Close()\n\n\tvar resp PaginatedResponse\n\t//TODO: need generic handling and ability to read from pages to get all datc... eventually\n\tdec := json.NewDecoder(body)\n\tif err = dec.Decode(&resp); err != nil {\n\t\tlog.Fatalf(\"Failed to decode: %v\\n\", err)\n\t\treturn\n\t}\n\n\tif err = json.Unmarshal(resp.Data, &cols); err != nil {\n\t\tlog.Fatalf(\"Failed to decode data: %v\\n\", err)\n\t\treturn\n\t}\n\n\treturn\n}", "func (d *dbBase) GetColumns(ctx context.Context, db dbQuerier, table string) (map[string][3]string, error) {\n\tcolumns := make(map[string][3]string)\n\tquery := d.ins.ShowColumnsQuery(table)\n\trows, err := db.QueryContext(ctx, query)\n\tif err != nil {\n\t\treturn columns, err\n\t}\n\n\tdefer rows.Close()\n\n\tfor rows.Next() {\n\t\tvar (\n\t\t\tname string\n\t\t\ttyp string\n\t\t\tnull string\n\t\t)\n\t\terr := rows.Scan(&name, &typ, &null)\n\t\tif err != nil {\n\t\t\treturn columns, err\n\t\t}\n\t\tcolumns[name] = [3]string{name, typ, null}\n\t}\n\n\treturn columns, nil\n}", "func (_m *RepositoryMock) GetColumns() []string {\n\tret := _m.Called()\n\n\tvar r0 []string\n\tif rf, ok := ret.Get(0).(func() []string); ok {\n\t\tr0 = rf()\n\t} else {\n\t\tif ret.Get(0) != nil {\n\t\t\tr0 = ret.Get(0).([]string)\n\t\t}\n\t}\n\n\treturn r0\n}", "func (m *Site) GetExternalColumns()([]ColumnDefinitionable) {\n val, err := m.GetBackingStore().Get(\"externalColumns\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]ColumnDefinitionable)\n }\n return nil\n}", "func GetColumns(tmap *gorp.TableMap) (columns []string) {\n\n\tfor index, _ := range tmap.Columns {\n\t\tcolumns = append(columns, tmap.Columns[index].ColumnName)\n\t}\n\treturn columns\n}", "func (d Dataset) GetColumns(a *config.AppContext) ([]models.Node, error) {\n\tds := models.Dataset(d)\n\treturn ds.GetColumns(a.Db)\n}", "func (qr *QueryResponse) Columns() []ColumnItem {\n\treturn qr.ColumnList\n}", "func (*__tbl_iba_servers) GetColumns() []string {\n\treturn []string{\"id\", \"name\", \"zex\", \"stort\", \"comment\"}\n}", "func (*__tbl_nodes) GetColumns() []string {\n\treturn []string{\"id\", \"parent_id\", \"description\", \"comment\", \"meta\", \"full_name\", \"directory_id\", \"signal_id\", \"created_at\", \"updated_at\", \"acl\"}\n}", "func (pg *PgSQL) GetColumns(table *Table) ([]*Column, error) {\n\tif err := pg.Db.Ping(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tquery := \"select column_name, column_default, is_nullable, data_type from information_schema.columns where table_schema = $1 and table_name = $2\"\n\trows, err := pg.Db.Query(query, table.Schema, table.Name)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcolumns := []*Column{}\n\n\tfor rows.Next() {\n\t\tc := new(Column)\n\t\ttmp := struct {\n\t\t\tName sql.NullString\n\t\t\tDefault sql.NullString\n\t\t\tNullable sql.NullString\n\t\t\tType sql.NullString\n\t\t}{}\n\n\t\terr := rows.Scan(&tmp.Name, &tmp.Default, &tmp.Nullable, &tmp.Type)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tc.Name = nullableToString(tmp.Name)\n\t\tc.Default = nullableToString(tmp.Default)\n\t\tc.Nullable = nullableToBool(tmp.Nullable)\n\t\tc.Type = nullableToString(tmp.Type)\n\n\t\tcolumns = append(columns, c)\n\t}\n\n\treturn columns, nil\n}", "func (s *DbRecorder) Columns(includeKeys bool) []string {\n\treturn s.colList(includeKeys, false)\n}", "func (*__tbl_sources) GetColumns() []string {\n\treturn []string{\"id\", \"parent_id\", \"name\", \"source_classes_id\", \"comment\"}\n}", "func (expr *MatchAgainst) ColumnsAsGetFields() []*GetField {\n\tvar ok bool\n\tfields := make([]*GetField, len(expr.Columns))\n\tfor i, col := range expr.Columns {\n\t\tfields[i], ok = col.(*GetField)\n\t\tif !ok {\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn fields\n}", "func (driver *SQLite3) GetColumns(db *sql.DB, t bartlett.Table) ([]string, error) {\n\tif driver.tables == nil {\n\t\tdriver.tables = make(map[string][]column)\n\t}\n\tvar (\n\t\tcreateQuery string\n\t\tout []string\n\t)\n\trows, err := sqrl.Select(`sql`).From(`sqlite_master`).Where(`name = ?`, t.Name).RunWith(db).Query()\n\tif err != nil {\n\t\treturn []string{}, err\n\t}\n\n\trows.Next() // We should only expect a single row here.\n\terr = rows.Scan(&createQuery)\n\tif err != nil {\n\t\treturn []string{}, err\n\t}\n\n\tdriver.tables[t.Name] = parseCreateTable(createQuery)\n\tfor _, col := range driver.tables[t.Name] {\n\t\tout = append(out, col.name)\n\t}\n\n\treturn out, err\n}", "func (ts *STableSpec) Columns() []IColumnSpec {\n\tif ts._columns == nil {\n\t\tval := reflect.Indirect(reflect.New(ts.structType))\n\t\tts.struct2TableSpec(val)\n\t}\n\treturn ts._columns\n}", "func (qr *QueryResult) Columns() []string {\n\treturn qr.columns\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetContentTypes gets the contentTypes property value. The collection of content types present in this list.
func (m *List) GetContentTypes()([]ContentTypeable) { return m.contentTypes }
[ "func (o *LastMileAccelerationOptions) GetContentTypes() []string {\n\tif o == nil || o.ContentTypes == nil {\n\t\tvar ret []string\n\t\treturn ret\n\t}\n\treturn *o.ContentTypes\n}", "func (m *Site) GetContentTypes()([]ContentTypeable) {\n val, err := m.GetBackingStore().Get(\"contentTypes\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]ContentTypeable)\n }\n return nil\n}", "func (web *Web) ContentTypes() *ContentTypes {\n\treturn NewContentTypes(\n\t\tweb.client,\n\t\tfmt.Sprintf(\"%s/ContentTypes\", web.endpoint),\n\t\tweb.config,\n\t)\n}", "func (m *List) SetContentTypes(value []ContentTypeable)() {\n m.contentTypes = value\n}", "func (set *ContentTypeSet) Types() (types []ContentType) {\n\tif set == nil || len(set.set) == 0 {\n\t\treturn []ContentType{}\n\t}\n\treturn append(make([]ContentType, 0, len(set.set)), set.set...)\n}", "func (ht HTMLContentTypeBinder) ContentTypes() []string {\n\treturn []string{\n\t\t\"application/html\",\n\t\t\"text/html\",\n\t\t\"application/x-www-form-urlencoded\",\n\t\t\"html\",\n\t}\n}", "func (_BaseAccessControlGroup *BaseAccessControlGroupCaller) ContentTypes(opts *bind.CallOpts) (struct {\n\tCategory uint8\n\tLength *big.Int\n}, error) {\n\tvar out []interface{}\n\terr := _BaseAccessControlGroup.contract.Call(opts, &out, \"contentTypes\")\n\n\toutstruct := new(struct {\n\t\tCategory uint8\n\t\tLength *big.Int\n\t})\n\tif err != nil {\n\t\treturn *outstruct, err\n\t}\n\n\toutstruct.Category = *abi.ConvertType(out[0], new(uint8)).(*uint8)\n\toutstruct.Length = *abi.ConvertType(out[1], new(*big.Int)).(**big.Int)\n\n\treturn *outstruct, err\n\n}", "func (c *Contentful) Types() (rc []byte, err error) {\n\turl := fmt.Sprintf(contentURLTemplate, c.ReadConfig.SpaceID, c.ReadConfig.Environment, c.ReadConfig.AccessToken, c.ReadConfig.Locale)\n\treturn c.get(url)\n}", "func (ycp *YamlContentParser) ContentTypes() []string {\n\treturn []string{\"text/x-yaml\", \"application/yaml\", \"text/yaml\", \"application/x-yaml\"}\n}", "func (m *Site) SetContentTypes(value []ContentTypeable)() {\n err := m.GetBackingStore().Set(\"contentTypes\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *SiteItemRequestBuilder) ContentTypes()(*ItemContentTypesRequestBuilder) {\n return NewItemContentTypesRequestBuilderInternal(m.pathParameters, m.requestAdapter)\n}", "func ContentTypes(types []string, blacklist bool) Option {\n\treturn func(c *config) error {\n\t\tc.contentTypes = []parsedContentType{}\n\t\tfor _, v := range types {\n\t\t\tmediaType, params, err := mime.ParseMediaType(v)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tc.contentTypes = append(c.contentTypes, parsedContentType{mediaType, params})\n\t\t}\n\t\tc.blacklist = blacklist\n\t\treturn nil\n\t}\n}", "func ContentTypes(contentTypes ...string) Option {\n\treturn ArrayOpt(\"content_types\", contentTypes...)\n}", "func (_AccessIndexor *AccessIndexorCaller) ContentTypes(opts *bind.CallOpts) (struct {\n\tCategory uint8\n\tLength *big.Int\n}, error) {\n\tvar out []interface{}\n\terr := _AccessIndexor.contract.Call(opts, &out, \"contentTypes\")\n\n\toutstruct := new(struct {\n\t\tCategory uint8\n\t\tLength *big.Int\n\t})\n\tif err != nil {\n\t\treturn *outstruct, err\n\t}\n\n\toutstruct.Category = *abi.ConvertType(out[0], new(uint8)).(*uint8)\n\toutstruct.Length = *abi.ConvertType(out[1], new(*big.Int)).(**big.Int)\n\n\treturn *outstruct, err\n\n}", "func (web *Web) AvailableContentTypes() *ContentTypes {\n\treturn NewContentTypes(\n\t\tweb.client,\n\t\tfmt.Sprintf(\"%s/AvailableContentTypes\", web.endpoint),\n\t\tweb.config,\n\t)\n}", "func (_BaseLibrary *BaseLibraryCaller) ContentTypes(opts *bind.CallOpts, arg0 *big.Int) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseLibrary.contract.Call(opts, &out, \"contentTypes\", arg0)\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func ContentCategoryType_Values() []string {\n\treturn []string{\n\t\tContentCategoryTypeImage,\n\t\tContentCategoryTypeDocument,\n\t\tContentCategoryTypePdf,\n\t\tContentCategoryTypeSpreadsheet,\n\t\tContentCategoryTypePresentation,\n\t\tContentCategoryTypeAudio,\n\t\tContentCategoryTypeVideo,\n\t\tContentCategoryTypeSourceCode,\n\t\tContentCategoryTypeOther,\n\t}\n}", "func ContentType_Values() []string {\n\treturn []string{\n\t\tContentTypeApplicationJson,\n\t\tContentTypeApplicationXml,\n\t}\n}", "func SetOfContentTypes(types ...ContentType) *ContentTypeSet {\n\tif len(types) == 0 {\n\t\treturn nil\n\t}\n\tset := &ContentTypeSet{\n\t\tset: make([]ContentType, 0, len(types)),\n\t\tpos: -1,\n\t}\nallTypes:\n\tfor _, t := range types {\n\t\t// Let's make sure we have not seen this type before.\n\t\tfor _, tt := range set.set {\n\t\t\tif tt == t {\n\t\t\t\t// Don't add it to the set, already exists\n\t\t\t\tcontinue allTypes\n\t\t\t}\n\t\t}\n\t\tset.set = append(set.set, t)\n\t}\n\tif len(set.set) == 0 {\n\t\treturn nil\n\t}\n\treturn set\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetDisplayName gets the displayName property value. The displayable title of the list.
func (m *List) GetDisplayName()(*string) { return m.displayName }
[ "func (m *UnifiedRoleAssignmentMultiple) GetDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"displayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *DeviceAndAppManagementAssignmentFilter) GetDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"displayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *ChecklistItem) GetDisplayName()(*string) {\n return m.displayName\n}", "func (s UserSet) DisplayName() string {\n\tres, _ := s.RecordCollection.Get(models.NewFieldName(\"DisplayName\", \"display_name\")).(string)\n\treturn res\n}", "func (m *BrowserSiteList) GetDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"displayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *UnifiedRoleDefinition) GetDisplayName()(*string) {\n return m.displayName\n}", "func (o *User) GetDisplayName() string {\n\tif o == nil || o.DisplayName == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.DisplayName\n}", "func (m *CompanyDetail) GetDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"displayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *IdentityProviderBase) GetDisplayName()(*string) {\n return m.displayName\n}", "func (o *SubscriptionRegistration) GetDisplayName() (value string, ok bool) {\n\tok = o != nil && o.bitmap_&4 != 0\n\tif ok {\n\t\tvalue = o.displayName\n\t}\n\treturn\n}", "func (m *TeamsApp) GetDisplayName()(*string) {\n return m.displayName\n}", "func (m *Group) GetDisplayName()(*string) {\n return m.displayName\n}", "func (m *RoleDefinition) GetDisplayName()(*string) {\n return m.displayName\n}", "func (m *DeviceCategory) GetDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"displayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *RelatedContact) GetDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"displayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (a *Action) GetDisplayName() string {\n\tif setting.UI.DefaultShowFullName {\n\t\ttrimmedFullName := strings.TrimSpace(a.GetActFullName())\n\t\tif len(trimmedFullName) > 0 {\n\t\t\treturn trimmedFullName\n\t\t}\n\t}\n\treturn a.ShortActUserName()\n}", "func (m *RemoteAssistancePartner) GetDisplayName()(*string) {\n return m.displayName\n}", "func (m *EducationRubric) GetDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"displayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *WorkforceIntegration) GetDisplayName()(*string) {\n return m.displayName\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetDrive gets the drive property value. Only present on document libraries. Allows access to the list as a [drive][] resource with [driveItems][driveItem].
func (m *List) GetDrive()(Driveable) { return m.drive }
[ "func (o *Block) GetDrive(ctx context.Context) (drive dbus.ObjectPath, err error) {\n\terr = o.object.CallWithContext(ctx, \"org.freedesktop.DBus.Properties.Get\", 0, InterfaceBlock, \"Drive\").Store(&drive)\n\treturn\n}", "func (m *Site) GetDrive()(Driveable) {\n val, err := m.GetBackingStore().Get(\"drive\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(Driveable)\n }\n return nil\n}", "func (m *Group) GetDrive()(Driveable) {\n return m.drive\n}", "func (m *User) GetDrive()(Driveable) {\n return m.drive\n}", "func (o *User) GetDrive() Drive {\n\tif o == nil || o.Drive == nil {\n\t\tvar ret Drive\n\t\treturn ret\n\t}\n\treturn *o.Drive\n}", "func (m *ItemReference) GetDriveType()(*string) {\n return m.driveType\n}", "func (m *ItemReference) GetDriveId()(*string) {\n return m.driveId\n}", "func (svc *StorageServiceOp) GetDrive(id string) (*Drive, *Response, error) {\n\tpath := fmt.Sprintf(\"%s/%s\", storageAllocPath, id)\n\n\treq, err := svc.client.NewRequest(\"GET\", path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(storageDriveRoot)\n\tresp, err := svc.client.Do(req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn root.Drive, resp, nil\n}", "func (ref FileView) Drive() resource.ID {\n\treturn ref.drive\n}", "func (repo Repository) Drive(driveID resource.ID) drivestream.DriveReference {\n\treturn Drive{\n\t\tdb: repo.db,\n\t\tdrive: driveID,\n\t}\n}", "func (o *MicrosoftGraphItemReference) GetDriveId() string {\n\tif o == nil || o.DriveId == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.DriveId\n}", "func (o *MicrosoftGraphItemReference) GetDriveType() string {\n\tif o == nil || o.DriveType == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.DriveType\n}", "func (ref Commit) Drive() resource.ID {\n\treturn ref.drive\n}", "func (o *StorageFlexUtilVirtualDrive) GetDriveType() string {\n\tif o == nil || o.DriveType == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.DriveType\n}", "func (device *PerformanceDCBricklet) GetDriveMode() (mode DriveMode, err error) {\n\tvar buf bytes.Buffer\n\n\tresultBytes, err := device.device.Get(uint8(FunctionGetDriveMode), buf.Bytes())\n\tif err != nil {\n\t\treturn mode, err\n\t}\n\tif len(resultBytes) > 0 {\n\t\tvar header PacketHeader\n\n\t\theader.FillFromBytes(resultBytes)\n\n\t\tif header.Length != 9 {\n\t\t\treturn mode, fmt.Errorf(\"Received packet of unexpected size %d, instead of %d\", header.Length, 9)\n\t\t}\n\n\t\tif header.ErrorCode != 0 {\n\t\t\treturn mode, DeviceError(header.ErrorCode)\n\t\t}\n\n\t\tresultBuf := bytes.NewBuffer(resultBytes[8:])\n\t\tbinary.Read(resultBuf, binary.LittleEndian, &mode)\n\n\t}\n\n\treturn mode, nil\n}", "func (m *Site) GetDrives()([]Driveable) {\n val, err := m.GetBackingStore().Get(\"drives\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]Driveable)\n }\n return nil\n}", "func (a *HyperflexApiService) GetHyperflexDriveList(ctx context.Context) ApiGetHyperflexDriveListRequest {\n\treturn ApiGetHyperflexDriveListRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (m *User) GetDrives()([]Driveable) {\n return m.drives\n}", "func (device *DCV2Bricklet) GetDriveMode() (mode DriveMode, err error) {\n\tvar buf bytes.Buffer\n\n\tresultBytes, err := device.device.Get(uint8(FunctionGetDriveMode), buf.Bytes())\n\tif err != nil {\n\t\treturn mode, err\n\t}\n\tif len(resultBytes) > 0 {\n\t\tvar header PacketHeader\n\n\t\theader.FillFromBytes(resultBytes)\n\n\t\tif header.Length != 9 {\n\t\t\treturn mode, fmt.Errorf(\"Received packet of unexpected size %d, instead of %d\", header.Length, 9)\n\t\t}\n\n\t\tif header.ErrorCode != 0 {\n\t\t\treturn mode, DeviceError(header.ErrorCode)\n\t\t}\n\n\t\tresultBuf := bytes.NewBuffer(resultBytes[8:])\n\t\tbinary.Read(resultBuf, binary.LittleEndian, &mode)\n\n\t}\n\n\treturn mode, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetItems gets the items property value. All items contained in the list.
func (m *List) GetItems()([]ListItemable) { return m.items }
[ "func (l *RestAPIList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (m *Site) GetItems()([]BaseItemable) {\n val, err := m.GetBackingStore().Get(\"items\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]BaseItemable)\n }\n return nil\n}", "func (out Outlooky) GetItems(folder *ole.IDispatch) *ole.IDispatch {\n\titems, err := out.CallMethod(folder, \"Items\")\n\tutil.Catch(err, \"Failed retrieving items.\")\n\n\treturn items.ToIDispatch()\n}", "func (l *IntegrationList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (api *API) ItemsGet(params Params) (res Items, err error) {\n\tif _, present := params[\"output\"]; !present {\n\t\tparams[\"output\"] = \"extend\"\n\t}\n\tresponse, err := api.CallWithError(\"item.get\", params)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treflector.MapsToStructs2(response.Result.([]interface{}), &res, reflector.Strconv, \"json\")\n\treturn\n}", "func (l *HealthCheckList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *AuthorizerList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (c *Client) GetItems(ctx context.Context, owner string) (Items, error) {\n\tresponse, err := c.sendRequest(ctx, owner, http.MethodGet, fmt.Sprintf(\"%s/%s\", c.storeBaseURL, c.bucket), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif response.Code != http.StatusOK {\n\t\tlevel.Error(c.getLogger(ctx)).Log(xlog.MessageKey(), \"Argus responded with non-200 response for GetItems request\",\n\t\t\t\"code\", response.Code, \"ErrorHeader\", response.ArgusErrorHeader)\n\t\treturn nil, fmt.Errorf(errStatusCodeFmt, response.Code, translateNonSuccessStatusCode(response.Code))\n\t}\n\n\tvar items Items\n\n\terr = json.Unmarshal(response.Body, &items)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"GetItems: %w: %s\", errJSONUnmarshal, err.Error())\n\t}\n\n\treturn items, nil\n}", "func (l *DomainNameList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *DeploymentList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *DocumentationVersionList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *ResourceList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *StageList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *IntegrationResponseList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *DocumentationPartList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (c *Client) GetItems(id int) (Item, error) {\n\tc.defaultify()\n\tvar item Item\n\tresp, err := http.Get(fmt.Sprintf(\"%s/item/%d.json\", c.apiBase, id))\n\tif err != nil {\n\t\treturn item, err\n\t}\n\tdefer resp.Body.Close()\n\tdec := json.NewDecoder(resp.Body)\n\terr = dec.Decode(&item)\n\tif err != nil {\n\t\treturn item, err\n\t}\n\treturn item, nil\n}", "func (l *MethodList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *ModelList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (l *MethodResponseList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetList gets the list property value. Provides additional details about the list.
func (m *List) GetList()(ListInfoable) { return m.list }
[ "func (client *Client) GetList(listID string) ([]Item, error) {\n\tbody, err := client.call(command{\"getList\", client.teamID, listID, \"\"})\n\treturn HandleGetList(body, err)\n}", "func (ls *membershipListStruct) GetList() []string {\n\tvar tmp []string\n\tls.mlLock.Lock()\n\tfor ID, entry := range ls.list {\n\t\tif entry.memIDListIndx != -1 {\n\t\t\ttmp = append(tmp, ID)\n\t\t}\n\t}\n\tls.mlLock.Unlock()\n\n\treturn tmp\n}", "func (a *Client) GetList(params *GetListParams, authInfo runtime.ClientAuthInfoWriter) (*GetListOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetListParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getList\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/contacts/lists/{listId}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetListReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetListOK), nil\n\n}", "func (client *ClientImpl) GetList(ctx context.Context, args GetListArgs) (*PickList, error) {\n\trouteValues := make(map[string]string)\n\tif args.ListId == nil {\n\t\treturn nil, &azuredevops.ArgumentNilError{ArgumentName: \"args.ListId\"}\n\t}\n\trouteValues[\"listId\"] = (*args.ListId).String()\n\n\tlocationId, _ := uuid.Parse(\"01e15468-e27c-4e20-a974-bd957dcccebc\")\n\tresp, err := client.Client.Send(ctx, http.MethodGet, locationId, \"6.0-preview.1\", routeValues, nil, nil, \"\", \"application/json\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar responseValue PickList\n\terr = client.Client.UnmarshalBody(resp, &responseValue)\n\treturn &responseValue, err\n}", "func (web *Web) GetList(listURI string) *List {\n\treturn NewList(\n\t\tweb.client,\n\t\tfmt.Sprintf(\"%s/GetList('%s')\", web.endpoint, checkGetRelativeURL(listURI, web.endpoint)),\n\t\tweb.config,\n\t)\n}", "func GetList(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ListState, opts ...pulumi.ResourceOption) (*List, error) {\n\tvar resource List\n\terr := ctx.ReadResource(\"aws-native:frauddetector:List\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (l *Lists) GetList(id string) (*ListOutput, error) {\n\tif id == \"\" {\n\t\treturn nil, errors.New(\"goengage: id is required\")\n\t}\n\n\treq, err := l.client.newRequest(http.MethodGet, fmt.Sprintf(\"/lists/%v\", id), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar output ListOutput\n\terr = l.client.makeRequest(req, &output)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &output, err\n}", "func (a *alphaMock) GetList(ctx context.Context, in *alpha.GetListRequest, opts ...grpc.CallOption) (*alpha.List, error) {\n\t// TODO(#2716): Implement me!\n\treturn nil, errors.Errorf(\"Unimplemented -- GetList coming soon\")\n}", "func GetList(w http.ResponseWriter, r *http.Request) {\n\tlist, err := watchlist.GetWatchList(r)\n\n\tif err != nil {\n\t\thttpext.AbortAPI(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\thttpext.JSON(w, list)\n}", "func (c *Client) GetList(board *trello.Board, name string) (*trello.List, error) {\n\tlists, err := board.GetLists(trello.Defaults())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, list := range lists {\n\t\tif list.Name == name {\n\t\t\treturn list, nil\n\t\t}\n\t}\n\treturn nil, fmt.Errorf(\"could not find list %s on board %s\", name, board.Name)\n}", "func (n NestedInteger) GetList() []*NestedInteger {\n\treturn nil\n}", "func (m *Drive) GetList()(Listable) {\n return m.list\n}", "func (pn *paxosNode) GetList(args *paxosrpc.GetListArgs, reply *paxosrpc.GetListReply) error {\n\tif value := pn.storage.GetList(args.Key); value != nil {\n\t\treply.L = value\n\t\treply.Status = paxosrpc.KeyFound\n\t} else {\n\t\treply.Status = paxosrpc.KeyNotFound\n\t}\n\treturn nil\n}", "func GetList(c Context) {\n\tres, err := db.SelectAllItems()\n\tif err != nil {\n\t\tc.JSON(http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, res)\n}", "func GetList() []string {\n\tvar reportList []string\n\tif config.UsingDB() {\n\t\treportList = getListSQL()\n\t} else {\n\t\tds := datastore.New()\n\t\treportList = ds.GetList(\"report\")\n\t}\n\treturn reportList\n}", "func (lr *ListResponse) GetList() [][]string {\n\treturn lr.rows\n}", "func (ctl *SaleCounterProductController) GetList() {\n\tviewType := ctl.Input().Get(\"view\")\n\tif viewType == \"\" || viewType == \"table\" {\n\t\tctl.Data[\"ViewType\"] = \"table\"\n\t}\n\tctl.PageAction = utils.MsgList\n\tctl.Data[\"tableId\"] = \"table-sale-counter-product\"\n\tctl.Layout = \"base/base_list_view.html\"\n\tctl.TplName = \"sale/sale_counter_product_list_search.html\"\n}", "func (client *GroupMgmtClient) listGet(\n\tpath string,\n\tqueryParams map[string]string,\n\tparams *param.GetParams,\n) (interface{}, error) {\n\t// build the url\n\turl := fmt.Sprintf(\"%s/%s/detail\", client.URL, path)\n\n\tresponse, err := client.Client.R().\n\t\tSetQueryParams(queryParams).\n\t\tSetHeader(\"X-Auth-Token\", client.SessionToken).\n\t\tGet(url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn processResponse(client, response, path, nil, params)\n}", "func getList(w io.Writer, r *http.Request) error {\n\tc := appengine.NewContext(r)\n\n\t// Get the list id from the URL.\n\tid := mux.Vars(r)[\"list\"]\n\n\t// Decode the obtained id into a datastore key.\n\tkey, err := datastore.DecodeKey(id)\n\tif err != nil {\n\t\treturn appErrorf(http.StatusBadRequest, \"invalid list id\")\n\t}\n\n\t// Fetch the list from the datastore.\n\tlist := &List{}\n\terr = datastore.Get(c, key, list)\n\tif err == datastore.ErrNoSuchEntity {\n\t\treturn appErrorf(http.StatusNotFound, \"list not found\")\n\t}\n\tif err != nil {\n\t\treturn fmt.Errorf(\"fetch list: %v\", err)\n\t}\n\n\t// Set the ID field with the id from the request url and encode the list.\n\tlist.ID = id\n\treturn json.NewEncoder(w).Encode(&list)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetOperations gets the operations property value. The collection of longrunning operations on the list.
func (m *List) GetOperations()([]RichLongRunningOperationable) { return m.operations }
[ "func (m *Site) GetOperations()([]RichLongRunningOperationable) {\n val, err := m.GetBackingStore().Get(\"operations\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]RichLongRunningOperationable)\n }\n return nil\n}", "func (o *MicrosoftGraphOnenote) GetOperations() []MicrosoftGraphOnenoteOperation {\n\tif o == nil || o.Operations == nil {\n\t\tvar ret []MicrosoftGraphOnenoteOperation\n\t\treturn ret\n\t}\n\treturn *o.Operations\n}", "func (e *Endpoint) GetOperations() []models.EndpointOperation {\n\treturn e.Operations\n}", "func GetOperations() ([]dtos.Operation, error) {\n\tvar ops []dtos.Operation\n\n\tresp, err := makeJSONRequest(\"GET\", apiURL+\"/operations\", http.NoBody)\n\tif err != nil {\n\t\treturn ops, errors.Append(err, ErrCannotConnect)\n\t}\n\n\tif err = evaluateResponseStatusCode(resp.StatusCode); err != nil {\n\t\treturn ops, err\n\t}\n\n\terr = readResponseBody(&ops, resp.Body)\n\n\treturn ops, err\n}", "func (m *Microservice) GetOperations(status string) (*c8y.OperationCollection, *c8y.Response, error) {\n\topt := &c8y.OperationCollectionOptions{\n\t\tStatus: status,\n\t\tAgentID: m.AgentID,\n\t\tPaginationOptions: c8y.PaginationOptions{\n\t\t\tPageSize: 5,\n\t\t\tWithTotalPages: false,\n\t\t},\n\t}\n\n\tdata, resp, err := m.Client.Operation.GetOperations(m.WithServiceUser(), opt)\n\treturn data, resp, err\n}", "func (m *Store) GetOperations(service string) ([]string, error) {\n\tm.RLock()\n\tdefer m.RUnlock()\n\tif operations, ok := m.operations[service]; ok {\n\t\tvar retMe []string\n\t\tfor ops := range operations {\n\t\t\tretMe = append(retMe, ops)\n\t\t}\n\t\treturn retMe, nil\n\t}\n\treturn []string{}, nil\n}", "func (client BaseClient) ListOperations(ctx context.Context) (result Operations, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/BaseClient.ListOperations\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.Response.Response != nil {\n\t\t\t\tsc = result.Response.Response.StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\treq, err := client.ListOperationsPreparer(ctx)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"serialconsole.BaseClient\", \"ListOperations\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.ListOperationsSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"serialconsole.BaseClient\", \"ListOperations\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.ListOperationsResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"serialconsole.BaseClient\", \"ListOperations\", resp, \"Failure responding to request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func (c *restClient) ListOperations(ctx context.Context, req *longrunningpb.ListOperationsRequest, opts ...gax.CallOption) *OperationIterator {\n\tit := &OperationIterator{}\n\treq = proto.Clone(req).(*longrunningpb.ListOperationsRequest)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tit.InternalFetch = func(pageSize int, pageToken string) ([]*longrunningpb.Operation, string, error) {\n\t\tresp := &longrunningpb.ListOperationsResponse{}\n\t\tif pageToken != \"\" {\n\t\t\treq.PageToken = pageToken\n\t\t}\n\t\tif pageSize > math.MaxInt32 {\n\t\t\treq.PageSize = math.MaxInt32\n\t\t} else if pageSize != 0 {\n\t\t\treq.PageSize = int32(pageSize)\n\t\t}\n\t\tbaseUrl, err := url.Parse(c.endpoint)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v/operations\", req.GetName())\n\n\t\tparams := url.Values{}\n\t\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\t\tif req.GetFilter() != \"\" {\n\t\t\tparams.Add(\"filter\", fmt.Sprintf(\"%v\", req.GetFilter()))\n\t\t}\n\t\tif req.GetPageSize() != 0 {\n\t\t\tparams.Add(\"pageSize\", fmt.Sprintf(\"%v\", req.GetPageSize()))\n\t\t}\n\t\tif req.GetPageToken() != \"\" {\n\t\t\tparams.Add(\"pageToken\", fmt.Sprintf(\"%v\", req.GetPageToken()))\n\t\t}\n\n\t\tbaseUrl.RawQuery = params.Encode()\n\n\t\t// Build HTTP headers from client and context metadata.\n\t\thds := append(c.xGoogHeaders, \"Content-Type\", \"application/json\")\n\t\theaders := gax.BuildHeaders(ctx, hds...)\n\t\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\t\tif settings.Path != \"\" {\n\t\t\t\tbaseUrl.Path = settings.Path\n\t\t\t}\n\t\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\thttpReq.Header = headers\n\n\t\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer httpRsp.Body.Close()\n\n\t\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}, opts...)\n\t\tif e != nil {\n\t\t\treturn nil, \"\", e\n\t\t}\n\t\tit.Response = resp\n\t\treturn resp.GetOperations(), resp.GetNextPageToken(), nil\n\t}\n\n\tfetch := func(pageSize int, pageToken string) (string, error) {\n\t\titems, nextPageToken, err := it.InternalFetch(pageSize, pageToken)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tit.items = append(it.items, items...)\n\t\treturn nextPageToken, nil\n\t}\n\n\tit.pageInfo, it.nextFunc = iterator.NewPageInfo(fetch, it.bufLen, it.takeBuf)\n\tit.pageInfo.MaxSize = int(req.GetPageSize())\n\tit.pageInfo.Token = req.GetPageToken()\n\n\treturn it\n}", "func (c *Controller) GetOperations() []operation.Handler {\n\treturn c.handlers\n}", "func (r *SpanReader) GetOperations(ctx context.Context, service string) ([]string, error){\n\treturn r.cache.LoadOperations(service)\n}", "func (m *Workbook) GetOperations()([]WorkbookOperationable) {\n return m.operations\n}", "func (m *List) SetOperations(value []RichLongRunningOperationable)() {\n m.operations = value\n}", "func (so *Operations) Operations() api.Operations {\n\treturn api.Operations(so)\n}", "func (m *Master) GetOperations(session *gocql.Session, hostname string) ([]ops.Operation, error) {\n\tvar operations []ops.Operation\n\tvar description, scriptName string\n\tvar attributes map[string]string\n\tq := `SELECT description, script_name, attributes FROM operations where hostname = ?`\n\titer := session.Query(q, hostname).Iter()\n\tfor iter.Scan(&description, &scriptName, &attributes) {\n\t\to := ops.Operation{\n\t\t\tDescription: description,\n\t\t\tScriptName: scriptName,\n\t\t\tAttributes: attributes,\n\t\t}\n\t\toperations = append(operations, o)\n\t}\n\tif err := iter.Close(); err != nil {\n\t\treturn []ops.Operation{}, fmt.Errorf(\"error getting operations from DB: %v\", err)\n\t}\n\n\treturn operations, nil\n}", "func (bq *InMemoryBuildQueue) ListOperations(ctx context.Context, request *buildqueuestate.ListOperationsRequest) (*buildqueuestate.ListOperationsResponse, error) {\n\tbq.enter(bq.clock.Now())\n\tdefer bq.leave()\n\n\t// Obtain operation names in sorted order.\n\tnameList := make([]string, 0, len(bq.operationsNameMap))\n\tfor name := range bq.operationsNameMap {\n\t\tnameList = append(nameList, name)\n\t}\n\tsort.Strings(nameList)\n\tpaginationInfo, endIndex := getPaginationInfo(len(nameList), request.PageSize, func(i int) bool {\n\t\treturn request.StartAfter == nil || nameList[i] > request.StartAfter.OperationName\n\t})\n\n\t// Extract status.\n\tnameListRegion := nameList[paginationInfo.StartIndex:endIndex]\n\toperations := make([]*buildqueuestate.OperationState, 0, len(nameListRegion))\n\tfor _, name := range nameListRegion {\n\t\to := bq.operationsNameMap[name]\n\t\toperations = append(operations, o.getOperationState(bq))\n\t}\n\treturn &buildqueuestate.ListOperationsResponse{\n\t\tOperations: operations,\n\t\tPaginationInfo: paginationInfo,\n\t}, nil\n}", "func ListOperations() ([]*op.Operation, error) {\n\tmessage := protocol.NewRequestListMessage()\n\terr := channel.Broadcast(message)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc1 := make(chan *protocol.ResponseList)\n\n\tonResponse := func(response *protocol.ResponseList) {\n\t\tc1 <- response\n\t}\n\n\tbus.SubscribeOnce(string(message.RequestList.ID), onResponse)\n\n\tselect {\n\tcase res := <-c1:\n\t\tif res.Result == protocol.ResponseOk {\n\t\t\treturn res.Operations, nil\n\t\t}\n\n\t\treturn nil, errors.New(string(res.Message))\n\tcase <-time.After(10 * time.Second):\n\t\tbus.Unsubscribe(string(message.RequestList.ID), onResponse)\n\t\treturn nil, errors.New(\"timeout\")\n\t}\n}", "func (o NamedRuleWithOperationsPatchOutput) Operations() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v NamedRuleWithOperationsPatch) []string { return v.Operations }).(pulumi.StringArrayOutput)\n}", "func Operations() (string, error) {\n\treturn makeRequest(\"operations\")\n}", "func (a *ImportsApiService) ListOperations(ctx context.Context) ApiListOperationsRequest {\n\treturn ApiListOperationsRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetSharepointIds gets the sharepointIds property value. Returns identifiers useful for SharePoint REST compatibility. Readonly.
func (m *List) GetSharepointIds()(SharepointIdsable) { return m.sharepointIds }
[ "func (m *Drive) GetSharePointIds()(SharepointIdsable) {\n return m.sharePointIds\n}", "func (o *MicrosoftGraphItemReference) GetSharepointIds() AnyOfmicrosoftGraphSharepointIds {\n\tif o == nil || o.SharepointIds == nil {\n\t\tvar ret AnyOfmicrosoftGraphSharepointIds\n\t\treturn ret\n\t}\n\treturn *o.SharepointIds\n}", "func (m *ItemReference) GetSharepointIds()(SharepointIdsable) {\n return m.sharepointIds\n}", "func (m *Site) GetSharepointIds()(SharepointIdsable) {\n val, err := m.GetBackingStore().Get(\"sharepointIds\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(SharepointIdsable)\n }\n return nil\n}", "func (m *ItemReference) SetSharepointIds(value SharepointIdsable)() {\n m.sharepointIds = value\n}", "func (m *List) SetSharepointIds(value SharepointIdsable)() {\n m.sharepointIds = value\n}", "func NewSharepointIds()(*SharepointIds) {\n m := &SharepointIds{\n }\n m.backingStore = ie8677ce2c7e1b4c22e9c3827ecd078d41185424dd9eeb92b7d971ed2d49a392e.BackingStoreFactoryInstance();\n m.SetAdditionalData(make(map[string]any))\n return m\n}", "func (o *MicrosoftGraphItemReference) SetSharepointIds(v AnyOfmicrosoftGraphSharepointIds) {\n\to.SharepointIds = &v\n}", "func (m *Site) SetSharepointIds(value SharepointIdsable)() {\n err := m.GetBackingStore().Set(\"sharepointIds\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *MicrosoftGraphListItem) HasSharepointIds() bool {\n\tif o != nil && o.SharepointIds != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (m *Drive) SetSharePointIds(value SharepointIdsable)() {\n m.sharePointIds = value\n}", "func (o *MicrosoftGraphItemReference) GetSharepointIdsOk() (AnyOfmicrosoftGraphSharepointIds, bool) {\n\tif o == nil || o.SharepointIds == nil {\n\t\tvar ret AnyOfmicrosoftGraphSharepointIds\n\t\treturn ret, false\n\t}\n\treturn *o.SharepointIds, true\n}", "func (o *MicrosoftGraphListItem) GetSharepointIdsOk() (AnyOfmicrosoftGraphSharepointIds, bool) {\n\tif o == nil || o.SharepointIds == nil {\n\t\tvar ret AnyOfmicrosoftGraphSharepointIds\n\t\treturn ret, false\n\t}\n\treturn *o.SharepointIds, true\n}", "func (o *MicrosoftGraphItemReference) HasSharepointIds() bool {\n\tif o != nil && o.SharepointIds != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func getIds() []string {\n\tclient := &http.Client{}\n\tvar ids []string\n\tsongRequest, err := http.NewRequest(\"GET\", \"https://api.spotify.com/v1/me/tracks?limit=50&offset=0\", nil)\n\tsongRequest.Header.Add(\"Authorization\", key)\n\tresponse, err := client.Do(songRequest)\n\tif err != nil {\n\t\tfmt.Println(\"Request failed with error:\", err)\n\t} else {\n\t\tdata, _ := ioutil.ReadAll(response.Body)\n\t\titems := gjson.Get(string(data), \"items\")\n\t\tfor i := 0; i < len(items.Array()); i++ {\n\t\t\ttrack := gjson.Get(items.Array()[i].String(), \"track\")\n\t\t\tid := gjson.Get(track.String(), \"id\")\n\t\t\tids = append(ids, id.String())\n\t\t}\n\t}\n\tids = append(ids, getPlaylistIds()...) // Calls to get song IDs from user playlists\n\treturn fixIds(ids)\n}", "func (m *SharepointIds) GetListItemId()(*string) {\n val, err := m.GetBackingStore().Get(\"listItemId\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *SharepointIds) GetListId()(*string) {\n val, err := m.GetBackingStore().Get(\"listId\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *DeviceManagementComplexSettingDefinition) GetPropertyDefinitionIds()([]string) {\n val, err := m.GetBackingStore().Get(\"propertyDefinitionIds\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]string)\n }\n return nil\n}", "func getPlaylistIds() []string {\n\tclient := &http.Client{}\n\n\tvar ids []string\n\turl := \"https://api.spotify.com/v1/users/\" + userId + \"/playlists?limit=50\"\n\tplaylistRequest, err := http.NewRequest(\"GET\", url, nil)\n\n\tplaylistRequest.Header.Add(\"Authorization\", key)\n\tresponse, err := client.Do(playlistRequest)\n\tif err != nil {\n\t\tfmt.Println(\"Request failed with error:\", err)\n\t} else {\n\t\tdata, _ := ioutil.ReadAll(response.Body)\n\t\titems := gjson.Get(string(data), \"items\")\n\t\tfor i := 0; i < len(items.Array()); i++ {\n\t\t\tid := gjson.Get(items.Array()[i].String(), \"id\")\n\t\t\tids = append(ids, id.String())\n\t\t}\n\t}\n\treturn getPlaylistSongIds(ids)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetSubscriptions gets the subscriptions property value. The set of subscriptions on the list.
func (m *List) GetSubscriptions()([]Subscriptionable) { return m.subscriptions }
[ "func (c *Client) GetSubscriptions(ctx context.Context) (*SubscriptionListResponse, error) {\n\tlist := &SubscriptionListResponse{}\n\n\terr := c.get(ctx, subscriptionEndpoint, list)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn list, err\n}", "func (c *Coinbene) GetSubscriptions() ([]wshandler.WebsocketChannelSubscription, error) {\n\treturn c.Websocket.GetSubscriptions(), nil\n}", "func (svc *UserDomService) GetSubscriptions(ctx context.Context, id string) ([]*domain.Subscription, error) {\n\tmainModel, err := svc.store.User.GetOne(ctx, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif mainModel.Subscriptions == nil {\n\t\treturn nil, omniErr.ErrResourceNotFound\n\t}\n\n\tfilters, err := filters.TransformModelsIDToFilters(mainModel.Subscriptions)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn svc.store.Subscription.GetList(ctx, filters)\n\n}", "func (a Accessor) GetSubscriptionList(service, servicePath string, subscriptions *[]Subscription) error {\n\treturn a.access(&AccessParameter{\n\t\tEpID: EntryPointIDs.Subscriptions,\n\t\tMethod: gohttp.HttpMethods.GET,\n\t\tService: service,\n\t\tServicePath: servicePath,\n\t\tPath: \"\",\n\t\tReceivedBody: subscriptions,\n\t})\n}", "func (subscriptions *Subscriptions) Get() ([]*SubscriptionInfo, error) {\n\tclient := NewHTTPClient(subscriptions.client)\n\tresp, err := client.Get(subscriptions.endpoint, subscriptions.config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdata, _ := NormalizeODataCollection(resp)\n\tvar subs []*SubscriptionInfo\n\tif err := json.Unmarshal(data, &subs); err != nil {\n\t\treturn nil, err\n\t}\n\treturn subs, nil\n}", "func (r *ServiceRegSubscriptions) GetSubscribedServices() []string {\n\tr.servicesLock.RLock()\n\tdefer r.servicesLock.RUnlock()\n\n\treturn r.subscribedServices\n}", "func (ss *SubscriptionsService) Get(ctx context.Context, cID, sID string) (res *Response, s *Subscription, err error) {\n\tu := fmt.Sprintf(\"v2/customers/%s/subscriptions/%s\", cID, sID)\n\n\tres, err = ss.client.get(ctx, u, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif err = json.Unmarshal(res.content, &s); err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (s *T) Subscriptions() <-chan map[string][]string {\n\treturn s.subscriptionsCh\n}", "func getSubscriptions(request router.Request) (int, []byte) {\n\n\tquery := datastore.NewQuery(SUBSCRIPTION_KEY).Filter(\"Project =\", request.GetPathParams()[\"project_id\"])\n\tsubscriptions := make([]Subscription, 0)\n\t_, err := query.GetAll(request.GetContext(), &subscriptions)\n\n\tif err != nil {\n\t\tlog.Errorf(request.GetContext(), \"Error retriving Subscriptions: %v\", err)\n\t\treturn http.StatusInternalServerError, []byte(err.Error())\n\t}\n\n\tsubscriptionBytes, err := json.MarshalIndent(subscriptions, \"\", \"\t\")\n\n\tif err != nil {\n\t\tlog.Errorf(request.GetContext(), \"Error retriving Subscriptions: %v\", err)\n\t\treturn http.StatusInternalServerError, []byte(err.Error())\n\t}\n\n\treturn http.StatusOK, subscriptionBytes\n\n}", "func (m *List) SetSubscriptions(value []Subscriptionable)() {\n m.subscriptions = value\n}", "func (*OpenConfigTelemetryServer) GetTelemetrySubscriptions(context.Context, *jpb.GetSubscriptionsRequest) (*jpb.GetSubscriptionsReply, error) {\n\treturn nil, nil\n}", "func (u User) GetSubscriptions(repos ...string) ([]Subscription, error) {\n\tresults := []Subscription{}\n\tif DB == nil {\n\t\treturn nil, fmt.Errorf(\"Failed to get subscriptions, invalid DB Connection\")\n\t}\n\tif len(repos) == 0 {\n\t\t// Get all subscriptions\n\t\terr := DB.Preload(\"EmailPreference\").Find(&results, \"user_id = ?\", u.ID).Error\n\t\treturn results, err\n\t}\n\tfor _, repo := range repos {\n\t\tvar sub Subscription\n\t\tif DB.Preload(\"EmailPreference\").Where(\"user_id = ? AND repo = ?\", u.ID, repo).First(&sub).RecordNotFound() == false {\n\t\t\tresults = append(results, sub)\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"No such subscription: %s\", repo)\n\t\t}\n\t}\n\treturn results, nil\n}", "func (r *SubscriptionsService) Get(customerId string, subscriptionId string) *SubscriptionsGetCall {\n\treturn &SubscriptionsGetCall{\n\t\ts: r.s,\n\t\tcustomerId: customerId,\n\t\tsubscriptionId: subscriptionId,\n\t\tcaller_: googleapi.JSONCall{},\n\t\tparams_: make(map[string][]string),\n\t\tpathTemplate_: \"customers/{customerId}/subscriptions/{subscriptionId}\",\n\t\tcontext_: googleapi.NoContext,\n\t}\n}", "func GetChatSubscriptions(db *models.DatabaseConfig, chatID int64) ([]*models.Subscription, error) {\n\n\tif db == nil {\n\t\tlog.Println(\"The DB model is nil\")\n\t\treturn nil, errors.New(\"the DB model passed is nil, can't operate\")\n\t}\n\n\tcursor, err := db.MongoClient.Collection(\"subscription\").Find(db.Ctx, bson.M{\"chatid\": chatID})\n\tif err != nil {\n\t\tlog.Println(\"There was an error trying to look for this chat's subscriptions: \", err)\n\t\treturn nil, err\n\t}\n\n\tsubs := make([]*models.Subscription, 0)\n\terr = cursor.All(db.Ctx, &subs)\n\tif err != nil {\n\t\tlog.Println(\"There was an error trying to decode subscriptions into a subscriptions slice: \", err)\n\t\treturn nil, err\n\t}\n\n\treturn subs, nil\n}", "func (r *ProjectsLocationsDataExchangesService) ListSubscriptions(resource string) *ProjectsLocationsDataExchangesListSubscriptionsCall {\n\tc := &ProjectsLocationsDataExchangesListSubscriptionsCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.resource = resource\n\treturn c\n}", "func (s *API) ListSubscriptions(status SubscriptionStatus) (data SubscriptionsResponse, err error) {\n\tif status == \"\" {\n\t\tstatus = SubscriptionStatusAll\n\t}\n\tendpoint := zoho.Endpoint{\n\t\tName: \"subscriptions\",\n\t\tURL: fmt.Sprintf(\"https://subscriptions.zoho.%s/api/v1/subscriptions\", s.ZohoTLD),\n\t\tMethod: zoho.HTTPGet,\n\t\tResponseData: &SubscriptionsResponse{},\n\t\tURLParameters: map[string]zoho.Parameter{\n\t\t\t\"filter_by\": zoho.Parameter(status),\n\t\t},\n\t\tHeaders: map[string]string{\n\t\t\tZohoSubscriptionsEndpointHeader: s.OrganizationID,\n\t\t},\n\t}\n\n\terr = s.Zoho.HTTPRequest(&endpoint)\n\tif err != nil {\n\t\treturn SubscriptionsResponse{}, fmt.Errorf(\"Failed to retrieve subscriptions: %s\", err)\n\t}\n\n\tif v, ok := endpoint.ResponseData.(*SubscriptionsResponse); ok {\n\t\treturn *v, nil\n\t}\n\n\treturn SubscriptionsResponse{}, fmt.Errorf(\"Data retrieved was not 'SubscriptionsResponse'\")\n}", "func (r *ProjectsLocationsDataExchangesListingsService) ListSubscriptions(resource string) *ProjectsLocationsDataExchangesListingsListSubscriptionsCall {\n\tc := &ProjectsLocationsDataExchangesListingsListSubscriptionsCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.resource = resource\n\treturn c\n}", "func (js *JetStream) GetAllSubscriptions() map[SubscriptionSubjectIdentifier]backendnats.Subscriber {\n\treturn js.subscriptions\n}", "func (api API) ListSubscriptions(cursor *Cursor, customerUUID string) (*Subscriptions, error) {\n\tresult := &Subscriptions{}\n\tpath := strings.Replace(subscriptionsEndpoint, \":customerUUID\", customerUUID, 1)\n\tquery := make([]interface{}, 0, 1)\n\tif cursor != nil {\n\t\tquery = append(query, *cursor)\n\t}\n\treturn result, api.list(path, result, query...)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetSystem gets the system property value. If present, indicates that this is a systemmanaged list. Readonly.
func (m *List) GetSystem()(SystemFacetable) { return m.system }
[ "func (client BaseClient) GetSystem(ctx context.Context, pathParameter string) (result System, err error) {\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: pathParameter,\n\t\t\tConstraints: []validation.Constraint{{Target: \"pathParameter\", Name: validation.Pattern, Rule: `.*`, Chain: nil}}}}); err != nil {\n\t\treturn result, validation.NewError(\"beacon.BaseClient\", \"GetSystem\", err.Error())\n\t}\n\n\treq, err := client.GetSystemPreparer(ctx, pathParameter)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetSystem\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.GetSystemSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetSystem\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.GetSystemResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetSystem\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (a *AllApiService) SystemPropertyGetSystemProperty(ctx _context.Context, body SystemPropertyGetSystemProperty) (SystemPropertyGetSystemPropertyResult, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue SystemPropertyGetSystemPropertyResult\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/systemProperty/getSystemProperty\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &body\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 200 {\n\t\t\tvar v SystemPropertyGetSystemPropertyResult\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 500 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func (cli *CLI) SystemList() {\n\tlist := nbv1.NooBaaList{}\n\terr := cli.Client.List(cli.Ctx, nil, &list)\n\tif meta.IsNoMatchError(err) {\n\t\tcli.Log.Warningf(\"CRD not installed.\\n\")\n\t\treturn\n\t}\n\tutil.Panic(err)\n\tif len(list.Items) == 0 {\n\t\tcli.Log.Printf(\"No systems found.\\n\")\n\t\treturn\n\t}\n\ttable := (&util.PrintTable{}).AddRow(\n\t\t\"NAMESPACE\",\n\t\t\"NAME\",\n\t\t\"PHASE\",\n\t\t\"MGMT-ENDPOINTS\",\n\t\t\"S3-ENDPOINTS\",\n\t\t\"IMAGE\",\n\t\t\"AGE\",\n\t)\n\tfor i := range list.Items {\n\t\ts := &list.Items[i]\n\t\ttable.AddRow(\n\t\t\ts.Namespace,\n\t\t\ts.Name,\n\t\t\tstring(s.Status.Phase),\n\t\t\tfmt.Sprint(s.Status.Services.ServiceMgmt.NodePorts),\n\t\t\tfmt.Sprint(s.Status.Services.ServiceS3.NodePorts),\n\t\t\ts.Status.ActualImage,\n\t\t\tsince(s.ObjectMeta.CreationTimestamp.Time),\n\t\t)\n\t}\n\tfmt.Print(table.String())\n}", "func (b *binding) SystemProperty(ctx context.Context, name string) (string, error) {\n\tres, err := b.Shell(\"getprop\", name).Call(ctx)\n\tif err != nil {\n\t\treturn \"\", log.Errf(ctx, err, \"getprop returned error: \\n%s\", err.Error())\n\t}\n\treturn res, nil\n}", "func (m *List) SetSystem(value SystemFacetable)() {\n m.system = value\n}", "func (m *Drive) GetSystem()(SystemFacetable) {\n return m.system\n}", "func (a *AllApiService) SystemPropertyGetSystemProperties(ctx _context.Context, body SystemPropertyGetSystemProperties) ([]SystemPropertyGetSystemPropertiesResultItem, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue []SystemPropertyGetSystemPropertiesResultItem\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/systemProperty/getSystemProperties\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &body\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 200 {\n\t\t\tvar v []SystemPropertyGetSystemPropertiesResultItem\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 500 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func (client *APIClient) GetSystem(systemName string) (system System, err error) {\n\tresponse, err := client.request(\"GET\", urlSystem(systemName), nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = utilities.FromJSON(response, &system)\n\treturn\n}", "func (m *Metrics) GetSystemMetrics(netif string) api.ResourceMetrics {\n\treturn api.ResourceMetrics{}\n}", "func (r *Radarr) GetSystemStatus() (*SystemStatus, error) {\n\tvar status SystemStatus\n\n\terr := r.GetInto(\"v3/system/status\", nil, &status)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"api.Get(system/status): %w\", err)\n\t}\n\n\treturn &status, nil\n}", "func getSystem(c config) (system, error) {\n\turl, err := getSystemURL(c)\n\tif err != nil {\n\t\treturn system{}, err\n\t}\n\tb, err := redfishGet(c, url)\n\tif err != nil {\n\t\treturn system{}, err\n\t}\n\tvar sys system\n\tif err := json.Unmarshal(b, &sys); err != nil {\n\t\treturn system{}, err\n\t}\n\treturn sys, nil\n}", "func (client BaseClient) GetSystems(ctx context.Context, tenant string) (result ListSystem, err error) {\n\treq, err := client.GetSystemsPreparer(ctx, tenant)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetSystems\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.GetSystemsSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetSystems\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.GetSystemsResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetSystems\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (o *Block) GetHintSystem(ctx context.Context) (hintSystem bool, err error) {\n\terr = o.object.CallWithContext(ctx, \"org.freedesktop.DBus.Properties.Get\", 0, InterfaceBlock, \"HintSystem\").Store(&hintSystem)\n\treturn\n}", "func (s *Systems) GetSystemResource(ctx context.Context, req *systemsproto.GetSystemsRequest) (*systemsproto.SystemsResponse, error) {\n\tvar resp systemsproto.SystemsResponse\n\tsessionToken := req.SessionToken\n\tauthResp := s.IsAuthorizedRPC(sessionToken, []string{common.PrivilegeLogin}, []string{})\n\tif authResp.StatusCode != http.StatusOK {\n\t\tlog.Error(\"error while trying to authenticate session\")\n\t\tfillSystemProtoResponse(&resp, authResp)\n\t\treturn &resp, nil\n\t}\n\tvar pc = systems.PluginContact{\n\t\tContactClient: pmbhandle.ContactPlugin,\n\t\tDevicePassword: common.DecryptWithPrivateKey,\n\t\tGetPluginStatus: scommon.GetPluginStatus,\n\t}\n\tdata := pc.GetSystemResource(req)\n\tfillSystemProtoResponse(&resp, data)\n\treturn &resp, nil\n}", "func (m *Manager) Get(stype SystemType) System {\n\ts, ok := m.systems[stype]\n\tif !ok {\n\t\treturn m.null\n\t}\n\treturn s\n}", "func (o *MyErratum) GetSystems() string {\n\tif o == nil || IsNil(o.Systems) {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Systems\n}", "func (o *IamServiceProviderAllOf) GetSystemOk() (*IamSystemRelationship, bool) {\n\tif o == nil || o.System == nil {\n\t\treturn nil, false\n\t}\n\treturn o.System, true\n}", "func (m *AndroidManagedStoreApp) GetIsSystemApp()(*bool) {\n val, err := m.GetBackingStore().Get(\"isSystemApp\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*bool)\n }\n return nil\n}", "func AvailableSystems() []System {\n\treturn systemRegistry\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetColumns sets the columns property value. The collection of field definitions for this list.
func (m *List) SetColumns(value []ColumnDefinitionable)() { m.columns = value }
[ "func (m *Site) SetColumns(value []ColumnDefinitionable)() {\n err := m.GetBackingStore().Set(\"columns\", value)\n if err != nil {\n panic(err)\n }\n}", "func (c *Creater) SetColumns(c1 []builder.Columns) builder.Creater {\n\n\tcolumnDefs := make([]string, 0, len(c1))\n\n\tfor _, item := range c1 {\n\t\tcolumnDefs = append(columnDefs, fmt.Sprintf(\"%s %s %s\", item.Name, item.Datatype, item.Constraint))\n\t}\n\n\tcolumns := strings.Join(columnDefs, seperator)\n\tc.sql.WriteString(fmt.Sprintf(\"%s %s %s\", \"(\", columns, \");\"))\n\n\treturn c\n}", "func (t *Table) SetColumns(columns []string) *Table {\n\tt.Clear(true)\n\tif t.showIndex {\n\t\tcolumns = append([]string{\"#\"}, columns...)\n\t\tif len(columns) >= 2 {\n\t\t\tt.sortCol = 1\n\t\t\tt.sortType = SortAsc\n\t\t}\n\t} else {\n\t\tif len(columns) >= 1 {\n\t\t\tt.sortCol = 0\n\t\t\tt.sortType = SortAsc\n\t\t}\n\t}\n\tfor i := 0; i < len(columns); i++ {\n\t\tcell := cview.NewTableCell(columns[i])\n\t\tif t.addCellFunc != nil {\n\t\t\tt.addCellFunc(cell, true, 0)\n\t\t}\n\t\tt.Table.SetCell(0, i, cell)\n\t}\n\tt.columns = columns\n\treturn t\n}", "func (info *CSVFileInfo) SetColumns(columns []string) *CSVFileInfo {\n\tinfo.Columns = columns\n\treturn info\n}", "func (o *InlineResponse20075Stats) SetColumns(v InlineResponse20075StatsColumns) {\n\to.Columns = &v\n}", "func (r *commandResult) SetColumns(cols []pgproto.FieldDescription) {\n\tr.conn.writerState.fi.registerCmd(r.pos)\n\tif r.descOpt == sql.NeedRowDesc {\n\t\tif err := r.conn.writeRowDescription(cols, r.formatCodes, &r.conn.writerState.buf); err != nil {\n\t\t\tgolog.Error(err.Error())\n\t\t}\n\t}\n}", "func SetColumns(names []string) {\n\tvar (\n\t\tn int\n\t\tcurColStr = ColumnsString()\n\t\tnewColumns = make([]*Column, len(GlobalColumns))\n\t)\n\n\tlock.Lock()\n\n\t// add enabled columns by name\n\tfor _, name := range names {\n\t\tnewColumns[n] = popColumn(name)\n\t\tnewColumns[n].Enabled = true\n\t\tn++\n\t}\n\n\t// extend with omitted columns as disabled\n\tfor _, col := range GlobalColumns {\n\t\tnewColumns[n] = col\n\t\tnewColumns[n].Enabled = false\n\t\tn++\n\t}\n\n\tGlobalColumns = newColumns\n\tlock.Unlock()\n\n\tlog.Noticef(\"config change [columns]: %s -> %s\", curColStr, ColumnsString())\n}", "func (o *SummaryResponse) SetColumns(v SummaryColumnResponse) {\n\to.Columns = &v\n}", "func (v *IconView) SetColumns(columns int) {\n\tC.gtk_icon_view_set_columns(v.native(), C.gint(columns))\n}", "func (m *Site) SetExternalColumns(value []ColumnDefinitionable)() {\n err := m.GetBackingStore().Set(\"externalColumns\", value)\n if err != nil {\n panic(err)\n }\n}", "func (e *EventTriggerPayload) Columns(columns ...string) {\n\tif len(columns) > 0 {\n\t\tvar cols []string\n\t\tfor i, _ := range columns {\n\t\t\tcols = append(cols, columns[i])\n\t\t}\n\t\te.Column = cols\n\t\treturn\n\t}\n\n\te.Column = \"*\"\n}", "func (r *Result) Columns(it func(Column)) {\n\tif r.set == nil {\n\t\treturn\n\t}\n\tfor _, m := range r.set.Columns {\n\t\tit(Column{\n\t\t\tName: m.Name,\n\t\t\tType: internal.TypeFromYDB(m.Type),\n\t\t})\n\t}\n}", "func (q *Query) Columns(columns ...Column) *Query {\n\tq.Parameter.setColumns(columns)\n\treturn q\n}", "func (b *Blueprint) Set(column string, allowed []string) *ColumnDefinition {\n\treturn b.addColumn(\"set\", column, &ColumnOptions{\n\t\tAllowed: allowed,\n\t})\n}", "func (ts *STableSpec) Columns() []IColumnSpec {\n\tif ts._columns == nil {\n\t\tval := reflect.Indirect(reflect.New(ts.structType))\n\t\tts.struct2TableSpec(val)\n\t}\n\treturn ts._columns\n}", "func Columns() *ColumnsType {\n\ttable := qbColumnsTable\n\treturn &ColumnsType{\n\t\tqbColumnsFColumnName.Copy(&table),\n\t\tqbColumnsFTableSchema.Copy(&table),\n\t\tqbColumnsFTableName.Copy(&table),\n\t\tqbColumnsFCharacterMaximumLength.Copy(&table),\n\t\t&table,\n\t}\n}", "func (g *Grid) SetLimitCols(limitCols bool) {\n\tcols := g.Cols()\n\tg.limitCols = limitCols\n\tif g.limitCols {\n\t\tg.j = func(j int) int { return j }\n\t\tg.jIsOut = func(j int) bool { return j < 0 || j >= cols }\n\t} else {\n\t\tg.j = func(j int) int { return ((j % cols) + cols) % cols }\n\t\tg.jIsOut = func(_ int) bool { return false }\n\t}\n}", "func (db *DB) InitColumns(param *Params) {\n\n\tvar (\n\t\tname = conf.Get[string](cons.ConfDBName)\n\t\ttables = []string{param.Table}\n\t)\n\n\ttables = append(tables, param.InnerTable...)\n\ttables = append(tables, param.LeftTable...)\n\n\tfor _, v := range tables {\n\t\tif v == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tif _, ok := TableCols[v]; ok {\n\t\t\tcontinue\n\t\t}\n\t\tvar columns []string\n\t\tparam.Data = &columns\n\t\ttb := TableOnly(v)\n\t\tdb.get(&GT{\n\t\t\tParams: &Params{Data: &columns},\n\t\t\tsql: \"SELECT COLUMN_NAME FROM `information_schema`.`COLUMNS` WHERE TABLE_NAME = ? and TABLE_SCHEMA = ?\",\n\t\t\tArgs: []any{tb, name},\n\t\t})\n\t\tTableCols[tb] = columns\n\t}\n}", "func (m *List) GetColumns()([]ColumnDefinitionable) {\n return m.columns\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetContentTypes sets the contentTypes property value. The collection of content types present in this list.
func (m *List) SetContentTypes(value []ContentTypeable)() { m.contentTypes = value }
[ "func (web *Web) ContentTypes() *ContentTypes {\n\treturn NewContentTypes(\n\t\tweb.client,\n\t\tfmt.Sprintf(\"%s/ContentTypes\", web.endpoint),\n\t\tweb.config,\n\t)\n}", "func (m *Site) SetContentTypes(value []ContentTypeable)() {\n err := m.GetBackingStore().Set(\"contentTypes\", value)\n if err != nil {\n panic(err)\n }\n}", "func ContentTypes(types []string, blacklist bool) Option {\n\treturn func(c *config) error {\n\t\tc.contentTypes = []parsedContentType{}\n\t\tfor _, v := range types {\n\t\t\tmediaType, params, err := mime.ParseMediaType(v)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tc.contentTypes = append(c.contentTypes, parsedContentType{mediaType, params})\n\t\t}\n\t\tc.blacklist = blacklist\n\t\treturn nil\n\t}\n}", "func (o *LastMileAccelerationOptions) GetContentTypes() []string {\n\tif o == nil || o.ContentTypes == nil {\n\t\tvar ret []string\n\t\treturn ret\n\t}\n\treturn *o.ContentTypes\n}", "func ContentTypes(contentTypes ...string) Option {\n\treturn ArrayOpt(\"content_types\", contentTypes...)\n}", "func (s *ChannelSpecification) SetSupportedContentTypes(v []*string) *ChannelSpecification {\n\ts.SupportedContentTypes = v\n\treturn s\n}", "func (ht HTMLContentTypeBinder) ContentTypes() []string {\n\treturn []string{\n\t\t\"application/html\",\n\t\t\"text/html\",\n\t\t\"application/x-www-form-urlencoded\",\n\t\t\"html\",\n\t}\n}", "func (o *ExtrasCustomLinksListParams) SetContentTypesIe(contentTypesIe *string) {\n\to.ContentTypesIe = contentTypesIe\n}", "func (set *ContentTypeSet) Types() (types []ContentType) {\n\tif set == nil || len(set.set) == 0 {\n\t\treturn []ContentType{}\n\t}\n\treturn append(make([]ContentType, 0, len(set.set)), set.set...)\n}", "func (s *InferenceSpecification) SetSupportedContentTypes(v []*string) *InferenceSpecification {\n\ts.SupportedContentTypes = v\n\treturn s\n}", "func SetOfContentTypes(types ...ContentType) *ContentTypeSet {\n\tif len(types) == 0 {\n\t\treturn nil\n\t}\n\tset := &ContentTypeSet{\n\t\tset: make([]ContentType, 0, len(types)),\n\t\tpos: -1,\n\t}\nallTypes:\n\tfor _, t := range types {\n\t\t// Let's make sure we have not seen this type before.\n\t\tfor _, tt := range set.set {\n\t\t\tif tt == t {\n\t\t\t\t// Don't add it to the set, already exists\n\t\t\t\tcontinue allTypes\n\t\t\t}\n\t\t}\n\t\tset.set = append(set.set, t)\n\t}\n\tif len(set.set) == 0 {\n\t\treturn nil\n\t}\n\treturn set\n}", "func (o *ExtrasSavedFiltersListParams) SetContentTypesIe(contentTypesIe *string) {\n\to.ContentTypesIe = contentTypesIe\n}", "func (o *ExtrasCustomLinksListParams) WithContentTypes(contentTypes *string) *ExtrasCustomLinksListParams {\n\to.SetContentTypes(contentTypes)\n\treturn o\n}", "func (ycp *YamlContentParser) ContentTypes() []string {\n\treturn []string{\"text/x-yaml\", \"application/yaml\", \"text/yaml\", \"application/x-yaml\"}\n}", "func (s *CaptureContentTypeHeader) SetCsvContentTypes(v []*string) *CaptureContentTypeHeader {\n\ts.CsvContentTypes = v\n\treturn s\n}", "func (o *ExtrasSavedFiltersListParams) WithContentTypes(contentTypes *string) *ExtrasSavedFiltersListParams {\n\to.SetContentTypes(contentTypes)\n\treturn o\n}", "func (m *List) GetContentTypes()([]ContentTypeable) {\n return m.contentTypes\n}", "func (m *Site) GetContentTypes()([]ContentTypeable) {\n val, err := m.GetBackingStore().Get(\"contentTypes\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]ContentTypeable)\n }\n return nil\n}", "func (m *SiteItemRequestBuilder) ContentTypes()(*ItemContentTypesRequestBuilder) {\n return NewItemContentTypesRequestBuilderInternal(m.pathParameters, m.requestAdapter)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetDrive sets the drive property value. Only present on document libraries. Allows access to the list as a [drive][] resource with [driveItems][driveItem].
func (m *List) SetDrive(value Driveable)() { m.drive = value }
[ "func (m *Group) SetDrive(value Driveable)() {\n m.drive = value\n}", "func (m *Site) SetDrive(value Driveable)() {\n err := m.GetBackingStore().Set(\"drive\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *User) SetDrive(value Driveable)() {\n m.drive = value\n}", "func (o *User) SetDrive(v AnyOfmicrosoftGraphDrive) {\n\to.Drive = &v\n}", "func (m *ItemReference) SetDriveType(value *string)() {\n m.driveType = value\n}", "func (m *ItemReference) SetDriveId(value *string)() {\n m.driveId = value\n}", "func (m *Site) SetDrives(value []Driveable)() {\n err := m.GetBackingStore().Set(\"drives\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *User) SetDrives(value []Driveable)() {\n m.drives = value\n}", "func (m *Group) SetDrives(value []Driveable)() {\n m.drives = value\n}", "func (h *stubDriveHandler) SetDrives(offset int64, d []models.Drive) {\n\th.drives = d\n\th.stubDriveIndex = offset\n}", "func (device *PerformanceDCBricklet) SetDriveMode(mode DriveMode) (err error) {\n\tvar buf bytes.Buffer\n\tbinary.Write(&buf, binary.LittleEndian, mode)\n\n\tresultBytes, err := device.device.Set(uint8(FunctionSetDriveMode), buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(resultBytes) > 0 {\n\t\tvar header PacketHeader\n\n\t\theader.FillFromBytes(resultBytes)\n\n\t\tif header.Length != 8 {\n\t\t\treturn fmt.Errorf(\"Received packet of unexpected size %d, instead of %d\", header.Length, 8)\n\t\t}\n\n\t\tif header.ErrorCode != 0 {\n\t\t\treturn DeviceError(header.ErrorCode)\n\t\t}\n\n\t\tbytes.NewBuffer(resultBytes[8:])\n\n\t}\n\n\treturn nil\n}", "func (device *DCV2Bricklet) SetDriveMode(mode DriveMode) (err error) {\n\tvar buf bytes.Buffer\n\tbinary.Write(&buf, binary.LittleEndian, mode)\n\n\tresultBytes, err := device.device.Set(uint8(FunctionSetDriveMode), buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(resultBytes) > 0 {\n\t\tvar header PacketHeader\n\n\t\theader.FillFromBytes(resultBytes)\n\n\t\tif header.Length != 8 {\n\t\t\treturn fmt.Errorf(\"Received packet of unexpected size %d, instead of %d\", header.Length, 8)\n\t\t}\n\n\t\tif header.ErrorCode != 0 {\n\t\t\treturn DeviceError(header.ErrorCode)\n\t\t}\n\n\t\tbytes.NewBuffer(resultBytes[8:])\n\n\t}\n\n\treturn nil\n}", "func (o *StorageFlexUtilVirtualDrive) SetDriveType(v string) {\n\to.DriveType = &v\n}", "func (repo Repository) Drive(driveID resource.ID) drivestream.DriveReference {\n\treturn Drive{\n\t\tdb: repo.db,\n\t\tdrive: driveID,\n\t}\n}", "func (o *MicrosoftGraphListItem) SetDriveItem(v AnyOfmicrosoftGraphDriveItem) {\n\to.DriveItem = &v\n}", "func (o *StoragePhysicalDisk) SetDriveFirmware(v string) {\n\to.DriveFirmware = &v\n}", "func (r *Rect) Drive(anchor int) *Rect {\n\tr.anchor = anchor\n\treturn r\n}", "func (ref FileView) Drive() resource.ID {\n\treturn ref.drive\n}", "func NewDrive()(*Drive) {\n m := &Drive{\n BaseItem: *NewBaseItem(),\n }\n odataTypeValue := \"#microsoft.graph.drive\";\n m.SetOdataType(&odataTypeValue);\n return m\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetList sets the list property value. Provides additional details about the list.
func (m *List) SetList(value ListInfoable)() { m.list = value }
[ "func (m *Drive) SetList(value Listable)() {\n m.list = value\n}", "func (a *ForjFlag) setList(ol *ForjObjectList, instance, field string) {\n\ta.list = ol\n\ta.setObjectField(ol.obj, field)\n\ta.setObjectInstance(instance)\n}", "func (o *AdminDeleteProfanityFilterParams) SetList(list string) {\n\to.List = list\n}", "func (c *Controller) setList(list []models.Category) {\n\tc.List = list\n\tc.parseNewID()\n}", "func (s *Storage) SetList(key string, value []string, expirationSec uint64) {\n\ts.set(key, value, expirationSec)\n}", "func (a *Any) SetList(l list.List) {\n\ta.rendering = l\n}", "func (c *UsageController) setList(list []models.Usage) {\n\tc.Usage = list\n\tc.parseNewID()\n}", "func (ds DeveloperStorage) SetDeveloperList(ctx sdk.Context, lst *DeveloperList) sdk.Error {\n\tstore := ctx.KVStore(ds.key)\n\tlistByte, err := ds.cdc.MarshalJSON(*lst)\n\tif err != nil {\n\t\treturn ErrFailedToMarshalDeveloperList(err)\n\t}\n\tstore.Set(GetDeveloperListKey(), listByte)\n\treturn nil\n}", "func (m *EntityMutation) SetListDate(t time.Time) {\n\tm.list_date = &t\n}", "func (pane *TaskPane) SetList(tasks []model.Task) {\n\tpane.ClearList()\n\tpane.tasks = tasks\n\n\tfor i := range pane.tasks {\n\t\tpane.addTaskToList(i)\n\t}\n}", "func (m *Site) SetLists(value []Listable)() {\n err := m.GetBackingStore().Set(\"lists\", value)\n if err != nil {\n panic(err)\n }\n}", "func (s *InputService4TestShapeInputService4TestCaseOperation1Input) SetListParam(v [][]byte) *InputService4TestShapeInputService4TestCaseOperation1Input {\n\ts.ListParam = v\n\treturn s\n}", "func (s *GetServersOutput) SetServerList(v []*Server) *GetServersOutput {\n\ts.ServerList = v\n\treturn s\n}", "func (vs VoteStorage) SetReferenceList(ctx sdk.Context, lst *ReferenceList) sdk.Error {\n\tstore := ctx.KVStore(vs.key)\n\tlstByte, err := vs.cdc.MarshalBinaryLengthPrefixed(*lst)\n\tif err != nil {\n\t\treturn ErrFailedToMarshalReferenceList(err)\n\t}\n\tstore.Set(getReferenceListKey(), lstByte)\n\treturn nil\n}", "func (m *SharepointIds) SetListId(value *string)() {\n err := m.GetBackingStore().Set(\"listId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (mr *MockUserCacheMockRecorder) SetList(arg0 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"SetList\", reflect.TypeOf((*MockUserCache)(nil).SetList), arg0)\n}", "func (s *ServerGroup) SetServerList(v []*Server) *ServerGroup {\n\ts.ServerList = v\n\treturn s\n}", "func (p *PoolSProperty) SetPropList(key, value string) {\n\tp.PropList = append(p.PropList, fmt.Sprintf(\"%s=%s\", key, value))\n}", "func (lbu *LoadBalanceUpdate) SetIPList(s string) *LoadBalanceUpdate {\n\tlbu.mutation.SetIPList(s)\n\treturn lbu\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetOperations sets the operations property value. The collection of longrunning operations on the list.
func (m *List) SetOperations(value []RichLongRunningOperationable)() { m.operations = value }
[ "func (o *MicrosoftGraphOnenote) SetOperations(v []MicrosoftGraphOnenoteOperation) {\n\to.Operations = &v\n}", "func (m *Site) SetOperations(value []RichLongRunningOperationable)() {\n err := m.GetBackingStore().Set(\"operations\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *ExternalConnection) SetOperations(value []ConnectionOperationable)() {\n m.operations = value\n}", "func (m *Workbook) SetOperations(value []WorkbookOperationable)() {\n m.operations = value\n}", "func (m *List) GetOperations()([]RichLongRunningOperationable) {\n return m.operations\n}", "func (c *restClient) ListOperations(ctx context.Context, req *longrunningpb.ListOperationsRequest, opts ...gax.CallOption) *OperationIterator {\n\tit := &OperationIterator{}\n\treq = proto.Clone(req).(*longrunningpb.ListOperationsRequest)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tit.InternalFetch = func(pageSize int, pageToken string) ([]*longrunningpb.Operation, string, error) {\n\t\tresp := &longrunningpb.ListOperationsResponse{}\n\t\tif pageToken != \"\" {\n\t\t\treq.PageToken = pageToken\n\t\t}\n\t\tif pageSize > math.MaxInt32 {\n\t\t\treq.PageSize = math.MaxInt32\n\t\t} else if pageSize != 0 {\n\t\t\treq.PageSize = int32(pageSize)\n\t\t}\n\t\tbaseUrl, err := url.Parse(c.endpoint)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v/operations\", req.GetName())\n\n\t\tparams := url.Values{}\n\t\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\t\tif req.GetFilter() != \"\" {\n\t\t\tparams.Add(\"filter\", fmt.Sprintf(\"%v\", req.GetFilter()))\n\t\t}\n\t\tif req.GetPageSize() != 0 {\n\t\t\tparams.Add(\"pageSize\", fmt.Sprintf(\"%v\", req.GetPageSize()))\n\t\t}\n\t\tif req.GetPageToken() != \"\" {\n\t\t\tparams.Add(\"pageToken\", fmt.Sprintf(\"%v\", req.GetPageToken()))\n\t\t}\n\n\t\tbaseUrl.RawQuery = params.Encode()\n\n\t\t// Build HTTP headers from client and context metadata.\n\t\thds := append(c.xGoogHeaders, \"Content-Type\", \"application/json\")\n\t\theaders := gax.BuildHeaders(ctx, hds...)\n\t\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\t\tif settings.Path != \"\" {\n\t\t\t\tbaseUrl.Path = settings.Path\n\t\t\t}\n\t\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\thttpReq.Header = headers\n\n\t\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer httpRsp.Body.Close()\n\n\t\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}, opts...)\n\t\tif e != nil {\n\t\t\treturn nil, \"\", e\n\t\t}\n\t\tit.Response = resp\n\t\treturn resp.GetOperations(), resp.GetNextPageToken(), nil\n\t}\n\n\tfetch := func(pageSize int, pageToken string) (string, error) {\n\t\titems, nextPageToken, err := it.InternalFetch(pageSize, pageToken)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tit.items = append(it.items, items...)\n\t\treturn nextPageToken, nil\n\t}\n\n\tit.pageInfo, it.nextFunc = iterator.NewPageInfo(fetch, it.bufLen, it.takeBuf)\n\tit.pageInfo.MaxSize = int(req.GetPageSize())\n\tit.pageInfo.Token = req.GetPageToken()\n\n\treturn it\n}", "func (req *UpsertObjectRequest) Operations(operations []Operation) *UpsertObjectRequest {\n\treq.operations = operations\n\treturn req\n}", "func (client BaseClient) ListOperations(ctx context.Context) (result Operations, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/BaseClient.ListOperations\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.Response.Response != nil {\n\t\t\t\tsc = result.Response.Response.StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\treq, err := client.ListOperationsPreparer(ctx)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"serialconsole.BaseClient\", \"ListOperations\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.ListOperationsSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"serialconsole.BaseClient\", \"ListOperations\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.ListOperationsResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"serialconsole.BaseClient\", \"ListOperations\", resp, \"Failure responding to request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func (req *UpsertRequest) Operations(operations []Operation) *UpsertRequest {\n\treq.operations = operations\n\treturn req\n}", "func (so *Operations) Operations() api.Operations {\n\treturn api.Operations(so)\n}", "func ListOperations() ([]*op.Operation, error) {\n\tmessage := protocol.NewRequestListMessage()\n\terr := channel.Broadcast(message)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc1 := make(chan *protocol.ResponseList)\n\n\tonResponse := func(response *protocol.ResponseList) {\n\t\tc1 <- response\n\t}\n\n\tbus.SubscribeOnce(string(message.RequestList.ID), onResponse)\n\n\tselect {\n\tcase res := <-c1:\n\t\tif res.Result == protocol.ResponseOk {\n\t\t\treturn res.Operations, nil\n\t\t}\n\n\t\treturn nil, errors.New(string(res.Message))\n\tcase <-time.After(10 * time.Second):\n\t\tbus.Unsubscribe(string(message.RequestList.ID), onResponse)\n\t\treturn nil, errors.New(\"timeout\")\n\t}\n}", "func (a *ImportsApiService) ListOperations(ctx context.Context) ApiListOperationsRequest {\n\treturn ApiListOperationsRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func Operations() (string, error) {\n\treturn makeRequest(\"operations\")\n}", "func (op *OperationRequest) SetOperationsEndpoint() *OperationRequest {\n\treturn op.setEndpoint(\"operations\")\n}", "func SetOps(ops []op.Operation) Options {\n\treturn func(p *Permission) error {\n\t\tif ops == nil {\n\t\t\treturn errors.ErrNilOps\n\t\t}\n\t\tp.Ops = ops\n\t\treturn nil\n\t}\n}", "func (o NamedRuleWithOperationsPatchOutput) Operations() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v NamedRuleWithOperationsPatch) []string { return v.Operations }).(pulumi.StringArrayOutput)\n}", "func NewOperations() *Operations {\n\treturn &Operations{}\n}", "func (bq *InMemoryBuildQueue) ListOperations(ctx context.Context, request *buildqueuestate.ListOperationsRequest) (*buildqueuestate.ListOperationsResponse, error) {\n\tbq.enter(bq.clock.Now())\n\tdefer bq.leave()\n\n\t// Obtain operation names in sorted order.\n\tnameList := make([]string, 0, len(bq.operationsNameMap))\n\tfor name := range bq.operationsNameMap {\n\t\tnameList = append(nameList, name)\n\t}\n\tsort.Strings(nameList)\n\tpaginationInfo, endIndex := getPaginationInfo(len(nameList), request.PageSize, func(i int) bool {\n\t\treturn request.StartAfter == nil || nameList[i] > request.StartAfter.OperationName\n\t})\n\n\t// Extract status.\n\tnameListRegion := nameList[paginationInfo.StartIndex:endIndex]\n\toperations := make([]*buildqueuestate.OperationState, 0, len(nameListRegion))\n\tfor _, name := range nameListRegion {\n\t\to := bq.operationsNameMap[name]\n\t\toperations = append(operations, o.getOperationState(bq))\n\t}\n\treturn &buildqueuestate.ListOperationsResponse{\n\t\tOperations: operations,\n\t\tPaginationInfo: paginationInfo,\n\t}, nil\n}", "func NewOperations(\n\tcg proxy.ClientGetter,\n\tcatalog catalogcontrollers.Interface,\n\trbac rbacv1controllers.Interface,\n\tcontentManager *content.Manager,\n\tpods corev1controllers.PodClient) *Operations {\n\treturn &Operations{\n\t\tcg: cg,\n\t\tcontentManager: contentManager,\n\t\tnamespace: namespaces.System,\n\t\tImpersonator: podimpersonation.New(\"helm-op\", cg, time.Hour, settings.FullShellImage),\n\t\tpods: pods,\n\t\tclusterRepos: catalog.ClusterRepo(),\n\t\tops: catalog.Operation(),\n\t\tapps: catalog.App(),\n\t\troleBindings: rbac.RoleBinding(),\n\t\troles: rbac.Role(),\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetSharepointIds sets the sharepointIds property value. Returns identifiers useful for SharePoint REST compatibility. Readonly.
func (m *List) SetSharepointIds(value SharepointIdsable)() { m.sharepointIds = value }
[ "func (m *ItemReference) SetSharepointIds(value SharepointIdsable)() {\n m.sharepointIds = value\n}", "func (m *Site) SetSharepointIds(value SharepointIdsable)() {\n err := m.GetBackingStore().Set(\"sharepointIds\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *MicrosoftGraphItemReference) SetSharepointIds(v AnyOfmicrosoftGraphSharepointIds) {\n\to.SharepointIds = &v\n}", "func (m *Drive) SetSharePointIds(value SharepointIdsable)() {\n m.sharePointIds = value\n}", "func (o *MicrosoftGraphItemReference) GetSharepointIds() AnyOfmicrosoftGraphSharepointIds {\n\tif o == nil || o.SharepointIds == nil {\n\t\tvar ret AnyOfmicrosoftGraphSharepointIds\n\t\treturn ret\n\t}\n\treturn *o.SharepointIds\n}", "func (m *Drive) GetSharePointIds()(SharepointIdsable) {\n return m.sharePointIds\n}", "func NewSharepointIds()(*SharepointIds) {\n m := &SharepointIds{\n }\n m.backingStore = ie8677ce2c7e1b4c22e9c3827ecd078d41185424dd9eeb92b7d971ed2d49a392e.BackingStoreFactoryInstance();\n m.SetAdditionalData(make(map[string]any))\n return m\n}", "func (m *ItemReference) GetSharepointIds()(SharepointIdsable) {\n return m.sharepointIds\n}", "func (m *List) GetSharepointIds()(SharepointIdsable) {\n return m.sharepointIds\n}", "func (m *Site) GetSharepointIds()(SharepointIdsable) {\n val, err := m.GetBackingStore().Get(\"sharepointIds\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(SharepointIdsable)\n }\n return nil\n}", "func (o *MicrosoftGraphListItem) HasSharepointIds() bool {\n\tif o != nil && o.SharepointIds != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *MicrosoftGraphItemReference) HasSharepointIds() bool {\n\tif o != nil && o.SharepointIds != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *MicrosoftGraphItemReference) GetSharepointIdsOk() (AnyOfmicrosoftGraphSharepointIds, bool) {\n\tif o == nil || o.SharepointIds == nil {\n\t\tvar ret AnyOfmicrosoftGraphSharepointIds\n\t\treturn ret, false\n\t}\n\treturn *o.SharepointIds, true\n}", "func (o *MicrosoftGraphListItem) GetSharepointIdsOk() (AnyOfmicrosoftGraphSharepointIds, bool) {\n\tif o == nil || o.SharepointIds == nil {\n\t\tvar ret AnyOfmicrosoftGraphSharepointIds\n\t\treturn ret, false\n\t}\n\treturn *o.SharepointIds, true\n}", "func (m *DeviceManagementComplexSettingDefinition) SetPropertyDefinitionIds(value []string)() {\n err := m.GetBackingStore().Set(\"propertyDefinitionIds\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *SharepointIds) SetListId(value *string)() {\n err := m.GetBackingStore().Set(\"listId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *SearchResponse_Body) SetIdList(v []*refs.ObjectID) {\n\tif m != nil {\n\t\tm.IdList = v\n\t}\n}", "func (m *SharepointIds) SetListItemId(value *string)() {\n err := m.GetBackingStore().Set(\"listItemId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *MicrosoftGraphItemReference) SetSharepointIdsExplicitNull(b bool) {\n\to.SharepointIds = nil\n\to.isExplicitNullSharepointIds = b\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetSubscriptions sets the subscriptions property value. The set of subscriptions on the list.
func (m *List) SetSubscriptions(value []Subscriptionable)() { m.subscriptions = value }
[ "func (mr *MockSessionMockRecorder) SetSubscriptions(arg0 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"SetSubscriptions\", reflect.TypeOf((*MockSession)(nil).SetSubscriptions), arg0)\n}", "func (m *MockSession) SetSubscriptions(arg0 []*nats.Subscription) {\n\tm.ctrl.T.Helper()\n\tm.ctrl.Call(m, \"SetSubscriptions\", arg0)\n}", "func (r *ProjectsLocationsDataExchangesService) ListSubscriptions(resource string) *ProjectsLocationsDataExchangesListSubscriptionsCall {\n\tc := &ProjectsLocationsDataExchangesListSubscriptionsCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.resource = resource\n\treturn c\n}", "func (em *EntityManager) UpdateSubscriptions(id EID) {\n\tem.UpdateComponentFlags(id)\n\n\tfor idx := range em.Subscriptions {\n\t\tallowed := em.Subscriptions[idx].Filter.Allow(em.ComponentFlags[id])\n\t\tem.Subscriptions[idx].dirty = true\n\t\tem.Subscriptions[idx].entityBitVector.Set(int(id), allowed)\n\t}\n}", "func (r *ProjectsLocationsDataExchangesListingsService) ListSubscriptions(resource string) *ProjectsLocationsDataExchangesListingsListSubscriptionsCall {\n\tc := &ProjectsLocationsDataExchangesListingsListSubscriptionsCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.resource = resource\n\treturn c\n}", "func (api API) ListSubscriptions(cursor *Cursor, customerUUID string) (*Subscriptions, error) {\n\tresult := &Subscriptions{}\n\tpath := strings.Replace(subscriptionsEndpoint, \":customerUUID\", customerUUID, 1)\n\tquery := make([]interface{}, 0, 1)\n\tif cursor != nil {\n\t\tquery = append(query, *cursor)\n\t}\n\treturn result, api.list(path, result, query...)\n}", "func NewSubscriptions(client *gosip.SPClient, endpoint string, config *RequestConfig) *Subscriptions {\n\treturn &Subscriptions{\n\t\tclient: client,\n\t\tendpoint: endpoint,\n\t\tconfig: config,\n\t}\n}", "func (router *Router) Subscriptions() {\n\tfor {\n\t\tvar sub *Subscription\n\t\tselect {\n\t\tcase sub = <-router.channels.subAdd:\n\t\t\trouter.elog.Logf(elog.LogLevelInfo2, \"SubAdd\")\n\t\tcase sub = <-router.channels.subMod:\n\t\t\trouter.elog.Logf(elog.LogLevelInfo2, \"SubMod\")\n\t\tcase sub = <-router.channels.subDel:\n\t\t\trouter.elog.Logf(elog.LogLevelInfo2, \"SubDel\")\n\t\t}\n\n\t\tif sub.SubID == 0 {\n\t\t\trouter.elog.Logf(elog.LogLevelError, \"FIXME: Use sub to keep compiler happy\")\n\t\t}\n\n\t}\n}", "func SetServerSubscription(s []string) func(*Server) error {\n\treturn func(c *Server) error {\n\t\tif s != nil {\n\t\t\tfor _, d := range s {\n\t\t\t\tc.subscriptionURLs = append(c.subscriptionURLs, d)\n\t\t\t}\n\t\t\treturn nil\n\t\t}\n\t\tc.subscriptionURLs = append(c.subscriptionURLs, \"http://joajgazyztfssty4w2on5oaqksz6tqoxbduy553y34mf4byv6gpq.b32.i2p/export/alive-hosts.txt\")\n\t\treturn nil\n\t}\n}", "func (s *T) Subscriptions() <-chan map[string][]string {\n\treturn s.subscriptionsCh\n}", "func (c *Client) ListSubscriptions(namespace string) (*v1alpha1.SubscriptionList, error) {\n\tif err := c.initClient(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tsubscriptionList := &v1alpha1.SubscriptionList{}\n\tif err := c.crClient.List(\n\t\tcontext.TODO(),\n\t\tsubscriptionList,\n\t\t&client.ListOptions{\n\t\t\tNamespace: namespace,\n\t\t},\n\t); err != nil {\n\t\treturn subscriptionList, err\n\t}\n\treturn subscriptionList, nil\n}", "func (a *StreamsApiService) UpdateSubscriptions(ctx _context.Context) ApiUpdateSubscriptionsRequest {\n\treturn ApiUpdateSubscriptionsRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (api API) ConnectSubscriptions(customerUUID string, subscriptions []Subscription) error {\n\tpath := strings.Replace(connectSubscriptionEndpoint, \":uuid\", customerUUID, 1)\n\treturn api.merge(path, Subscriptions{\n\t\tSubscriptions: subscriptions,\n\t})\n}", "func (s *CreateBudgetActionInput) SetSubscribers(v []*Subscriber) *CreateBudgetActionInput {\n\ts.Subscribers = v\n\treturn s\n}", "func (sns *SNS) ListSubscriptions(NextToken *string) (resp *ListSubscriptionsResp, err error) {\n\tresp = &ListSubscriptionsResp{}\n\tparams := makeParams(\"ListSubscriptions\")\n\tif NextToken != nil {\n\t\tparams[\"NextToken\"] = *NextToken\n\t}\n\terr = sns.query(nil, nil, params, resp)\n\treturn\n}", "func (s *UpdateBudgetActionInput) SetSubscribers(v []*Subscriber) *UpdateBudgetActionInput {\n\ts.Subscribers = v\n\treturn s\n}", "func (c *Client) ListSubscriptions(params *ListSubscriptionsParams, opts ...Option) (SubscriptionLister, error) {\n\tpath, err := c.InterpolatePath(\"/subscriptions\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trequestOptions := NewRequestOptions(opts...)\n\tpath = BuildURL(path, params)\n\treturn NewSubscriptionList(c, path, requestOptions), nil\n}", "func (svc *UserDomService) ReplaceSubscriptions(ctx context.Context, id string, relationData []*domain.Subscription) error {\n\treturn svc.store.User.ReplaceSubscriptions(ctx, id, relationData)\n}", "func (s *CreateNotificationInput) SetSubscribers(v []*Subscriber) *CreateNotificationInput {\n\ts.Subscribers = v\n\treturn s\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetSystem sets the system property value. If present, indicates that this is a systemmanaged list. Readonly.
func (m *List) SetSystem(value SystemFacetable)() { m.system = value }
[ "func (o *RoleWithAccess) SetSystem(v bool) {\n\to.System = &v\n}", "func (a *IssueManagement) SetSystem(value string) {\n\tcopy := value\n\ta.System = &copy\n\n}", "func (o *IamServiceProviderAllOf) SetSystem(v IamSystemRelationship) {\n\to.System = &v\n}", "func (m *Drive) SetSystem(value SystemFacetable)() {\n m.system = value\n}", "func (b *binding) SetSystemProperty(ctx context.Context, name, value string) error {\n\tif len(value) == 0 {\n\t\tvalue = `\"\"`\n\t}\n\tres, err := b.Shell(\"setprop\", name, value).Call(ctx)\n\tif res != \"\" {\n\t\treturn log.Errf(ctx, nil, \"setprop returned error: \\n%s\", res)\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (m *ProvisioningObjectSummary) SetTargetSystem(value ProvisioningSystemable)() {\n m.targetSystem = value\n}", "func (a *Dependency) SetSystemPath(value string) {\n\tcopy := value\n\ta.SystemPath = &copy\n\n}", "func RegisterSystem(scope parens.Scope) error {\n\treturn registerList(scope, system)\n}", "func RegisterSystem(scope parser.Scope) error {\n\treturn registerList(scope, system)\n}", "func (m *AccessPackageResourceScope) SetOriginSystem(value *string)() {\n err := m.GetBackingStore().Set(\"originSystem\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *AndroidManagedStoreApp) SetIsSystemApp(value *bool)() {\n err := m.GetBackingStore().Set(\"isSystemApp\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *MyErratum) SetSystems(v string) {\n\to.Systems = &v\n}", "func (m *ProvisioningObjectSummary) SetSourceSystem(value ProvisioningSystemable)() {\n m.sourceSystem = value\n}", "func (c *Client) UpdateSystem(system *System) error {\n\titem := reflect.ValueOf(system).Elem()\n\tid, err := c.GetItemHandle(\"system\", system.Name)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn c.updateCobblerFields(\"system\", item, id)\n}", "func (inst *ValidateSafetyDepositBoxV2) SetSystemAccount(system ag_solanago.PublicKey) *ValidateSafetyDepositBoxV2 {\n\tinst.AccountMetaSlice[16] = ag_solanago.Meta(system)\n\treturn inst\n}", "func (o *HyperflexHxPlatformDatastoreConfigDtAllOf) SetSystemDatastore(v bool) {\n\to.SystemDatastore = &v\n}", "func SetSystemLogLevel(l level.Level) {\n\tSystemLogLevel = l\n}", "func (cu *CounterUpdate) SetNetworkManagerSystem(s string) *CounterUpdate {\n\tcu.mutation.SetNetworkManagerSystem(s)\n\treturn cu\n}", "func (o *SystemDiagnosticsEntity) SetSystemDiagnostics(v SystemDiagnosticsDTO) {\n\to.SystemDiagnostics = &v\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Chair specifies the participation role for the calendar user specified by the property is CHAIR.
func Chair(v string) parameter.Parameter { return Other("CHAIR") }
[ "func (v *victorianFurnitureFactory) CreateChair() Chairer {\n\treturn &victorianChair{}\n}", "func chamberSpeakerRole(s *discordgo.Session, channelID string) (*discordgo.Role, error) {\n\tchamber, ok := Chambers[channelID]\n\tif !ok {\n\t\treturn nil, ERR_NOT_A_CHAMBER\n\t}\n\n\tch, err := s.State.Channel(channelID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.State.Role(ch.GuildID, chamber.SpeakerRole)\n}", "func (hba *HostBasedAuthentication) Role(name string) *HostBasedAuthentication {\n\thba.user = \"+\" + hba.quote(name)\n\treturn hba\n}", "func desiredRole(name string, contour *operatorv1alpha1.Contour) *rbacv1.Role {\n\trole := &rbacv1.Role{\n\t\tTypeMeta: metav1.TypeMeta{\n\t\t\tKind: \"Role\",\n\t\t},\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tNamespace: contour.Spec.Namespace.Name,\n\t\t\tName: name,\n\t\t},\n\t}\n\tgroupAll := []string{\"\"}\n\tverbCU := []string{\"create\", \"update\"}\n\tsecret := rbacv1.PolicyRule{\n\t\tVerbs: verbCU,\n\t\tAPIGroups: groupAll,\n\t\tResources: []string{\"secrets\"},\n\t}\n\trole.Rules = []rbacv1.PolicyRule{secret}\n\trole.Labels = map[string]string{\n\t\toperatorv1alpha1.OwningContourNameLabel: contour.Name,\n\t\toperatorv1alpha1.OwningContourNsLabel: contour.Namespace,\n\t}\n\treturn role\n}", "func CadenceWorkerActor(worker CadenceWorker) (execute func() error, interrupt func(error)) {\n\tcloseCh := make(chan struct{})\n\n\treturn func() error {\n\t\t\terr := worker.Start()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t<-closeCh\n\n\t\t\treturn nil\n\t\t}, func(error) {\n\t\t\tworker.Stop()\n\t\t\tclose(closeCh)\n\t\t}\n}", "func (e *ClusterElector) Role() Role {\n\tif (e.state & stateLeaderBootStrapping) != 0 {\n\t\treturn RoleUnstable\n\t}\n\treturn e.role\n}", "func (j *AuroraJob) Role(role string) Job {\n\tj.jobConfig.Key.Role = role\n\n\t// Will be deprecated\n\tidentity := &aurora.Identity{User: role}\n\tj.jobConfig.Owner = identity\n\tj.jobConfig.TaskConfig.Owner = identity\n\treturn j\n}", "func (c *TiFlashComponent) Role() string {\n\treturn ComponentTiFlash\n}", "func (d Dispatcher) Chord(jrs []string, callbackJr string) (string, error) {\n\t//TODO: send result to message broker\n\tvar requests []job.Request\n\tfor _, jr := range jrs {\n\t\tvar request job.Request\n\t\terr := helpers.Deserialize([]byte(jr), &request)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\trequests = append(requests, request)\n\t}\n\tvar callbackRequest job.Request\n\terr := helpers.Deserialize([]byte(callbackJr), &callbackRequest)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tc, err := chord.NewChord(requests, callbackRequest, d.GetBC(), d.GetJobPQ(), d.GetJC())\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tc.Dispatch()\n\tresult, err := helpers.Serialize(c.Result())\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(result), nil\n}", "func (c *Config) Role() int {\n\trole := c.Get(\"role\", \"follower\")\n\tswitch role {\n\tcase \"follower\":\n\t\treturn FOLLOWER\n\tcase \"leader\":\n\t\treturn LEADER\n\tdefault:\n\t\tlog.Panic(\"Invalid role: %s.\", role)\n\t}\n\treturn LEADER\n}", "func (_BREM *BREMCaller) ROLEAUDITOR(opts *bind.CallOpts) (string, error) {\n\tvar (\n\t\tret0 = new(string)\n\t)\n\tout := ret0\n\terr := _BREM.contract.Call(opts, out, \"ROLE_AUDITOR\")\n\treturn *ret0, err\n}", "func (r PractitionerRole) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(struct {\n\t\tOtherPractitionerRole\n\t\tResourceType string `json:\"resourceType\"`\n\t}{\n\t\tOtherPractitionerRole: OtherPractitionerRole(r),\n\t\tResourceType: \"PractitionerRole\",\n\t})\n}", "func (user *User) SetAsParticipant() {\n\tuser.Role = UserRoleParticipant\n}", "func chamberMemberRole(s *discordgo.Session, channelID string) (*discordgo.Role, error) {\n\tchamber, ok := Chambers[channelID]\n\tif !ok {\n\t\treturn nil, ERR_NOT_A_CHAMBER\n\t}\n\n\tch, err := s.State.Channel(channelID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.State.Role(ch.GuildID, chamber.MemberRole)\n}", "func ChoriaPlugin() plugin.Pluggable {\n\treturn mcorpc.NewChoriaAgentPlugin(metadata, New)\n}", "func (_Harberger *HarbergerCallerSession) BURNROLE() ([32]byte, error) {\n\treturn _Harberger.Contract.BURNROLE(&_Harberger.CallOpts)\n}", "func LeagueByChallenger(region string, queueType string) (league League, err error) {\n\tif !IsKeySet() {\n\t\treturn league, ErrAPIKeyNotSet\n\t}\n\targs := fmt.Sprintf(\n\t\t\"type=%v&api_key=%v\",\n\t\tqueueType,\n\t\tapikey)\n\turl := fmt.Sprintf(\n\t\t\"%v/lol/%v/v2.3/league/challenger?%v\",\n\t\tBaseURL,\n\t\tregion,\n\t\targs)\n\terr = requestAndUnmarshal(url, &league)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn league, err\n}", "func (s *Store) LeaderCh() <-chan bool {\n\ts.mu.RLock()\n\tdefer s.mu.RUnlock()\n\tassert(s.raft != nil, \"cannot retrieve leadership channel when closed\")\n\treturn s.raft.LeaderCh()\n}", "func (_Userable *UserableCaller) ROLEAUDITOR(opts *bind.CallOpts) (string, error) {\n\tvar (\n\t\tret0 = new(string)\n\t)\n\tout := ret0\n\terr := _Userable.contract.Call(opts, out, \"ROLE_AUDITOR\")\n\treturn *ret0, err\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
ReqParticipant specifies the participation role for the calendar user specified by the property is a required participant, REQPARTICIPANT.
func ReqParticipant() parameter.Parameter { return Other("REQ-PARTICIPANT") }
[ "func OptParticipant() parameter.Parameter {\n\treturn Other(\"OPT-PARTICIPANT\")\n}", "func (s *Item) SetParticipantRole(v string) *Item {\n\ts.ParticipantRole = &v\n\treturn s\n}", "func (s *SentimentFilter) SetParticipantRole(v string) *SentimentFilter {\n\ts.ParticipantRole = &v\n\treturn s\n}", "func (s *ParticipantTimerConfiguration) SetParticipantRole(v string) *ParticipantTimerConfiguration {\n\ts.ParticipantRole = &v\n\treturn s\n}", "func (c *participantsRESTClient) UpdateParticipant(ctx context.Context, req *dialogflowpb.UpdateParticipantRequest, opts ...gax.CallOption) (*dialogflowpb.Participant, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tbody := req.GetParticipant()\n\tjsonReq, err := m.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2/%v\", req.GetParticipant().GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\tif req.GetUpdateMask() != nil {\n\t\tupdateMask, err := protojson.Marshal(req.GetUpdateMask())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tparams.Add(\"updateMask\", string(updateMask[1:len(updateMask)-1]))\n\t}\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"participant.name\", url.QueryEscape(req.GetParticipant().GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).UpdateParticipant[0:len((*c.CallOptions).UpdateParticipant):len((*c.CallOptions).UpdateParticipant)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &dialogflowpb.Participant{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"PATCH\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (duo *DatumUpdateOne) SetParticipant(p *Participant) *DatumUpdateOne {\n\treturn duo.SetParticipantID(p.ID)\n}", "func (du *DatumUpdate) SetParticipant(p *Participant) *DatumUpdate {\n\treturn du.SetParticipantID(p.ID)\n}", "func (r *ChannelsReportSpamRequest) GetParticipant() (value InputPeerClass) {\n\tif r == nil {\n\t\treturn\n\t}\n\treturn r.Participant\n}", "func (user *User) IsParticipant() bool {\n\treturn user.Role == UserRoleParticipant\n}", "func (c *ApiService) UpdateParticipant(ConferenceSid string, CallSid string, params *UpdateParticipantParams) (*ApiV2010Participant, error) {\n\tpath := \"/2010-04-01/Accounts/{AccountSid}/Conferences/{ConferenceSid}/Participants/{CallSid}.json\"\n\tif params != nil && params.PathAccountSid != nil {\n\t\tpath = strings.Replace(path, \"{\"+\"AccountSid\"+\"}\", *params.PathAccountSid, -1)\n\t} else {\n\t\tpath = strings.Replace(path, \"{\"+\"AccountSid\"+\"}\", c.requestHandler.Client.AccountSid(), -1)\n\t}\n\tpath = strings.Replace(path, \"{\"+\"ConferenceSid\"+\"}\", ConferenceSid, -1)\n\tpath = strings.Replace(path, \"{\"+\"CallSid\"+\"}\", CallSid, -1)\n\n\tdata := url.Values{}\n\theaders := make(map[string]interface{})\n\n\tif params != nil && params.Muted != nil {\n\t\tdata.Set(\"Muted\", fmt.Sprint(*params.Muted))\n\t}\n\tif params != nil && params.Hold != nil {\n\t\tdata.Set(\"Hold\", fmt.Sprint(*params.Hold))\n\t}\n\tif params != nil && params.HoldUrl != nil {\n\t\tdata.Set(\"HoldUrl\", *params.HoldUrl)\n\t}\n\tif params != nil && params.HoldMethod != nil {\n\t\tdata.Set(\"HoldMethod\", *params.HoldMethod)\n\t}\n\tif params != nil && params.AnnounceUrl != nil {\n\t\tdata.Set(\"AnnounceUrl\", *params.AnnounceUrl)\n\t}\n\tif params != nil && params.AnnounceMethod != nil {\n\t\tdata.Set(\"AnnounceMethod\", *params.AnnounceMethod)\n\t}\n\tif params != nil && params.WaitUrl != nil {\n\t\tdata.Set(\"WaitUrl\", *params.WaitUrl)\n\t}\n\tif params != nil && params.WaitMethod != nil {\n\t\tdata.Set(\"WaitMethod\", *params.WaitMethod)\n\t}\n\tif params != nil && params.BeepOnExit != nil {\n\t\tdata.Set(\"BeepOnExit\", fmt.Sprint(*params.BeepOnExit))\n\t}\n\tif params != nil && params.EndConferenceOnExit != nil {\n\t\tdata.Set(\"EndConferenceOnExit\", fmt.Sprint(*params.EndConferenceOnExit))\n\t}\n\tif params != nil && params.Coaching != nil {\n\t\tdata.Set(\"Coaching\", fmt.Sprint(*params.Coaching))\n\t}\n\tif params != nil && params.CallSidToCoach != nil {\n\t\tdata.Set(\"CallSidToCoach\", *params.CallSidToCoach)\n\t}\n\n\tresp, err := c.requestHandler.Post(c.baseURL+path, data, headers)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer resp.Body.Close()\n\n\tps := &ApiV2010Participant{}\n\tif err := json.NewDecoder(resp.Body).Decode(ps); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ps, err\n}", "func (e *EncryptedChatRequested) GetParticipantID() (value int) {\n\treturn e.ParticipantID\n}", "func ScheduleChangeRequestActorPRecipient() *ScheduleChangeRequestActor {\n\tv := ScheduleChangeRequestActorVRecipient\n\treturn &v\n}", "func (u *UpdateChannelParticipant) SetPrevParticipant(value ChannelParticipantClass) {\n\tu.Flags.Set(0)\n\tu.PrevParticipant = value\n}", "func (p *Poll) IsParticipantConfirmed(mail, poll string) bool {\n\n\tquery := fmt.Sprintf(\"SELECT participate FROM participate WHERE mail = '%v' AND poll = '%v'\", mail, poll)\n\n\trows := p.db.QueryRow(query)\n\n\tvar participate string\n\n\terr := rows.Scan(&participate)\n\n\tswitch {\n\tcase err == sql.ErrNoRows:\n\t\treturn false\n\tcase err != nil:\n\t\tlog.Println(err)\n\t\treturn false\n\tdefault:\n\t\tvalue, err := strconv.ParseBool(participate)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\t\treturn value\n\t}\n\n}", "func (user *User) SetAsParticipant() {\n\tuser.Role = UserRoleParticipant\n}", "func (s *DealService) DeleteParticipant(ctx context.Context, dealID int, participantID int) (*Response, error) {\n\turi := fmt.Sprintf(\"/deals/%v/participants/%v\", dealID, participantID)\n\treq, err := s.client.NewRequest(http.MethodDelete, uri, nil, nil)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.client.Do(ctx, req, nil)\n}", "func (*Appointment_Participant) Descriptor() ([]byte, []int) {\n\treturn file_proto_google_fhir_proto_dstu2_resources_proto_rawDescGZIP(), []int{2, 0}\n}", "func ScheduleChangeRequestActorPSender() *ScheduleChangeRequestActor {\n\tv := ScheduleChangeRequestActorVSender\n\treturn &v\n}", "func CreateConfirmFabricConsortiumMemberRequest() (request *ConfirmFabricConsortiumMemberRequest) {\n\trequest = &ConfirmFabricConsortiumMemberRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Baas\", \"2018-12-21\", \"ConfirmFabricConsortiumMember\", \"baas\", \"openAPI\")\n\treturn\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
OptParticipant specifies the participation role for the calendar user specified by the property is an optional participant, OPTPARTICIPANT.
func OptParticipant() parameter.Parameter { return Other("OPT-PARTICIPANT") }
[ "func ReqParticipant() parameter.Parameter {\n\treturn Other(\"REQ-PARTICIPANT\")\n}", "func (s *Item) SetParticipantRole(v string) *Item {\n\ts.ParticipantRole = &v\n\treturn s\n}", "func (s *ParticipantTimerConfiguration) SetParticipantRole(v string) *ParticipantTimerConfiguration {\n\ts.ParticipantRole = &v\n\treturn s\n}", "func (s *SentimentFilter) SetParticipantRole(v string) *SentimentFilter {\n\ts.ParticipantRole = &v\n\treturn s\n}", "func (duo *DatumUpdateOne) SetParticipant(p *Participant) *DatumUpdateOne {\n\treturn duo.SetParticipantID(p.ID)\n}", "func (du *DatumUpdate) SetParticipant(p *Participant) *DatumUpdate {\n\treturn du.SetParticipantID(p.ID)\n}", "func (user *User) IsParticipant() bool {\n\treturn user.Role == UserRoleParticipant\n}", "func (user *User) SetAsParticipant() {\n\tuser.Role = UserRoleParticipant\n}", "func (m *MeetingParticipants) SetOrganizer(value MeetingParticipantInfoable)() {\n m.organizer = value\n}", "func (c *ApiService) UpdateParticipant(ConferenceSid string, CallSid string, params *UpdateParticipantParams) (*ApiV2010Participant, error) {\n\tpath := \"/2010-04-01/Accounts/{AccountSid}/Conferences/{ConferenceSid}/Participants/{CallSid}.json\"\n\tif params != nil && params.PathAccountSid != nil {\n\t\tpath = strings.Replace(path, \"{\"+\"AccountSid\"+\"}\", *params.PathAccountSid, -1)\n\t} else {\n\t\tpath = strings.Replace(path, \"{\"+\"AccountSid\"+\"}\", c.requestHandler.Client.AccountSid(), -1)\n\t}\n\tpath = strings.Replace(path, \"{\"+\"ConferenceSid\"+\"}\", ConferenceSid, -1)\n\tpath = strings.Replace(path, \"{\"+\"CallSid\"+\"}\", CallSid, -1)\n\n\tdata := url.Values{}\n\theaders := make(map[string]interface{})\n\n\tif params != nil && params.Muted != nil {\n\t\tdata.Set(\"Muted\", fmt.Sprint(*params.Muted))\n\t}\n\tif params != nil && params.Hold != nil {\n\t\tdata.Set(\"Hold\", fmt.Sprint(*params.Hold))\n\t}\n\tif params != nil && params.HoldUrl != nil {\n\t\tdata.Set(\"HoldUrl\", *params.HoldUrl)\n\t}\n\tif params != nil && params.HoldMethod != nil {\n\t\tdata.Set(\"HoldMethod\", *params.HoldMethod)\n\t}\n\tif params != nil && params.AnnounceUrl != nil {\n\t\tdata.Set(\"AnnounceUrl\", *params.AnnounceUrl)\n\t}\n\tif params != nil && params.AnnounceMethod != nil {\n\t\tdata.Set(\"AnnounceMethod\", *params.AnnounceMethod)\n\t}\n\tif params != nil && params.WaitUrl != nil {\n\t\tdata.Set(\"WaitUrl\", *params.WaitUrl)\n\t}\n\tif params != nil && params.WaitMethod != nil {\n\t\tdata.Set(\"WaitMethod\", *params.WaitMethod)\n\t}\n\tif params != nil && params.BeepOnExit != nil {\n\t\tdata.Set(\"BeepOnExit\", fmt.Sprint(*params.BeepOnExit))\n\t}\n\tif params != nil && params.EndConferenceOnExit != nil {\n\t\tdata.Set(\"EndConferenceOnExit\", fmt.Sprint(*params.EndConferenceOnExit))\n\t}\n\tif params != nil && params.Coaching != nil {\n\t\tdata.Set(\"Coaching\", fmt.Sprint(*params.Coaching))\n\t}\n\tif params != nil && params.CallSidToCoach != nil {\n\t\tdata.Set(\"CallSidToCoach\", *params.CallSidToCoach)\n\t}\n\n\tresp, err := c.requestHandler.Post(c.baseURL+path, data, headers)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer resp.Body.Close()\n\n\tps := &ApiV2010Participant{}\n\tif err := json.NewDecoder(resp.Body).Decode(ps); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ps, err\n}", "func NonParticipant() parameter.Parameter {\n\treturn Other(\"NON-PARTICIPANT\")\n}", "func (p *Poll) IsParticipantConfirmed(mail, poll string) bool {\n\n\tquery := fmt.Sprintf(\"SELECT participate FROM participate WHERE mail = '%v' AND poll = '%v'\", mail, poll)\n\n\trows := p.db.QueryRow(query)\n\n\tvar participate string\n\n\terr := rows.Scan(&participate)\n\n\tswitch {\n\tcase err == sql.ErrNoRows:\n\t\treturn false\n\tcase err != nil:\n\t\tlog.Println(err)\n\t\treturn false\n\tdefault:\n\t\tvalue, err := strconv.ParseBool(participate)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\t\treturn value\n\t}\n\n}", "func (s *DealService) DeleteParticipant(ctx context.Context, dealID int, participantID int) (*Response, error) {\n\turi := fmt.Sprintf(\"/deals/%v/participants/%v\", dealID, participantID)\n\treq, err := s.client.NewRequest(http.MethodDelete, uri, nil, nil)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.client.Do(ctx, req, nil)\n}", "func (u *UpdateChannelParticipant) SetPrevParticipant(value ChannelParticipantClass) {\n\tu.Flags.Set(0)\n\tu.PrevParticipant = value\n}", "func (c *participantsRESTClient) UpdateParticipant(ctx context.Context, req *dialogflowpb.UpdateParticipantRequest, opts ...gax.CallOption) (*dialogflowpb.Participant, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tbody := req.GetParticipant()\n\tjsonReq, err := m.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2/%v\", req.GetParticipant().GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\tif req.GetUpdateMask() != nil {\n\t\tupdateMask, err := protojson.Marshal(req.GetUpdateMask())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tparams.Add(\"updateMask\", string(updateMask[1:len(updateMask)-1]))\n\t}\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"participant.name\", url.QueryEscape(req.GetParticipant().GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).UpdateParticipant[0:len((*c.CallOptions).UpdateParticipant):len((*c.CallOptions).UpdateParticipant)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &dialogflowpb.Participant{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"PATCH\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (p *Poll) IsParticipant(mail, poll string) bool {\n\n\tquery := fmt.Sprintf(\"SELECT id FROM participate WHERE mail = '%v' AND poll = '%v'\", mail, poll)\n\n\trows := p.db.QueryRow(query)\n\n\tvar id string\n\n\terr := rows.Scan(&id)\n\n\tswitch {\n\tcase err == sql.ErrNoRows:\n\t\treturn false\n\tcase err != nil:\n\t\tlog.Println(err)\n\t\treturn false\n\tdefault:\n\t\treturn true\n\t}\n\n}", "func (store *Store) DebugSetParticipant(ctx context.Context, p *model.Participant) error {\n\tkey := participantKey(p.ID)\n\t_, err := store.dsClient.Put(ctx, key, p)\n\treturn err\n}", "func DelegateCandidateOption() AccountCreationOption {\n\treturn func(account *Account) error {\n\t\taccount.isCandidate = true\n\t\treturn nil\n\t}\n}", "func (s *Item) SetParticipantId(v string) *Item {\n\ts.ParticipantId = &v\n\treturn s\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
NonParticipant specifies the participation role for the calendar user specified by the property is a nonparticipant, NONPARTICIPANT.
func NonParticipant() parameter.Parameter { return Other("NON-PARTICIPANT") }
[ "func NonInteractiveNEQ(v bool) predicate.Event {\n\treturn predicate.Event(func(s *sql.Selector) {\n\t\ts.Where(sql.NEQ(s.C(FieldNonInteractive), v))\n\t})\n}", "func NonInteractive(v bool) predicate.Event {\n\treturn predicate.Event(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldNonInteractive), v))\n\t})\n}", "func (rec *RawEventCreate) SetNonInteractive(b bool) *RawEventCreate {\n\trec.mutation.SetNonInteractive(b)\n\treturn rec\n}", "func (ec *EventCreate) SetNonInteractive(b bool) *EventCreate {\n\tec.mutation.SetNonInteractive(b)\n\treturn ec\n}", "func OutcomeOverviewParticipantsNotIn(vs ...string) predicate.OutcomeOverview {\n\tv := make([]interface{}, len(vs))\n\tfor i := range v {\n\t\tv[i] = vs[i]\n\t}\n\treturn predicate.OutcomeOverview(func(s *sql.Selector) {\n\t\t// if not arguments were provided, append the FALSE constants,\n\t\t// since we can't apply \"IN ()\". This will make this predicate falsy.\n\t\tif len(v) == 0 {\n\t\t\ts.Where(sql.False())\n\t\t\treturn\n\t\t}\n\t\ts.Where(sql.NotIn(s.C(FieldOutcomeOverviewParticipants), v...))\n\t})\n}", "func ReplicaTypeNonVoter() *ReplicaType {\n\tt := NON_VOTER\n\treturn &t\n}", "func OutcomeOverviewParticipantsNEQ(v string) predicate.OutcomeOverview {\n\treturn predicate.OutcomeOverview(func(s *sql.Selector) {\n\t\ts.Where(sql.NEQ(s.C(FieldOutcomeOverviewParticipants), v))\n\t})\n}", "func (s *Rule) SetNonTalkTimeFilter(v *NonTalkTimeFilter) *Rule {\n\ts.NonTalkTimeFilter = v\n\treturn s\n}", "func (cb CommitteeBits) FilterNonParticipants(committee []ValidatorIndex) []ValidatorIndex {\n\tbitLen := cb.BitLen()\n\tout := committee[:0]\n\tif bitLen != uint64(len(committee)) {\n\t\tpanic(\"committee mismatch, bitfield length does not match\")\n\t}\n\tfor i := uint64(0); i < bitLen; i++ {\n\t\tif !cb.GetBit(i) {\n\t\t\tout = append(out, committee[i])\n\t\t}\n\t}\n\treturn out\n}", "func OptParticipant() parameter.Parameter {\n\treturn Other(\"OPT-PARTICIPANT\")\n}", "func Not(e TemporalExpression) NotExpression {\n\treturn NotExpression{e}\n}", "func (dao *GkvDB) RemoveNonParticipantChannel(channel common.Hash) error {\n\tvar m nonParticipantChannel\n\terr := dao.getKeyValueToBucket(models.BucketChannel, channel[:], &m)\n\tif err != nil {\n\t\treturn models.GeneratDBError(err)\n\t}\n\terr = dao.removeKeyValueFromBucket(models.BucketChannel, channel[:])\n\treturn models.GeneratDBError(err)\n}", "func (user *User) IsParticipant() bool {\n\treturn user.Role == UserRoleParticipant\n}", "func (me TAssociationTypeType) IsNonProfit() bool { return me.String() == \"Non-Profit\" }", "func not(e semantic.Expression) semantic.Expression {\n\treturn &semantic.UnaryOp{Type: semantic.BoolType, Expression: e, Operator: ast.OpNot}\n}", "func NonPrimarySelector(name string) labels.Selector {\n\ts := labels.SelectorFromSet(labels.Set{constants.ClusterLabel: name})\n\trequirement, _ := labels.NewRequirement(constants.LabelClusterRole, selection.NotIn, []string{constants.ClusterRolePrimary})\n\treturn s.Add(*requirement)\n}", "func (self *Event) NotPitchingTeamIterator() model.Iterator {\n\treturn EventTeams.Filter(\"Event\", self.ID).Filter(\"Pitching\", false).SortFunc(compareTeamNames)\n}", "func NonInteractiveEQ(v bool) predicate.Event {\n\treturn predicate.Event(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldNonInteractive), v))\n\t})\n}", "func notActiveRule(name string) model.MutableRule {\n\trule := ruleapi.NewRule(name)\n\trule.AddCondition(\"c1\", []string{\"orgstatus\"}, notOrgActive, nil)\n\trule.SetAction(notActiveAction)\n\treturn rule\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Other specifies some other participation role for the calendar user specified by the property.
func Other(v string) parameter.Parameter { return parameter.Single(ROLE, v) }
[ "func (o *W2) SetOther(v string) {\n\to.Other.Set(&v)\n}", "func (this *activityGroupStruct) Other() OtherActivity {\n\to := &this.other\n\treturn o\n}", "func (me TxsdActuate) IsOther() bool { return me.String() == \"other\" }", "func (me TxsdIncidentPurpose) IsOther() bool { return me.String() == \"other\" }", "func (me TxsdActuate) IsOther() bool { return me == \"other\" }", "func (r *Restriction) UsedOther(tx *sqlx.Tx, id mtid.MTID) error {\n\tjs, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\treturn mytokenrepohelper.IncreaseTokenUsageOther(tx, id, js)\n}", "func NonParticipant() parameter.Parameter {\n\treturn Other(\"NON-PARTICIPANT\")\n}", "func (o *W2) GetOther() string {\n\tif o == nil || o.Other.Get() == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Other.Get()\n}", "func (o SubnetworkOutput) Role() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Subnetwork) pulumi.StringOutput { return v.Role }).(pulumi.StringOutput)\n}", "func DenyIfStatusChangedByOther() privacy.MutationRule {\n\tpolicy := privacy.TaskMutationRuleFunc(func(ctx context.Context, m *ent.TaskMutation) error {\n\t\t// Skip if the mutation does not change the task status.\n\t\tif _, exists := m.Status(); !exists {\n\t\t\treturn privacy.Skip\n\t\t}\n\t\tview, ok := viewer.FromContext(ctx).(*viewer.UserViewer)\n\t\t// Skip if the viewer is an admin (or an app).\n\t\tif !ok || view.Admin() {\n\t\t\treturn privacy.Skip\n\t\t}\n\t\tid, ok := m.ID()\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"missing task id\")\n\t\t}\n\t\towner, err := m.Client().User.Query().Where(user.HasTasksWith(task.ID(id))).Only(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Deny the mutation, if the viewer is not the owner.\n\t\tif owner.ID != view.User.ID {\n\t\t\treturn privacy.Denyf(\"viewer %d is not allowed to change the task status\", view.User.ID)\n\t\t}\n\t\treturn privacy.Skip\n\t})\n\treturn privacy.OnMutationOperation(policy, ent.OpUpdateOne)\n}", "func (p Player) Other() Player {\n\tif p == Black {\n\t\treturn White\n\t} else {\n\t\treturn Black\n\t}\n}", "func (me TPublicationTypeTypes) IsOther() bool { return me.String() == \"Other\" }", "func (proc *ConsensusProcess) currentRole() Role {\n\tif proc.oracle.Eligible(hashInstanceAndK(proc.instanceId, proc.k), proc.expectedCommitteeSize(proc.k), proc.signing.Verifier().String(), proc.roleProof()) {\n\t\tif proc.currentRound() == Round2 {\n\t\t\treturn Leader\n\t\t}\n\t\treturn Active\n\t}\n\n\treturn Passive\n}", "func (e *ClusterElector) Role() Role {\n\tif (e.state & stateLeaderBootStrapping) != 0 {\n\t\treturn RoleUnstable\n\t}\n\treturn e.role\n}", "func (r *RolePolicy) Role() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"role\"])\n}", "func (me TactionType) IsOther() bool { return me.String() == \"other\" }", "func EmailRolePUnknown() *EmailRole {\n\tv := EmailRoleVUnknown\n\treturn &v\n}", "func EmailRolePRecipient() *EmailRole {\n\tv := EmailRoleVRecipient\n\treturn &v\n}", "func EmailRolePUnknownFutureValue() *EmailRole {\n\tv := EmailRoleVUnknownFutureValue\n\treturn &v\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
CommonPool returns the balance of the global common pool.
func (s *ImmutableState) CommonPool() (*quantity.Quantity, error) { _, value := s.Snapshot.Get(commonPoolKeyFmt.Encode()) if value == nil { return &quantity.Quantity{}, nil } var q quantity.Quantity if err := cbor.Unmarshal(value, &q); err != nil { return nil, err } return &q, nil }
[ "func (m *Acknowledgment) PoolBalance() uint64 {\n\treturn m.Terms.GetAmount()\n}", "func (p *P2C) Balance(key string) (string, error) {\n\tp.Lock()\n\tdefer p.Unlock()\n\n\tif len(p.hosts) == 0 {\n\t\treturn \"\", liblb.ErrNoHost\n\t}\n\n\t// chosen host\n\tvar host string\n\n\tvar n1, n2 string\n\n\tif len(key) > 0 {\n\t\tn1, n2 = p.hash(key)\n\t} else {\n\t\tn1 = p.hosts[p.rndm.Intn(len(p.hosts))].name\n\t\tn2 = p.hosts[p.rndm.Intn(len(p.hosts))].name\n\t}\n\n\thost = n2\n\n\tif p.loadMap[n1].load <= p.loadMap[n2].load {\n\t\thost = n1\n\t}\n\n\tp.loadMap[host].load++\n\treturn host, nil\n}", "func (sch *Scheduler) CalClusterBalance(podUsed *[PHYNUM][DIMENSION]float64, podReq []PodRequest) {\n\t//cal the pod sum and used rate\n\tpodLen := len(podReq)\n\tvar podNum [PHYNUM]int\n\tvar podSum int\n\tfor i := 0; i < podLen; i++ {\n\t\tif podReq[i].nodeName != -1 {\n\t\t\tpodSum++\n\t\t\tpodNum[podReq[i].nodeName]++\n\t\t}\n\t}\n\n\tvar podIdle [PHYNUM]float64\n\tvar resIdle [PHYNUM][DIMENSION]float64\n\tvar podVal float64\n\tvar resVal [DIMENSION]float64 // cal the sum and mean value\n\n\tfor i := 0; i < PHYNUM; i++ {\n\t\tpodIdle[i] = 1.0 - (float64)(podNum[i])/(float64)(podSum)\n\t\tpodVal = podVal + podIdle[i]\n\t\tfor j := 0; j < DIMENSION; j++ {\n\t\t\tresIdle[i][j] = (sch.reTotal[j] - podUsed[i][j]) / sch.reTotal[j]\n\t\t\tresVal[j] = resVal[j] + resIdle[i][j]\n\t\t}\n\t}\n\t// cal the balance value\n\tpodMean := podVal / (float64)(podSum)\n\tvar resMean [DIMENSION]float64\n\tfor j := 0; j < DIMENSION; j++ {\n\t\tresMean[j] = resVal[j] / (float64)(PHYNUM)\n\t}\n\tvar baIdle float64\n\tfor i := 0; i < PHYNUM; i++ {\n\t\tfor j := 0; j < DIMENSION; j++ {\n\t\t\tbaIdle = baIdle + math.Pow((resIdle[i][j]-resMean[j]), 2)\n\t\t}\n\t\tbaIdle = baIdle + math.Pow((podIdle[i]-podMean), 2)\n\t}\n\tbaIdle = math.Sqrt(baIdle)\n\tfmt.Printf(\"The balance value is %.3f \\n\", baIdle)\n}", "func (_Bindings *BindingsSession) BalanceOf(owner common.Address) (*big.Int, error) {\n\treturn _Bindings.Contract.BalanceOf(&_Bindings.CallOpts, owner)\n}", "func (c *rpcclient) smartBalance(ctx context.Context, ec *ethConn, assetID uint32, addr common.Address) (bal *big.Int, err error) {\n\ttip, err := c.blockNumber(ctx)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"blockNumber error: %v\", err)\n\t}\n\n\t// We need to subtract and pending outgoing value, but ignore any pending\n\t// incoming value since that can't be spent until mined. So we can't using\n\t// PendingBalanceAt or BalanceAt by themselves.\n\t// We'll iterate tx pool transactions and subtract any value and fees being\n\t// sent from this account. The rpc.Client doesn't expose the\n\t// txpool_contentFrom => (*TxPool).ContentFrom RPC method, for whatever\n\t// reason, so we'll have to use CallContext and copy the mimic the\n\t// internal RPCTransaction type.\n\tvar txs map[string]map[string]*RPCTransaction\n\tif err := ec.caller.CallContext(ctx, &txs, \"txpool_contentFrom\", addr); err != nil {\n\t\treturn nil, fmt.Errorf(\"contentFrom error: %w\", err)\n\t}\n\n\tif assetID == BipID {\n\t\tethBalance, err := ec.BalanceAt(ctx, addr, big.NewInt(int64(tip)))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\toutgoingEth := new(big.Int)\n\t\tfor _, group := range txs { // 2 groups, pending and queued\n\t\t\tfor _, tx := range group {\n\t\t\t\toutgoingEth.Add(outgoingEth, tx.Value.ToInt())\n\t\t\t\tgas := new(big.Int).SetUint64(uint64(tx.Gas))\n\t\t\t\tif tx.GasPrice != nil && tx.GasPrice.ToInt().Cmp(bigZero) > 0 {\n\t\t\t\t\toutgoingEth.Add(outgoingEth, new(big.Int).Mul(gas, tx.GasPrice.ToInt()))\n\t\t\t\t} else if tx.GasFeeCap != nil {\n\t\t\t\t\toutgoingEth.Add(outgoingEth, new(big.Int).Mul(gas, tx.GasFeeCap.ToInt()))\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, fmt.Errorf(\"cannot find fees for tx %s\", tx.Hash)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn ethBalance.Sub(ethBalance, outgoingEth), nil\n\t}\n\n\t// For tokens, we'll do something similar, but with checks for pending txs\n\t// that transfer tokens or pay to the swap contract.\n\t// Can't use withTokener because we need to use the same ethConn due to\n\t// txPoolSupported being used to decide between {smart/dumb}Balance.\n\ttkn := ec.tokens[assetID]\n\tif tkn == nil {\n\t\treturn nil, fmt.Errorf(\"no tokener for asset ID %d\", assetID)\n\t}\n\tbal, err = tkn.balanceOf(ctx, addr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, group := range txs {\n\t\tfor _, rpcTx := range group {\n\t\t\tto := *rpcTx.To\n\t\t\tif to == tkn.tokenAddr {\n\t\t\t\tif sent := tkn.transferred(rpcTx.Input); sent != nil {\n\t\t\t\t\tbal.Sub(bal, sent)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif to == tkn.contractAddr {\n\t\t\t\tif swapped := tkn.swapped(rpcTx.Input); swapped != nil {\n\t\t\t\t\tbal.Sub(bal, swapped)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn bal, nil\n}", "func (_ERC20Basic *ERC20BasicCaller) BalanceOf(opts *bind.CallOpts, _who common.Address) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _ERC20Basic.contract.Call(opts, out, \"balanceOf\", _who)\n\treturn *ret0, err\n}", "func (_Htlc *HtlcCaller) Balance(opts *bind.CallOpts) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _Htlc.contract.Call(opts, out, \"balance\")\n\treturn *ret0, err\n}", "func (_MainnetCryptoCardsContract *MainnetCryptoCardsContractCaller) BalanceOf(opts *bind.CallOpts, owner common.Address) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _MainnetCryptoCardsContract.contract.Call(opts, out, \"balanceOf\", owner)\n\treturn *ret0, err\n}", "func (_NectarToken *NectarTokenCaller) BalanceOf(opts *bind.CallOpts, _owner common.Address) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _NectarToken.contract.Call(opts, out, \"balanceOf\", _owner)\n\treturn *ret0, err\n}", "func (_TokenContractFunctions *TokenContractFunctionsCaller) BalanceOf(opts *bind.CallOpts, _owner common.Address) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _TokenContractFunctions.contract.Call(opts, out, \"balanceOf\", _owner)\n\treturn *ret0, err\n}", "func (m *Multicall) GetBalances(\n\tblock uint64,\n\tpool string,\n\ttokens []string,\n) ([]common.Address, []*big.Int, error) {\n\tpoolAddr := common.HexToAddress(pool)\n\tvar tokenAddrs = make([]common.Address, len(tokens))\n\tfor i, tok := range tokens {\n\t\ttokenAddrs[i] = common.HexToAddress(tok)\n\t}\n\treturn m.binding.GetBalances(\n\t\t&bind.CallOpts{\n\t\t\tContext: m.ctx,\n\t\t\tBlockNumber: new(big.Int).SetUint64(block),\n\t\t},\n\t\tpoolAddr,\n\t\ttokenAddrs,\n\t)\n}", "func (_Token *tokenCaller) BalanceOf(opts *bind.CallOpts, _owner common.Address) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _Token.contract.Call(opts, out, \"balanceOf\", _owner)\n\treturn *ret0, err\n}", "func (w *worker) staticHostAccountBalance() (types.Currency, error) {\n\t// Sanity check - only one account balance check should be running at a\n\t// time.\n\tif !atomic.CompareAndSwapUint64(&w.atomicAccountBalanceCheckRunning, 0, 1) {\n\t\tw.renter.log.Critical(\"account balance is being checked in two threads concurrently\")\n\t}\n\tdefer atomic.StoreUint64(&w.atomicAccountBalanceCheckRunning, 0)\n\n\t// Get a stream.\n\tstream, err := w.staticNewStream()\n\tif err != nil {\n\t\treturn types.ZeroCurrency, err\n\t}\n\tdefer func() {\n\t\tif err := stream.Close(); err != nil {\n\t\t\tw.renter.log.Println(\"ERROR: failed to close stream\", err)\n\t\t}\n\t}()\n\n\t// write the specifier\n\terr = modules.RPCWrite(stream, modules.RPCAccountBalance)\n\tif err != nil {\n\t\treturn types.ZeroCurrency, err\n\t}\n\n\t// send price table uid\n\tpt := w.staticPriceTable().staticPriceTable\n\terr = modules.RPCWrite(stream, pt.UID)\n\tif err != nil {\n\t\treturn types.ZeroCurrency, err\n\t}\n\n\t// provide payment\n\terr = w.renter.hostContractor.ProvidePayment(stream, w.staticHostPubKey, modules.RPCAccountBalance, pt.AccountBalanceCost, w.staticAccount.staticID, pt.HostBlockHeight)\n\tif err != nil {\n\t\t// If the error could be caused by a revision number mismatch,\n\t\t// signal it by setting the flag.\n\t\tif errCausedByRevisionMismatch(err) {\n\t\t\tw.staticSetSuspectRevisionMismatch()\n\t\t\tw.staticWake()\n\t\t}\n\t\treturn types.ZeroCurrency, err\n\t}\n\n\t// prepare the request.\n\tabr := modules.AccountBalanceRequest{Account: w.staticAccount.staticID}\n\terr = modules.RPCWrite(stream, abr)\n\tif err != nil {\n\t\treturn types.ZeroCurrency, err\n\t}\n\n\t// read the response\n\tvar resp modules.AccountBalanceResponse\n\terr = modules.RPCRead(stream, &resp)\n\tif err != nil {\n\t\treturn types.ZeroCurrency, err\n\t}\n\treturn resp.Balance, nil\n}", "func (_GameJam *GameJamCaller) Balance(opts *bind.CallOpts) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _GameJam.contract.Call(opts, out, \"balance\")\n\treturn *ret0, err\n}", "func (_Token *TokenCaller) BalanceOf(opts *bind.CallOpts, src common.Address) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _Token.contract.Call(opts, out, \"balanceOf\", src)\n\treturn *ret0, err\n}", "func GetCommonPool(w http.ResponseWriter, r *http.Request) {\n\n\t// Add header so that received knows they're receiving JSON\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\n\t// Retrieving name of node from query request\n\tnodeName := r.URL.Query().Get(\"name\")\n\tconfirmation, socket := checkNodeName(nodeName)\n\tif confirmation == false {\n\n\t\t// Stop code here no need to establish connection and reply\n\t\tjson.NewEncoder(w).Encode(responses.ErrorResponse{\n\t\t\tError: \"Node name requested doesn't exist\"})\n\t\treturn\n\t}\n\n\t// Retrieving height from query request\n\trecvHeight := r.URL.Query().Get(\"height\")\n\theight := checkHeight(recvHeight)\n\tif height == -1 {\n\n\t\t// Stop code here no need to establish connection and reply\n\t\tjson.NewEncoder(w).Encode(responses.ErrorResponse{\n\t\t\tError: \"Unexepcted value found, height needs to be string of int!\"})\n\t\treturn\n\t}\n\n\t// Attempt to load connection with staking client\n\tconnection, so := loadStakingClient(socket)\n\n\t// Close connection once code underneath executes\n\tdefer connection.Close()\n\n\t// If null object was retrieved send response\n\tif so == nil {\n\n\t\t// Stop code here faild to establish connection and reply\n\t\tjson.NewEncoder(w).Encode(responses.ErrorResponse{\n\t\t\tError: \"Failed to establish connection using socket : \" + socket})\n\t\treturn\n\t}\n\n\t// Return common pool at specific block height\n\tcommonPool, err := so.CommonPool(context.Background(), height)\n\tif err != nil {\n\t\tjson.NewEncoder(w).Encode(responses.ErrorResponse{\n\t\t\tError: \"Failed to get Common Pool!\"})\n\n\t\tlgr.Error.Println(\n\t\t\t\"Request at /api/staking/commonpool failed to retrieve common \"+\n\t\t\t\t\"pool : \", err)\n\t\treturn\n\t}\n\n\tlgr.Info.Println(\"Request at /api/staking/commonpool responding with \" +\n\t\t\"Common Pool!\")\n\tjson.NewEncoder(w).Encode(responses.QuantityResponse{Quantity: commonPool})\n}", "func ConsumableBalance(color ledgerstate.Color, consumables ...*ConsumableOutput) uint64 {\n\tret := uint64(0)\n\tfor _, out := range consumables {\n\t\tret += out.ConsumableBalance(color)\n\t}\n\treturn ret\n}", "func (_Lmc *LmcCaller) BalanceOf(opts *bind.CallOpts, _userAddress common.Address) (*big.Int, error) {\n\tvar out []interface{}\n\terr := _Lmc.contract.Call(opts, &out, \"balanceOf\", _userAddress)\n\n\tif err != nil {\n\t\treturn *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\n\treturn out0, err\n\n}", "func Balance() string {\n\tserver := servers[i]\n\ti++\n\n\t// reset the counter and start from the beginning\n\t// if we reached the end of servers\n\tif i >= len(servers) {\n\t\ti = 0\n\t}\n\treturn server\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
EscrowBalance returns the escrow balance for the ID.
func (s *ImmutableState) EscrowBalance(id signature.PublicKey) *quantity.Quantity { account := s.Account(id) return account.Escrow.Active.Balance.Clone() }
[ "func (o *MortgageLiability) GetEscrowBalance() float64 {\n\tif o == nil || o.EscrowBalance.Get() == nil {\n\t\tvar ret float64\n\t\treturn ret\n\t}\n\n\treturn *o.EscrowBalance.Get()\n}", "func (o *MortgageLiability) SetEscrowBalance(v float64) {\n\to.EscrowBalance.Set(&v)\n}", "func GetAccountBalanceById(w http.ResponseWriter, r *http.Request) {\n\tparams := mux.Vars(r)\n\n\taccountID, erro := strconv.ParseUint(params[\"accountID\"], 10, 64)\n\tif erro != nil {\n\t\tresponses.Error(w, http.StatusBadRequest, erro)\n\t\treturn\n\t}\n\n\tdb, erro := database.Connect()\n\tif erro != nil {\n\t\tresponses.Error(w, http.StatusInternalServerError, erro)\n\t\treturn\n\t}\n\tdefer db.Close()\n\n\trepository := repositories.NewAccountRepository(db)\n\taccount, erro := repository.FindBalanceById(accountID)\n\tif erro != nil {\n\t\tresponses.Error(w, http.StatusInternalServerError, erro)\n\t\treturn\n\t}\n\n\tresponses.JSON(w, http.StatusOK, account)\n}", "func (o *MortgageLiability) GetEscrowBalanceOk() (*float64, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn o.EscrowBalance.Get(), o.EscrowBalance.IsSet()\n}", "func (eth *EthClient) GetERC20Balance(address common.Address, contractAddress common.Address) (*big.Int, error) {\n\tresult := \"\"\n\tnumLinkBigInt := new(big.Int)\n\tfunctionSelector := models.HexToFunctionSelector(\"0x70a08231\") // balanceOf(address)\n\tdata, err := utils.ConcatBytes(functionSelector.Bytes(), common.LeftPadBytes(address.Bytes(), utils.EVMWordByteLen))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\targs := callArgs{\n\t\tTo: contractAddress,\n\t\tData: data,\n\t}\n\terr = eth.Call(&result, \"eth_call\", args, \"latest\")\n\tif err != nil {\n\t\treturn numLinkBigInt, err\n\t}\n\tnumLinkBigInt.SetString(result, 0)\n\treturn numLinkBigInt, nil\n}", "func (w *Wallet) GetBalance(address string, id schema.ID, asset Asset) (xdr.Int64, []string, error) {\n\tif address == \"\" {\n\t\terr := fmt.Errorf(\"trying to get the balance of an empty address. this should never happen\")\n\t\tlog.Warn().Err(err).Send()\n\t\treturn 0, nil, err\n\t}\n\n\tvar total xdr.Int64\n\thorizonClient, err := w.GetHorizonClient()\n\tif err != nil {\n\t\treturn 0, nil, err\n\t}\n\n\tcursor := \"\"\n\n\ttxReq := horizonclient.TransactionRequest{\n\t\tForAccount: address,\n\t\tCursor: cursor,\n\t\tLimit: stellarPageLimit,\n\t}\n\n\tlog.Info().Str(\"address\", address).Msg(\"fetching balance for address\")\n\ttxes, err := horizonClient.Transactions(txReq)\n\tif err != nil {\n\t\treturn 0, nil, errors.Wrap(err, \"could not get transactions\")\n\t}\n\n\tdonors := make(map[string]struct{})\n\tfor len(txes.Embedded.Records) != 0 {\n\t\tfor _, tx := range txes.Embedded.Records {\n\t\t\tif tx.Memo == strconv.FormatInt(int64(id), 10) {\n\t\t\t\teffectsReq := horizonclient.EffectRequest{\n\t\t\t\t\tForTransaction: tx.Hash,\n\t\t\t\t}\n\t\t\t\teffects, err := horizonClient.Effects(effectsReq)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error().Err(err).Msgf(\"failed to get transaction effects\")\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\t// first check if we have been paid\n\t\t\t\tvar isFunding bool\n\t\t\t\tfor _, effect := range effects.Embedded.Records {\n\t\t\t\t\tif effect.GetAccount() != address {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tif effect.GetType() == \"account_credited\" {\n\t\t\t\t\t\tcreditedEffect := effect.(horizoneffects.AccountCredited)\n\t\t\t\t\t\tif creditedEffect.Asset.Code != asset.Code() ||\n\t\t\t\t\t\t\tcreditedEffect.Asset.Issuer != asset.Issuer() {\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\t\t\t\t\t\tparsedAmount, err := amount.Parse(creditedEffect.Amount)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\t\t\t\t\t\tisFunding = true\n\t\t\t\t\t\ttotal += parsedAmount\n\t\t\t\t\t} else if effect.GetType() == \"account_debited\" {\n\t\t\t\t\t\tdebitedEffect := effect.(horizoneffects.AccountDebited)\n\t\t\t\t\t\tif debitedEffect.Asset.Code != asset.Code() ||\n\t\t\t\t\t\t\tdebitedEffect.Asset.Issuer != asset.Issuer() {\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\t\t\t\t\t\tparsedAmount, err := amount.Parse(debitedEffect.Amount)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\t\t\t\t\t\tisFunding = false\n\t\t\t\t\t\ttotal -= parsedAmount\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif isFunding {\n\t\t\t\t\t// we don't need to verify the asset here anymore, since this\n\t\t\t\t\t// flag is only toggled on after that check passed in the loop\n\t\t\t\t\t// above\n\t\t\t\t\tfor _, effect := range effects.Embedded.Records {\n\t\t\t\t\t\tif effect.GetType() == \"account_debited\" && effect.GetAccount() != address {\n\t\t\t\t\t\t\tdonors[effect.GetAccount()] = struct{}{}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tcursor = tx.PagingToken()\n\t\t}\n\n\t\t// if the amount of records fetched is smaller than the page limit\n\t\t// we can assume we are on the last page and we break to prevent another\n\t\t// call to horizon\n\t\tif len(txes.Embedded.Records) < stellarPageLimit {\n\t\t\tbreak\n\t\t}\n\n\t\ttxReq.Cursor = cursor\n\t\tlog.Info().Str(\"address\", address).Msgf(\"fetching balance for address with cursor: %s\", cursor)\n\t\ttxes, err = horizonClient.Transactions(txReq)\n\t\tif err != nil {\n\t\t\treturn 0, nil, errors.Wrap(err, \"could not get transactions\")\n\t\t}\n\t}\n\n\tdonorList := []string{}\n\tfor donor := range donors {\n\t\tdonorList = append(donorList, donor)\n\t}\n\tlog.Info().\n\t\tInt64(\"balance\", int64(total)).\n\t\tStr(\"address\", address).\n\t\tInt64(\"id\", int64(id)).Msgf(\"status of balance for reservation\")\n\treturn total, donorList, nil\n}", "func (c *Client) RetrieveBalance(\n\tctx context.Context,\n\tid string,\n) (*BalanceResource, error) {\n\towner, _, err := NormalizedOwnerAndTokenFromID(ctx, id)\n\tif err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\t_, host, err := UsernameAndMintHostFromAddress(ctx, owner)\n\tif err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\n\treq, err := http.NewRequest(\"GET\",\n\t\tFullMintURL(ctx,\n\t\t\thost, fmt.Sprintf(\"/balances/%s\", id), url.Values{}).String(), nil)\n\tif err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\treq.Header.Add(\"Mint-Protocol-Version\", ProtocolVersion)\n\tr, err := c.httpClient.Do(req)\n\tif err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\tdefer r.Body.Close()\n\n\tvar raw svc.Resp\n\tif err := json.NewDecoder(r.Body).Decode(&raw); err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\n\tif r.StatusCode != http.StatusOK && r.StatusCode != http.StatusCreated {\n\t\tvar e errors.ConcreteUserError\n\t\terr = raw.Extract(\"error\", &e)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Trace(err)\n\t\t}\n\t\treturn nil, errors.Trace(ErrMintClient{\n\t\t\tr.StatusCode, e.ErrCode, e.ErrMessage,\n\t\t})\n\t}\n\n\tvar balance BalanceResource\n\tif err := raw.Extract(\"balance\", &balance); err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\n\treturn &balance, nil\n}", "func (_Withdrawable *WithdrawableSession) GetDepositedBalance(arg0 common.Address, arg1 common.Address) (*big.Int, error) {\n\treturn _Withdrawable.Contract.GetDepositedBalance(&_Withdrawable.CallOpts, arg0, arg1)\n}", "func Erc20Balance(userAddress string, contractAddress string) (*BalanceAmount, error) {\n\tvar resp string\n\tparams := map[string]string{\"to\": contractAddress, \"data\": utils.PaddingData(ERC20MethodBalanceOf, userAddress)}\n\terr := endpointsManager.RPC(&resp, \"eth_call\", params, \"latest\")\n\tif err != nil {\n\t\tcommon.Logger.Debug(err)\n\t\treturn &BalanceAmount{}, err\n\t}\n\n\tamount := \"0\"\n\tt, ok := utils.ParseBig256(resp)\n\tif ok && t != nil {\n\t\tamount = t.Text(10)\n\t}\n\treturn &BalanceAmount{Amount: amount}, err\n}", "func (_DevUtils *DevUtilsTransactorSession) GetBalance(ownerAddress common.Address, assetData []byte) (*types.Transaction, error) {\n\treturn _DevUtils.Contract.GetBalance(&_DevUtils.TransactOpts, ownerAddress, assetData)\n}", "func (_DevUtils *DevUtilsSession) GetBalance(ownerAddress common.Address, assetData []byte) (*types.Transaction, error) {\n\treturn _DevUtils.Contract.GetBalance(&_DevUtils.TransactOpts, ownerAddress, assetData)\n}", "func (c Client) Get(params *stripe.BalanceParams) (*stripe.Balance, error) {\n\tbalance := &stripe.Balance{}\n\terr := c.B.Call(http.MethodGet, \"/v1/balance\", c.Key, params, balance)\n\treturn balance, err\n}", "func (_Sigmacore *SigmacoreSession) GetBalance(token common.Address) (*big.Int, error) {\n\treturn _Sigmacore.Contract.GetBalance(&_Sigmacore.CallOpts, token)\n}", "func (m *Monzgo) Balance(accountID string) (*Balance, error) {\n\trequestParams := make(map[string]string)\n\trequestParams[\"account_id\"] = accountID\n\n\tbalance := &Balance{}\n\tif err := m.request(\"GET\", \"balance\", balance, requestParams); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn balance, nil\n}", "func (a Account) GetBalance() int {\n\treturn a.balance\n}", "func (r BankRepo) DecrementAccountBalance(accountID int, sum float64) (err error) {\n\treturn r.IncrementAccountBalance(accountID, -sum)\n}", "func (_Registry *RegistryCaller) BalanceOf(opts *bind.CallOpts, account common.Address, id *big.Int) (*big.Int, error) {\n\tvar out []interface{}\n\terr := _Registry.contract.Call(opts, &out, \"balanceOf\", account, id)\n\n\tif err != nil {\n\t\treturn *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\n\treturn out0, err\n\n}", "func GetBalance(ctx context.Context, table api.BatchBalanceTable, addr types.AccAddress, key uint64) (*api.BatchBalance, error) {\n\tbal, err := table.Get(ctx, addr, key)\n\tif err != nil {\n\t\tif !ormerrors.IsNotFound(err) {\n\t\t\treturn nil, err\n\t\t}\n\t\tbal = &api.BatchBalance{\n\t\t\tBatchKey: key,\n\t\t\tAddress: addr,\n\t\t\tTradableAmount: \"0\",\n\t\t\tRetiredAmount: \"0\",\n\t\t\tEscrowedAmount: \"0\",\n\t\t}\n\t}\n\treturn bal, nil\n}", "func (s *Client) GetBalance(ctx context.Context, scripthash string) (GetBalanceResult, error) {\n\tvar resp GetBalanceResp\n\n\terr := s.request(ctx, \"blockchain.scripthash.get_balance\", []interface{}{scripthash}, &resp)\n\tif err != nil {\n\t\treturn GetBalanceResult{}, err\n\t}\n\n\treturn resp.Result, err\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SlashEscrow slashes the escrow balance and the escrowbutundergoingdebonding balance of the account, transferring it to the global common pool, returning true iff the amount actually slashed is > 0. WARNING: This is an internal routine to be used to implement staking policy, and MUST NOT be exposed outside of backend implementations.
func (s *MutableState) SlashEscrow(ctx *abci.Context, fromID signature.PublicKey, amount *quantity.Quantity) (bool, error) { commonPool, err := s.CommonPool() if err != nil { return false, fmt.Errorf("staking: failed to query common pool for slash: %w", err) } from := s.Account(fromID) // Compute the amount we need to slash each pool. The amount is split // between the pools based on relative total balance. total := from.Escrow.Active.Balance.Clone() if err = total.Add(&from.Escrow.Debonding.Balance); err != nil { return false, fmt.Errorf("staking: compute total balance: %w", err) } var slashed quantity.Quantity if err = slashPool(&slashed, &from.Escrow.Active, amount, total); err != nil { return false, errors.Wrap(err, "slashing active escrow") } if err = slashPool(&slashed, &from.Escrow.Debonding, amount, total); err != nil { return false, errors.Wrap(err, "slashing debonding escrow") } if slashed.IsZero() { return false, nil } totalSlashed := slashed.Clone() if err = quantity.Move(commonPool, &slashed, totalSlashed); err != nil { return false, errors.Wrap(err, "moving tokens to common pool") } s.SetCommonPool(commonPool) s.SetAccount(fromID, from) if !ctx.IsCheckOnly() { ev := cbor.Marshal(&staking.TakeEscrowEvent{ Owner: fromID, Tokens: *totalSlashed, }) ctx.EmitEvent(api.NewEventBuilder(AppName).Attribute(KeyTakeEscrow, ev)) } return true, nil }
[ "func buildEscrowContract(pkPayer, pkRedeemer []byte, locktime int64) ([]byte, error) {\n\tb := txscript.NewScriptBuilder()\n\n\tb.AddOp(txscript.OP_IF) // Normal redeem path\n\t{\n\t\t// Check 2-of-2 multisig.\n\t\tb.AddOp(txscript.OP_2)\n\t\tb.AddData(pkPayer)\n\t\tb.AddData(pkRedeemer)\n\t\tb.AddOp(txscript.OP_2)\n\t\tb.AddOp(txscript.OP_CHECKMULTISIG)\n\t}\n\tb.AddOp(txscript.OP_ELSE) // Refund path\n\t{\n\t\t// Verify locktime and drop it off the stack (which is not done\n\t\t// by CLTV).\n\t\tb.AddInt64(locktime)\n\t\tb.AddOp(txscript.OP_CHECKLOCKTIMEVERIFY)\n\t\tb.AddOp(txscript.OP_DROP)\n\n\t\t// Verify our signature is being used to redeem the output.\n\t\tb.AddData(pkPayer)\n\t\tb.AddOp(txscript.OP_CHECKSIG)\n\t}\n\tb.AddOp(txscript.OP_ENDIF)\n\n\treturn b.Script()\n}", "func EnoughBalance(color ledgerstate.Color, amount uint64, consumables ...*ConsumableOutput) bool {\n\tconsumable := ConsumableBalance(color, consumables...)\n\treturn consumable >= amount\n}", "func (h *Hand) IsBust() bool {\n\treturn util.MinInt(h.Scores()) > 21\n}", "func (d *HostLowerDeposit) Disrupt(s string) bool {\n\treturn s == \"lowerDeposit\"\n}", "func (o *MortgageLiability) GetEscrowBalanceOk() (*float64, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn o.EscrowBalance.Get(), o.EscrowBalance.IsSet()\n}", "func HalfCarryAdd(val1 byte, val2 byte) bool {\n\treturn (val1&0xF)+(val2&0xF) > 0xF\n}", "func canFlowerSoftlock(g graph.Graph) error {\n\t// first check if cucco has been reached\n\tcucco := g[\"spring banana cucco\"]\n\tif cucco.Mark != graph.MarkTrue {\n\t\treturn nil\n\t}\n\n\t// temporarily make entrance and bush items unavailable\n\tdisabledNodes := append(g[\"remove flower sustainable\"].Parents)\n\tdisabledParents := make([][]*graph.Node, len(disabledNodes))\n\tfor i, node := range disabledNodes {\n\t\tdisabledParents[i] = node.Parents\n\t\tnode.ClearParents()\n\t}\n\tdefer func() {\n\t\tfor i, node := range disabledNodes {\n\t\t\tnode.AddParents(disabledParents[i]...)\n\t\t}\n\t}()\n\n\t// see if you can still reach the exit\n\tg.ClearMarks()\n\tif cucco.GetMark(cucco, nil) == graph.MarkTrue {\n\t\treturn errors.New(\"cucco softlock\")\n\t}\n\treturn nil\n}", "func (act *Action_4_DiamondCreate) IsBurning90PersentTxFees() bool {\n\tif uint32(act.Number) > DiamondCreateBurning90PercentTxFeesAboveNumber {\n\t\t// 90% of the cost of destroying this transaction from the 30001 diamond\n\t\treturn true\n\t}\n\treturn false\n}", "func (o *MortgageLiability) SetEscrowBalance(v float64) {\n\to.EscrowBalance.Set(&v)\n}", "func (_DepositContract *DepositContractCaller) IsSlashed(opts *bind.CallOpts, account common.Address) (bool, error) {\n\tvar out []interface{}\n\terr := _DepositContract.contract.Call(opts, &out, \"isSlashed\", account)\n\n\tif err != nil {\n\t\treturn *new(bool), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(bool)).(*bool)\n\n\treturn out0, err\n\n}", "func (e *Ethereum) Ping() bool {\n\n\t_, err := e.GetBalance(BURN_ADDRESS)\n\tif err != nil {\n\t\tlog.Debug(err)\n\t\treturn false\n\t}\n\treturn true\n}", "func prepareSignedGuardContractForShardOffSign(ctx context.Context, ss *storage.FileContracts, shard *storage.Shard, n *core.IpfsNode, retryCalling bool) error {\n\t// \"/storage/upload/getcontractbatch\" and \"/storage/upload/signedbatch\" handlers perform responses\n\t// to SDK application's requests and sets each `shard.HalfSignedEscrowContract` with signed bytes.\n\t// The corresponding endpoint for `signedbatch` closes \"ss.OfflineSignChan\" to broadcast\n\t// Here we wait for the broadcast signal.\n\tselect {\n\tcase <-ss.OfflineCB.OfflineSignEscrowChan:\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\t}\n\tvar err error\n\tshard.HalfSignedEscrowContract, err = escrow.SignContractAndMarshalOffSign(shard.UnsignedEscrowContract, shard.SignedBytes, nil, true)\n\tif err != nil {\n\t\tlog.Error(\"sign escrow contract and maorshal failed \")\n\t\treturn err\n\t}\n\n\t// Output for this function is set here\n\t//shard.HalfSignedEscrowContract = halfSignedEscrowContract\n\tisLast, err := ss.IncrementAndCompareOffSignReadyShards(len(ss.ShardInfo))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif isLast {\n\t\tss.SetOffSignReadyShards(0)\n\t\tcurrentStatus := ss.GetCurrentStatus()\n\t\tif currentStatus != storage.InitSignProcessForEscrowStatus {\n\t\t\treturn fmt.Errorf(\"current status %d does not match expected InitSignProcessForEscrowStatus\", currentStatus)\n\t\t}\n\t\tss.UpdateSessionStatus(currentStatus, true, nil) // call this instead of SendSessionStatusChan() as it is before \"initStatus\"\n\t}\n\treturn nil\n}", "func (a *FakeAccount) Check() bool {\n\tb, err := stellarnet.ParseStellarAmount(a.balance.Amount)\n\trequire.NoError(a.T, err)\n\tminimumReserve, err := stellarnet.ParseStellarAmount(\"1.0\")\n\trequire.NoError(a.T, err)\n\tswitch {\n\tcase b == 0:\n\t\treturn false\n\tcase b < 0:\n\t\trequire.Fail(a.T, \"account has negative balance\", \"%v\", a.accountID)\n\tcase b < minimumReserve:\n\t\trequire.Fail(a.T, \"account has less than the minimum balance\", \"%v < %v %v\",\n\t\t\tstellarnet.StringFromStellarAmount(b), stellarnet.StringFromStellarAmount(minimumReserve), a.accountID)\n\tdefault:\n\t\treturn true\n\t}\n\n\treturn b != 0\n}", "func isBalanceSufficient(payer common.Address, cache *storage.CacheDB, config *smartcontract.Config, store store.LedgerStore, gas uint64) (uint64, error) {\n\tbalance, err := getBalanceFromNative(config, cache, store, payer)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tif balance < gas {\n\t\treturn 0, fmt.Errorf(\"payer gas insufficient, need %d , only have %d\", gas, balance)\n\t}\n\treturn balance, nil\n}", "func (con *Contract) BuildRefundTx() error {\n\tvar err error\n\n\t// XXX: temporary compat with the old code\n\tif con.EscrowTx == nil {\n\t\tvar tx wire.MsgTx\n\t\terr = tx.Deserialize(bytes.NewReader(con.EscrowBytes))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to deserialize escrow tx: %v\", err)\n\t\t}\n\t\tcon.EscrowTx = &tx\n\t}\n\n\tcontractOutPoint := wire.OutPoint{\n\t\tHash: con.EscrowTx.TxHash(),\n\t\tIndex: ^uint32(0),\n\t}\n\tfor i, o := range con.EscrowTx.TxOut {\n\t\tif bytes.Equal(o.PkScript, con.EscrowPayScript) {\n\t\t\tcontractOutPoint.Index = uint32(i)\n\t\t\tbreak\n\t\t}\n\t}\n\tif contractOutPoint.Index == ^uint32(0) {\n\t\treturn errors.New(\"contract tx does not contain a P2SH contract payment\")\n\t}\n\n\trefundOutScript, err := txscript.PayToAddrScript(con.RefundAddr)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttx := wire.NewMsgTx()\n\ttx.LockTime = uint32(con.LockTime)\n\ttx.AddTxOut(wire.NewTxOut(0, refundOutScript)) // amount set below\n\trefundSize := estimateRefundSerializeSize(con.EscrowScript,\n\t\ttx.TxOut)\n\trefundFee := txrules.FeeForSerializeSize(feePerKb, refundSize)\n\ttx.TxOut[0].Value = con.EscrowTx.TxOut[contractOutPoint.Index].Value -\n\t\tint64(refundFee)\n\tif txrules.IsDustOutput(tx.TxOut[0], feePerKb) {\n\t\treturn fmt.Errorf(\"refund output value of %v is dust\",\n\t\t\tdcrutil.Amount(tx.TxOut[0].Value))\n\t}\n\n\ttxIn := wire.NewTxIn(&contractOutPoint, nil)\n\ttxIn.Sequence = 0\n\ttx.AddTxIn(txIn)\n\n\tvar buf bytes.Buffer\n\tbuf.Grow(tx.SerializeSize())\n\ttx.Serialize(&buf)\n\n\tcon.RefundTx = tx\n\tcon.RefundBytes = buf.Bytes()\n\n\treturn nil\n}", "func TestSlashAtNegativeHeight(t *testing.T) {\n\tapp, ctx, _, _ := bootstrapSlashTest(t, 10)\n\tconsAddr := sdk.ConsAddress(PKs[0].Address())\n\tfraction := sdk.NewDecWithPrec(5, 1)\n\n\tbondedPool := app.StakingKeeper.GetBondedPool(ctx)\n\toldBondedPoolBalances := app.BankKeeper.GetAllBalances(ctx, bondedPool.GetAddress())\n\n\tvalidator, found := app.StakingKeeper.GetValidatorByConsAddr(ctx, consAddr)\n\trequire.True(t, found)\n\tapp.StakingKeeper.Slash(ctx, consAddr, -2, 10, fraction)\n\n\t// read updated state\n\tvalidator, found = app.StakingKeeper.GetValidatorByConsAddr(ctx, consAddr)\n\trequire.True(t, found)\n\n\t// end block\n\tapplyValidatorSetUpdates(t, ctx, app.StakingKeeper, 1)\n\n\tvalidator, found = app.StakingKeeper.GetValidator(ctx, validator.GetOperator())\n\trequire.True(t, found)\n\t// power decreased\n\trequire.Equal(t, int64(5), validator.GetConsensusPower(app.StakingKeeper.PowerReduction(ctx)))\n\n\t// pool bonded shares decreased\n\tnewBondedPoolBalances := app.BankKeeper.GetAllBalances(ctx, bondedPool.GetAddress())\n\tdiffTokens := oldBondedPoolBalances.Sub(newBondedPoolBalances).AmountOf(app.StakingKeeper.BondDenom(ctx))\n\trequire.Equal(t, app.StakingKeeper.TokensFromConsensusPower(ctx, 5).String(), diffTokens.String())\n}", "func (act *Action_5_DiamondTransfer) IsBurning90PersentTxFees() bool {\n\treturn false\n}", "func (_DepositContract *DepositContractCallerSession) IsSlashed(account common.Address) (bool, error) {\n\treturn _DepositContract.Contract.IsSlashed(&_DepositContract.CallOpts, account)\n}", "func (s *Ethereum) shouldPreserve(block *types.Block) bool {\n\t// The reason we need to disable the self-reorg preserving for clique\n\t// is it can be probable to introduce a deadlock.\n\t//\n\t// e.g. If there are 7 available signers\n\t//\n\t// r1 A\n\t// r2 B\n\t// r3 C\n\t// r4 D\n\t// r5 A [X] F G\n\t// r6 [X]\n\t//\n\t// In the round5, the inturn signer E is offline, so the worst case\n\t// is A, F and G sign the block of round5 and reject the block of opponents\n\t// and in the round6, the last available signer B is offline, the whole\n\t// network is stuck.\n\tif _, ok := s.engine.(*clique.Clique); ok {\n\t\treturn false\n\t}\n\treturn s.isLocalBlock(block)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
TransferFromCommon transfers up to the amount from the global common pool to the general balance of the account, returning true iff the amount transferred is > 0. WARNING: This is an internal routine to be used to implement incentivization policy, and MUST NOT be exposed outside of backend implementations.
func (s *MutableState) TransferFromCommon(ctx *abci.Context, toID signature.PublicKey, amount *quantity.Quantity) (bool, error) { commonPool, err := s.CommonPool() if err != nil { return false, errors.Wrap(err, "staking: failed to query common pool for transfer") } to := s.Account(toID) transfered, err := quantity.MoveUpTo(&to.General.Balance, commonPool, amount) if err != nil { return false, errors.Wrap(err, "staking: failed to transfer from common pool") } ret := !transfered.IsZero() if ret { s.SetCommonPool(commonPool) s.SetAccount(toID, to) if !ctx.IsCheckOnly() { ev := cbor.Marshal(&staking.TransferEvent{ // XXX: Reserve an id for the common pool? To: toID, Tokens: *transfered, }) ctx.EmitEvent(api.NewEventBuilder(AppName).Attribute(KeyTransfer, ev)) } } return ret, nil }
[ "func CanTransfer(db StateDB, addr common.Address, amount *big.Int) bool {\n\treturn db.GetBalance(addr).Cmp(amount) >= 0\n}", "func CanTransfer(db vm.StateDB, addr types.AddressHash, amount *big.Int) bool {\n\treturn db.GetBalance(addr).Cmp(amount) >= 0\n}", "func (s *StorageInMemory) Transfer(accountFrom, accountTo storage.AccountID, amountToTransfer storage.AccountBalance) error {\n\tbalanceFrom, err := s.getBalance(accountFrom)\n\tif err != nil {\n\t\treturn err\n\t}\n\tbalanceTo, err := s.getBalance(accountTo)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbalanceFrom.mu.Lock()\n\tbalanceTo.mu.Lock()\n\tdefer balanceFrom.mu.Unlock()\n\tdefer balanceTo.mu.Unlock()\n\n\tif balanceFrom.amount < amountToTransfer {\n\t\treturn ErrNotEnoughBalance\n\t}\n\t// todo del (для отладки)\n\t// fmt.Println(\"операция: \", balanceFrom.amount, balanceTo.amount, balanceFrom.amount+balanceTo.amount)\n\tbalanceFrom.amount -= amountToTransfer\n\tbalanceTo.amount += amountToTransfer\n\treturn nil\n}", "func (_Sigmacore *SigmacoreTransactor) Transfer(opts *bind.TransactOpts, dst common.Address, amt *big.Int) (*types.Transaction, error) {\n\treturn _Sigmacore.contract.Transact(opts, \"transfer\", dst, amt)\n}", "func (unitImpl *UnitImpl) Transfer(unit stardash.Unit, amount int64, material string) bool {\n\treturn unitImpl.RunOnServer(\"transfer\", map[string]interface{}{\n\t\t\"unit\": unit,\n\t\t\"amount\": amount,\n\t\t\"material\": material,\n\t}).(bool)\n}", "func (_Sigmacore *SigmacoreTransactorSession) Transfer(dst common.Address, amt *big.Int) (*types.Transaction, error) {\n\treturn _Sigmacore.Contract.Transfer(&_Sigmacore.TransactOpts, dst, amt)\n}", "func (reqctx *requestContext) transferAllowedFunds(target isc.AgentID, transfer ...*isc.Assets) *isc.Assets {\n\tif reqctx.vm.isCoreAccount(target) {\n\t\t// if the target is one of core contracts, assume target is the common account\n\t\ttarget = accounts.CommonAccount()\n\t}\n\n\tvar toMove *isc.Assets\n\tif len(transfer) == 0 {\n\t\ttoMove = reqctx.allowanceAvailable()\n\t} else {\n\t\ttoMove = transfer[0]\n\t}\n\n\treqctx.spendAllowedBudget(toMove) // panics if not enough\n\n\tcaller := reqctx.Caller() // have to take it here because callCore changes that\n\n\t// if the caller is a core contract, funds should be taken from the common account\n\tif reqctx.vm.isCoreAccount(caller) {\n\t\tcaller = accounts.CommonAccount()\n\t}\n\treqctx.callCore(accounts.Contract, func(s kv.KVStore) {\n\t\tif err := accounts.MoveBetweenAccounts(s, caller, target, toMove); err != nil {\n\t\t\tpanic(vm.ErrNotEnoughFundsForAllowance)\n\t\t}\n\t})\n\treturn reqctx.allowanceAvailable()\n}", "func (_SmartToken *SmartTokenTransactor) Transfer(opts *bind.TransactOpts, to common.Address, tokens *big.Int) (*types.Transaction, error) {\n\treturn _SmartToken.contract.Transact(opts, \"transfer\", to, tokens)\n}", "func (_TokenContractFunctions *TokenContractFunctionsTransactor) Transfer(opts *bind.TransactOpts, _to common.Address, _value *big.Int) (*types.Transaction, error) {\n\treturn _TokenContractFunctions.contract.Transact(opts, \"transfer\", _to, _value)\n}", "func (_ProjectWallet *ProjectWalletSession) Transfer(_receiver common.Address, _amt *big.Int) (*types.Transaction, error) {\n\treturn _ProjectWallet.Contract.Transfer(&_ProjectWallet.TransactOpts, _receiver, _amt)\n}", "func (c *CheckPointExecutor) Transfer(ctx context.Context, src, dst string, download bool, limit int, compress bool) (err error) {\n\tpoint := checkpoint.Acquire(ctx, scpPoint, map[string]any{\n\t\t\"host\": c.config.Host,\n\t\t\"port\": c.config.Port,\n\t\t\"user\": c.config.User,\n\t\t\"src\": src,\n\t\t\"dst\": dst,\n\t\t\"download\": download,\n\t\t\"limit\": limit,\n\t\t\"compress\": compress,\n\t})\n\tdefer func() {\n\t\tpoint.Release(err,\n\t\t\tzap.String(\"host\", c.config.Host),\n\t\t\tzap.Int(\"port\", c.config.Port),\n\t\t\tzap.String(\"user\", c.config.User),\n\t\t\tzap.String(\"src\", src),\n\t\t\tzap.String(\"dst\", dst),\n\t\t\tzap.Bool(\"download\", download))\n\t}()\n\tif point.Hit() != nil {\n\t\treturn nil\n\t}\n\n\treturn c.Executor.Transfer(ctx, src, dst, download, limit, compress)\n}", "func Transfer(db StateDB, sender, recipient common.Address, amount *big.Int) {\n\tdb.SubBalance(sender, amount)\n\tdb.AddBalance(recipient, amount)\n}", "func (_ProjectWallet *ProjectWalletTransactorSession) Transfer(_receiver common.Address, _amt *big.Int) (*types.Transaction, error) {\n\treturn _ProjectWallet.Contract.Transfer(&_ProjectWallet.TransactOpts, _receiver, _amt)\n}", "func TransferPay(openid, tradeno, ip string, amount int64) bool {\n\tvar req TransferRequest\n\treq.MchAppid = InquiryAppid\n\treq.Mchid = InquiryMerID\n\treq.NonceStr = util.GenSalt()\n\treq.PartnerTradeNo = tradeno\n\treq.Openid = openid\n\treq.CheckName = \"NO_CHECK\"\n\treq.Amount = amount\n\treq.Desc = \"提现\"\n\treq.SpbillCreateIP = ip\n\treq.Sign = calcTransferSign(req, InquiryMerKey)\n\n\treturn transfer(req)\n}", "func (_Supercoin *SupercoinTransactor) Transfer(opts *bind.TransactOpts, _to common.Address, _value *big.Int) (*types.Transaction, error) {\n\treturn _Supercoin.contract.Transact(opts, \"transfer\", _to, _value)\n}", "func (p *Poloniex) TransferBalance(ctx context.Context, currency, from, to string, amount float64) (bool, error) {\n\tvalues := url.Values{}\n\tresult := GenericResponse{}\n\n\tvalues.Set(\"currency\", currency)\n\tvalues.Set(\"amount\", strconv.FormatFloat(amount, 'f', -1, 64))\n\tvalues.Set(\"fromAccount\", from)\n\tvalues.Set(\"toAccount\", to)\n\n\terr := p.SendAuthenticatedHTTPRequest(ctx, exchange.RestSpot, http.MethodPost, poloniexTransferBalance, values, &result)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tif result.Error != \"\" && result.Success != 1 {\n\t\treturn false, errors.New(result.Error)\n\t}\n\n\treturn true, nil\n}", "func (c *ContaCorrente) Transferir(valorDaTransferencia float64, contaDestino *ContaCorrente) bool {\n\tif valorDaTransferencia < c.Saldo && valorDaTransferencia > 0 {\n\t\tc.Saldo -= valorDaTransferencia\n\t\tcontaDestino.Depositar(valorDaTransferencia)\n\t\treturn true\n\t} else {\n\t\treturn false\n\t}\n}", "func Transfer(fromAcct Account, toAcct Account, amount Money) error {\n\tif err := fromAcct.Withdraw(amount); err == nil {\n\t\tif depErr := toAcct.Deposit(amount); depErr == nil {\n\t\t\tfmt.Printf(\"Transfered %f from %s to %+v\", amount, fromAcct, toAcct)\n\t\t} else {\n\t\t\t//return the root cause\n\t\t\treturn depErr\n\t\t}\n\t} else {\n\t\t//return the root cause\n\t\treturn err\n\t}\n\treturn nil\n}", "func (bank *Bank) continuousMoneyTransfer() {\n\tfor {\n\t\tfrom := makeAccountID(rand.Intn(bank.numAccounts))\n\t\tto := makeAccountID(rand.Intn(bank.numAccounts))\n\t\t// Continue when from == to\n\t\tif bytes.Equal(from, to) {\n\t\t\tcontinue\n\t\t}\n\t\texchangeAmount := rand.Int63n(100)\n\t\t// transferMoney transfers exchangeAmount between the two accounts\n\t\ttransferMoney := func(runner client.Runner) error {\n\t\t\tbatchRead := &client.Batch{}\n\t\t\tbatchRead.Get(from, to)\n\t\t\tif err := runner.Run(batchRead); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif batchRead.Results[0].Err != nil {\n\t\t\t\treturn batchRead.Results[0].Err\n\t\t\t}\n\t\t\t// Read from value.\n\t\t\tfromAccount := &Account{}\n\t\t\terr := fromAccount.decode(batchRead.Results[0].Rows[0].ValueBytes())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t// Ensure there is enough cash.\n\t\t\tif fromAccount.Balance < exchangeAmount {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\t// Read to value.\n\t\t\ttoAccount := &Account{}\n\t\t\terrRead := toAccount.decode(batchRead.Results[0].Rows[1].ValueBytes())\n\t\t\tif errRead != nil {\n\t\t\t\treturn errRead\n\t\t\t}\n\t\t\t// Update both accounts.\n\t\t\tbatchWrite := &client.Batch{}\n\t\t\tfromAccount.Balance -= exchangeAmount\n\t\t\ttoAccount.Balance += exchangeAmount\n\t\t\tif fromValue, err := fromAccount.encode(); err != nil {\n\t\t\t\treturn err\n\t\t\t} else if toValue, err := toAccount.encode(); err != nil {\n\t\t\t\treturn err\n\t\t\t} else {\n\t\t\t\tbatchWrite.Put(fromValue, toValue)\n\t\t\t}\n\t\t\treturn runner.Run(batchWrite)\n\t\t}\n\t\tif *useTransaction {\n\t\t\tif err := bank.db.Tx(func(tx *client.Tx) error { return transferMoney(tx) }); err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\t\t} else if err := transferMoney(bank.db); err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tatomic.AddInt32(&bank.numTransfers, 1)\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
AddRewards computes and transfers a staking reward to active escrow accounts. If an error occurs, the pool and affected accounts are left in an invalid state. This may fail due to the common pool running out of tokens. In this case, the returned error's cause will be `staking.ErrInsufficientBalance`, and it should be safe for the caller to roll back to an earlier state tree and continue from there.
func (s *MutableState) AddRewards(time epochtime.EpochTime, factor *quantity.Quantity, accounts []signature.PublicKey) error { steps, err := s.RewardSchedule() if err != nil { return err } var activeStep *staking.RewardStep for _, step := range steps { if time < step.Until { activeStep = &step break } } if activeStep == nil { // We're past the end of the schedule. return nil } commonPool, err := s.CommonPool() if err != nil { return errors.Wrap(err, "loading common pool") } for _, id := range accounts { ent := s.Account(id) q := ent.Escrow.Active.Balance.Clone() // Multiply first. if err := q.Mul(factor); err != nil { return errors.Wrap(err, "multiplying by reward factor") } if err := q.Mul(&activeStep.Scale); err != nil { return errors.Wrap(err, "multiplying by reward step scale") } if err := q.Quo(staking.RewardAmountDenominator); err != nil { return errors.Wrap(err, "dividing by reward amount denominator") } if q.IsZero() { continue } var com *quantity.Quantity rate := ent.Escrow.CommissionSchedule.CurrentRate(time) if rate != nil { com = q.Clone() // Multiply first. if err := com.Mul(rate); err != nil { return errors.Wrap(err, "multiplying by commission rate") } if err := com.Quo(staking.CommissionRateDenominator); err != nil { return errors.Wrap(err, "dividing by commission rate denominator") } if err := q.Sub(com); err != nil { return errors.Wrap(err, "subtracting commission") } } if !q.IsZero() { if err := quantity.Move(&ent.Escrow.Active.Balance, commonPool, q); err != nil { return errors.Wrap(err, "transferring to active escrow balance from common pool") } } if com != nil && !com.IsZero() { delegation := s.Delegation(id, id) if err := ent.Escrow.Active.Deposit(&delegation.Shares, commonPool, com); err != nil { return errors.Wrap(err, "depositing commission") } s.SetDelegation(id, id, delegation) } s.SetAccount(id, ent) } s.SetCommonPool(commonPool) return nil }
[ "func accumulateRewards(config *params.ChainConfig, state *state.DB, header *types.Header) {\n\t// TODO: implement mining rewards\n}", "func (_RandomBeacon *RandomBeaconTransactor) WithdrawRewards(opts *bind.TransactOpts, stakingProvider common.Address) (*types.Transaction, error) {\n\treturn _RandomBeacon.contract.Transact(opts, \"withdrawRewards\", stakingProvider)\n}", "func (path *Path) AddRewards(rewards map[*Reward]int) {\n\tfor key, value := range rewards {\n\t\tpath.rewards[key] += value\n\t}\n}", "func (_RandomBeacon *RandomBeaconTransactorSession) WithdrawRewards(stakingProvider common.Address) (*types.Transaction, error) {\n\treturn _RandomBeacon.Contract.WithdrawRewards(&_RandomBeacon.TransactOpts, stakingProvider)\n}", "func (_SfcContract *SfcContractTransactorSession) StashRewards(delegator common.Address, toValidatorID *big.Int) (*types.Transaction, error) {\n\treturn _SfcContract.Contract.StashRewards(&_SfcContract.TransactOpts, delegator, toValidatorID)\n}", "func (_XStaking *XStakingCallerSession) Rewards(arg0 common.Address) (*big.Int, error) {\n\treturn _XStaking.Contract.Rewards(&_XStaking.CallOpts, arg0)\n}", "func (_Stakingbindings *StakingbindingsCallerSession) Rewards(arg0 common.Address) (*big.Int, error) {\n\treturn _Stakingbindings.Contract.Rewards(&_Stakingbindings.CallOpts, arg0)\n}", "func (_SfcContract *SfcContractTransactor) StashRewards(opts *bind.TransactOpts, delegator common.Address, toValidatorID *big.Int) (*types.Transaction, error) {\n\treturn _SfcContract.contract.Transact(opts, \"stashRewards\", delegator, toValidatorID)\n}", "func (_SfcContract *SfcContractTransactor) ClaimRewards(opts *bind.TransactOpts, toValidatorID *big.Int) (*types.Transaction, error) {\n\treturn _SfcContract.contract.Transact(opts, \"claimRewards\", toValidatorID)\n}", "func (_XStaking *XStakingCaller) Rewards(opts *bind.CallOpts, arg0 common.Address) (*big.Int, error) {\n\tvar out []interface{}\n\terr := _XStaking.contract.Call(opts, &out, \"rewards\", arg0)\n\n\tif err != nil {\n\t\treturn *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\n\treturn out0, err\n\n}", "func (_Token *TokenTransactor) SetupRewards(opts *bind.TransactOpts, multiplier *big.Int, anualRewardRates []*big.Int, lowerBounds []*big.Int, upperBounds []*big.Int) (*types.Transaction, error) {\n\treturn _Token.contract.Transact(opts, \"setupRewards\", multiplier, anualRewardRates, lowerBounds, upperBounds)\n}", "func AccumulateRewardsAndCountSigs(\n\tbc engine.ChainReader, state *state.DB,\n\theader *block.Header, beaconChain engine.ChainReader,\n) (reward.Reader, error) {\n\tblockNum := header.Number().Uint64()\n\tcurrentHeader := beaconChain.CurrentHeader()\n\tnowEpoch, blockNow := currentHeader.Epoch(), currentHeader.Number()\n\n\tif blockNum == 0 {\n\t\t// genesis block has no parent to reward.\n\t\treturn network.EmptyPayout, nil\n\t}\n\n\tif bc.Config().IsStaking(header.Epoch()) &&\n\t\tbc.CurrentHeader().ShardID() != shard.BeaconChainShardID {\n\t\treturn network.EmptyPayout, nil\n\n\t}\n\n\t// After staking\n\tif headerE := header.Epoch(); bc.Config().IsStaking(headerE) &&\n\t\tbc.CurrentHeader().ShardID() == shard.BeaconChainShardID {\n\t\tutils.AnalysisStart(\"accumulateRewardBeaconchainSelfPayout\", nowEpoch, blockNow)\n\t\tdefaultReward := network.BaseStakedReward\n\n\t\t// Following is commented because the new econ-model has a flat-rate block reward\n\t\t// of 28 ONE per block assuming 4 shards and 8s block time:\n\t\t//// TODO Use cached result in off-chain db instead of full computation\n\t\t//_, percentageStaked, err := network.WhatPercentStakedNow(\n\t\t//\tbeaconChain, header.Time().Int64(),\n\t\t//)\n\t\t//if err != nil {\n\t\t//\treturn network.EmptyPayout, err\n\t\t//}\n\t\t//howMuchOff, adjustBy := network.Adjustment(*percentageStaked)\n\t\t//defaultReward = defaultReward.Add(adjustBy)\n\t\t//utils.Logger().Info().\n\t\t//\tStr(\"percentage-token-staked\", percentageStaked.String()).\n\t\t//\tStr(\"how-much-off\", howMuchOff.String()).\n\t\t//\tStr(\"adjusting-by\", adjustBy.String()).\n\t\t//\tStr(\"block-reward\", defaultReward.String()).\n\t\t//\tMsg(\"dynamic adjustment of block-reward \")\n\n\t\t// If too much is staked, then possible to have negative reward,\n\t\t// not an error, just a possible economic situation, hence we return\n\t\tif defaultReward.IsNegative() {\n\t\t\treturn network.EmptyPayout, nil\n\t\t}\n\n\t\tnewRewards, beaconP, shardP :=\n\t\t\tbig.NewInt(0), []reward.Payout{}, []reward.Payout{}\n\n\t\t// Take care of my own beacon chain committee, _ is missing, for slashing\n\t\tmembers, payable, missing, err := ballotResultBeaconchain(beaconChain, header)\n\t\tif err != nil {\n\t\t\treturn network.EmptyPayout, err\n\t\t}\n\t\tsubComm := shard.Committee{shard.BeaconChainShardID, members}\n\n\t\tif err := availability.IncrementValidatorSigningCounts(\n\t\t\tbeaconChain,\n\t\t\tsubComm.StakedValidators(),\n\t\t\tstate,\n\t\t\tpayable,\n\t\t\tmissing,\n\t\t); err != nil {\n\t\t\treturn network.EmptyPayout, err\n\t\t}\n\t\tvotingPower, err := lookupVotingPower(\n\t\t\theaderE, &subComm,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn network.EmptyPayout, err\n\t\t}\n\n\t\tallSignersShare := numeric.ZeroDec()\n\t\tfor j := range payable {\n\t\t\tvoter := votingPower.Voters[payable[j].BLSPublicKey]\n\t\t\tif !voter.IsHarmonyNode {\n\t\t\t\tvoterShare := voter.OverallPercent\n\t\t\t\tallSignersShare = allSignersShare.Add(voterShare)\n\t\t\t}\n\t\t}\n\t\tfor beaconMember := range payable {\n\t\t\t// TODO Give out whatever leftover to the last voter/handle\n\t\t\t// what to do about share of those that didn't sign\n\t\t\tblsKey := payable[beaconMember].BLSPublicKey\n\t\t\tvoter := votingPower.Voters[blsKey]\n\t\t\tif !voter.IsHarmonyNode {\n\t\t\t\tsnapshot, err := bc.ReadValidatorSnapshot(voter.EarningAccount)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t}\n\t\t\t\tdue := defaultReward.Mul(\n\t\t\t\t\tvoter.OverallPercent.Quo(allSignersShare),\n\t\t\t\t).RoundInt()\n\t\t\t\tnewRewards.Add(newRewards, due)\n\n\t\t\t\tshares, err := lookupDelegatorShares(snapshot)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t}\n\t\t\t\tif err := state.AddReward(snapshot.Validator, due, shares); err != nil {\n\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t}\n\t\t\t\tbeaconP = append(beaconP, reward.Payout{\n\t\t\t\t\tShardID: shard.BeaconChainShardID,\n\t\t\t\t\tAddr: voter.EarningAccount,\n\t\t\t\t\tNewlyEarned: due,\n\t\t\t\t\tEarningKey: voter.Identity,\n\t\t\t\t})\n\t\t\t}\n\t\t}\n\t\tutils.AnalysisEnd(\"accumulateRewardBeaconchainSelfPayout\", nowEpoch, blockNow)\n\n\t\tutils.AnalysisStart(\"accumulateRewardShardchainPayout\", nowEpoch, blockNow)\n\t\t// Handle rewards for shardchain\n\t\tif cxLinks := header.CrossLinks(); len(cxLinks) > 0 {\n\t\t\tcrossLinks := types.CrossLinks{}\n\t\t\tif err := rlp.DecodeBytes(cxLinks, &crossLinks); err != nil {\n\t\t\t\treturn network.EmptyPayout, err\n\t\t\t}\n\n\t\t\ttype slotPayable struct {\n\t\t\t\tshard.Slot\n\t\t\t\tpayout *big.Int\n\t\t\t\tbucket int\n\t\t\t\tindex int\n\t\t\t\tshardID uint32\n\t\t\t}\n\n\t\t\ttype slotMissing struct {\n\t\t\t\tshard.Slot\n\t\t\t\tbucket int\n\t\t\t\tindex int\n\t\t\t}\n\n\t\t\tallPayables, allMissing := []slotPayable{}, []slotMissing{}\n\n\t\t\tfor i := range crossLinks {\n\n\t\t\t\tcxLink := crossLinks[i]\n\t\t\t\tepoch, shardID := cxLink.Epoch(), cxLink.ShardID()\n\t\t\t\tif !bc.Config().IsStaking(epoch) {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tshardState, err := bc.ReadShardState(epoch)\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t}\n\n\t\t\t\tsubComm, err := shardState.FindCommitteeByID(shardID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t}\n\n\t\t\t\tpayableSigners, missing, err := availability.BlockSigners(\n\t\t\t\t\tcxLink.Bitmap(), subComm,\n\t\t\t\t)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t}\n\n\t\t\t\tstaked := subComm.StakedValidators()\n\t\t\t\tif err := availability.IncrementValidatorSigningCounts(\n\t\t\t\t\tbeaconChain, staked, state, payableSigners, missing,\n\t\t\t\t); err != nil {\n\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t}\n\n\t\t\t\tvotingPower, err := lookupVotingPower(\n\t\t\t\t\tepoch, subComm,\n\t\t\t\t)\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t}\n\n\t\t\t\tallSignersShare := numeric.ZeroDec()\n\t\t\t\tfor j := range payableSigners {\n\t\t\t\t\tvoter := votingPower.Voters[payableSigners[j].BLSPublicKey]\n\t\t\t\t\tif !voter.IsHarmonyNode {\n\t\t\t\t\t\tvoterShare := voter.OverallPercent\n\t\t\t\t\t\tallSignersShare = allSignersShare.Add(voterShare)\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tfor j := range payableSigners {\n\t\t\t\t\tvoter := votingPower.Voters[payableSigners[j].BLSPublicKey]\n\t\t\t\t\tif !voter.IsHarmonyNode && !voter.OverallPercent.IsZero() {\n\t\t\t\t\t\tdue := defaultReward.Mul(\n\t\t\t\t\t\t\tvoter.OverallPercent.Quo(allSignersShare),\n\t\t\t\t\t\t)\n\t\t\t\t\t\tallPayables = append(allPayables, slotPayable{\n\t\t\t\t\t\t\tSlot: payableSigners[j],\n\t\t\t\t\t\t\tpayout: due.TruncateInt(),\n\t\t\t\t\t\t\tbucket: i,\n\t\t\t\t\t\t\tindex: j,\n\t\t\t\t\t\t\tshardID: shardID,\n\t\t\t\t\t\t})\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tfor j := range missing {\n\t\t\t\t\tallMissing = append(allMissing, slotMissing{\n\t\t\t\t\t\tSlot: missing[j],\n\t\t\t\t\t\tbucket: i,\n\t\t\t\t\t\tindex: j,\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tresultsHandle := make([][]slotPayable, len(crossLinks))\n\t\t\tfor i := range resultsHandle {\n\t\t\t\tresultsHandle[i] = []slotPayable{}\n\t\t\t}\n\n\t\t\tfor _, payThem := range allPayables {\n\t\t\t\tbucket := payThem.bucket\n\t\t\t\tresultsHandle[bucket] = append(resultsHandle[bucket], payThem)\n\t\t\t}\n\n\t\t\t// Check if any errors and sort each bucket to enforce order\n\t\t\tfor bucket := range resultsHandle {\n\t\t\t\tsort.SliceStable(resultsHandle[bucket],\n\t\t\t\t\tfunc(i, j int) bool {\n\t\t\t\t\t\treturn resultsHandle[bucket][i].index < resultsHandle[bucket][j].index\n\t\t\t\t\t},\n\t\t\t\t)\n\t\t\t}\n\n\t\t\t// Finally do the pay\n\t\t\tfor bucket := range resultsHandle {\n\t\t\t\tfor payThem := range resultsHandle[bucket] {\n\t\t\t\t\tpayable := resultsHandle[bucket][payThem]\n\t\t\t\t\tsnapshot, err := bc.ReadValidatorSnapshot(\n\t\t\t\t\t\tpayable.EcdsaAddress,\n\t\t\t\t\t)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t\t}\n\t\t\t\t\tdue := resultsHandle[bucket][payThem].payout\n\t\t\t\t\tnewRewards.Add(newRewards, due)\n\n\t\t\t\t\tshares, err := lookupDelegatorShares(snapshot)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t\t}\n\t\t\t\t\tif err := state.AddReward(snapshot.Validator, due, shares); err != nil {\n\t\t\t\t\t\treturn network.EmptyPayout, err\n\t\t\t\t\t}\n\t\t\t\t\tshardP = append(shardP, reward.Payout{\n\t\t\t\t\t\tShardID: payable.shardID,\n\t\t\t\t\t\tAddr: payable.EcdsaAddress,\n\t\t\t\t\t\tNewlyEarned: due,\n\t\t\t\t\t\tEarningKey: payable.BLSPublicKey,\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t\tutils.AnalysisEnd(\"accumulateRewardShardchainPayout\", nowEpoch, blockNow)\n\t\t\treturn network.NewStakingEraRewardForRound(\n\t\t\t\tnewRewards, missing, beaconP, shardP,\n\t\t\t), nil\n\t\t}\n\t\treturn network.EmptyPayout, nil\n\t}\n\n\t// Before staking\n\tparentHeader := bc.GetHeaderByHash(header.ParentHash())\n\tif parentHeader == nil {\n\t\treturn network.EmptyPayout, errors.Errorf(\n\t\t\t\"cannot find parent block header in DB at parent hash %s\",\n\t\t\theader.ParentHash().Hex(),\n\t\t)\n\t}\n\tif parentHeader.Number().Cmp(common.Big0) == 0 {\n\t\t// Parent is an epoch block,\n\t\t// which is not signed in the usual manner therefore rewards nothing.\n\t\treturn network.EmptyPayout, nil\n\t}\n\tparentShardState, err := bc.ReadShardState(parentHeader.Epoch())\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(\n\t\t\terr, \"cannot read shard state at epoch %v\", parentHeader.Epoch(),\n\t\t)\n\t}\n\t_, signers, _, err := availability.BallotResult(\n\t\tparentHeader, header, parentShardState, header.ShardID(),\n\t)\n\n\tif err != nil {\n\t\treturn network.EmptyPayout, err\n\t}\n\n\ttotalAmount := big.NewInt(0)\n\n\t{\n\t\tlast := big.NewInt(0)\n\t\tcount := big.NewInt(int64(len(signers)))\n\t\tfor i, account := range signers {\n\t\t\tcur := big.NewInt(0)\n\t\t\tcur.Mul(network.BlockReward, big.NewInt(int64(i+1))).Div(cur, count)\n\t\t\tdiff := big.NewInt(0).Sub(cur, last)\n\t\t\tstate.AddBalance(account.EcdsaAddress, diff)\n\t\t\ttotalAmount.Add(totalAmount, diff)\n\t\t\tlast = cur\n\t\t}\n\t}\n\n\tif totalAmount.Cmp(network.BlockReward) != 0 {\n\t\tutils.Logger().Error().\n\t\t\tInt64(\"block-reward\", network.BlockReward.Int64()).\n\t\t\tInt64(\"total-amount-paid-out\", totalAmount.Int64()).\n\t\t\tMsg(\"Total paid out was not equal to block-reward\")\n\t\treturn nil, errors.Wrapf(\n\t\t\tnetwork.ErrPayoutNotEqualBlockReward, \"payout \"+totalAmount.String(),\n\t\t)\n\t}\n\n\treturn network.NewPreStakingEraRewarded(totalAmount), nil\n}", "func (c *gRPCClient) AccountRewards(address gosmtypes.Address, offset uint32, maxResults uint32) ([]*apitypes.Reward, uint32, error) {\n\tgsc := c.getGlobalStateServiceClient()\n\tresp, err := gsc.AccountDataQuery(context.Background(), &apitypes.AccountDataQueryRequest{\n\t\tFilter: &apitypes.AccountDataFilter{\n\t\t\tAccountId: &apitypes.AccountId{Address: address.Bytes()},\n\t\t\tAccountDataFlags: uint32(apitypes.AccountDataFlag_ACCOUNT_DATA_FLAG_REWARD),\n\t\t},\n\n\t\tMaxResults: maxResults,\n\t\tOffset: offset,\n\t})\n\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\n\trewards := make([]*apitypes.Reward, 0)\n\n\tfor _, data := range resp.AccountItem {\n\t\tr := data.GetReward()\n\t\tif r != nil {\n\t\t\trewards = append(rewards, r)\n\t\t}\n\t}\n\n\treturn rewards, resp.TotalResults, nil\n}", "func (_Stakingbindings *StakingbindingsCaller) Rewards(opts *bind.CallOpts, arg0 common.Address) (*big.Int, error) {\n\tvar out []interface{}\n\terr := _Stakingbindings.contract.Call(opts, &out, \"rewards\", arg0)\n\n\tif err != nil {\n\t\treturn *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\n\treturn out0, err\n\n}", "func (c *Coinbase) AddReward(output *Output) {\n\toutput.EncryptedMask = make([]byte, 1)\n\tc.Rewards = append(c.Rewards, output)\n}", "func (_SfcContract *SfcContractTransactor) RestakeRewards(opts *bind.TransactOpts, toValidatorID *big.Int) (*types.Transaction, error) {\n\treturn _SfcContract.contract.Transact(opts, \"restakeRewards\", toValidatorID)\n}", "func (_SfcContract *SfcContractTransactorSession) ClaimRewards(toValidatorID *big.Int) (*types.Transaction, error) {\n\treturn _SfcContract.Contract.ClaimRewards(&_SfcContract.TransactOpts, toValidatorID)\n}", "func (_Token *TokenTransactorSession) SetupRewards(multiplier *big.Int, anualRewardRates []*big.Int, lowerBounds []*big.Int, upperBounds []*big.Int) (*types.Transaction, error) {\n\treturn _Token.Contract.SetupRewards(&_Token.TransactOpts, multiplier, anualRewardRates, lowerBounds, upperBounds)\n}", "func (fac *FMintAccount) RewardsEarned() (hexutil.Big, error) {\n\treturn fac.repo.FMintRewardsEarned(&fac.Address)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
AddRewardSingleAttenuated computes, scales, and transfers a staking reward to an active escrow account.
func (s *MutableState) AddRewardSingleAttenuated(time epochtime.EpochTime, factor *quantity.Quantity, attenuationNumerator, attenuationDenominator int, account signature.PublicKey) error { steps, err := s.RewardSchedule() if err != nil { return err } var activeStep *staking.RewardStep for _, step := range steps { if time < step.Until { activeStep = &step break } } if activeStep == nil { // We're past the end of the schedule. return nil } var numQ, denQ quantity.Quantity if err = numQ.FromInt64(int64(attenuationNumerator)); err != nil { return errors.Wrapf(err, "importing attenuation numerator %d", attenuationNumerator) } if err = denQ.FromInt64(int64(attenuationDenominator)); err != nil { return errors.Wrapf(err, "importing attenuation denominator %d", attenuationDenominator) } commonPool, err := s.CommonPool() if err != nil { return errors.Wrap(err, "loading common pool") } ent := s.Account(account) q := ent.Escrow.Active.Balance.Clone() // Multiply first. if err := q.Mul(factor); err != nil { return errors.Wrap(err, "multiplying by reward factor") } if err := q.Mul(&activeStep.Scale); err != nil { return errors.Wrap(err, "multiplying by reward step scale") } if err := q.Mul(&numQ); err != nil { return errors.Wrap(err, "multiplying by attenuation numerator") } if err := q.Quo(staking.RewardAmountDenominator); err != nil { return errors.Wrap(err, "dividing by reward amount denominator") } if err := q.Quo(&denQ); err != nil { return errors.Wrap(err, "dividing by attenuation denominator") } if q.IsZero() { return nil } var com *quantity.Quantity rate := ent.Escrow.CommissionSchedule.CurrentRate(time) if rate != nil { com = q.Clone() // Multiply first. if err := com.Mul(rate); err != nil { return errors.Wrap(err, "multiplying by commission rate") } if err := com.Quo(staking.CommissionRateDenominator); err != nil { return errors.Wrap(err, "dividing by commission rate denominator") } if err := q.Sub(com); err != nil { return errors.Wrap(err, "subtracting commission") } } if !q.IsZero() { if err := quantity.Move(&ent.Escrow.Active.Balance, commonPool, q); err != nil { return errors.Wrap(err, "transferring to active escrow balance from common pool") } } if com != nil && !com.IsZero() { delegation := s.Delegation(account, account) if err := ent.Escrow.Active.Deposit(&delegation.Shares, commonPool, com); err != nil { return errors.Wrap(err, "depositing commission") } s.SetDelegation(account, account, delegation) } s.SetAccount(account, ent) s.SetCommonPool(commonPool) return nil }
[ "func (k Keeper) CalculateSingleReward(oldIndex, newIndex, sourceShares sdk.Dec) (sdkmath.Int, error) {\n\tincrease := newIndex.Sub(oldIndex)\n\tif increase.IsNegative() {\n\t\treturn sdkmath.Int{}, errorsmod.Wrapf(types.ErrDecreasingRewardFactor, \"old: %v, new: %v\", oldIndex, newIndex)\n\t}\n\treward := increase.Mul(sourceShares).RoundInt()\n\treturn reward, nil\n}", "func (c *Coinbase) AddReward(output *Output) {\n\toutput.EncryptedMask = make([]byte, 1)\n\tc.Rewards = append(c.Rewards, output)\n}", "func (a Actor) AwardBlockReward(rt vmr.Runtime, params *AwardBlockRewardParams) *adt.EmptyValue {\n\trt.ValidateImmediateCallerIs(builtin.SystemActorAddr)\n\tAssertMsg(rt.CurrentBalance().GreaterThanEqual(params.GasReward),\n\t\t\"actor current balance %v insufficient to pay gas reward %v\", rt.CurrentBalance(), params.GasReward)\n\n\tAssertMsg(params.TicketCount > 0, \"cannot give block reward for zero tickets\")\n\n\tminer, ok := rt.ResolveAddress(params.Miner)\n\tif !ok {\n\t\trt.Abortf(exitcode.ErrIllegalState, \"failed to resolve given owner address\")\n\t}\n\n\tpriorBalance := rt.CurrentBalance()\n\n\tvar penalty abi.TokenAmount\n\tvar st State\n\trt.State().Transaction(&st, func() interface{} {\n\t\tblockReward := a.computeBlockReward(&st, big.Sub(priorBalance, params.GasReward), params.TicketCount)\n\t\ttotalReward := big.Add(blockReward, params.GasReward)\n\n\t\t// Cap the penalty at the total reward value.\n\t\tpenalty = big.Min(params.Penalty, totalReward)\n\n\t\t// Reduce the payable reward by the penalty.\n\t\trewardPayable := big.Sub(totalReward, penalty)\n\n\t\tAssertMsg(big.Add(rewardPayable, penalty).LessThanEqual(priorBalance),\n\t\t\t\"reward payable %v + penalty %v exceeds balance %v\", rewardPayable, penalty, priorBalance)\n\n\t\t// Record new reward into reward map.\n\t\tif rewardPayable.GreaterThan(abi.NewTokenAmount(0)) {\n\t\t\tnewReward := Reward{\n\t\t\t\tStartEpoch: rt.CurrEpoch(),\n\t\t\t\tEndEpoch: rt.CurrEpoch() + rewardVestingPeriod,\n\t\t\t\tValue: rewardPayable,\n\t\t\t\tAmountWithdrawn: abi.NewTokenAmount(0),\n\t\t\t\tVestingFunction: rewardVestingFunction,\n\t\t\t}\n\t\t\tif err := st.addReward(adt.AsStore(rt), miner, &newReward); err != nil {\n\t\t\t\trt.Abortf(exitcode.ErrIllegalState, \"failed to add reward to rewards map: %w\", err)\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n\n\t// Burn the penalty amount.\n\t_, code := rt.Send(builtin.BurntFundsActorAddr, builtin.MethodSend, nil, penalty)\n\tbuiltin.RequireSuccess(rt, code, \"failed to send penalty to BurntFundsActor\")\n\n\treturn nil\n}", "func (d *Dao) AddReward(c context.Context, iRewardID int64, uid int64, iSource int64, iRoomid int64, iLifespan int64) (err error) {\n\t//aReward, _ := getRewardConfByLid(iRewardID)\n\n\tm, _ := time.ParseDuration(fmt.Sprintf(\"+%dh\", iLifespan))\n\n\targ := &AnchorTaskModel.AnchorReward{\n\t\tUid: uid,\n\t\tRewardId: iRewardID,\n\t\tRoomid: iRoomid,\n\t\tSource: iSource,\n\t\tAchieveTime: xtime.Time(time.Now().Unix()),\n\t\tExpireTime: xtime.Time(time.Now().Add(m).Unix()),\n\t\tStatus: model.RewardUnUsed,\n\t}\n\n\t//spew.Dump\n\t// (arg)\n\tif err := d.orm.Create(arg).Error; err != nil {\n\t\tlog.Error(\"addReward(%v) error(%v)\", arg, err)\n\t\treturn err\n\t}\n\n\tif err := d.SetNewReward(c, uid, int64(1)); err != nil {\n\t\tlog.Error(\"addRewardMc(%v) error(%v)\", uid, err)\n\t}\n\n\tif err := d.SetHasReward(c, uid, int64(1)); err != nil {\n\t\tlog.Error(\"SetHasReward(%v) error(%v)\", uid, err)\n\t}\n\n\tlog.Info(\"addReward (%v) succ\", arg)\n\n\treturn\n}", "func ApplyRewardTx(tx *types.Transaction, statedb *state.Statedb) (*types.Receipt, error) {\n\tstatedb.CreateAccount(tx.Data.To)\n\tstatedb.AddBalance(tx.Data.To, tx.Data.Amount)\n\n\thash, err := statedb.Hash()\n\tif err != nil {\n\t\treturn nil, errors.NewStackedError(err, \"failed to get statedb root hash\")\n\t}\n\n\treceipt := &types.Receipt{\n\t\tTxHash: tx.Hash,\n\t\tPostState: hash,\n\t}\n\n\treturn receipt, nil\n}", "func (t *trusteeImpl) NewMiningRewardTx(block consensus.Block) *consensus.Transaction {\n\tvar tx *consensus.Transaction\n\t// build list of miner nodes for uncle blocks\n\tuncleMiners := make([][]byte, len(block.UncleMiners()))\n\tfor i, uncleMiner := range block.UncleMiners() {\n\t\tuncleMiners[i] = uncleMiner\n\t}\n\t\n\tops := make([]Op, 1 + len(uncleMiners))\n\t// first add self's mining reward\n\tops[0] = *t.myReward\n\t\n\t// now add award for each uncle\n\tfor i, uncleMiner := range uncleMiners {\n\t\top := NewOp(OpReward)\n\t\top.Params[ParamUncle] = bytesToHexString(uncleMiner)\n\t\top.Params[ParamAward] = UncleAward\n\t\tops[i+1] = *op \n\t}\n\t// serialize ops into payload\n\tif payload,err := common.Serialize(ops); err != nil {\n\t\tt.log.Error(\"Failed to serialize ops into payload: %s\", err)\n\t\treturn nil\n\t} else {\n\t\t// make a signed transaction out of payload\n\t\tif signature := t.sign(payload); len(signature) > 0 {\n\t\t\t// return the signed transaction\n\t\t\ttx = consensus.NewTransaction(payload, signature, t.myAddress)\n\t\t\tblock.AddTransaction(tx)\n\t\t\tt.process(block, tx)\n\t\t}\n\t}\n\treturn tx\n}", "func (_XStaking *XStakingFilterer) FilterRewardAdded(opts *bind.FilterOpts) (*XStakingRewardAddedIterator, error) {\n\n\tlogs, sub, err := _XStaking.contract.FilterLogs(opts, \"RewardAdded\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &XStakingRewardAddedIterator{contract: _XStaking.contract, event: \"RewardAdded\", logs: logs, sub: sub}, nil\n}", "func (_Stakingbindings *StakingbindingsFilterer) FilterRewardAdded(opts *bind.FilterOpts) (*StakingbindingsRewardAddedIterator, error) {\n\n\tlogs, sub, err := _Stakingbindings.contract.FilterLogs(opts, \"RewardAdded\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &StakingbindingsRewardAddedIterator{contract: _Stakingbindings.contract, event: \"RewardAdded\", logs: logs, sub: sub}, nil\n}", "func (_XStaking *XStakingFilterer) WatchRewardAdded(opts *bind.WatchOpts, sink chan<- *XStakingRewardAdded) (event.Subscription, error) {\n\n\tlogs, sub, err := _XStaking.contract.WatchLogs(opts, \"RewardAdded\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(XStakingRewardAdded)\n\t\t\t\tif err := _XStaking.contract.UnpackLog(event, \"RewardAdded\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (s *MutableState) AddRewards(time epochtime.EpochTime, factor *quantity.Quantity, accounts []signature.PublicKey) error {\n\tsteps, err := s.RewardSchedule()\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar activeStep *staking.RewardStep\n\tfor _, step := range steps {\n\t\tif time < step.Until {\n\t\t\tactiveStep = &step\n\t\t\tbreak\n\t\t}\n\t}\n\tif activeStep == nil {\n\t\t// We're past the end of the schedule.\n\t\treturn nil\n\t}\n\n\tcommonPool, err := s.CommonPool()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"loading common pool\")\n\t}\n\n\tfor _, id := range accounts {\n\t\tent := s.Account(id)\n\n\t\tq := ent.Escrow.Active.Balance.Clone()\n\t\t// Multiply first.\n\t\tif err := q.Mul(factor); err != nil {\n\t\t\treturn errors.Wrap(err, \"multiplying by reward factor\")\n\t\t}\n\t\tif err := q.Mul(&activeStep.Scale); err != nil {\n\t\t\treturn errors.Wrap(err, \"multiplying by reward step scale\")\n\t\t}\n\t\tif err := q.Quo(staking.RewardAmountDenominator); err != nil {\n\t\t\treturn errors.Wrap(err, \"dividing by reward amount denominator\")\n\t\t}\n\n\t\tif q.IsZero() {\n\t\t\tcontinue\n\t\t}\n\n\t\tvar com *quantity.Quantity\n\t\trate := ent.Escrow.CommissionSchedule.CurrentRate(time)\n\t\tif rate != nil {\n\t\t\tcom = q.Clone()\n\t\t\t// Multiply first.\n\t\t\tif err := com.Mul(rate); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"multiplying by commission rate\")\n\t\t\t}\n\t\t\tif err := com.Quo(staking.CommissionRateDenominator); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"dividing by commission rate denominator\")\n\t\t\t}\n\n\t\t\tif err := q.Sub(com); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"subtracting commission\")\n\t\t\t}\n\t\t}\n\n\t\tif !q.IsZero() {\n\t\t\tif err := quantity.Move(&ent.Escrow.Active.Balance, commonPool, q); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"transferring to active escrow balance from common pool\")\n\t\t\t}\n\t\t}\n\n\t\tif com != nil && !com.IsZero() {\n\t\t\tdelegation := s.Delegation(id, id)\n\n\t\t\tif err := ent.Escrow.Active.Deposit(&delegation.Shares, commonPool, com); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"depositing commission\")\n\t\t\t}\n\n\t\t\ts.SetDelegation(id, id, delegation)\n\t\t}\n\n\t\ts.SetAccount(id, ent)\n\t}\n\n\ts.SetCommonPool(commonPool)\n\n\treturn nil\n}", "func (d *Dao) UseReward(id int64, usePlat string) (rst bool, err error) {\n\tif err := d.orm.\n\t\tModel(&model.AnchorReward{}).\n\t\tWhere(\"id=?\", id).\n\t\tUpdate(map[string]interface{}{\"status\": model.RewardUsed, \"use_plat\": usePlat, \"use_time\": xtime.Time(time.Now().Unix())}).Error; err != nil {\n\t\tlog.Error(\"useReward (%v) error(%v)\", id, err)\n\t\treturn rst, err\n\t}\n\trst = true\n\treturn\n}", "func (_Staking *StakingCallerSession) AccumulatedRewardRate() (*big.Int, error) {\n\treturn _Staking.Contract.AccumulatedRewardRate(&_Staking.CallOpts)\n}", "func (_Stakingbindings *StakingbindingsFilterer) WatchRewardAdded(opts *bind.WatchOpts, sink chan<- *StakingbindingsRewardAdded) (event.Subscription, error) {\n\n\tlogs, sub, err := _Stakingbindings.contract.WatchLogs(opts, \"RewardAdded\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(StakingbindingsRewardAdded)\n\t\t\t\tif err := _Stakingbindings.contract.UnpackLog(event, \"RewardAdded\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (_BondedECDSAKeep *BondedECDSAKeepTransactor) DistributeETHReward(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _BondedECDSAKeep.contract.Transact(opts, \"distributeETHReward\")\n}", "func (_Minter *MinterTransactorSession) CreateReward(_fracNum *big.Int, _fracDenom *big.Int) (*types.Transaction, error) {\n\treturn _Minter.Contract.CreateReward(&_Minter.TransactOpts, _fracNum, _fracDenom)\n}", "func (_Minter *MinterSession) CreateReward(_fracNum *big.Int, _fracDenom *big.Int) (*types.Transaction, error) {\n\treturn _Minter.Contract.CreateReward(&_Minter.TransactOpts, _fracNum, _fracDenom)\n}", "func (as AccountStorage) SetReward(ctx sdk.Context, accKey types.AccountKey, reward *Reward) sdk.Error {\n\tstore := ctx.KVStore(as.key)\n\trewardByte, err := as.cdc.MarshalJSON(*reward)\n\tif err != nil {\n\t\treturn ErrFailedToMarshalReward(err)\n\t}\n\tstore.Set(getRewardKey(accKey), rewardByte)\n\treturn nil\n}", "func (_BondedECDSAKeep *BondedECDSAKeepSession) DistributeETHReward() (*types.Transaction, error) {\n\treturn _BondedECDSAKeep.Contract.DistributeETHReward(&_BondedECDSAKeep.TransactOpts)\n}", "func (_Staking *StakingCaller) AccumulatedRewardRate(opts *bind.CallOpts) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _Staking.contract.Call(opts, out, \"accumulatedRewardRate\")\n\treturn *ret0, err\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
NewMutableState creates a new mutable staking state wrapper.
func NewMutableState(tree *iavl.MutableTree) *MutableState { inner := &abci.ImmutableState{Snapshot: tree.ImmutableTree} return &MutableState{ ImmutableState: &ImmutableState{inner}, tree: tree, } }
[ "func newLockState() *lockState {\n\treturn &lockState{\n\t\tlocks: make(map[string]string),\n\t}\n}", "func newSyncState(startBlock, syncedTo *BlockStamp) *syncState {\n\n\treturn &syncState{\n\t\tstartBlock: *startBlock,\n\t\tsyncedTo: *syncedTo,\n\t}\n}", "func NewState() *glua.LState {\n\t// Create a new lua state\n\tstate := glua.NewState(\n\t\tglua.Options{\n\t\t\tIncludeGoStackTrace: util.Config.Configuration.IsDev(),\n\t\t},\n\t)\n\n\t// Set castro metatables\n\tGetApplicationState(state)\n\n\t// Return the lua state\n\treturn state\n}", "func NewSecurityActionState()(*SecurityActionState) {\n m := &SecurityActionState{\n }\n m.backingStore = ie8677ce2c7e1b4c22e9c3827ecd078d41185424dd9eeb92b7d971ed2d49a392e.BackingStoreFactoryInstance();\n m.SetAdditionalData(make(map[string]any))\n return m\n}", "func NewState(v interface{}) State {\n\n\ts := &state{}\n\ts.i = v\n\treturn s\n\n}", "func NewState() *State {\n\treturn &State{C.tgl_state_alloc()}\n}", "func Newstate(player int) *State {\n\ts := State{player: player}\n\treturn &s\n}", "func NewState() *State {\n\tglobals := newModule()\n\t\n\t// Provide a self reference for use when declaring global script variables.\n\tglobals.modules.add(\"global\", globals)\n\t\n\treturn &State{\n\t\tglobal: globals,\n\t\tOutput: os.Stdout,\n\t}\n}", "func (p *luaStatePool) New() *glua.LState {\n\t// Create a new lua state\n\tstate := glua.NewState(\n\t\tglua.Options{\n\t\t\tIncludeGoStackTrace: util.Config.Configuration.IsDev(),\n\t\t},\n\t)\n\n\t// Set castro metatables\n\tGetApplicationState(state)\n\n\t// Return the lua state\n\treturn state\n}", "func NewState(name string) State {\n\ts := State{\n\t\tName: name,\n\t\tIsMoved: false,\n\t\tChain: \"\",\n\t}\n\n\treturn s\n}", "func newEvalState() *evalState {\n\treturn &evalState{\n\t\tstatus: structs.EvalStatusPending,\n\t\tallocs: make(map[string]*allocState),\n\t}\n}", "func New() State {\n\ts := new(viewState)\n\ts.viewChangeStartedFinished = sync.NewCond(s)\n\ts.activeViewReleased = sync.NewCond(s)\n\ts.viewChangeUnblocked = sync.NewCond(s)\n\n\treturn s\n}", "func (st *State) Clone() *State {\n\tst2 := NewState()\n\tif len(st.vars) != 0 {\n\t\tst2.vars = make(Vars)\n\t\tst2.unused = make(map[string]struct{})\n\t}\n\tfor k, v := range st.vars {\n\t\tst2.vars[k] = v\n\t}\n\tfor k := range st.unused {\n\t\tst2.unused[k] = struct{}{}\n\t}\n\tif len(st.states) != 0 {\n\t\tst2.states = make(map[string][]*State)\n\t}\n\tfor k, v := range st.states {\n\t\tst2.states[k] = v\n\t}\n\treturn st2\n}", "func MakeState(hash []byte) *State {\n var s *State = new(State)\n s.coeff = make(map[string]float64)\n s.off = make(map[string]float64)\n s.coeff[\"+\"] = 1\n s.coeff[\"*\"] = 1\n s.coeff[\"-\"] = 1\n s.coeff[\"/\"] = 1\n s.coeff[\"%\"] = 1\n s.coeff[\"++\"] = 1\n s.coeff[\"--\"] = 1\n s.coeff[\"==\"] = 1\n s.coeff[\"<\"] = 1\n s.coeff[\"<=\"] = 1\n s.coeff[\">\"] = 1\n s.coeff[\">=\"] = 1\n s.coeff[\"neg\"] = 1\n s.coeff[\"and\"] = 1\n s.coeff[\"or\"] = 1\n s.coeff[\"pow\"] = 1\n s.coeff[\"sqrt\"] = 1\n s.coeff[\"abs\"] = 1\n s.coeff[\"sign\"] = 1\n s.coeff[\"ceil\"] = 1\n s.coeff[\"floor\"] = 1\n s.coeff[\"round\"] = 1\n s.coeff[\"min\"] = 1\n s.coeff[\"max\"] = 1\n s.coeff[\"sin\"] = 1\n s.coeff[\"cos\"] = 1\n s.coeff[\"asin\"] = 1\n s.coeff[\"acos\"] = 1\n s.coeff[\"tan\"] = 1\n s.coeff[\"atan\"] = 1\n s.coeff[\"sinh\"] = 1\n s.coeff[\"cosh\"] = 1\n s.coeff[\"asinh\"] = 1\n s.coeff[\"acosh\"] = 1\n s.coeff[\"tanh\"] = 1\n s.coeff[\"atanh\"] = 1\n s.coeff[\"log\"] = 1\n s.coeff[\"log2\"] = 1\n s.coeff[\"log10\"] = 1\n s.coeff[\"exp\"] = 1\n s.coeff[\"exp2\"] = 1\n s.coeff[\"assign\"] = 1\n s.coeff[\"append\"] = 1\n s.coeff[\"delete\"] = 1\n s.SetBlockHash(hash)\n s.Reset()\n return s\n}", "func New(storage types.StateStorage) types.State {\n\treturn &stateManager{\n\t\tstorage: storage,\n\t}\n}", "func newStudentState(content *Content) *studentState {\n\tstate := &studentState{content: content}\n\tstate.reset()\n\treturn state\n}", "func NewState(t *testing.T) (*Mocker, *state.State) {\n\tm, se := NewSession(t)\n\n\tst, err := state.NewFromSession(se, new(state.NoopStore))\n\tif err != nil {\n\t\tpanic(err) // this should never happen\n\t}\n\n\treturn m, st\n}", "func newCompactionState(maxFileSize uint32, snapshot version.Snapshot, compaction *version.Compaction) *compactionState {\n\treturn &compactionState{\n\t\tmaxFileSize: maxFileSize,\n\t\tsnapshot: snapshot,\n\t\tcompaction: compaction,\n\t}\n}", "func NewSts() *StatefulSet { return &StatefulSet{sts: &v1.StatefulSet{}} }" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
NewMerkleBlobAccess creates an adapter that validates that blobs read from and written to storage correspond with the digest that is used for identification. It ensures that the size and the SHA256 based checksum match. This is used to ensure clients cannot corrupt the CAS and that if corruption were to occur, use of corrupted data is prevented.
func NewMerkleBlobAccess(blobAccess BlobAccess) BlobAccess { return &merkleBlobAccess{ BlobAccess: blobAccess, } }
[ "func NewFromBlob(blob []byte) Digest {\n\th := HashFn.New()\n\th.Write(blob)\n\tarr := h.Sum(nil)\n\treturn Digest{Hash: hex.EncodeToString(arr[:]), Size: int64(len(blob))}\n}", "func NewFlatBlobAccess(keyLocationMap KeyLocationMap, locationBlobMap LocationBlobMap, digestKeyFormat digest.KeyFormat, lock *sync.RWMutex, storageType string, capabilitiesProvider capabilities.Provider) blobstore.BlobAccess {\n\tflatBlobAccessPrometheusMetrics.Do(func() {\n\t\tprometheus.MustRegister(flatBlobAccessRefreshes)\n\t})\n\n\treturn &flatBlobAccess{\n\t\tProvider: capabilitiesProvider,\n\n\t\tkeyLocationMap: keyLocationMap,\n\t\tlocationBlobMap: locationBlobMap,\n\t\tdigestKeyFormat: digestKeyFormat,\n\t\tlock: lock,\n\n\t\trefreshesGet: flatBlobAccessRefreshes.WithLabelValues(storageType, \"Get\"),\n\t\trefreshesGetFromComposite: flatBlobAccessRefreshes.WithLabelValues(storageType, \"GetFromComposite\"),\n\t\trefreshesFindMissing: flatBlobAccessRefreshes.WithLabelValues(storageType, \"FindMissing\"),\n\t}\n}", "func NewHierarchicalCASBlobAccess(keyLocationMap KeyLocationMap, locationBlobMap LocationBlobMap, lock *sync.RWMutex, capabilitiesProvider capabilities.Provider) blobstore.BlobAccess {\n\treturn &hierarchicalCASBlobAccess{\n\t\tProvider: capabilitiesProvider,\n\n\t\tkeyLocationMap: keyLocationMap,\n\t\tlocationBlobMap: locationBlobMap,\n\t\tlock: lock,\n\t}\n}", "func NewCompletenessCheckingBlobAccess(actionCache, contentAddressableStorage blobstore.BlobAccess, batchSize, maximumMessageSizeBytes int, maximumTotalTreeSizeBytes int64) blobstore.BlobAccess {\n\treturn &completenessCheckingBlobAccess{\n\t\tBlobAccess: actionCache,\n\t\tcontentAddressableStorage: contentAddressableStorage,\n\t\tbatchSize: batchSize,\n\t\tmaximumMessageSizeBytes: maximumMessageSizeBytes,\n\t\tmaximumTotalTreeSizeBytes: maximumTotalTreeSizeBytes,\n\t}\n}", "func NewBlobDigestCalculator() *BlobDigestCalculator {\n\treturn &BlobDigestCalculator{\n\t\th: sha256.New(),\n\t}\n}", "func newBlobCache(size int) *blobCache {\n\tc := &blobCache{\n\t\tfree: size,\n\t\tsize: size,\n\t}\n\n\t// NewLRU wants us to specify some max. number of entries, else it errors.\n\t// The actual maximum will be smaller than size/cacheOverhead, because we\n\t// evict entries (RemoveOldest in add) to maintain our size bound.\n\tmaxEntries := size / cacheOverhead\n\tlru, err := simplelru.NewLRU(maxEntries, c.evict)\n\tif err != nil {\n\t\tpanic(err) // Can only be maxEntries <= 0.\n\t}\n\tc.c = lru\n\n\treturn c\n}", "func FromBlob(blob []byte) *repb.Digest {\n\tsha256Arr := sha256.Sum256(blob)\n\treturn mustNew(hex.EncodeToString(sha256Arr[:]), int64(len(blob)))\n}", "func (c *Client) ReadBlob(ctx context.Context, d digest.Digest) ([]byte, error) {\n\treturn c.readBlob(ctx, d.Hash, d.Size, 0, 0)\n}", "func (c *Client) digestBlob(b []byte) *pb.Digest {\n\tsum := c.sum(b)\n\treturn &pb.Digest{\n\t\tHash: hex.EncodeToString(sum[:]),\n\t\tSizeBytes: int64(len(b)),\n\t}\n}", "func TestReaderRoot(t *testing.T) {\n\tmt := CreateMerkleTester(t)\n\tbytes8 := []byte{0, 1, 2, 3, 4, 5, 6, 7}\n\treader := bytes.NewReader(bytes8)\n\troot, err := ReaderRoot(reader, sha256.New(), 1)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif bytes.Compare(root, mt.roots[8]) != 0 {\n\t\tt.Error(\"ReaderRoot returned the wrong root\")\n\t}\n}", "func TestTree_hash(t *testing.T) {\n\t// MD5\n\thashFuncMd5 := md5.New()\n\tdoctree, err := NewDocumentTree(TreeOptions{Hash: hashFuncMd5, Salts: NewSaltForTest})\n\tassert.Nil(t, err)\n\terr = doctree.AddLeavesFromDocument(&documentspb.LongDocumentExample)\n\tassert.Nil(t, err)\n\terr = doctree.Generate()\n\tassert.Nil(t, err)\n\n\texpectedRootHash := []byte{0xea, 0xa2, 0x2c, 0xc4, 0x1b, 0x91, 0x96, 0x23, 0x66, 0xc6, 0xa0, 0x8f, 0xaa, 0x49, 0xc0, 0xe8}\n\tassert.Equal(t, expectedRootHash, doctree.rootHash)\n\n\t// No hash func set\n\tdoctreeNoHash, err := NewDocumentTree(TreeOptions{Salts: NewSaltForTest})\n\tassert.Nil(t, err)\n\terr = doctreeNoHash.AddLeavesFromDocument(&documentspb.LongDocumentExample)\n\tassert.NotNil(t, err)\n\tassert.EqualError(t, err, \"hash is not set\")\n\n\t// SHA256\n\tdoctreeSha256, err := NewDocumentTree(TreeOptions{Hash: sha256Hash, Salts: NewSaltForTest})\n\tassert.Nil(t, err)\n\terr = doctreeSha256.AddLeavesFromDocument(&documentspb.LongDocumentExample)\n\tassert.Nil(t, err)\n\terr = doctreeSha256.Generate()\n\texpectedRootHash = []byte{0xe1, 0xee, 0x59, 0x40, 0xb8, 0x2c, 0x2b, 0xb4, 0x44, 0xa0, 0x4e, 0xe2, 0x3, 0x87, 0x27, 0xe8, 0x3a, 0xaa, 0xfd, 0xb0, 0x77, 0x70, 0x56, 0x5a, 0x5c, 0x40, 0xb3, 0x57, 0x14, 0x3d, 0xf0, 0xb5}\n\tassert.Equal(t, expectedRootHash, doctreeSha256.rootHash)\n}", "func New(ag agent.Agent, validators []types.Address, storage storage.Storage) (*MerkleSyncer, error) {\n\tctx, cancel := context.WithCancel(context.Background())\n\n\treturn &MerkleSyncer{\n\t\twrapperC: make(chan *pb.MerkleWrapper, wrapperCNumber),\n\t\tagent: ag,\n\t\tvalidators: validators,\n\t\tstorage: storage,\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t}, nil\n}", "func NewBlob(bs blob.Store) graph.Storage { return storage{bs: bs} }", "func NewMetricsBlobAccess(blobAccess BlobAccess, clock clock.Clock, storageType, backendType string) BlobAccess {\n\tblobAccessOperationsPrometheusMetrics.Do(func() {\n\t\tprometheus.MustRegister(blobAccessOperationsBlobSizeBytes)\n\t\tprometheus.MustRegister(blobAccessOperationsFindMissingBatchSize)\n\t\tprometheus.MustRegister(blobAccessOperationsDurationSeconds)\n\t})\n\n\treturn &metricsBlobAccess{\n\t\tblobAccess: blobAccess,\n\t\tclock: clock,\n\n\t\tgetBlobSizeBytes: blobAccessOperationsBlobSizeBytes.WithLabelValues(storageType, backendType, \"Get\"),\n\t\tgetDurationSeconds: blobAccessOperationsDurationSeconds.MustCurryWith(map[string]string{\"storage_type\": storageType, \"backend_type\": backendType, \"operation\": \"Get\"}),\n\t\tgetFromCompositeBlobSizeBytes: blobAccessOperationsBlobSizeBytes.WithLabelValues(storageType, backendType, \"GetFromComposite\"),\n\t\tgetFromCompositeDurationSeconds: blobAccessOperationsDurationSeconds.MustCurryWith(map[string]string{\"storage_type\": storageType, \"backend_type\": backendType, \"operation\": \"GetFromComposite\"}),\n\t\tputBlobSizeBytes: blobAccessOperationsBlobSizeBytes.WithLabelValues(storageType, backendType, \"Put\"),\n\t\tputDurationSeconds: blobAccessOperationsDurationSeconds.MustCurryWith(map[string]string{\"storage_type\": storageType, \"backend_type\": backendType, \"operation\": \"Put\"}),\n\t\tfindMissingBatchSize: blobAccessOperationsFindMissingBatchSize.WithLabelValues(storageType, backendType),\n\t\tfindMissingDurationSeconds: blobAccessOperationsDurationSeconds.MustCurryWith(map[string]string{\"storage_type\": storageType, \"backend_type\": backendType, \"operation\": \"FindMissing\"}),\n\t\tgetCapabilitiesSeconds: blobAccessOperationsDurationSeconds.MustCurryWith(map[string]string{\"storage_type\": storageType, \"backend_type\": backendType, \"operation\": \"GetCapabilities\"}),\n\t}\n}", "func (te *TreeEntry) Blob() *Blob {\n\treturn &Blob{\n\t\tID: te.ID,\n\t\tname: te.Name(),\n\t\tsize: te.size,\n\t\tgotSize: te.sized,\n\t\trepo: te.ptree.repo,\n\t}\n}", "func NewBlob(r io.Reader) (Blob, error) {\n\tb, err := ioutil.ReadAll(r)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to read data\")\n\t}\n\treturn &dataBlob{\n\t\tdata: b,\n\t}, nil\n}", "func newMerkleTree(height, n uint32) merkleTree {\n\treturn merkleTreeFromBuf(make([]byte, ((1<<height)-1)*n), height, n)\n}", "func (c *containerdCAS) ReadBlob(blobHash string) (io.Reader, error) {\n\tshaDigest := digest.Digest(blobHash)\n\t_, err := contentStore.Info(ctrdCtx, shaDigest)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"ReadBlob: Exception getting info of blob: %s. %s\", blobHash, err.Error())\n\t}\n\treaderAt, err := contentStore.ReaderAt(ctrdCtx, spec.Descriptor{Digest: shaDigest})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"ReadBlob: Exception while reading blob: %s. %s\", blobHash, err.Error())\n\t}\n\treturn content.NewReader(readerAt), nil\n}", "func (sr *immutableRef) setBlob(ctx context.Context, desc ocispec.Descriptor) error {\n\tif _, ok := leases.FromContext(ctx); !ok {\n\t\treturn errors.Errorf(\"missing lease requirement for setBlob\")\n\t}\n\n\tdiffID, err := diffIDFromDescriptor(desc)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif _, err := sr.cm.ContentStore.Info(ctx, desc.Digest); err != nil {\n\t\treturn err\n\t}\n\n\tsr.mu.Lock()\n\tdefer sr.mu.Unlock()\n\n\tif getChainID(sr.md) != \"\" {\n\t\treturn nil\n\t}\n\n\tif err := sr.finalize(ctx, true); err != nil {\n\t\treturn err\n\t}\n\n\tp := sr.parent\n\tvar parentChainID digest.Digest\n\tvar parentBlobChainID digest.Digest\n\tif p != nil {\n\t\tpInfo := p.Info()\n\t\tif pInfo.ChainID == \"\" || pInfo.BlobChainID == \"\" {\n\t\t\treturn errors.Errorf(\"failed to set blob for reference with non-addressable parent\")\n\t\t}\n\t\tparentChainID = pInfo.ChainID\n\t\tparentBlobChainID = pInfo.BlobChainID\n\t}\n\n\tif err := sr.cm.LeaseManager.AddResource(ctx, leases.Lease{ID: sr.ID()}, leases.Resource{\n\t\tID: desc.Digest.String(),\n\t\tType: \"content\",\n\t}); err != nil {\n\t\treturn err\n\t}\n\n\tqueueDiffID(sr.md, diffID.String())\n\tqueueBlob(sr.md, desc.Digest.String())\n\tchainID := diffID\n\tblobChainID := imagespecidentity.ChainID([]digest.Digest{desc.Digest, diffID})\n\tif parentChainID != \"\" {\n\t\tchainID = imagespecidentity.ChainID([]digest.Digest{parentChainID, chainID})\n\t\tblobChainID = imagespecidentity.ChainID([]digest.Digest{parentBlobChainID, blobChainID})\n\t}\n\tqueueChainID(sr.md, chainID.String())\n\tqueueBlobChainID(sr.md, blobChainID.String())\n\tqueueMediaType(sr.md, desc.MediaType)\n\tqueueBlobSize(sr.md, desc.Size)\n\tif err := sr.md.Commit(); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Restart restarts the application
func Restart() { log.Println("An error has occured, restarting the app") file, _ := osext.Executable() syscall.Exec(file, os.Args, os.Environ()) }
[ "func Restart(args ...string) {\n logger.Log(fmt.Sprintf(\"Restarting %v\\n\", process))\n Stop(args...)\n Start(args...)\n}", "func restart() {\n\tdefaultWorker.stop()\n\tdefaultWorker = newWorker()\n\tgo defaultWorker.start()\n}", "func (r *Runner) Restart(application *config.Application) {\n\tif cmd, ok := r.cmds[application.Name]; ok {\n\t\tpgid, err := syscall.Getpgid(cmd.Process.Pid)\n\t\tif err == nil {\n\t\t\tsyscall.Kill(-pgid, 15)\n\t\t}\n\t}\n\n\tgo r.Run(application)\n}", "func restart() {\n\tfmt.Println(\"Config change detected, restarting\")\n}", "func (s *Syncthing) Restart(ctx context.Context) error {\n\t_, err := s.APICall(ctx, \"rest/system/restart\", \"POST\", 200, nil, true, nil, false, 3)\n\treturn err\n}", "func (a API) Restart(cmd *None) (e error) {\n\tRPCHandlers[\"restart\"].Call <-API{a.Ch, cmd, nil}\n\treturn\n}", "func (pomo *Pomo) Restart() {\n\tpomo.SetDuration(DEFAULT_DURATION)\n}", "func (a *App) Restart(w io.Writer) error {\n\ta.Log(\"executing hook to restart\", \"tsuru\")\n\terr := a.preRestart(w)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = write(w, []byte(\"\\n ---> Restarting your app\\n\"))\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = a.run(\"/var/lib/tsuru/hooks/restart\", w)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn a.posRestart(w)\n}", "func (ac *appsClient) Restart(ctx context.Context, namespace, name string) error {\n\t_, err := ac.coreClient.Transform(ctx, namespace, name, func(app *v1alpha1.App) error {\n\t\tapp.Spec.Template.UpdateRequests++\n\t\treturn nil\n\t})\n\n\treturn err\n}", "func (r *Runner) restart(fileName string) {\n\tif r.running {\n\t\tr.kill()\n\t\tr.removeFile()\n\t}\n\n\tr.start <- fileName\n}", "func (f *ForestCustodian) restart(q *msg.Request, mr *msg.Result) {\n\t// stop the running TreeKeeper\n\tif f.stop(q, mr); !mr.IsOK() {\n\t\treturn\n\t}\n\tmr.Code = 0 // reset result code\n\n\t// load the tree again\n\tif err := f.loadSomaTree(q); err != nil {\n\t\tmr.ServerError(err)\n\t\treturn\n\t}\n\tmr.OK()\n}", "func Restart() {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tinternalPanicHandler.Done()\n\tinternalPanicHandler = NewHandler(internalPanicHandler.handle)\n}", "func (cg *CandlesGroup) restart() {\n\ttime.Sleep(5 * time.Second)\n\tif err := cg.wsClient.Exit(); err != nil {\n\t\tlog.Println(\"[BITFINEX] Error destroying connection: \", err)\n\t}\n\tcg.Start(cg.bus.outChannel)\n}", "func (h *Haproxy) restart() error {\n\tcmd := exec.Cmd{\n\t\tPath: h.cmdPath,\n\t\tArgs: append([]string{h.cmdPath, \"-f\", h.confFile, \"-D\", \"-p\", h.pidFile, \"-sf\"}, h.readPids()...),\n\t\tStdout: os.Stdout,\n\t\tStderr: os.Stderr,\n\t}\n\tglog.Infof(\"starting haproxy...\")\n\tif err := cmd.Start(); err != nil {\n\t\treturn fmt.Errorf(\"failed to start haproxy: %v\", err)\n\t}\n\tif err := cmd.Wait(); err != nil {\n\t\treturn fmt.Errorf(\"failed to wait haproxy: %v\", err)\n\t}\n\tglog.Infof(\"haproxy restarted\")\n\treturn nil\n}", "func restartCons() {\n\tfor _, inst := range getInstances() {\n\t\tif inst.Running {\n\t\t\tgo startRecordedWebConsole(inst.Instance)\n\t\t}\n\t}\n}", "func TriggerAppRestart(appName string) error {\n\treturn Restart(appName)\n}", "func (Tests) Restart(ctx context.Context) {\n\tmg.SerialCtxDeps(ctx,\n\t\tTests.Stop,\n\t\tTests.Start,\n\t)\n}", "func (a *App) preRestart(w io.Writer) error {\n\tif err := a.loadHooks(); err != nil {\n\t\treturn err\n\t}\n\treturn a.runHook(w, a.hooks.PreRestart, \"pre-restart\")\n}", "func (m *Machine) Restart() error {\n\tm.State = driver.Running\n\tfmt.Printf(\"Restart %s: %s\\n\", m.Name, m.State)\n\treturn nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
IsInputNode returns whether Node is InputNode
func (inNode *InputNode) IsInputNode() bool { return true }
[ "func (w *Wire) IsInput() bool {\n\treturn w.input == nil\n}", "func (neuron *Neuron) HasInput(e NeuronIndex) bool {\n\tfor _, ni := range neuron.InputNodes {\n\t\tif ni == e {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (r *RunRequest) IsInput() bool {\n\treturn true\n}", "func (Process) IsNode() {}", "func (Node) Is(typ string) bool { return boolResult }", "func IsInputType(ttype Type) bool {\n\tswitch GetNamed(ttype).(type) {\n\tcase *Scalar, *Enum, *InputObject:\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func (Feature) IsNode() {}", "func IsNode(elmt Element) bool {\n\tif elmt == nil {\n\t\treturn false\n\t}\n\treturn elmt.GetElementType() == ElementTypeNode\n}", "func (Project) IsNode() {}", "func inputSelector(n *Node) bool {\n\treturn n.Type() == ElementNode && (n.Name() == \"input\" || n.Name() == \"select\" || n.Name() == \"textarea\" || n.Name() == \"button\")\n}", "func NewInputNode(inStream msgstream.MsgStream, nodeName string, maxQueueLength int32, maxParallelism int32, role string, nodeID int64, collectionID int64, dataType string) *InputNode {\n\tbaseNode := BaseNode{}\n\tbaseNode.SetMaxQueueLength(maxQueueLength)\n\tbaseNode.SetMaxParallelism(maxParallelism)\n\n\treturn &InputNode{\n\t\tBaseNode: baseNode,\n\t\tinStream: inStream,\n\t\tname: nodeName,\n\t\trole: role,\n\t\tnodeID: nodeID,\n\t\tcollectionID: collectionID,\n\t\tdataType: dataType,\n\t}\n}", "func (g *GPIOControllerPCF8574T) IsInput(index int) bool {\n\tif index < 0 || index > pinCount {\n\t\tfmt.Printf(\"Input out of range for gpio: %d\", index)\n\t\treturn false\n\t}\n\n\treturn false\n}", "func IsOperationInputAPropertyDefinition(ctx context.Context, deploymentID, nodeTemplateImpl, typeNameImpl, operationName, inputName string) (bool, error) {\n\tvar typeOrNodeTemplate string\n\tif nodeTemplateImpl == \"\" {\n\t\ttypeOrNodeTemplate = typeNameImpl\n\t} else {\n\t\ttypeOrNodeTemplate = nodeTemplateImpl\n\t}\n\toperationDef, interfaceDef, err := getOperationAndInterfaceDefinitions(ctx, deploymentID, nodeTemplateImpl, typeNameImpl, operationName)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif operationDef != nil {\n\t\tinput, is := operationDef.Inputs[inputName]\n\t\tif is && &input != nil {\n\t\t\treturn input.PropDef != nil, nil\n\t\t}\n\t}\n\n\tif interfaceDef != nil {\n\t\tinput, is := interfaceDef.Inputs[inputName]\n\t\tif is && &input != nil {\n\t\t\treturn input.PropDef != nil, nil\n\t\t}\n\t}\n\treturn false, errors.Errorf(\"failed to find input with name:%q for operation:%q and node template/type:%q\", inputName, operationName, typeOrNodeTemplate)\n}", "func (r Role) IsNode() bool {\n\treturn r == RoleNode || r == RoleAll\n}", "func IsNodeIn(fi *v1alpha1.FileIntegrity, nodeName string, annotation string) bool {\n\tif fi.Annotations == nil {\n\t\treturn false\n\t}\n\tif nodeList, has := fi.Annotations[annotation]; has {\n\t\t// If the annotation is empty, we assume all nodes are in reinit\n\t\tif nodeList == \"\" {\n\t\t\treturn true\n\t\t}\n\t\tfor _, node := range strings.Split(nodeList, \",\") {\n\t\t\tif node == nodeName {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func isAcceptingInput(oldFn reflect.Value) bool {\n\treturn oldFn.Type().NumIn() == 1\n}", "func (MatchedNode) Is(typ string) bool { return boolResult }", "func (this *Node) IsMe(nodeName string) bool {\n\treturn this.NodeInfo.Name == nodeName\n}", "func (n *Network) AddInputNode(node *NNode) {\n\tn.Inputs = append(n.Inputs, node)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
InStream returns the internal MsgStream
func (inNode *InputNode) InStream() msgstream.MsgStream { return inNode.inStream }
[ "func NewInMessage() *InMessage {\n\treturn &InMessage{\n\t\tstorage: make([]byte, bufSize),\n\t}\n}", "func (gi *Invoker) StreamRecv(param *common.Params) error {\n\treturn nil\n}", "func (a *accountStream) OldStream() io.Reader {\n\treturn a.in\n}", "func (c *requestInterceptedClient) GetStream() rpcc.Stream { return c.Stream }", "func (acc *Account) OldStream() io.Reader {\n\tacc.mu.Lock()\n\tdefer acc.mu.Unlock()\n\treturn acc.in\n}", "func (c *authRequiredClient) GetStream() rpcc.Stream { return c.Stream }", "func (ws *WrappedStream) Stream() net.Stream {\n\treturn ws.stream\n}", "func (gi *Invoker) StreamRecv(param *common.Params) error {\n\t//gloryPkg := newGloryRequestPackage(\"\", param.MethodName, uint64(common.StreamSendPkg), param.Seq)\n\t//gloryPkg.Params = append(gloryPkg.Params, param.Value)\n\t//gloryPkg.Header.ChanOffset = param.ChanOffset\n\t//gloryPkg.Header.Seq = param.Seq\n\t//if err := gloryPkg.sendToConn(gi.gloryConnClient, gi.handler); err != nil {\n\t//\tlog.Error(\"StreamRecv: gloryPkg.sendToConn(gi.conn, gi.handler) err =\", err)\n\t//\treturn GloryErrorConnErr\n\t//}\n\treturn nil\n}", "func (c *downloadWillBeginClient) GetStream() rpcc.Stream { return c.Stream }", "func (as *AerospikeSink) In() chan<- interface{} {\n\treturn as.in\n}", "func (c *webTransportClosedClient) GetStream() rpcc.Stream { return c.Stream }", "func (s *MessengerDiffServerCallStub) RecvStream() interface {\n\tAdvance() bool\n\tValue() []string\n\tErr() error\n} {\n\treturn implMessengerDiffServerCallRecv{s}\n}", "func (s *MessengerPushServerCallStub) RecvStream() interface {\n\tAdvance() bool\n\tValue() []byte\n\tErr() error\n} {\n\treturn implMessengerPushServerCallRecv{s}\n}", "func (fs *Ipfs) GetStream(path string) (io.ReadCloser, error) {\n\tp := ipath.New(path)\n\tunixfs := fs.coreAPI.Unixfs()\n\tnode, err := unixfs.Get(context.Background(), p)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// node should be files.File\n\tfile, ok := node.(files.File)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"path is not a file: '%s'\", path)\n\t}\n\n\treturn file, nil\n}", "func (notifee *Notifee) OpenedStream(network.Network, network.Stream) {}", "func (r *TemplateReader) ReadStream(\n\ton templater.OnDataStream,\n\tstopCh <-chan struct{},\n) error {\n\treturn nil\n}", "func (c *Client) GetInMessage(from string, idx uint64) ([][]byte, error) {\n\tvar blockNum *big.Int\n\tif err := retry.Retry(func(attempt uint) error {\n\t\tvar err error\n\t\tblockNum, err = c.session.GetInMessage(from, idx)\n\t\tif err != nil {\n\t\t\tlogger.Error(\"get in message\", \"err\", err.Error())\n\t\t}\n\t\treturn err\n\t}); err != nil {\n\t\tlogger.Error(\"retry error in GetInMessage\", \"err\", err.Error())\n\t}\n\n\treturn [][]byte{blockNum.Bytes()}, nil\n}", "func InMessage(messengerID, msg, stringBuffer string) (outServerMsg string, err error) {\n\tif msg == \"info\" {\n\t\toutServerMsg, err = controlsystemhome.GetInfoControlSystemHomeInterfaces()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tstr, errr := controlled.GetInfoControlledsString()\n\t\tif errr != nil {\n\t\t\terr = errr\n\t\t\treturn\n\t\t}\n\t\toutServerMsg += \"\\n\" + str\n\t\treturn\n\t}\n\n\toutServerMsg, err = commandrecord.UsedTextCommand(msg, stringBuffer)\n\treturn\n}", "func (h *hijackedIOStreamer) stream(ctx context.Context) error {\n\trestoreInput, err := h.setupInput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to setup input stream: %s\", err)\n\t}\n\n\tdefer restoreInput()\n\n\toutputDone := h.beginOutputStream(restoreInput)\n\tinputDone, detached := h.beginInputStream(restoreInput)\n\n\tselect {\n\tcase err := <-outputDone:\n\t\treturn err\n\tcase <-inputDone:\n\t\t// Input stream has closed.\n\t\tif h.outputStream != nil || h.errorStream != nil {\n\t\t\t// Wait for output to complete streaming.\n\t\t\tselect {\n\t\t\tcase err := <-outputDone:\n\t\t\t\treturn err\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn ctx.Err()\n\t\t\t}\n\t\t}\n\t\treturn nil\n\tcase err := <-detached:\n\t\t// Got a detach key sequence.\n\t\treturn err\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
NewInputNode composes an InputNode with provided MsgStream, name and parameters
func NewInputNode(inStream msgstream.MsgStream, nodeName string, maxQueueLength int32, maxParallelism int32, role string, nodeID int64, collectionID int64, dataType string) *InputNode { baseNode := BaseNode{} baseNode.SetMaxQueueLength(maxQueueLength) baseNode.SetMaxParallelism(maxParallelism) return &InputNode{ BaseNode: baseNode, inStream: inStream, name: nodeName, role: role, nodeID: nodeID, collectionID: collectionID, dataType: dataType, } }
[ "func NewInput(Name string, Type string, Repr msgs.Representation, Chan string, Default msgs.Message) Input {\n\n\t// Validates if the message-type is registered\n\tif !msgs.IsMessageTypeRegistered(Type) {\n\t\terrorString := fmt.Sprintf(\"The '%s' message type has not been registered!\", Type)\n\t\tpanic(errorString)\n\t}\n\n\t// Validates if the representation format is supported\n\tif !msgs.DoesMessageTypeImplementsRepresentation(Type, Repr) {\n\t\terrorString := fmt.Sprintf(\"'%s' message-type does not implement codec for '%s' representation format\", Type, Repr)\n\t\tpanic(errorString)\n\t}\n\n\treturn Input{IO: IO{Name: Name, Type: Type, Representation: Repr, Channel: Chan, Message: Default}, DefaultMessage: Default}\n}", "func NewInput(\n\tcfg *common.Config,\n\toutlet channel.Connector,\n\tcontext input.Context,\n) (input.Input, error) {\n\n\tout, err := outlet(cfg, context.DynamicFields)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tforwarder := harvester.NewForwarder(out)\n\n\tconfig := defaultConfig\n\terr = cfg.Unpack(&config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcb := func(data []byte, metadata inputsource.NetworkMetadata) {\n\t\tevent := createEvent(data, metadata)\n\t\tforwarder.Send(event)\n\t}\n\n\tsplitFunc := tcp.SplitFunc([]byte(config.LineDelimiter))\n\tif splitFunc == nil {\n\t\treturn nil, fmt.Errorf(\"unable to create splitFunc for delimiter %s\", config.LineDelimiter)\n\t}\n\n\tserver, err := tcp.New(&config.Config, splitFunc, cb)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Input{\n\t\tserver: server,\n\t\tstarted: false,\n\t\toutlet: out,\n\t\tconfig: &config,\n\t\tlog: logp.NewLogger(\"tcp input\").With(\"address\", config.Config.Host),\n\t}, nil\n}", "func NewInput(input *Synapse) *Neuron {\n\treturn &Neuron{\n\t\tInputs: []*Synapse{input},\n\t\tOutputs: []*Synapse{},\n\t\tFunction: func(inputs, outputs []*Synapse) {\n\t\t\tfor _, s := range outputs {\n\t\t\t\ts.Value = inputs[0].Value\n\t\t\t}\n\t\t},\n\t}\n}", "func NewInput(client *hdfs.Client, path string, excludes []string) (*Input, error) {\n\tsources, err := getInfo(client, path, excludes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Input{\n\t\tpath: path,\n\t\tclient: client,\n\t\tsources: sources,\n\t\tindex: 0,\n\t}, nil\n}", "func NewInput(r io.Reader) Input {\r\n\treturn Input{\r\n\t\treader: bufio.NewReader(r),\r\n\t}\r\n}", "func (r *ReactorGraph) CreateInput(value int) InputCell {\n\treturn &Node{value: value, dependencies: make([]*Node, 0)}\n}", "func (m *Master) NewInput(a *NewInputArgs, r *int) error {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\n\ts := m.slaves[a.ID]\n\tif s == nil {\n\t\treturn errors.New(\"unknown slave\")\n\t}\n\n\tart := Artifact{a.Data, a.Prio, false}\n\tif !m.corpus.add(art) {\n\t\treturn nil\n\t}\n\tm.lastInput = time.Now()\n\t// Queue the input for sending to every slave.\n\tfor _, s1 := range m.slaves {\n\t\ts1.pending = append(s1.pending, MasterInput{a.Data, a.Prio, execCorpus, true, s1 != s})\n\t}\n\n\treturn nil\n}", "func (w *Watcher) NewInput(p ProducerFunc) *Input {\n\tret := &Input{\n\t\tProducer: p,\n\t\tLogFunc: w.LogFunc,\n\t}\n\tw.Inputs = append(w.Inputs, ret)\n\treturn ret\n}", "func NewInput() inputT {\n\tcntr := ctr.Increment()\n\tt := inputT{\n\t\tName: fmt.Sprintf(\"input_%v\", cntr),\n\t\tType: \"text\",\n\t\tLabel: trl.S{\"en\": fmt.Sprintf(\"Label %v\", cntr), \"de\": fmt.Sprintf(\"Titel %v\", cntr)},\n\t\tDesc: trl.S{\"en\": \"Description\", \"de\": \"Beschreibung\"},\n\t}\n\treturn t\n}", "func NewInput(conn *rpcc.Conn) *Input {\n\treturn &Input{conn: conn}\n}", "func (pub *Publisher) CreateInput(nodeHWID string, inputType types.InputType, instance string,\n\tsetCommandHandler func(input *types.InputDiscoveryMessage, sender string, value string)) *types.InputDiscoveryMessage {\n\tinput := pub.inputFromSetCommands.CreateInput(nodeHWID, inputType, instance, setCommandHandler)\n\treturn input\n}", "func (p *Parser) NewInput(s io.RuneScanner) {\n\tselect {\n\tcase p.AddInput <- s:\n\tcase <-p.reqStop:\n\t}\n}", "func NewInputComponent(parent *Entity) *InputComponent {\n\tinputComponent := &InputComponent{\n\t\tID: \"input\",\n\t\tParent: parent,\n\t}\n\treturn inputComponent\n}", "func NewInput(uri string) (*Input, error) {\n\tdialer := websocket.Dialer{\n\t\tHandshakeTimeout: 10 * time.Second,\n\t\tTLSClientConfig: &tls.Config{\n\t\t\tInsecureSkipVerify: true,\n\t\t},\n\t\tNetDial: (&net.Dialer{\n\t\t\tTimeout: time.Second * 5,\n\t\t}).Dial,\n\t}\n\tws, resp, err := dialer.Dial(uri, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err = resp.Body.Close(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Input{ws: ws}, nil\n}", "func NewInput(r io.Reader) *Input {\n\tvar b []byte\n\tif r != nil {\n\t\tif buffer, ok := r.(interface {\n\t\t\tBytes() []byte\n\t\t}); ok {\n\t\t\tb = buffer.Bytes()\n\t\t} else {\n\t\t\tvar err error\n\t\t\tb, err = ioutil.ReadAll(r)\n\t\t\tif err != nil {\n\t\t\t\treturn &Input{\n\t\t\t\t\tbuf: nullBuffer,\n\t\t\t\t\terr: err,\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn NewInputBytes(b)\n}", "func readInput(r io.Reader) Node {\n\tdata, err := ioutil.ReadAll(r)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdata = bytes.Trim(data, \"^$ \\n\") // remove extraneous symbols\n\tnode, i := parseSequence(data, 0)\n\tif i < len(data) {\n\t\tpanic(fmt.Sprintf(\"parse error at offset %d\", i))\n\t}\n\treturn node\n}", "func NewIncoming(x, pred interface{}) (*ast.Incoming, error) {\n\txx, err := NewValue(&ast.TypeDummy{}, x)\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\tpp, err := NewValue(&ast.TypeDummy{}, pred)\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\tp, ok := pp.(ast.NamedValue)\n\tif !ok {\n\t\treturn nil, errors.Errorf(\"invalid predecessor type; expected ast.NamedValue, got %T\", pp)\n\t}\n\treturn &ast.Incoming{X: xx, Pred: p}, nil\n}", "func NewInput() *BeegoInput {\n\treturn &BeegoInput{\n\t\tpnames: make([]string, 0, maxParam),\n\t\tpvalues: make([]string, 0, maxParam),\n\t\tdata: make(map[interface{}]interface{}),\n\t}\n}", "func (factory *IdentityInputFactory) NewInput(hash string, mgr *InputManager) Input {\n\tif factory.input.Hash() != hash {\n\t\tpanic(fmt.Errorf(\n\t\t\t\"Invalid hash for IdentityInputFactory. Expected %s got %s\",\n\t\t\tfactory.input.Hash(),\n\t\t\thash,\n\t\t))\n\t}\n\treturn factory.input\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
ODataSQLQuery builds a SQL like query based on OData 2.0 specification
func ODataSQLQuery(query url.Values, table string, column string, db *sql.DB) (*sql.Rows, error) { // Parse url values queryMap, err := parser.ParseURLValues(query) if err != nil { return nil, errors.Wrap(ErrInvalidInput, err.Error()) } var finalQuery strings.Builder // SELECT clause finalQuery.WriteString(buildSelectClause(queryMap, column)) // FROM clause finalQuery.WriteString(" FROM ") finalQuery.WriteString(pq.QuoteIdentifier(table)) // WHERE clause if queryMap[parser.Filter] != nil { finalQuery.WriteString(" WHERE ") filterQuery, _ := queryMap[parser.Filter].(*parser.ParseNode) filterClause, err := applyFilter(filterQuery, column) if err != nil { return nil, errors.Wrap(ErrInvalidInput, err.Error()) } finalQuery.WriteString(filterClause) } // Order by if queryMap[parser.OrderBy] != nil { finalQuery.WriteString(buildOrderBy(queryMap, column)) } // Limit & Offset finalQuery.WriteString(buildLimitSkipClause(queryMap)) rows, err := db.Query(finalQuery.String()) if err != nil { return nil, err } return rows, nil }
[ "func (qb *QueryBuilder) CreateQuery(e1 entities.Entity, e2 entities.Entity, id interface{}, qo *odata.QueryOptions) (string, *QueryParseInfo) {\n\tet1 := e1.GetEntityType()\n\tet2 := e1.GetEntityType()\n\tif e2 != nil { // 2nd entity is given, this means get e1 by e2\n\t\tet2 = e2.GetEntityType()\n\t}\n\n\teo := &odata.ExpandOperation{\n\t\tQueryOptions: qo,\n\t}\n\tqpi := &QueryParseInfo{}\n\tqpi.Init(et1, 0, nil, eo)\n\n\tif qo != nil && !qo.QueryExpand.IsNil() {\n\t\tqpi.SubEntities = make([]*QueryParseInfo, 0)\n\t\tif len(qo.QueryExpand.Operations) > 0 {\n\t\t\tqb.constructQueryParseInfo(qo.QueryExpand.Operations, qpi, qpi)\n\t\t}\n\t}\n\n\tqueryString := fmt.Sprintf(\"SELECT %s FROM %s %s\", qb.getSelect(e1, qo, qpi, true, true, false, false, \"\"), qb.tables[et1], qb.createJoin(e1, e2, false, qo, qpi, \"\"))\n\tif id != nil {\n\t\tif e2 == nil {\n\t\t\tqueryString = fmt.Sprintf(\"%s WHERE %s = %v\", queryString, selectMappings[et2][idField], id)\n\t\t} else {\n\t\t\tqueryString = fmt.Sprintf(\"%s WHERE %s.%s = %v\", queryString, tableMappings[et2], asMappings[et2][idField], id)\n\t\t}\n\t}\n\n\tif qo != nil && !qo.QueryFilter.IsNil() {\n\t\tif id != nil {\n\t\t\tqueryString = fmt.Sprintf(\"%s AND %s\", queryString, qb.getFilterQueryString(et1, qo, false))\n\t\t} else {\n\t\t\tqueryString = fmt.Sprintf(\"%s %s\", queryString, qb.getFilterQueryString(et1, qo, true))\n\t\t}\n\t}\n\n\tqueryString = fmt.Sprintf(\"%s ORDER BY %s\", queryString, qb.getOrderBy(et1, qo))\n\tqueryString = fmt.Sprintf(\"%s LIMIT %s OFFSET %s\", queryString, qb.getLimit(qo), qb.getOffset(qo))\n\n\tfmt.Printf(\"%s\\n\", queryString)\n\treturn queryString, qpi\n}", "func buildQuery(op ops.Operator, qb *queryBuilder) error {\n\tswitch op := op.(type) {\n\tcase *Table:\n\t\tbuildTable(op, qb)\n\tcase *Projection:\n\t\treturn buildProjection(op, qb)\n\tcase *ApplyJoin:\n\t\treturn buildApplyJoin(op, qb)\n\tcase *Filter:\n\t\treturn buildFilter(op, qb)\n\tcase *Horizon:\n\t\tif op.TableId != nil {\n\t\t\treturn buildDerived(op, qb)\n\t\t}\n\t\treturn buildHorizon(op, qb)\n\tcase *Limit:\n\t\treturn buildLimit(op, qb)\n\tcase *Ordering:\n\t\treturn buildOrdering(op, qb)\n\tcase *Aggregator:\n\t\treturn buildAggregation(op, qb)\n\tdefault:\n\t\treturn vterrors.VT13001(fmt.Sprintf(\"do not know how to turn %T into SQL\", op))\n\t}\n\treturn nil\n}", "func generateDTWhereQuery(dtFields []dtColumn) string {\n\twhereQuery := fmt.Sprintf(\"%s like ? \", dtFields[0].dbColumnName)\n\n\tfor _, field := range dtFields[1:] {\n\t\twhereQuery += fmt.Sprintf(\"OR %s like ? \", field.dbColumnName)\n\t}\n\n\treturn whereQuery\n}", "func (g *NgGrid) GenerateSql(tx gorp.SqlExecutor, fields []string) error {\n\n\tvar sq bytes.Buffer\n\tsq.WriteString(\"select count(*) \")\n\tsq.WriteString(\"from ( %s ) as t\")\n\tquery := fmt.Sprintf(sq.String(), g.MainQuery)\n\n\ttotalCount, err := tx.SelectInt(query)\n\tg.TotalCount = totalCount\n\tg.FilterCount = g.TotalCount\n\n\tif g.TotalCount == 0 {\n\t\treturn err\n\t}\n\n\tsq.Reset()\n\tif fields == nil {\n\t\tsq.WriteString(\"select * \")\n\t} else {\n\t\tsq.WriteString(fmt.Sprintf(\"select %s \", strings.Join(fields, \",\")))\n\t}\n\n\tsq.WriteString(\"from \")\n\tsq.WriteString(\"(select row_number() over(order by %s %s) as rownum, \")\n\tsq.WriteString(\"t.* \")\n\tsq.WriteString(\"from \")\n\tsq.WriteString(\"( %s ) as t) as t \")\n\tsq.WriteString(\"where t.rownum between %d and %d \")\n\n\tif g.SortField == \"\" {\n\t\treturn errors.New(\"Query sortField parameter is missing\")\n\t}\n\n\tsortDirection := \"asc\"\n\tif strings.ToLower(g.SortDirection) == \"desc\" {\n\t\tsortDirection = \"desc\"\n\t}\n\tg.SortDirection = sortDirection\n\n\t//if the sort field refers to an inner object e.g. bomItem.line.name, then we should consider only the latest part\n\tsf := strings.Split(g.SortField, \".\")\n\tg.SortField = sf[len(sf)-1]\n\n\tg.GeneratedQuery = fmt.Sprintf(sq.String(), strings.ToLower(g.SortField), g.SortDirection, g.MainQuery, g.FromRow(), g.ToRow())\n\n\tglog.V(4).Infoln(g.GeneratedQuery)\n\n\treturn err\n}", "func (rawQuery *SearchRawQuery) ToSQLQuery(namespace string) *SearchSQLQuery {\n\tvar q string\n\tvar args []interface{}\n\n\tswitch namespace {\n\tcase SearchNamespaceAccounts:\n\t\tq = \"SELECT id, balance, data FROM current_balances\"\n\tcase SearchNamespaceTransactions:\n\t\tq = `SELECT id, timestamp, data,\n\t\t\t\t\tarray_to_json(ARRAY(\n\t\t\t\t\t\tSELECT lines.account_id FROM lines\n\t\t\t\t\t\t\tWHERE transaction_id=transactions.id\n\t\t\t\t\t\t\tORDER BY lines.account_id\n\t\t\t\t\t)) AS account_array,\n\t\t\t\t\tarray_to_json(ARRAY(\n\t\t\t\t\t\tSELECT lines.delta FROM lines\n\t\t\t\t\t\t\tWHERE transaction_id=transactions.id\n\t\t\t\t\t\t\tORDER BY lines.account_id\n\t\t\t\t\t)) AS delta_array\n\t\t\tFROM transactions`\n\tdefault:\n\t\treturn nil\n\t}\n\n\t// Process must queries\n\tvar mustWhere []string\n\tmustClause := rawQuery.Query.MustClause\n\tfieldsWhere, fieldsArgs := convertFieldsToSQL(mustClause.Fields)\n\tmustWhere = append(mustWhere, fieldsWhere...)\n\targs = append(args, fieldsArgs...)\n\n\ttermsWhere, termsArgs := convertTermsToSQL(mustClause.Terms)\n\tmustWhere = append(mustWhere, termsWhere...)\n\targs = append(args, termsArgs...)\n\n\trangesWhere, rangesArgs := convertRangesToSQL(mustClause.RangeItems)\n\tmustWhere = append(mustWhere, rangesWhere...)\n\targs = append(args, rangesArgs...)\n\n\t// Process should queries\n\tvar shouldWhere []string\n\tshouldClause := rawQuery.Query.ShouldClause\n\tfieldsWhere, fieldsArgs = convertFieldsToSQL(shouldClause.Fields)\n\tshouldWhere = append(shouldWhere, fieldsWhere...)\n\targs = append(args, fieldsArgs...)\n\n\ttermsWhere, termsArgs = convertTermsToSQL(shouldClause.Terms)\n\tshouldWhere = append(shouldWhere, termsWhere...)\n\targs = append(args, termsArgs...)\n\n\trangesWhere, rangesArgs = convertRangesToSQL(shouldClause.RangeItems)\n\tshouldWhere = append(shouldWhere, rangesWhere...)\n\targs = append(args, rangesArgs...)\n\n\tvar offset = rawQuery.Offset\n\tvar limit = rawQuery.Limit\n\n\tif len(mustWhere) == 0 && len(shouldWhere) == 0 {\n\t\treturn &SearchSQLQuery{sql: q, args: args}\n\t}\n\n\tq += \" WHERE \"\n\tif len(mustWhere) != 0 {\n\t\tq += \"(\" + strings.Join(mustWhere, \" AND \") + \")\"\n\t\tif len(shouldWhere) != 0 {\n\t\t\tq += \" AND \"\n\t\t}\n\t}\n\n\tif len(shouldWhere) != 0 {\n\t\tq += \"(\" + strings.Join(shouldWhere, \" OR \") + \")\"\n\t}\n\n\tif namespace == SearchNamespaceTransactions {\n\t\tif rawQuery.SortTime == SortDescByTime {\n\t\t\tq += \" ORDER BY timestamp DESC\"\n\t\t} else {\n\t\t\tq += \" ORDER BY timestamp\"\n\t\t}\n\t}\n\n\tif offset > 0 {\n\t\tq += \" OFFSET \" + strconv.Itoa(offset) + \" \"\n\t}\n\tif limit > 0 {\n\t\tq += \" LIMIT \" + strconv.Itoa(limit)\n\t}\n\n\tq = enumerateSQLPlacholder(q)\n\treturn &SearchSQLQuery{sql: q, args: args}\n}", "func buildDatastoreQuery(kind string, prefix string) *datastore.Query {\n\t\n\tvar q *datastore.Query\n\n\tif prefix != \"\" {\n\t\tkeystart := datastore.NameKey(kind, prefix, nil)\n\t\tkeyendstr := prefix+\"\\ufffd\"\n\t\tkeyend := datastore.NameKey(kind, keyendstr, nil)\n\t\tq = datastore.NewQuery(kind).Filter(\"__key__ >=\", keystart).\n\t\tFilter(\"__key__ <\", keyend)\n\t} else {\n\t\tq = datastore.NewQuery(kind)\n\t}\n\treturn q\n}", "func ConcatQuery(c *Context, values ...interface{}) string {\n\ts := strings.Builder{}\n\n\tfor _, val := range values {\n\t\tswitch v := val.(type) {\n\t\tcase (Field):\n\t\t\ts.WriteString(v.QueryString(c))\n\t\tcase (Condition):\n\t\t\ts.WriteString(v(c))\n\t\tcase (SelectQuery):\n\t\t\tsql, _ := v.SQL(SQLBuilder{Context: c})\n\t\t\ts.WriteString(getSubQuerySQL(sql))\n\t\tcase (string):\n\t\t\ts.WriteString(v)\n\t\tdefault:\n\t\t\tpanic(`Can only use strings, Fields, Conditions and SelectQueries to build SQL`)\n\t\t}\n\t}\n\treturn s.String()\n}", "func (q *DataQuerySQL) sql() (s string, e error) {\n\tif q.baseClass == \"\" {\n\t\treturn \"\", errors.New(\"No base class\")\n\t}\n\n\t// columns\n\tsql := \"select \"\n\tif len(q.columns) == 0 {\n\t\tsql += \"* \"\n\t} else {\n\t\tsql += \"\\\"\" + strings.Join(q.columns, \"\\\",\\\"\") + \"\\\" \"\n\t}\n\n\t// Tables. This is basically a join of all tables from base DataObject thru to the table for the class, and all\n\t// tables for subclasses. This will have been precalculated, so it's trivial here.\n\tbaseClass := dbMetadata.GetClass(q.baseClass)\n\tsql += \"from \" + baseClass.defaultFrom\n\n\t// where clause\n\tsql += \" where \" + baseClass.defaultWhere\n\tif len(q.where) > 0 {\n\t\tsql += \" and \" + strings.Join(q.where, \" and \")\n\t}\n\n\tif q.orderBy != \"\" {\n\t\tsql += \" order by \" + q.orderBy\n\t}\n\n\tif q.start >= 0 {\n\t\tsql += \" limit \" + strconv.Itoa(q.start) + \", \" + strconv.Itoa(q.limit)\n\t}\n\t//\tfmt.Printf(\"query is %s\\n\", sql)\n\treturn sql, nil\n}", "func (f NetworkServerFilters) SQL() string {\n\tvar filters []string\n\n\tif f.OrganizationID != 0 {\n\t\tfilters = append(filters, \"sp.organization_id = :organization_id\")\n\t}\n\n\tif len(filters) == 0 {\n\t\treturn \"\"\n\t}\n\n\treturn \"where \" + strings.Join(filters, \" and \")\n}", "func BuildQueryDef(introspect interface{}, ed metadata.EntityDef, cm metadata.ColumnMapper) QueryDef {\n\tselectColumns := strings.Join(cm.Columns(), \", \")\n\n\tqd := sqlQueryDef{\n\t\tselectColumns: selectColumns,\n\t\tfindOne: fmt.Sprintf(findOneByAttributeTemplate, selectColumns, ed.FullTableName(), ed.PKColumn()),\n\t\tfindAll: fmt.Sprintf(findAllTemplate, selectColumns, ed.FullTableName(), ed.DefaultSort()),\n\t\tinsert: generateInsertStatement(ed.Schema(), ed.Table(), cm),\n\t\tupdate: generateUpdateStatement(ed.Schema(), ed.Table(), cm, introspect),\n\t\tdelete: generateDeleteStatement(ed, introspect),\n\t}\n\n\tqueryDefRegistry[ed.Name()] = qd\n\n\treturn qd\n}", "func queryConstructor(params map[string]string) string {\n\tquery := \"SELECT songs.artist, songs.song, genres.name, songs.length FROM songs LEFT OUTER JOIN genres ON songs.genre = genres.ID AND songs.genre LIKE genres.id\"\n\tparameters := parameterFilter(params)\n\n\tif len(parameters) > 0 {\n\t\tquery += \" WHERE \"\n\t}\n\n\tfor column, value := range parameters {\n\t\tquery += parameterConstructor(column, value)\n\t\tdelete(parameters, column)\n\t\tif len(parameters) > 0 {\n\t\t\tquery += \" AND \"\n\t\t}\n\t}\n\n\treturn query\n}", "func (w *Wrapper) buildQuery() {\n\tw.query += w.buildDuplicate()\n\tw.query += w.buildJoin()\n\tw.query += w.buildWhere(\"WHERE\")\n\tw.query += w.buildWhere(\"HAVING\")\n\tw.query += w.buildOrderBy()\n\tw.query += w.buildGroupBy()\n\tw.query += w.buildLimit()\n\n\t_, afterOptions := w.buildQueryOptions()\n\tw.query += afterOptions\n\tw.query = strings.TrimSpace(w.query)\n}", "func (df *DateFilter) BuildQuery(t *time.Time) string {\n\tvar bf bytes.Buffer\n\n\tif df.Attribute == \"\" {\n\t\tdf.defaultAttribute()\n\t}\n\n\tif t == nil {\n\t\tn := time.Now()\n\t\tt = &n\n\t}\n\n\tbf.WriteString(string(df.Attribute))\n\tif df.Custom != \"\" {\n\t\tbf.WriteString(df.Custom)\n\t} else {\n\t\tbf.WriteString(\">=\")\n\t\tbf.WriteString(df.getDateAPIFormat(t))\n\t}\n\n\treturn bf.String()\n}", "func (a *AuditSrv) buildSearchWhereClause(searchParms *globalUtils.AuditSearchParams) (string, []interface{}, error) {\n\tsqlWhereClause := \" where 1=1\"\n\tvar values []interface{}\n\n\ti := 1\n\tif searchParms.ObjectName != \"\" {\n\t\tsqlWhereClause += fmt.Sprintf(\" AND audit.objectname = $%d\", i)\n\t\tvalues = append(values, searchParms.ObjectName)\n\t\ti++\n\t}\n\tif searchParms.ObjectId != \"\" {\n\t\tsqlWhereClause += fmt.Sprintf(\" AND audit.objectid = $%d\", i)\n\t\tvalues = append(values, searchParms.ObjectId)\n\t\ti++\n\t}\n\tif !searchParms.ActionTimeStart.IsZero() {\n\t\tsqlWhereClause += fmt.Sprintf(\" AND audit.actiontime >= $%d\", i)\n\t\tvalues = append(values, searchParms.ActionTimeStart)\n\t\ti++\n\t}\n\tif !searchParms.ActionTimeEnd.IsZero() {\n\t\tsqlWhereClause += fmt.Sprintf(\" AND audit.actiontime <= $%d\", i)\n\t\tvalues = append(values, searchParms.ActionTimeEnd)\n\t\t//i++\n\t}\n\treturn sqlWhereClause, values, nil\n}", "func (m *Mongo) BuildQuery(graph, sub, pred string, obj interface{}, overrides *Overrides) bson.M {\n\tquery := bson.M{\"g\": graph}\n\n\tsEmpty := isEmpty(sub)\n\tpEmpty := isEmpty(pred)\n\toEmpty := isEmpty(obj)\n\n\tswitch {\n\tcase graph == \"\":\n\t\t// all items in collection, never executed\n\tcase sEmpty && pEmpty && oEmpty:\n\t\t// nil nil nil\n\tcase !sEmpty && pEmpty && oEmpty:\n\t\t// sub nil nil\n\t\tquery[\"s\"] = sub\n\tcase !sEmpty && !pEmpty && oEmpty:\n\t\t// sub pred nil\n\t\tquery[\"s\"] = sub\n\t\tquery[\"p\"] = pred\n\tcase !sEmpty && pEmpty && !oEmpty:\n\t\t// sub nil obj\n\t\tquery[\"s\"] = sub\n\t\tquery[\"o\"] = obj\n\tcase !sEmpty && !pEmpty && !oEmpty:\n\t\t// sub pred obj\n\t\tquery[\"s\"] = sub\n\t\tquery[\"p\"] = pred\n\t\tquery[\"o\"] = obj\n\tcase sEmpty && !pEmpty && oEmpty:\n\t\t// nil pred nil\n\t\tquery[\"p\"] = pred\n\tcase sEmpty && pEmpty && !oEmpty:\n\t\t// nil nil obj\n\t\tquery[\"o\"] = obj\n\tcase sEmpty && !pEmpty && !oEmpty:\n\t\t// nil pred obj\n\t\tquery[\"p\"] = pred\n\t\tquery[\"o\"] = obj\n\t}\n\tif overrides != nil {\n\t\tif len(overrides.Subs) > 0 {\n\t\t\tquery[\"s\"] = bson.M{\"$in\": overrides.Subs}\n\t\t}\n\t\tif len(overrides.Preds) > 0 {\n\t\t\tquery[\"p\"] = bson.M{\"$in\": overrides.Preds}\n\t\t}\n\t\tif len(overrides.Objs) > 0 {\n\t\t\tquery[\"o\"] = bson.M{\"$in\": overrides.Objs}\n\t\t}\n\t}\n\treturn query\n}", "func (q VariadicQuery) AppendSQL(buf *strings.Builder, args *[]interface{}) {\n\tif q.Operator == \"\" {\n\t\tq.Operator = QueryUnion\n\t}\n\tswitch len(q.Queries) {\n\tcase 0:\n\t\tbreak\n\tcase 1:\n\t\tq.Queries[0].AppendSQL(buf, args)\n\tdefault:\n\t\tif q.Nested {\n\t\t\tbuf.WriteString(\"(\")\n\t\t}\n\t\tfor i, query := range q.Queries {\n\t\t\tif i > 0 {\n\t\t\t\tbuf.WriteString(\" \")\n\t\t\t\tbuf.WriteString(string(q.Operator))\n\t\t\t\tbuf.WriteString(\" \")\n\t\t\t}\n\t\t\tswitch v := query.(type) {\n\t\t\tcase nil:\n\t\t\t\tbuf.WriteString(\"NULL\")\n\t\t\tcase VariadicQuery:\n\t\t\t\tv.Nested = true\n\t\t\t\tv.AppendSQL(buf, args)\n\t\t\tdefault:\n\t\t\t\tv.AppendSQL(buf, args)\n\t\t\t}\n\t\t}\n\t\tif q.Nested {\n\t\t\tbuf.WriteString(\")\")\n\t\t}\n\t}\n}", "func (w *QueryWrapper) ToQuery() (string, []interface{}) {\n\tclauses := make([]string, 0, w.queryLen+2)\n\tw.binds = make([]interface{}, 0, w.bindsLen)\n\n\tclauses = append(clauses, \"SELECT\")\n\n\tif len(w.distinct) != 0 {\n\t\tclauses = append(clauses, \"DISTINCT\", strings.Join(w.distinct, \", \"))\n\t} else if len(w.columns) != 0 {\n\t\tclauses = append(clauses, strings.Join(w.columns, \", \"))\n\t} else {\n\t\tclauses = append(clauses, \"*\")\n\t}\n\n\tclauses = append(clauses, \"FROM\", w.table)\n\n\tif len(w.joins) != 0 {\n\t\tclauses = append(clauses, w.joins...)\n\t}\n\n\tif w.where != nil {\n\t\tclauses = append(clauses, \"WHERE\", w.where.query)\n\t\tw.binds = append(w.binds, w.where.args...)\n\t}\n\n\tif w.group != \"\" {\n\t\tclauses = append(clauses, \"GROUP BY\", w.group)\n\t}\n\n\tif w.having != nil {\n\t\tclauses = append(clauses, \"HAVING\", w.having.query)\n\t\tw.binds = append(w.binds, w.having.args...)\n\t}\n\n\tif w.order != \"\" {\n\t\tclauses = append(clauses, \"ORDER BY\", w.order)\n\t}\n\n\tif w.offset != 0 {\n\t\tclauses = append(clauses, \"OFFSET\", strconv.Itoa(w.offset))\n\t}\n\n\tif w.limit != 0 {\n\t\tclauses = append(clauses, \"LIMIT\", strconv.Itoa(w.limit))\n\t}\n\n\tquery, binds, err := sqlx.In(strings.Join(clauses, \" \"), w.binds...)\n\n\tif err != nil {\n\t\tlogger.Error(\"yiigo: build 'IN' query error\", zap.Error(err))\n\n\t\treturn \"\", nil\n\t}\n\n\tquery = sqlx.Rebind(sqlx.BindType(string(w.driver)), query)\n\n\tif debug {\n\t\tlogger.Info(query, zap.Any(\"binds\", binds))\n\t}\n\n\treturn query, binds\n}", "func (f ServiceProfileFilters) SQL() string {\n\tvar filters []string\n\n\tif f.UserID != 0 {\n\t\tfilters = append(filters, \"u.id = :user_id\")\n\t}\n\n\tif f.OrganizationID != 0 {\n\t\tfilters = append(filters, \"sp.organization_id = :organization_id\")\n\t}\n\n\tif f.NetworkServerID != 0 {\n\t\tfilters = append(filters, \"sp.network_server_id = :network_server_id\")\n\t}\n\n\tif len(filters) == 0 {\n\t\treturn \"\"\n\t}\n\n\treturn \"where \" + strings.Join(filters, \" and \")\n}", "func (f FUOTADeploymentFilters) SQL() string {\n\tvar filters []string\n\tvar nullDevEUI lorawan.EUI64\n\n\tif f.DevEUI != nullDevEUI {\n\t\tfilters = append(filters, \"fdd.dev_eui = :dev_eui\")\n\t}\n\n\tif f.ApplicationID != 0 {\n\t\tfilters = append(filters, \"d.application_id = :application_id\")\n\t}\n\n\tif len(filters) == 0 {\n\t\treturn \"\"\n\t}\n\n\treturn \"where \" + strings.Join(filters, \" and \")\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
ODataCount returns the number of rows from a table
func ODataCount(db *sql.DB, table string) (int, error) { var count int selectStmt := fmt.Sprintf("SELECT count(*) FROM %s", pq.QuoteIdentifier(table)) row := db.QueryRow(selectStmt) err := row.Scan(&count) if err != nil { return 0, err } return count, nil }
[ "func (db *Database) Count(obj interface{}) (count int64, err error) {\n\tcount, err = gosql.Model(obj).Count()\n\treturn count, err\n}", "func (dp *ArangodbDataProvider) Count(collectionName string, filter string) (int, error) {\n\n\tresponse, responseError := dp.execute(collectionName, http.MethodGet, nil, map[string]string{\"filter\": filter}, []string{\"count\"})\n\n\tif responseError != nil {\n\t\tlogger.Errorf(dp.Context, \"Arangodb Data Provider\", \"execution error %s\", responseError.Error())\n\t\treturn 0, responseError\n\t}\n\n\treturn util.HTTPBodyAsInt64(response.Body), nil\n\n}", "func (s *RepositoryService) Count(rs app.RequestScope) (int64, error) {\n\treturn s.dao.Count(rs.DB())\n}", "func (ds *DatastoreModel) Count(filter string) (int, error) {\n\treturn ds.DataProvider.Count(ds.Kind, filter)\n}", "func (q serviceQuery) Count(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to count service rows\")\n\t}\n\n\treturn count, nil\n}", "func (s *NewsService) Count(rs app.RequestScope) (int, error) {\n\treturn s.dao.Count(rs)\n}", "func (q automodRuleDatumQuery) Count(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to count automod_rule_data rows\")\n\t}\n\n\treturn count, nil\n}", "func Count(db *sql.DB, table string) int {\n\tvar count int\n\tq := fmt.Sprintf(`SELECT COUNT(*) FROM %s`, pq.QuoteIdentifier(table))\n\terr := db.QueryRow(q).Scan(&count)\n\tbhlindex.Check(err)\n\treturn count\n}", "func (q oauthClientQuery) Count(exec boil.Executor) (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRow(exec).Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to count oauth_clients rows\")\n\t}\n\n\treturn count, nil\n}", "func Count(i interface{}) (int, error) {\r\n\ts, err := GetSession()\r\n\tif err != nil {\r\n\t\treturn 0, err\r\n\t}\r\n\tdefer s.Close()\r\n\r\n\tcoll := GetColl(s, typeName(i))\r\n\r\n\treturn coll.Count()\r\n}", "func (q kvstoreQuery) Count(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to count kvstore rows\")\n\t}\n\n\treturn count, nil\n}", "func (q stockcollectionpropQuery) Count() (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"chado: failed to count stockcollectionprop rows\")\n\t}\n\n\treturn count, nil\n}", "func Count(collection string, query interface{}) (int, error) {\n\n\tsession, db, err := GetGlobalSessionFactory().GetSession()\n\tif err != nil {\n\t\tgrip.Errorf(\"error establishing db connection: %+v\", err)\n\n\t\treturn 0, err\n\t}\n\tdefer session.Close()\n\n\treturn db.C(collection).Find(query).Count()\n}", "func (q shelfQuery) Count() (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to count shelf rows\")\n\t}\n\n\treturn count, nil\n}", "func (q utxoQuery) Count(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to count utxo rows\")\n\t}\n\n\treturn count, nil\n}", "func (r repository) Count(ctx context.Context) (int64, error) {\n\tvar count int64\n\terr := r.db.With(ctx).Select(\"COUNT(*)\").From(\"deposit\").Row(&count)\n\treturn count, err\n}", "func (q *Query) Count(ctx context.Context) (int, error) {\n\tif q.dq == nil {\n\t\tq.dq = datastore.NewQuery(q.entity)\n\t}\n\treturn q.dq.Count(ctx)\n}", "func (s *Store) Count(key storage.Key) (int, error) {\n\tkeys := util.BytesPrefix([]byte(key.Namespace() + separator))\n\titer := s.db.NewIterator(keys, nil)\n\n\tvar c int\n\tfor iter.Next() {\n\t\tc++\n\t}\n\n\titer.Release()\n\n\treturn c, iter.Error()\n}", "func (q featureQuery) Count() (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"chado: failed to count feature rows\")\n\t}\n\n\treturn count, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
post handleDBGettokenizedcards receive and handle the request from client, access DB, and web
func handleDBPostGettokenizedcards(w http.ResponseWriter, r *http.Request) { defer func() { db.Connection.Close(nil) }() var errorGeneral string var errorGeneralNbr string var requestData modelito.RequestTokenizedCards errorGeneral="" requestData, errorGeneral=obtainPostParmsGettokenizedcards(r,errorGeneral) //logicrequest_post.go ////////////////////////////////////////////////process business rules /// START if errorGeneral=="" { errorGeneral,errorGeneralNbr= ProcessGettokenizedcards(w , requestData) } /// END if errorGeneral!=""{ //send error response if any //prepare an error JSON Response, if any log.Print("CZ STEP Get the ERROR response JSON ready") /// START fieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr) ////////// write the response (ERROR) w.Header().Set("Content-Type", "application/json") w.Write(fieldDataBytesJson) if(err!=nil){ } } }
[ "func handleDBGeneratetokenized(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n var requestData modelito.RequestTokenized\n var errorGeneral string\n var errorGeneralNbr string\n \n errorGeneral=\"\"\n requestData,errorGeneral =obtainParmsGeneratetokenized(r,errorGeneral)\n\n\n\t////////////////////////////////////////////////validate parms\n\t/// START\n \n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= ProcessGeneratetokenized(w , requestData)\n\t}\n\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func v4handleDBPostProcesspayment(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n var errorGeneral string\n var errorGeneralNbr string\n var requestData modelito.RequestPayment\n \n errorGeneral=\"\"\nrequestData,errorGeneral =obtainPostParmsProcessPayment(r,errorGeneral) //logicrequest_post.go\n\n\t////////////////////////////////////////////////validate parms\n\t/// START\n\t////////////////////////////////////////////////validate parms\n\t/// START\n \n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= v4ProcessProcessPayment(w , requestData) //logicbusiness.go \n\t}\n\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func HandleGetDatabaseConnectionState(adminMan *admin.Manager, modules *modules.Modules) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\t// Get the JWT token from header\n\t\ttoken := utils.GetTokenFromHeader(r)\n\n\t\tdefer utils.CloseTheCloser(r.Body)\n\n\t\t// Check if the request is authorised\n\t\tif err := adminMan.IsTokenValid(token); err != nil {\n\t\t\t_ = utils.SendErrorResponse(w, http.StatusUnauthorized, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\t// Create a context of execution\n\t\tctx, cancel := context.WithTimeout(r.Context(), 60*time.Second)\n\t\tdefer cancel()\n\n\t\tvars := mux.Vars(r)\n\t\tdbAlias := vars[\"dbAlias\"]\n\n\t\tcrud := modules.DB()\n\t\tconnState := crud.GetConnectionState(ctx, dbAlias)\n\n\t\t_ = utils.SendResponse(w, http.StatusOK, model.Response{Result: connState})\n\t}\n}", "func v4handleDBProcesspayment(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n\n var errorGeneral string\n var\terrorGeneralNbr string\n var requestData modelito.RequestPayment\n errorGeneral=\"\"\nrequestData,errorGeneral =obtainParmsProcessPayment(r,errorGeneral)\n\n\t////////////////////////////////////////////////validate parms\n\t/// START\n \n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= v4ProcessProcessPayment(w , requestData) //logicbusiness.go \n\t}\n\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func DeckCardsGET(db *sqlx.DB, ctx *gin.Context) {\n\n // parse id param\n var deckIDString string = strings.ToLower(ctx.Param(\"id\"))\n\n _deckID, err := strconv.ParseUint(deckIDString, 10, 32)\n if err != nil {\n ctx.JSON(http.StatusBadRequest, gin.H{\n \"status\": http.StatusBadRequest,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"given id is invalid\",\n })\n ctx.Error(err)\n return\n }\n var deckID uint = uint(_deckID)\n\n // parse page query\n var pageQueryString string = ctx.DefaultQuery(\"page\", \"1\")\n _page, err := strconv.ParseUint(pageQueryString, 10, 32)\n if err != nil || _page <= 0 {\n ctx.JSON(http.StatusBadRequest, gin.H{\n \"status\": http.StatusBadRequest,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"given page query param is invalid\",\n })\n ctx.Error(err)\n return\n }\n var page uint = uint(_page)\n\n // parse per_page query\n var perpageQueryString string = ctx.DefaultQuery(\"per_page\", \"25\")\n _per_page, err := strconv.ParseUint(perpageQueryString, 10, 32)\n if err != nil || _per_page <= 0 {\n ctx.JSON(http.StatusBadRequest, gin.H{\n \"status\": http.StatusBadRequest,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"given per_page query param is invalid\",\n })\n ctx.Error(err)\n return\n }\n var per_page uint = uint(_per_page)\n\n // parse sort order\n var orderQueryString string = strings.ToUpper(ctx.DefaultQuery(\"order\", \"DESC\"))\n\n switch {\n case orderQueryString == \"DESC\":\n case orderQueryString == \"ASC\":\n default:\n ctx.JSON(http.StatusBadRequest, gin.H{\n \"status\": http.StatusBadRequest,\n \"developerMessage\": \"invalid order query\",\n \"userMessage\": \"invalid order query\",\n })\n return\n }\n\n // parse sort metric query\n var sortQueryString string = ctx.DefaultQuery(\"sort\", \"reviewed_at\")\n\n var query PipeInput\n switch {\n case sortQueryString == \"created_at\":\n query = FETCH_CARDS_BY_DECK_SORT_CREATED_QUERY(orderQueryString)\n case sortQueryString == \"updated_at\":\n query = FETCH_CARDS_BY_DECK_SORT_UPDATED_QUERY(orderQueryString)\n case sortQueryString == \"title\":\n query = FETCH_CARDS_BY_DECK_SORT_TITLE_QUERY(orderQueryString)\n case sortQueryString == \"reviewed_at\":\n query = FETCH_CARDS_BY_DECK_REVIEWED_DATE_QUERY(orderQueryString)\n case sortQueryString == \"times_reviewed\":\n query = FETCH_CARDS_BY_DECK_TIMES_REVIEWED_QUERY(orderQueryString)\n default:\n ctx.JSON(http.StatusBadRequest, gin.H{\n \"status\": http.StatusBadRequest,\n \"developerMessage\": \"invalid sort query\",\n \"userMessage\": \"invalid sort query\",\n })\n return\n }\n\n // verify deck id exists\n _, err = GetDeck(db, deckID)\n\n switch {\n case err == ErrDeckNoSuchDeck:\n ctx.JSON(http.StatusNotFound, gin.H{\n \"status\": http.StatusNotFound,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"cannot find deck by id\",\n })\n ctx.Error(err)\n return\n case err != nil:\n ctx.JSON(http.StatusInternalServerError, gin.H{\n \"status\": http.StatusInternalServerError,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"unable to retrieve deck\",\n })\n ctx.Error(err)\n return\n }\n\n // fetch cards\n var cards *([]CardRow)\n cards, err = CardsByDeck(db, query, deckID, page, per_page)\n\n switch {\n case err == ErrCardNoCardsByDeck:\n ctx.JSON(http.StatusNotFound, gin.H{\n \"status\": http.StatusNotFound,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"deck has no cards\",\n })\n ctx.Error(err)\n return\n case err == ErrCardPageOutOfBounds:\n ctx.JSON(http.StatusBadRequest, gin.H{\n \"status\": http.StatusBadRequest,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"page is out of bound\",\n })\n ctx.Error(err)\n return\n case err != nil:\n ctx.JSON(http.StatusInternalServerError, gin.H{\n \"status\": http.StatusInternalServerError,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"unable to retrieve cards for deck\",\n })\n ctx.Error(err)\n return\n }\n\n var response []gin.H = make([]gin.H, 0, len(*cards))\n\n for _, cr := range *cards {\n\n // fetch card score\n var fetchedCardScore *CardScoreRow\n fetchedCardScore, err = GetCardScoreRecord(db, cr.ID)\n if err != nil {\n ctx.JSON(http.StatusInternalServerError, gin.H{\n \"status\": http.StatusInternalServerError,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"unable to retrieve card score record\",\n })\n ctx.Error(err)\n return\n }\n\n var fetchedStashes []uint\n fetchedStashes, err = StashesByCard(db, cr.ID)\n if err != nil {\n ctx.JSON(http.StatusInternalServerError, gin.H{\n \"status\": http.StatusInternalServerError,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"unable to retrieve card stashes\",\n })\n ctx.Error(err)\n return\n }\n\n var cardrow gin.H = CardRowToResponse(db, &cr)\n var cardscore gin.H = CardScoreToResponse(fetchedCardScore)\n\n foo := MergeResponses(\n &cardrow,\n &gin.H{\"review\": cardscore},\n &gin.H{\"stashes\": fetchedStashes},\n )\n response = append(response, foo)\n }\n\n ctx.JSON(http.StatusOK, response)\n}", "func Db_access_list(w http.ResponseWriter, r *http.Request) {\n\n///\n/// show d.b. access list inf. on web\n///\n\n process3.Db_access_list(w , r )\n\n}", "func (a *App) handleRequest(handler func(storage.Database, http.ResponseWriter, *http.Request)) func(http.ResponseWriter, *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\thandler(a.DB, w, r)\n\t}\n}", "func HandleGetPreparedQuery(adminMan *admin.Manager, syncMan *syncman.Manager) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\t// Get the JWT token from header\n\t\ttoken := utils.GetTokenFromHeader(r)\n\n\t\t// Check if the request is authorised\n\t\tif err := adminMan.IsTokenValid(token); err != nil {\n\t\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\t_ = json.NewEncoder(w).Encode(map[string]string{\"error\": err.Error()})\n\t\t\treturn\n\t\t}\n\t\tctx, cancel := context.WithTimeout(r.Context(), 10*time.Second)\n\t\tdefer cancel()\n\t\t// get project id and dbType from url\n\t\tvars := mux.Vars(r)\n\t\tprojectID := vars[\"project\"]\n\t\tdbAlias := \"\"\n\t\tdbAliasQuery, exists := r.URL.Query()[\"dbAlias\"]\n\t\tif exists {\n\t\t\tdbAlias = dbAliasQuery[0]\n\t\t}\n\t\tidQuery, exists := r.URL.Query()[\"id\"]\n\t\tid := \"\"\n\t\tif exists {\n\t\t\tid = idQuery[0]\n\t\t}\n\t\tresult, err := syncMan.GetPreparedQuery(ctx, projectID, dbAlias, id)\n\t\tif err != nil {\n\t\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\t_ = json.NewEncoder(w).Encode(map[string]string{\"error\": err.Error()})\n\t\t\treturn\n\t\t}\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.WriteHeader(http.StatusOK)\n\t\t_ = json.NewEncoder(w).Encode(model.Response{Result: result})\n\t}\n}", "func (m *DDLServiceMgr) handleListDropInstanceTokens(w http.ResponseWriter, r *http.Request) {\n\n\tif !m.validateAuth(w, r) {\n\t\tlogging.Errorf(\"DDLServiceMgr::handleListDropInstanceTokens Validation Failure req: %v\", common.GetHTTPReqInfo(r))\n\t\treturn\n\t}\n\n\tif r.Method == \"GET\" {\n\n\t\tlogging.Infof(\"DDLServiceMgr::handleListDropInstanceTokens Processing Request req: %v\", common.GetHTTPReqInfo(r))\n\n\t\tnumRetries := 8\n\t\tdeleteTokens, err := mc.ListAndFetchAllDropInstanceCommandToken(numRetries)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"DDLServiceMgr::handleListDropInstanceTokens Error %v in ListAndFetchAllDropInstanceCommandToken. req: %v\", err, common.GetHTTPReqInfo(r))\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\tw.Write([]byte(err.Error() + \"\\n\"))\n\t\t\treturn\n\t\t}\n\n\t\tlist := &mc.DropInstanceCommandTokenList{}\n\t\tlist.Tokens = make([]mc.DropInstanceCommandToken, 0, len(deleteTokens))\n\n\t\tfor _, token := range deleteTokens {\n\t\t\tlist.Tokens = append(list.Tokens, *token)\n\t\t}\n\n\t\tbuf, err := mc.MarshallDropInstanceCommandTokenList(list)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"DDLServiceMgr::handleListDropInstanceTokens Error %v in MarshallDropInstanceCommandTokenList. req: %v\", err, common.GetHTTPReqInfo(r))\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\tw.Write([]byte(err.Error() + \"\\n\"))\n\t\t\treturn\n\t\t}\n\n\t\tw.WriteHeader(http.StatusOK)\n\t\tw.Write(buf)\n\t}\n}", "func DataRetrievalHandler(reader fcrserver.FCRServerRequestReader, writer fcrserver.FCRServerResponseWriter, request *fcrmessages.FCRReqMsg) error {\n\tlogging.Debug(\"Handle data retrieval\")\n\t// Get core structure\n\tc := core.GetSingleInstance()\n\tc.MsgSigningKeyLock.RLock()\n\tdefer c.MsgSigningKeyLock.RUnlock()\n\n\t// Message decoding\n\tnonce, senderID, offer, accountAddr, voucher, err := fcrmessages.DecodeDataRetrievalRequest(request)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Error in decoding payload: %v\", err.Error())\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\n\t// Verify signature\n\tif request.VerifyByID(senderID) != nil {\n\t\t// Verify by signing key\n\t\tgwInfo := c.PeerMgr.GetGWInfo(senderID)\n\t\tif gwInfo == nil {\n\t\t\t// Not found, try sync once\n\t\t\tgwInfo = c.PeerMgr.SyncGW(senderID)\n\t\t\tif gwInfo == nil {\n\t\t\t\terr = fmt.Errorf(\"Error in obtaining information for gateway %v\", senderID)\n\t\t\t\tlogging.Error(err.Error())\n\t\t\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t\t\t}\n\t\t}\n\t\tif request.Verify(gwInfo.MsgSigningKey, gwInfo.MsgSigningKeyVer) != nil {\n\t\t\t// Try update\n\t\t\tgwInfo = c.PeerMgr.SyncGW(senderID)\n\t\t\tif gwInfo == nil || request.Verify(gwInfo.MsgSigningKey, gwInfo.MsgSigningKeyVer) != nil {\n\t\t\t\terr = fmt.Errorf(\"Error in verifying request from gateway %v: %v\", senderID, err.Error())\n\t\t\t\tlogging.Error(err.Error())\n\t\t\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Check payment\n\trefundVoucher := \"\"\n\treceived, lane, err := c.PaymentMgr.Receive(accountAddr, voucher)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Error in receiving voucher %v:\", err.Error())\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\tif lane != 1 {\n\t\terr = fmt.Errorf(\"Not correct lane received expect 1 got %v:\", lane)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\texpected := big.NewInt(0).Add(c.Settings.SearchPrice, offer.GetPrice())\n\tif received.Cmp(expected) < 0 {\n\t\t// Short payment\n\t\t// Refund money\n\t\tif received.Cmp(c.Settings.SearchPrice) <= 0 {\n\t\t\t// No refund\n\t\t} else {\n\t\t\tvar ierr error\n\t\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\t\tif ierr != nil {\n\t\t\t\t// This should never happen\n\t\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t\t}\n\t\t}\n\t\terr = fmt.Errorf(\"Short payment received, expect %v got %v, refund voucher %v\", expected.String(), received.String(), refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\n\t// Payment is fine, verify offer\n\tif offer.Verify(c.OfferSigningPubKey) != nil {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Fail to verify the offer signature, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Verify offer merkle proof\n\tif offer.VerifyMerkleProof() != nil {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Fail to verify the offer merkle proof, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Verify offer expiry\n\tif offer.HasExpired() {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Offer has expired, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Offer is verified. Respond\n\t// First get the tag\n\ttag := c.OfferMgr.GetTagByCID(offer.GetSubCID())\n\t// Second read the data\n\tdata, err := ioutil.ReadFile(filepath.Join(c.Settings.RetrievalDir, tag))\n\tif err != nil {\n\t\t// Refund money, internal error, refund all\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, received)\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Internal error in finding the content, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Third encoding response\n\tresponse, err := fcrmessages.EncodeDataRetrievalResponse(nonce, tag, data)\n\tif err != nil {\n\t\t// Refund money, internal error, refund all\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, received)\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Internal error in encoding the response, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\tc.OfferMgr.IncrementCIDAccessCount(offer.GetSubCID())\n\n\treturn writer.Write(response, c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n}", "func handleRequests(cfg datastructures.Configuration, mgoClient *mgo.Session, redisClient *redis.Client) {\n\tm := func(ctx *fasthttp.RequestCtx) {\n\t\tif cfg.SSL.Enabled {\n\t\t\tlog.Debug(\"handleRequests | SSL is enabled!\")\n\t\t}\n\t\thttputils.SecureRequest(ctx, cfg.SSL.Enabled)\n\t\tctx.Response.Header.Set(\"AuthentiGo\", \"$v0.2.1\")\n\n\t\t// Avoid to print stats for the expvar handler\n\t\tif strings.Compare(string(ctx.Path()), \"/stats\") != 0 {\n\t\t\tlog.Info(\"\\n|REQUEST --> \", ctx, \" \\n|Headers: \", ctx.Request.Header.String(), \"| Body: \", string(ctx.PostBody()))\n\t\t}\n\n\t\tswitch string(ctx.Path()) {\n\t\tcase \"/middleware\":\n\t\t\tmiddleware(ctx, redisClient)\n\t\tcase \"/benchmark\":\n\t\t\tfastBenchmarkHTTP(ctx) // Benchmark API\n\t\tcase \"/auth/login\":\n\t\t\tAuthLoginWrapper(ctx, mgoClient, redisClient, cfg) // Login functionality [Test purpouse]\n\t\tcase \"/auth/register\":\n\t\t\tAuthRegisterWrapper(ctx, mgoClient, cfg) // Register an user into the DB [Test purpouse]\n\t\tcase \"/auth/delete\":\n\t\t\tDeleteCustomerHTTP(ctx, cfg.Mongo.Users.DB, cfg.Mongo.Users.Collection, redisClient, mgoClient)\n\t\tcase \"/auth/verify\":\n\t\t\tVerifyCookieFromRedisHTTP(ctx, redisClient) // Verify if an user is authorized to use the service\n\t\tcase \"/test/crypt\":\n\t\t\tCryptDataHTTPWrapper(ctx)\n\t\tcase \"/test/decrypt\":\n\t\t\tDecryptDataHTTPWrapper(ctx)\n\t\tcase \"/stats\":\n\t\t\texpvarhandler.ExpvarHandler(ctx)\n\t\tdefault:\n\t\t\t_, err := ctx.WriteString(\"The url \" + string(ctx.URI().RequestURI()) + string(ctx.QueryArgs().QueryString()) + \" does not exist :(\\n\")\n\t\t\tcommonutils.Check(err, \"handleRequests\")\n\t\t\tctx.Response.SetStatusCode(404)\n\t\t\tfastBenchmarkHTTP(ctx)\n\t\t}\n\t}\n\t// ==== GZIP HANDLER ====\n\t// The gzipHandler will serve a compress request only if the client request it with headers (Content-Type: gzip, deflate)\n\tgzipHandler := fasthttp.CompressHandlerLevel(m, fasthttp.CompressBestSpeed) // Compress data before sending (if requested by the client)\n\tlog.Info(\"HandleRequests | Binding services to @[\", cfg.Host, \":\", cfg.Port)\n\n\t// ==== SSL HANDLER + GZIP if requested ====\n\tif cfg.SSL.Enabled {\n\t\thttputils.ListAndServerSSL(cfg.Host, cfg.SSL.Path, cfg.SSL.Cert, cfg.SSL.Key, cfg.Port, gzipHandler)\n\t}\n\t// ==== Simple GZIP HANDLER ====\n\thttputils.ListAndServerGZIP(cfg.Host, cfg.Port, gzipHandler)\n\n\tlog.Trace(\"HandleRequests | STOP\")\n}", "func DeckGET(db *sqlx.DB, ctx *gin.Context) {\n\n var err error\n\n // parse id param\n var deckIDString string = strings.ToLower(ctx.Param(\"id\"))\n\n var fetchedDeckRow *DeckRow\n\n // fetch deck row from the db\n\n if deckIDString == \"root\" {\n fetchedDeckRow, err = GetRootDeck(db)\n\n if err != nil {\n ctx.JSON(http.StatusInternalServerError, gin.H{\n \"status\": http.StatusInternalServerError,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"unable to retrieve root\",\n })\n ctx.Error(err)\n return\n }\n } else {\n _deckID, err := strconv.ParseUint(deckIDString, 10, 32)\n if err != nil {\n ctx.JSON(http.StatusBadRequest, gin.H{\n \"status\": http.StatusBadRequest,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"given id is invalid\",\n })\n ctx.Error(err)\n return\n }\n\n fetchedDeckRow, err = GetDeck(db, uint(_deckID))\n switch {\n case err == ErrDeckNoSuchDeck:\n ctx.JSON(http.StatusNotFound, gin.H{\n \"status\": http.StatusNotFound,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"cannot find deck by id\",\n })\n ctx.Error(err)\n return\n case err != nil:\n ctx.JSON(http.StatusInternalServerError, gin.H{\n \"status\": http.StatusInternalServerError,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"unable to retrieve deck\",\n })\n ctx.Error(err)\n return\n }\n }\n\n // fetch children\n var children []uint\n children, err = GetDeckChildren(db, fetchedDeckRow.ID)\n switch {\n case err == ErrDeckNoChildren:\n children = []uint{}\n case err != nil:\n ctx.JSON(http.StatusInternalServerError, gin.H{\n \"status\": http.StatusInternalServerError,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"unable to retrieve deck children\",\n })\n ctx.Error(err)\n return\n }\n\n // fetch parent\n var parentID uint\n var hasParent bool = true\n parentID, err = GetDeckParent(db, fetchedDeckRow.ID)\n switch {\n case err == ErrDeckHasNoParent:\n parentID = 0\n hasParent = false\n case err != nil:\n ctx.JSON(http.StatusInternalServerError, gin.H{\n \"status\": http.StatusInternalServerError,\n \"developerMessage\": err.Error(),\n \"userMessage\": \"unable to retrieve parent deck\",\n })\n ctx.Error(err)\n return\n }\n\n ctx.JSON(http.StatusOK, DeckResponse(&gin.H{\n \"id\": fetchedDeckRow.ID,\n \"name\": fetchedDeckRow.Name,\n \"description\": fetchedDeckRow.Description,\n \"children\": children,\n \"parent\": parentID,\n \"hasParent\": hasParent,\n }))\n}", "func (h *Handler) serveAuthenticateDBUser(w http.ResponseWriter, r *http.Request) {}", "func dbPostHandler(w http.ResponseWriter, r *http.Request) {\r\n\tdefer r.Body.Close()\r\n\tdec := json.NewDecoder(r.Body)\r\n\tentry := &Entry{}\r\n\tif err := dec.Decode(entry); err != nil {\r\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\r\n\t\treturn\r\n\t}\r\n\tdbLock.Lock()\r\n\tdefer dbLock.Unlock()\r\n\tdb[entry.Key] = entry.Value\r\n\tsendResponse(entry, w)\r\n}", "func (sr *sapmReceiver) handleRequest(req *http.Request) error {\n\tsapm, err := sapmprotocol.ParseTraceV2Request(req)\n\t// errors processing the request should return http.StatusBadRequest\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tctx := sr.obsrecv.StartTracesOp(req.Context())\n\n\ttd, err := jaeger.ProtoToTraces(sapm.Batches)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif sr.config.AccessTokenPassthrough {\n\t\tif accessToken := req.Header.Get(splunk.SFxAccessTokenHeader); accessToken != \"\" {\n\t\t\trSpans := td.ResourceSpans()\n\t\t\tfor i := 0; i < rSpans.Len(); i++ {\n\t\t\t\trSpan := rSpans.At(i)\n\t\t\t\tattrs := rSpan.Resource().Attributes()\n\t\t\t\tattrs.PutStr(splunk.SFxAccessTokenLabel, accessToken)\n\t\t\t}\n\t\t}\n\t}\n\n\t// pass the trace data to the next consumer\n\terr = sr.nextConsumer.ConsumeTraces(ctx, td)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"error passing trace data to next consumer: %w\", err)\n\t}\n\n\tsr.obsrecv.EndTracesOp(ctx, \"protobuf\", td.SpanCount(), err)\n\treturn err\n}", "func TokenizeHandler(request events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// get pan\n\t// tokenize\n\t// store in db\n\t// return token\n\n\treturn events.APIGatewayProxyResponse{\n\t\tBody: \"Tokenize\",\n\t\tStatusCode: 200,\n\t}, nil\n}", "func HandleRequest(query []byte, conn *DatabaseConnection) {\n\tlog.Printf(\"Handling raw query: %s\", query)\n\tlog.Printf(\"Parsing request...\")\n\trequest, err := grammar.ParseRequest(query)\n\tlog.Printf(\"Parsed request\")\n\tvar response grammar.Response\n\n\tif err != nil {\n\t\tlog.Printf(\"Error in request parsing! %s\", err.Error())\n\t\tresponse.Type = grammar.UNKNOWN_TYPE_RESPONSE\n\t\tresponse.Status = grammar.RESP_STATUS_ERR_INVALID_QUERY\n\t\tresponse.Data = err.Error()\n\t\tconn.Write(grammar.GetBufferFromResponse(response))\n\t}\n\n\tswitch request.Type {\n\tcase grammar.AUTH_REQUEST:\n\t\t// AUTH {username} {password}\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_AUTH_REQUEST, false, false)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in AUTH request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\t\tusername := request.RequestData[0]\n\t\tpassword := request.RequestData[1]\n\t\t// bucketname := tokens[2]\n\t\tlog.Printf(\"Client wants to authenticate.<username>:<password> %s:%s\", username, password)\n\n\t\tauthRequest := AuthRequest{Username: username, Password: password, Conn: conn}\n\t\tresponse = processAuthRequest(authRequest)\n\tcase grammar.SET_REQUEST:\n\t\t// SET {key} {value} [ttl] [nooverride]\n\t\trequest.Type = grammar.SET_RESPONSE\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_SET_REQUEST, true, true)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in SET request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tkey := request.RequestData[0]\n\t\tvalue := request.RequestData[1]\n\t\tlog.Printf(\"Setting %s:%s\", key, value)\n\t\tsetRequest := SetRequest{Key: key, Value: value, Conn: conn}\n\t\tresponse = processSetRequest(setRequest)\n\n\tcase grammar.GET_REQUEST:\n\t\t// GET {key}\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_GET_REQUEST, true, true)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in GET request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tkey := request.RequestData[0]\n\t\tlog.Printf(\"Client wants to get key '%s'\", key)\n\t\tgetRequest := GetRequest{Key: key, Conn: conn}\n\t\tresponse = processGetRequest(getRequest)\n\n\tcase grammar.DELETE_REQUEST:\n\t\t// DELETE {key}\n\t\tlog.Println(\"Client wants to delete a bucket/key\")\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_DELETE_REQUEST, true, true)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in DELETE request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\t\t// TODO implement\n\tcase grammar.CREATE_BUCKET_REQUEST:\n\t\tlog.Println(\"Client wants to create a bucket\")\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_CREATE_BUCKET_REQUEST, true, false)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in CREATE bucket request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tbucketName := request.RequestData[0]\n\t\tcreateBucketRequest := CreateBucketRequest{BucketName: bucketName, Conn: conn}\n\n\t\tresponse = processCreateBucketRequest(createBucketRequest)\n\tcase grammar.CREATE_USER_REQUEST:\n\t\tlog.Printf(\"Client wants to create a user\")\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_CREATE_USER_REQUEST, false, false)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in CREATE user request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tusername := request.RequestData[0]\n\t\tpassword := request.RequestData[1]\n\t\tcreateUserRequest := CreateUserRequest{Username: username, Password: password, Conn: conn}\n\n\t\tresponse = processCreateUserRequest(createUserRequest)\n\tcase grammar.USE_REQUEST:\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_USE_REQUEST, true, false)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in USE request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tbucketname := request.RequestData[0]\n\t\tif bucketname == SALTS_BUCKET || bucketname == USERS_BUCKET {\n\t\t\tresponse.Status = grammar.RESP_STATUS_ERR_UNAUTHORIZED\n\t\t\tbreak\n\t\t}\n\n\t\tuseRequest := UseRequest{BucketName: bucketname, Conn: conn}\n\t\tresponse = processUseRequest(useRequest)\n\tdefault:\n\t\tlog.Printf(illegalRequestTemplate, request.Type)\n\t\tresponse.Type = grammar.UNKNOWN_TYPE_RESPONSE\n\t\tresponse.Status = grammar.RESP_STATUS_ERR_UNKNOWN_COMMAND\n\t}\n\tif response.Status != 0 {\n\t\tlog.Printf(\"Error in request. status: %d\", response.Status)\n\t}\n\tconn.Write(grammar.GetBufferFromResponse(response))\n\tlog.Printf(\"Wrote buffer: %s to client\", grammar.GetBufferFromResponse(response))\n\n}", "func CommissaireHandler(w http.ResponseWriter, r *http.Request) {\n\tvalidToken := true //checkJwt(w,r)\n\tif (validToken){\n\t\tvars := mux.Vars(r)\n\t\tcomID := vars[\"commissaireID\"]\n\t\n\t\tw.Header().Set(\"Content-Type\", \"application/json;charset=UTF-8\")\n\t\n\t\tif !IsValidUCIID(comID) {\n\t\t\tw.WriteHeader(http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t\n\t\tcommissaire, err := Models.GetCommissaire(comID, connection())\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusNotFound)\n\t\t} else {\n\t\t\tw.WriteHeader(http.StatusOK)\t\t\t\n\t\t}\n\t\n\t\tif err := json.NewEncoder(w).Encode(commissaire); err != nil {\n\t\t\thttp.Error(w, err.Error(), 500)\n\t\t\treturn\n\t\t}\n\t} else {\n\t\tw.WriteHeader(http.StatusUnauthorized)\n\t}\n}", "func (httpServer *HttpServer) handleListRewardAmount(params interface{}, closeChan <-chan struct{}) (interface{}, *rpcservice.RPCError) {\n\tresult := httpServer.databaseService.ListRewardAmount()\n\treturn result, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
/////////////////////////////v4 /////////////////////////////v4 v4handleDBProcesspayment receive and handle the request from client, access DB
func v4handleDBPostProcesspayment(w http.ResponseWriter, r *http.Request) { defer func() { db.Connection.Close(nil) }() var errorGeneral string var errorGeneralNbr string var requestData modelito.RequestPayment errorGeneral="" requestData,errorGeneral =obtainPostParmsProcessPayment(r,errorGeneral) //logicrequest_post.go ////////////////////////////////////////////////validate parms /// START ////////////////////////////////////////////////validate parms /// START if errorGeneral=="" { errorGeneral,errorGeneralNbr= v4ProcessProcessPayment(w , requestData) //logicbusiness.go } if errorGeneral!=""{ //send error response if any //prepare an error JSON Response, if any log.Print("CZ STEP Get the ERROR response JSON ready") /// START fieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr) ////////// write the response (ERROR) w.Header().Set("Content-Type", "application/json") w.Write(fieldDataBytesJson) if(err!=nil){ } } }
[ "func v4handleDBProcesspayment(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n\n var errorGeneral string\n var\terrorGeneralNbr string\n var requestData modelito.RequestPayment\n errorGeneral=\"\"\nrequestData,errorGeneral =obtainParmsProcessPayment(r,errorGeneral)\n\n\t////////////////////////////////////////////////validate parms\n\t/// START\n \n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= v4ProcessProcessPayment(w , requestData) //logicbusiness.go \n\t}\n\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func handleDBPostGettokenizedcards(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n var errorGeneral string\n var errorGeneralNbr string\n \n \tvar requestData modelito.RequestTokenizedCards\n\n errorGeneral=\"\"\n requestData, errorGeneral=obtainPostParmsGettokenizedcards(r,errorGeneral) //logicrequest_post.go\n\n\t////////////////////////////////////////////////process business rules\n\t/// START\n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= ProcessGettokenizedcards(w , requestData)\n\t}\n\t/// END\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func handleDBGeneratetokenized(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n var requestData modelito.RequestTokenized\n var errorGeneral string\n var errorGeneralNbr string\n \n errorGeneral=\"\"\n requestData,errorGeneral =obtainParmsGeneratetokenized(r,errorGeneral)\n\n\n\t////////////////////////////////////////////////validate parms\n\t/// START\n \n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= ProcessGeneratetokenized(w , requestData)\n\t}\n\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func (ctx *Context) PaymentDB(ros ...dbRequestReadOnly) *sql.DB {\n\tvar ro bool\n\tif len(ros) > 0 {\n\t\tfor _, r := range ros {\n\t\t\tif r {\n\t\t\t\tro = true\n\t\t\t}\n\t\t}\n\t}\n\tif !ro {\n\t\treturn ctx.paymentDBWrite\n\t}\n\tif ctx.paymentDBReadOnly == nil {\n\t\treturn ctx.paymentDBWrite\n\t}\n\treturn ctx.paymentDBReadOnly\n}", "func HandlePayments(client *http.Client, strapiURL string, strapiToken string, db *sql.DB, w http.ResponseWriter, r *http.Request) {\n\tvar customerOrder CustomerOrder\n\tbodyBytes, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = json.Unmarshal(bodyBytes, &customerOrder)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer r.Body.Close()\n\turl := fmt.Sprintf(\"%v/restaurants/payment/%v\", strapiURL, customerOrder.RestaurantID)\n\treq, err := http.NewRequest(http.MethodGet, url, nil)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treq.Header.Set(\"Authorization\", \"Bearer \"+strapiToken)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer resp.Body.Close()\n\tvar strapiResponse []Restaurant\n\n\tbodyBytes, err = ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = json.Unmarshal(bodyBytes, &strapiResponse)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\trestaurant := strapiResponse[0]\n\n\ttotal := calculateTotalPriceInPence(customerOrder, restaurant)\n\tsc := stripeClient.New(restaurant.Credential.TestPrivateKey, nil)\n\n\tparams := &stripe.PaymentIntentParams{\n\t\tAmount: stripe.Int64(total),\n\t\tCurrency: stripe.String(string(stripe.CurrencyGBP)),\n\t\tSetupFutureUsage: stripe.String(string(stripe.PaymentIntentSetupFutureUsageOffSession)),\n\t\tStatementDescriptor: stripe.String(restaurant.Name[:21]),\n\t}\n\tparams.AddMetadata(\"order_id\", customerOrder.UUID)\n\tpi, _ := sc.PaymentIntents.New(params)\n\n\tcustomerResponse := CheckoutData{\n\t\tClientSecret: pi.ClientSecret,\n\t}\n\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(customerResponse)\n\n}", "func DataRetrievalHandler(reader fcrserver.FCRServerRequestReader, writer fcrserver.FCRServerResponseWriter, request *fcrmessages.FCRReqMsg) error {\n\tlogging.Debug(\"Handle data retrieval\")\n\t// Get core structure\n\tc := core.GetSingleInstance()\n\tc.MsgSigningKeyLock.RLock()\n\tdefer c.MsgSigningKeyLock.RUnlock()\n\n\t// Message decoding\n\tnonce, senderID, offer, accountAddr, voucher, err := fcrmessages.DecodeDataRetrievalRequest(request)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Error in decoding payload: %v\", err.Error())\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\n\t// Verify signature\n\tif request.VerifyByID(senderID) != nil {\n\t\t// Verify by signing key\n\t\tgwInfo := c.PeerMgr.GetGWInfo(senderID)\n\t\tif gwInfo == nil {\n\t\t\t// Not found, try sync once\n\t\t\tgwInfo = c.PeerMgr.SyncGW(senderID)\n\t\t\tif gwInfo == nil {\n\t\t\t\terr = fmt.Errorf(\"Error in obtaining information for gateway %v\", senderID)\n\t\t\t\tlogging.Error(err.Error())\n\t\t\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t\t\t}\n\t\t}\n\t\tif request.Verify(gwInfo.MsgSigningKey, gwInfo.MsgSigningKeyVer) != nil {\n\t\t\t// Try update\n\t\t\tgwInfo = c.PeerMgr.SyncGW(senderID)\n\t\t\tif gwInfo == nil || request.Verify(gwInfo.MsgSigningKey, gwInfo.MsgSigningKeyVer) != nil {\n\t\t\t\terr = fmt.Errorf(\"Error in verifying request from gateway %v: %v\", senderID, err.Error())\n\t\t\t\tlogging.Error(err.Error())\n\t\t\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Check payment\n\trefundVoucher := \"\"\n\treceived, lane, err := c.PaymentMgr.Receive(accountAddr, voucher)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Error in receiving voucher %v:\", err.Error())\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\tif lane != 1 {\n\t\terr = fmt.Errorf(\"Not correct lane received expect 1 got %v:\", lane)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\texpected := big.NewInt(0).Add(c.Settings.SearchPrice, offer.GetPrice())\n\tif received.Cmp(expected) < 0 {\n\t\t// Short payment\n\t\t// Refund money\n\t\tif received.Cmp(c.Settings.SearchPrice) <= 0 {\n\t\t\t// No refund\n\t\t} else {\n\t\t\tvar ierr error\n\t\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\t\tif ierr != nil {\n\t\t\t\t// This should never happen\n\t\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t\t}\n\t\t}\n\t\terr = fmt.Errorf(\"Short payment received, expect %v got %v, refund voucher %v\", expected.String(), received.String(), refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\n\t// Payment is fine, verify offer\n\tif offer.Verify(c.OfferSigningPubKey) != nil {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Fail to verify the offer signature, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Verify offer merkle proof\n\tif offer.VerifyMerkleProof() != nil {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Fail to verify the offer merkle proof, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Verify offer expiry\n\tif offer.HasExpired() {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Offer has expired, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Offer is verified. Respond\n\t// First get the tag\n\ttag := c.OfferMgr.GetTagByCID(offer.GetSubCID())\n\t// Second read the data\n\tdata, err := ioutil.ReadFile(filepath.Join(c.Settings.RetrievalDir, tag))\n\tif err != nil {\n\t\t// Refund money, internal error, refund all\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, received)\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Internal error in finding the content, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Third encoding response\n\tresponse, err := fcrmessages.EncodeDataRetrievalResponse(nonce, tag, data)\n\tif err != nil {\n\t\t// Refund money, internal error, refund all\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, received)\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Internal error in encoding the response, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\tc.OfferMgr.IncrementCIDAccessCount(offer.GetSubCID())\n\n\treturn writer.Write(response, c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n}", "func (_BaseContent *BaseContentTransactor) ProcessRequestPayment(opts *bind.TransactOpts, request_ID *big.Int, payee common.Address, label string, amount *big.Int) (*types.Transaction, error) {\n\treturn _BaseContent.contract.Transact(opts, \"processRequestPayment\", request_ID, payee, label, amount)\n}", "func ProcessStripePayment(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"NOT IMPLEMENTED\"})\n}", "func (s *Server) handleDashboardPaymentView() http.HandlerFunc {\n\tvar o sync.Once\n\tvar tpl *template.Template\n\n\t//steps on the page\n\tsteps := struct {\n\t\tStepDel string\n\t\tStepMarkPaid string\n\t}{\n\t\tStepDel: \"stepDel\",\n\t\tStepMarkPaid: \"stepMarkPaid\",\n\t}\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tctx, logger := GetLogger(s.getCtx(r))\n\t\to.Do(func() {\n\t\t\ttpl = s.loadWebTemplateDashboard(ctx, \"payment-view.html\")\n\t\t})\n\t\tctx, provider, data, errs, ok := s.createTemplateDataDashboard(w, r.WithContext(ctx), tpl, true)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tdata[TplParamActiveNav] = provider.GetURLPayments()\n\t\tdata[TplParamSteps] = steps\n\n\t\t//load the booking\n\t\tnow := data[TplParamCurrentTime].(time.Time)\n\t\tvar paymentUI *paymentUI\n\t\tbookIDStr := r.FormValue(URLParams.BookID)\n\t\tif bookIDStr != \"\" {\n\t\t\tctx, book, ok := s.loadTemplateBook(w, r.WithContext(ctx), tpl, data, errs, bookIDStr, false, false)\n\t\t\tif !ok {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tdata[TplParamFormAction] = book.GetURLPaymentView()\n\n\t\t\t//load the service\n\t\t\tctx, _, ok = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, book.Service.ID, now)\n\t\t\tif !ok {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t//probe for a payment\n\t\t\tctx, payment, err := LoadPaymentByProviderIDAndSecondaryIDAndType(ctx, s.getDB(), provider.ID, book.ID, PaymentTypeBooking)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"load payment\", \"error\", err, \"id\", book.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif payment == nil {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpaymentUI = s.createPaymentUI(payment)\n\t\t} else {\n\t\t\t//load the payment directly\n\t\t\tidStr := r.FormValue(URLParams.PaymentID)\n\t\t\tif idStr == \"\" {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tid := uuid.FromStringOrNil(idStr)\n\t\t\tif id == uuid.Nil {\n\t\t\t\tlogger.Errorw(\"invalid uuid\", \"id\", idStr)\n\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tctx, payment, err := LoadPaymentByID(ctx, s.getDB(), &id)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"load payment\", \"error\", err, \"id\", id)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpaymentUI = s.createPaymentUI(payment)\n\t\t\tdata[TplParamFormAction] = paymentUI.GetURLView()\n\n\t\t\t//probe for a booking\n\t\t\tctx, book, ok := s.loadTemplateBook(w, r.WithContext(ctx), tpl, data, errs, payment.SecondaryID.String(), false, false)\n\t\t\tif ok {\n\t\t\t\tctx, _, _ = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, book.Service.ID, now)\n\t\t\t} else if paymentUI.ServiceID != \"\" {\n\t\t\t\tsvcID := uuid.FromStringOrNil(paymentUI.ServiceID)\n\t\t\t\tif svcID == uuid.Nil {\n\t\t\t\t\tlogger.Errorw(\"invalid uuid\", \"id\", paymentUI.ServiceID)\n\t\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tctx, _, _ = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, &svcID, now)\n\t\t\t}\n\t\t}\n\t\tdata[TplParamPayment] = paymentUI\n\n\t\t//set-up the confirmation\n\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgPaymentMarkPaid)\n\t\tdata[TplParamConfirmSubmitName] = URLParams.Step\n\t\tdata[TplParamConfirmSubmitValue] = steps.StepMarkPaid\n\n\t\t//check the method\n\t\tif r.Method == http.MethodGet {\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\n\t\t//process the step\n\t\tstep := r.FormValue(URLParams.Step)\n\t\tswitch step {\n\t\tcase steps.StepDel:\n\t\t\tctx, err := DeletePayment(ctx, s.getDB(), paymentUI.ID)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"delete payment\", \"error\", err, \"id\", paymentUI.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase steps.StepMarkPaid:\n\t\t\tctx, err := UpdatePaymentDirectCapture(ctx, s.getDB(), paymentUI.ID, &now)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"update payment captured\", \"error\", err, \"id\", paymentUI.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\tdefault:\n\t\t\tlogger.Errorw(\"invalid step\", \"id\", paymentUI.ID, \"step\", step)\n\t\t\ts.SetCookieErr(w, Err)\n\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\t\ts.SetCookieMsg(w, MsgUpdateSuccess)\n\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t}\n}", "func (db DB) handleNewTransaction(log *logrus.Entry, paymentChannel paymentchannel.Channel, order *m.UserOrder, iTx *m.IncomingTransaction, denomAmount *big.Int) (err error) {\n\t//check order amount\n\toa := new(big.Int)\n\toa.SetString(order.ExchangeCurrencyDenominationAmount, 0)\n\n\tcmp := oa.Cmp(denomAmount)\n\tif cmp == -1 || cmp == 1 {\n\t\tif cmp == -1 {\n\t\t\t//order amount < denomAmount\n\t\t\torder.OrderStatus = m.OrderStatusOverPay\n\t\t} else if cmp == 1 {\n\t\t\torder.OrderStatus = m.OrderStatusUnderPay\n\t\t}\n\n\t\t_, err = order.Update(db, boil.Whitelist(m.UserOrderColumns.OrderStatus, m.UserOrderColumns.UpdatedAt))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = paymentChannel.TransferAmount(order, iTx, denomAmount, iTx.SenderAddress, m.TransactionStatusRefund, iTx.BTCSRCOutIndex)\n\t\tif err != nil {\n\t\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error refunding wrong amount\")\n\t\t}\n\t\treturn nil\n\t}\n\n\t//amount payed is exactly the amount bought. we can check/update, if there are coins left\n\tph := new(m.IcoPhase)\n\n\tsqlStr := querying.GetSQLKeyString(`update @ico_phase set @tokens_left=@tokens_left-$1, @updated_at=current_timestamp where id=$2 and @ico_phase_status=$3 and\n\t\t start_time<=current_timestamp and end_time>=current_timestamp and @tokens_left>=$4 returning *`,\n\t\tmap[string]string{\n\t\t\t\"@ico_phase\": m.TableNames.IcoPhase,\n\t\t\t\"@tokens_left\": m.IcoPhaseColumns.TokensLeft,\n\t\t\t\"@updated_at\": m.IcoPhaseColumns.UpdatedAt,\n\t\t})\n\n\terr = queries.Raw(sqlStr, order.TokenAmount, order.IcoPhaseID, m.IcoPhaseStatusActive, order.TokenAmount).Bind(nil, db, ph)\n\tif err != nil {\n\t\t//something is not ok\n\t\t//either left token-amount is to small, or phase is already gone... we will read the data againe and check\n\t\tif err != sql.ErrNoRows {\n\t\t\t// log error\n\t\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error selecting phasedata\")\n\t\t\tiTx.Status = m.TransactionStatusError\n\t\t\tif _, err := iTx.Update(db, boil.Whitelist(m.IncomingTransactionColumns.Status, m.IncomingTransactionColumns.UpdatedAt)); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\n\t\tph, err = m.IcoPhases(qm.Where(\"id=?\", order.IcoPhaseID)).One(db)\n\t\tif err != nil {\n\t\t\tiTx.Status = m.TransactionStatusError\n\t\t\tif _, err := iTx.Update(db, boil.Whitelist(m.IncomingTransactionColumns.Status, m.IncomingTransactionColumns.UpdatedAt)); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\n\t\tif ph.TokensLeft < order.TokenAmount {\n\t\t\torder.OrderStatus = m.OrderStatusNoCoinsLeft\n\t\t}\n\t\tif ph.IcoPhaseStatus != m.IcoPhaseStatusActive {\n\t\t\torder.OrderStatus = m.OrderStatusPhaseExpired\n\t\t}\n\t\tif _, err := order.Update(db, boil.Whitelist(m.UserOrderColumns.OrderStatus, m.UserOrderColumns.UpdatedAt)); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := paymentChannel.TransferAmount(order, iTx, denomAmount, iTx.SenderAddress, m.TransactionStatusRefund, iTx.BTCSRCOutIndex); err != nil {\n\t\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error refunding wrong phase_status or tokenamount\")\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}\n\n\t//everything seems ok -> update the order but first re-check current status\n\terr = order.Reload(db)\n\tif err != nil {\n\t\tiTx.Status = m.TransactionStatusError\n\t\tif _, err := iTx.Update(db, boil.Whitelist(m.IncomingTransactionColumns.Status, m.IncomingTransactionColumns.UpdatedAt)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn err\n\t}\n\n\tif order.OrderStatus != m.OrderStatusPaymentReceived {\n\t\t//order changed meanwhile, eg from second process, we refund and exit\n\t\tif err := paymentChannel.TransferAmount(order, iTx, denomAmount, iTx.SenderAddress, m.TransactionStatusRefund, iTx.BTCSRCOutIndex); err != nil {\n\t\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error refunding wrong order status\")\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n\n\torder.OrderStatus = m.OrderStatusWaitingUserTransaction\n\t_, err = order.Update(db, boil.Whitelist(m.UserOrderColumns.OrderStatus, m.UserOrderColumns.UpdatedAt))\n\tif err != nil {\n\t\tiTx.Status = m.TransactionStatusError\n\t\tif _, err := iTx.Update(db, boil.Whitelist(m.IncomingTransactionColumns.Status, m.IncomingTransactionColumns.UpdatedAt)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn err\n\t}\n\n\t//TODO: move amount to payout-account in payment network\n\n\t//check all user orders and if one is payed, set flag, if not, remove flag\n\tuser, err := m.UserProfiles(qm.Where(\"id=?\", order.UserID)).One(db)\n\tif err != nil {\n\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error selecting user profile-payment-status\")\n\t\treturn nil\n\t}\n\n\tcnt, err := m.UserOrders(qm.Where(\"user_id=? and order_status=?\", order.UserID, m.OrderStatusWaitingUserTransaction)).Count(db)\n\tif cnt > 0 {\n\t\tuser.PaymentState = m.PaymentStateOpen\n\t} else {\n\t\tuser.PaymentState = m.PaymentStateClose\n\t}\n\t_, err = user.Update(db, boil.Whitelist(m.UserProfileColumns.PaymentState, m.UserProfileColumns.UpdatedAt))\n\tif err != nil {\n\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error updating user profile-payment-status\")\n\t}\n\n\treturn nil\n}", "func (a *App) handleRequest(handler func(storage.Database, http.ResponseWriter, *http.Request)) func(http.ResponseWriter, *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\thandler(a.DB, w, r)\n\t}\n}", "func HandleInsert(w http.ResponseWriter, r *http.Request) {\n\tvar payment models.Payment\n\n\tdecoder := json.NewDecoder(r.Body)\n\tif err := decoder.Decode(&payment); err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(\"Invalid request payload. If you have specified an ID, remove it and let the system generate one.\"))\n\t\treturn\n\t}\n\n\tdefer r.Body.Close()\n\n\t_, err := db.InsertPayment(payment)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(err.Error()))\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusOK)\n}", "func processTxHandler(w http.ResponseWriter, r *http.Request) {\n\tif r.URL.Path != \"/processTx/\" {\n\t\thttp.NotFound(w, r)\n\t\treturn\n\t}\n\n\tif r.Method != \"POST\" { // expecting POST method\n\t\thttp.Error(w, \"Invalid request method.\", 405)\n\t\treturn\n\t}\n\n\tdecoder := json.NewDecoder(r.Body)\n\tvar txIn TxInput\n\n\terr := decoder.Decode(&txIn)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tdefer r.Body.Close()\n\n\t// fmt.Printf(\"\\nTX input:\\n%+v\\n\", txIn)\n\n\ttxResultStr := processTx(&txIn)\n\n\tfmt.Fprintf(w, \"%s\", txResultStr)\n}", "func BobPurchaseDataAPIHandler(w http.ResponseWriter, r *http.Request) {\n\tLog := Logger.NewSessionLogger()\n\n\tLog.Infof(\"start purchase data...\")\n\tvar plog PodLog\n\tplog.Result = LOG_RESULT_FAILED\n\tplog.Operation = LOG_OPERATION_TYPE_BOB_TX\n\tdefer func() {\n\t\terr := insertLogToDB(plog)\n\t\tif err != nil {\n\t\t\tLog.Warnf(\"insert log error! %v\", err)\n\t\t\treturn\n\t\t}\n\t\tnodeRecovery(w, Log)\n\t}()\n\n\trequestData := r.FormValue(\"request_data\")\n\tvar data RequestData\n\terr := json.Unmarshal([]byte(requestData), &data)\n\tif err != nil {\n\t\tLog.Warnf(\"invalid parameter. data=%v, err=%v\", requestData, err)\n\t\tfmt.Fprintf(w, RESPONSE_INCOMPLETE_PARAM)\n\t\treturn\n\t}\n\tLog.Debugf(\"success to parse request data. data=%v\", requestData)\n\n\tif data.MerkleRoot == \"\" || data.AliceIP == \"\" || data.AliceAddr == \"\" || data.BulletinFile == \"\" || data.PubPath == \"\" {\n\t\tLog.Warnf(\"invalid parameter. merkleRoot=%v, AliceIP=%v, AliceAddr=%v, bulletinFile=%v, PubPath=%v\",\n\t\t\tdata.MerkleRoot, data.AliceIP, data.AliceAddr, data.BulletinFile, data.PubPath)\n\t\tfmt.Fprintf(w, RESPONSE_INCOMPLETE_PARAM)\n\t\treturn\n\t}\n\tLog.Debugf(\"read parameters. merkleRoot=%v, AliceIP=%v, AliceAddr=%v, bulletinFile=%v, PubPath=%v\",\n\t\tdata.MerkleRoot, data.AliceIP, data.AliceAddr, data.BulletinFile, data.PubPath)\n\n\tplog.Detail = fmt.Sprintf(\"merkleRoot=%v, AliceIP=%v, AliceAddr=%v, bulletinFile=%v, PubPath=%v\",\n\t\tdata.MerkleRoot, data.AliceIP, data.AliceAddr, data.BulletinFile, data.PubPath)\n\n\tbulletin, err := readBulletinFile(data.BulletinFile, Log)\n\tif err != nil {\n\t\tLog.Warnf(\"failed to read bulletin File. err=%v\", err)\n\t\tfmt.Fprintf(w, RESPONSE_PURCHASE_FAILED)\n\t\treturn\n\t}\n\tplog.Detail = fmt.Sprintf(\"%v, merkle root=%v,\", plog.Detail, bulletin.SigmaMKLRoot)\n\n\tLog.Debugf(\"step0: prepare for transaction...\")\n\tvar params = BobConnParam{data.AliceIP, data.AliceAddr, bulletin.Mode, data.SubMode, data.OT, data.UnitPrice, \"\", bulletin.SigmaMKLRoot}\n\tnode, conn, params, err := preBobConn(params, ETHKey, Log)\n\tif err != nil {\n\t\tLog.Warnf(\"failed to prepare net for transaction. err=%v\", err)\n\t\tfmt.Fprintf(w, RESPONSE_PURCHASE_FAILED)\n\t\treturn\n\t}\n\tdefer func() {\n\t\tif err := node.Close(); err != nil {\n\t\t\tfmt.Errorf(\"failed to close client node: %v\", err)\n\t\t}\n\t\tif err := conn.Close(); err != nil {\n\t\t\tLog.Errorf(\"failed to close connection on client side: %v\", err)\n\t\t}\n\t}()\n\tLog.Debugf(\"[%v]step0: success to establish connecting session with Alice. Alice IP=%v, Alice address=%v\", params.SessionID, params.AliceIPAddr, params.AliceAddr)\n\tplog.Detail = fmt.Sprintf(\"%v, sessionID=%v,\", plog.Detail, params.SessionID)\n\tplog.SessionId = params.SessionID\n\n\tvar tx BobTransaction\n\ttx.SessionID = params.SessionID\n\ttx.Status = TRANSACTION_STATUS_START\n\ttx.Bulletin = bulletin\n\ttx.AliceIP = params.AliceIPAddr\n\ttx.AliceAddr = params.AliceAddr\n\ttx.Mode = params.Mode\n\ttx.SubMode = params.SubMode\n\ttx.OT = params.OT\n\ttx.UnitPrice = params.UnitPrice\n\ttx.BobAddr = fmt.Sprintf(\"%v\", ETHKey.Address.Hex())\n\n\tLog.Debugf(\"[%v]step0: success to prepare for transaction...\", params.SessionID)\n\ttx.Status = TRANSACTION_STATUS_START\n\terr = insertBobTxToDB(tx)\n\tif err != nil {\n\t\tLog.Warnf(\"failed to save transaction to db for Bob. err=%v\", err)\n\t\tfmt.Fprintf(w, fmt.Sprintf(RESPONSE_TRANSACTION_FAILED, \"failed to save transaction to db for Bob.\"))\n\t\treturn\n\t}\n\n\tvar response string\n\tif tx.Mode == TRANSACTION_MODE_PLAIN_POD {\n\t\tswitch tx.SubMode {\n\t\tcase TRANSACTION_SUB_MODE_COMPLAINT:\n\t\t\tif tx.OT {\n\t\t\t\tresponse = BobTxForPOC(node, ETHKey, tx, data.Demands, data.Phantoms, data.BulletinFile, data.PubPath, Log)\n\t\t\t} else {\n\t\t\t\tresponse = BobTxForPC(node, ETHKey, tx, data.Demands, data.BulletinFile, data.PubPath, Log)\n\t\t\t}\n\t\tcase TRANSACTION_SUB_MODE_ATOMIC_SWAP:\n\t\t\tresponse = BobTxForPAS(node, ETHKey, tx, data.Demands, data.BulletinFile, data.PubPath, Log)\n\t\t}\n\t} else if tx.Mode == TRANSACTION_MODE_TABLE_POD {\n\t\tswitch tx.SubMode {\n\t\tcase TRANSACTION_SUB_MODE_COMPLAINT:\n\t\t\tif tx.OT {\n\t\t\t\tresponse = BobTxForTOC(node, ETHKey, tx, data.Demands, data.Phantoms, data.BulletinFile, data.PubPath, Log)\n\t\t\t} else {\n\t\t\t\tresponse = BobTxForTC(node, ETHKey, tx, data.Demands, data.BulletinFile, data.PubPath, Log)\n\t\t\t}\n\t\tcase TRANSACTION_SUB_MODE_ATOMIC_SWAP:\n\t\t\tresponse = BobTxForTAS(node, ETHKey, tx, data.Demands, data.BulletinFile, data.PubPath, Log)\n\t\tcase TRANSACTION_SUB_MODE_VRF:\n\t\t\tif tx.OT {\n\t\t\t\tresponse = BobTxForTOQ(node, ETHKey, tx, data.KeyName, data.KeyValue, data.PhantomKeyValue, data.BulletinFile, data.PubPath, Log)\n\t\t\t} else {\n\t\t\t\tresponse = BobTxForTQ(node, ETHKey, tx, data.KeyName, data.KeyValue, data.BulletinFile, data.PubPath, Log)\n\t\t\t}\n\t\t}\n\t}\n\tvar resp Response\n\terr = json.Unmarshal([]byte(response), &resp)\n\tif err != nil {\n\t\tLog.Warnf(\"failed to parse response. response=%v, err=%v\", response, err)\n\t\tfmt.Fprintf(w, RESPONSE_FAILED_TO_RESPONSE)\n\t\treturn\n\t}\n\tif resp.Code == \"0\" {\n\t\tplog.Result = LOG_RESULT_SUCCESS\n\t}\n\tLog.Debugf(\"[%v]the transaction finish. merkel root=%v, response=%v\", params.SessionID, bulletin.SigmaMKLRoot, response)\n\tfmt.Fprintf(w, response)\n\treturn\n}", "func (p *Politeiawww) handleInvoiceDetails(w http.ResponseWriter, r *http.Request) {\n\tlog.Tracef(\"handleInvoiceDetails\")\n\n\t// Get the invoice details command\n\tvar pd cms.InvoiceDetails\n\t// get version from query string parameters\n\terr := util.ParseGetParams(r, &pd)\n\tif err != nil {\n\t\tRespondWithError(w, r, 0, \"handleInvoiceDetails: ParseGetParams\",\n\t\t\twww.UserError{\n\t\t\t\tErrorCode: www.ErrorStatusInvalidInput,\n\t\t\t})\n\t\treturn\n\t}\n\n\t// Get invoice token from path parameters\n\tpathParams := mux.Vars(r)\n\tpd.Token = pathParams[\"token\"]\n\n\tuser, err := p.sessions.GetSessionUser(w, r)\n\tif err != nil {\n\t\tif !errors.Is(err, sessions.ErrSessionNotFound) {\n\t\t\tRespondWithError(w, r, 0,\n\t\t\t\t\"handleInvoiceDetails: getSessionUser %v\", err)\n\t\t\treturn\n\t\t}\n\t}\n\treply, err := p.processInvoiceDetails(pd, user)\n\tif err != nil {\n\t\tRespondWithError(w, r, 0,\n\t\t\t\"handleInvoiceDetails: processInvoiceDetails %v\", err)\n\t\treturn\n\t}\n\n\t// Reply with the proposal details.\n\tutil.RespondWithJSON(w, http.StatusOK, reply)\n}", "func (p *Politeiawww) handlePayInvoices(w http.ResponseWriter, r *http.Request) {\n\tlog.Tracef(\"handlePayInvoices\")\n\n\tuser, err := p.sessions.GetSessionUser(w, r)\n\tif err != nil {\n\t\tRespondWithError(w, r, 0,\n\t\t\t\"handlePayInvoices: getSessionUser %v\", err)\n\t\treturn\n\t}\n\n\treply, err := p.processPayInvoices(r.Context(), user)\n\tif err != nil {\n\t\tRespondWithError(w, r, 0, \"handlePayInvoices: processPayInvoices %v\",\n\t\t\terr)\n\t\treturn\n\t}\n\n\tutil.RespondWithJSON(w, http.StatusOK, reply)\n}", "func (_obj *DoPayment) Dispatch(ctx context.Context, _val interface{}, req *requestf.RequestPacket, resp *requestf.ResponsePacket, withContext bool) (err error) {\n\t_is := codec.NewReader(tools.Int8ToByte(req.SBuffer))\n\t_os := codec.NewBuffer()\n\tswitch req.SFuncName {\n\tcase \"echoHello\":\n\t\terr := echoHello(ctx, _val, _os, _is, withContext)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\tcase \"health\":\n\t\terr := health(ctx, _val, _os, _is, withContext)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\tcase \"authorise\":\n\t\terr := authorise(ctx, _val, _os, _is, withContext)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\tdefault:\n\t\treturn fmt.Errorf(\"func mismatch\")\n\t}\n\tvar _status map[string]string\n\ts, ok := current.GetResponseStatus(ctx)\n\tif ok && s != nil {\n\t\t_status = s\n\t}\n\tvar _context map[string]string\n\tc, ok := current.GetResponseContext(ctx)\n\tif ok && c != nil {\n\t\t_context = c\n\t}\n\t*resp = requestf.ResponsePacket{\n\t\tIVersion: 1,\n\t\tCPacketType: 0,\n\t\tIRequestId: req.IRequestId,\n\t\tIMessageType: 0,\n\t\tIRet: 0,\n\t\tSBuffer: tools.ByteToInt8(_os.ToBytes()),\n\t\tStatus: _status,\n\t\tSResultDesc: \"\",\n\t\tContext: _context,\n\t}\n\treturn nil\n}", "func (h CreatePaymentRequestHandler) Handle(params paymentrequestop.CreatePaymentRequestParams) middleware.Responder {\n\t// TODO: authorization to create payment request\n\n\treturn h.AuditableAppContextFromRequestWithErrors(params.HTTPRequest,\n\t\tfunc(appCtx appcontext.AppContext) (middleware.Responder, error) {\n\n\t\t\tpayload := params.Body\n\t\t\tif payload == nil {\n\t\t\t\terr := apperror.NewBadDataError(\"Invalid payment request: params Body is nil\")\n\t\t\t\terrPayload := payloads.ClientError(handlers.SQLErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\t\t\t\tappCtx.Logger().Error(err.Error(), zap.Any(\"payload\", errPayload))\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestBadRequest().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tappCtx.Logger().Info(\"primeapi.CreatePaymentRequestHandler info\", zap.String(\"pointOfContact\", params.Body.PointOfContact))\n\n\t\t\tmoveTaskOrderIDString := payload.MoveTaskOrderID.String()\n\t\t\tmtoID, err := uuid.FromString(moveTaskOrderIDString)\n\t\t\tif err != nil {\n\t\t\t\tappCtx.Logger().Error(\"Invalid payment request: params MoveTaskOrderID cannot be converted to a UUID\",\n\t\t\t\t\tzap.String(\"MoveTaskOrderID\", moveTaskOrderIDString), zap.Error(err))\n\t\t\t\t// create a custom verrs for returning a 422\n\t\t\t\tverrs :=\n\t\t\t\t\t&validate.Errors{Errors: map[string][]string{\n\t\t\t\t\t\t\"move_id\": {\"id cannot be converted to UUID\"},\n\t\t\t\t\t},\n\t\t\t\t\t}\n\t\t\t\terrPayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tisFinal := false\n\t\t\tif payload.IsFinal != nil {\n\t\t\t\tisFinal = *payload.IsFinal\n\t\t\t}\n\n\t\t\tpaymentRequest := models.PaymentRequest{\n\t\t\t\tIsFinal: isFinal,\n\t\t\t\tMoveTaskOrderID: mtoID,\n\t\t\t}\n\n\t\t\t// Build up the paymentRequest.PaymentServiceItems using the incoming payload to offload Swagger data coming\n\t\t\t// in from the API. These paymentRequest.PaymentServiceItems will be used as a temp holder to process the incoming API data\n\t\t\tvar verrs *validate.Errors\n\t\t\tpaymentRequest.PaymentServiceItems, verrs, err = h.buildPaymentServiceItems(appCtx, payload)\n\n\t\t\tif err != nil || verrs.HasAny() {\n\n\t\t\t\tappCtx.Logger().Error(\"could not build service items\", zap.Error(err))\n\t\t\t\t// TODO: do not bail out before creating the payment request, we need the failed record\n\t\t\t\t// we should create the failed record and store it as failed with a rejection\n\t\t\t\terrPayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tcreatedPaymentRequest, err := h.PaymentRequestCreator.CreatePaymentRequestCheck(appCtx, &paymentRequest)\n\t\t\tif err != nil {\n\t\t\t\tappCtx.Logger().Error(\"Error creating payment request\", zap.Error(err))\n\t\t\t\tswitch e := err.(type) {\n\t\t\t\tcase apperror.InvalidCreateInputError:\n\t\t\t\t\tverrs := e.ValidationErrors\n\t\t\t\t\tdetail := err.Error()\n\t\t\t\t\tpayload := payloads.ValidationError(detail, h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(payload), err\n\n\t\t\t\tcase apperror.NotFoundError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.NotFoundMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestNotFound().WithPayload(payload), err\n\t\t\t\tcase apperror.ConflictError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.ConflictErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestConflict().WithPayload(payload), err\n\t\t\t\tcase apperror.InvalidInputError:\n\t\t\t\t\tpayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), &validate.Errors{})\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(payload), err\n\t\t\t\tcase apperror.QueryError:\n\t\t\t\t\tif e.Unwrap() != nil {\n\t\t\t\t\t\t// If you can unwrap, log the internal error (usually a pq error) for better debugging\n\t\t\t\t\t\tappCtx.Logger().Error(\"primeapi.CreatePaymentRequestHandler query error\", zap.Error(e.Unwrap()))\n\t\t\t\t\t}\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestInternalServerError().WithPayload(\n\t\t\t\t\t\tpayloads.InternalServerError(nil, h.GetTraceIDFromRequest(params.HTTPRequest))), err\n\n\t\t\t\tcase *apperror.BadDataError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.BadRequestErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestBadRequest().WithPayload(payload), err\n\t\t\t\tdefault:\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestInternalServerError().WithPayload(\n\t\t\t\t\t\tpayloads.InternalServerError(nil, h.GetTraceIDFromRequest(params.HTTPRequest))), err\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturnPayload := payloads.PaymentRequest(createdPaymentRequest)\n\t\t\tappCtx.Logger().Info(\"Successful payment request creation for mto ID\", zap.String(\"moveID\", moveTaskOrderIDString))\n\t\t\treturn paymentrequestop.NewCreatePaymentRequestCreated().WithPayload(returnPayload), nil\n\t\t})\n}", "func ProcessPaymentRequested(ctx fsm.Context, environment ClientDealEnvironment, deal rm.ClientDealState) error {\n\t// If the unseal payment hasn't been made, we need to send funds\n\tif deal.UnsealPrice.GreaterThan(deal.UnsealFundsPaid) {\n\t\tlog.Debugf(\"client: payment needed: unseal price %d > unseal paid %d\",\n\t\t\tdeal.UnsealPrice, deal.UnsealFundsPaid)\n\t\treturn ctx.Trigger(rm.ClientEventSendFunds)\n\t}\n\n\t// If all bytes received have been paid for, we don't need to send funds\n\tif deal.BytesPaidFor >= deal.TotalReceived {\n\t\tlog.Debugf(\"client: no payment needed: bytes paid for %d >= bytes received %d\",\n\t\t\tdeal.BytesPaidFor, deal.TotalReceived)\n\t\treturn nil\n\t}\n\n\t// Not all bytes received have been paid for\n\n\t// If all blocks have been received we need to send a final payment\n\tif deal.AllBlocksReceived {\n\t\tlog.Debugf(\"client: payment needed: all blocks received, bytes paid for %d < bytes received %d\",\n\t\t\tdeal.BytesPaidFor, deal.TotalReceived)\n\t\treturn ctx.Trigger(rm.ClientEventSendFunds)\n\t}\n\n\t// Payments are made in intervals, as bytes are received from the provider.\n\t// If the number of bytes received is at or above the size of the current\n\t// interval, we need to send a payment.\n\tif deal.TotalReceived >= deal.CurrentInterval {\n\t\tlog.Debugf(\"client: payment needed: bytes received %d >= interval %d, bytes paid for %d < bytes received %d\",\n\t\t\tdeal.TotalReceived, deal.CurrentInterval, deal.BytesPaidFor, deal.TotalReceived)\n\t\treturn ctx.Trigger(rm.ClientEventSendFunds)\n\t}\n\n\tlog.Debugf(\"client: no payment needed: received %d < interval %d (paid for %d)\",\n\t\tdeal.TotalReceived, deal.CurrentInterval, deal.BytesPaidFor)\n\treturn nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
/////////////////////////////v4 /////////////////////////////v4 v4handleDBProcesspayment receive and handle the request from client, access DB
func v4handleDBProcesspayment(w http.ResponseWriter, r *http.Request) { defer func() { db.Connection.Close(nil) }() var errorGeneral string var errorGeneralNbr string var requestData modelito.RequestPayment errorGeneral="" requestData,errorGeneral =obtainParmsProcessPayment(r,errorGeneral) ////////////////////////////////////////////////validate parms /// START if errorGeneral=="" { errorGeneral,errorGeneralNbr= v4ProcessProcessPayment(w , requestData) //logicbusiness.go } if errorGeneral!=""{ //send error response if any //prepare an error JSON Response, if any log.Print("CZ STEP Get the ERROR response JSON ready") /// START fieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr) ////////// write the response (ERROR) w.Header().Set("Content-Type", "application/json") w.Write(fieldDataBytesJson) if(err!=nil){ } } }
[ "func v4handleDBPostProcesspayment(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n var errorGeneral string\n var errorGeneralNbr string\n var requestData modelito.RequestPayment\n \n errorGeneral=\"\"\nrequestData,errorGeneral =obtainPostParmsProcessPayment(r,errorGeneral) //logicrequest_post.go\n\n\t////////////////////////////////////////////////validate parms\n\t/// START\n\t////////////////////////////////////////////////validate parms\n\t/// START\n \n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= v4ProcessProcessPayment(w , requestData) //logicbusiness.go \n\t}\n\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func handleDBPostGettokenizedcards(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n var errorGeneral string\n var errorGeneralNbr string\n \n \tvar requestData modelito.RequestTokenizedCards\n\n errorGeneral=\"\"\n requestData, errorGeneral=obtainPostParmsGettokenizedcards(r,errorGeneral) //logicrequest_post.go\n\n\t////////////////////////////////////////////////process business rules\n\t/// START\n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= ProcessGettokenizedcards(w , requestData)\n\t}\n\t/// END\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func handleDBGeneratetokenized(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n var requestData modelito.RequestTokenized\n var errorGeneral string\n var errorGeneralNbr string\n \n errorGeneral=\"\"\n requestData,errorGeneral =obtainParmsGeneratetokenized(r,errorGeneral)\n\n\n\t////////////////////////////////////////////////validate parms\n\t/// START\n \n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= ProcessGeneratetokenized(w , requestData)\n\t}\n\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func (ctx *Context) PaymentDB(ros ...dbRequestReadOnly) *sql.DB {\n\tvar ro bool\n\tif len(ros) > 0 {\n\t\tfor _, r := range ros {\n\t\t\tif r {\n\t\t\t\tro = true\n\t\t\t}\n\t\t}\n\t}\n\tif !ro {\n\t\treturn ctx.paymentDBWrite\n\t}\n\tif ctx.paymentDBReadOnly == nil {\n\t\treturn ctx.paymentDBWrite\n\t}\n\treturn ctx.paymentDBReadOnly\n}", "func HandlePayments(client *http.Client, strapiURL string, strapiToken string, db *sql.DB, w http.ResponseWriter, r *http.Request) {\n\tvar customerOrder CustomerOrder\n\tbodyBytes, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = json.Unmarshal(bodyBytes, &customerOrder)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer r.Body.Close()\n\turl := fmt.Sprintf(\"%v/restaurants/payment/%v\", strapiURL, customerOrder.RestaurantID)\n\treq, err := http.NewRequest(http.MethodGet, url, nil)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treq.Header.Set(\"Authorization\", \"Bearer \"+strapiToken)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer resp.Body.Close()\n\tvar strapiResponse []Restaurant\n\n\tbodyBytes, err = ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = json.Unmarshal(bodyBytes, &strapiResponse)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\trestaurant := strapiResponse[0]\n\n\ttotal := calculateTotalPriceInPence(customerOrder, restaurant)\n\tsc := stripeClient.New(restaurant.Credential.TestPrivateKey, nil)\n\n\tparams := &stripe.PaymentIntentParams{\n\t\tAmount: stripe.Int64(total),\n\t\tCurrency: stripe.String(string(stripe.CurrencyGBP)),\n\t\tSetupFutureUsage: stripe.String(string(stripe.PaymentIntentSetupFutureUsageOffSession)),\n\t\tStatementDescriptor: stripe.String(restaurant.Name[:21]),\n\t}\n\tparams.AddMetadata(\"order_id\", customerOrder.UUID)\n\tpi, _ := sc.PaymentIntents.New(params)\n\n\tcustomerResponse := CheckoutData{\n\t\tClientSecret: pi.ClientSecret,\n\t}\n\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(customerResponse)\n\n}", "func DataRetrievalHandler(reader fcrserver.FCRServerRequestReader, writer fcrserver.FCRServerResponseWriter, request *fcrmessages.FCRReqMsg) error {\n\tlogging.Debug(\"Handle data retrieval\")\n\t// Get core structure\n\tc := core.GetSingleInstance()\n\tc.MsgSigningKeyLock.RLock()\n\tdefer c.MsgSigningKeyLock.RUnlock()\n\n\t// Message decoding\n\tnonce, senderID, offer, accountAddr, voucher, err := fcrmessages.DecodeDataRetrievalRequest(request)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Error in decoding payload: %v\", err.Error())\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\n\t// Verify signature\n\tif request.VerifyByID(senderID) != nil {\n\t\t// Verify by signing key\n\t\tgwInfo := c.PeerMgr.GetGWInfo(senderID)\n\t\tif gwInfo == nil {\n\t\t\t// Not found, try sync once\n\t\t\tgwInfo = c.PeerMgr.SyncGW(senderID)\n\t\t\tif gwInfo == nil {\n\t\t\t\terr = fmt.Errorf(\"Error in obtaining information for gateway %v\", senderID)\n\t\t\t\tlogging.Error(err.Error())\n\t\t\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t\t\t}\n\t\t}\n\t\tif request.Verify(gwInfo.MsgSigningKey, gwInfo.MsgSigningKeyVer) != nil {\n\t\t\t// Try update\n\t\t\tgwInfo = c.PeerMgr.SyncGW(senderID)\n\t\t\tif gwInfo == nil || request.Verify(gwInfo.MsgSigningKey, gwInfo.MsgSigningKeyVer) != nil {\n\t\t\t\terr = fmt.Errorf(\"Error in verifying request from gateway %v: %v\", senderID, err.Error())\n\t\t\t\tlogging.Error(err.Error())\n\t\t\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Check payment\n\trefundVoucher := \"\"\n\treceived, lane, err := c.PaymentMgr.Receive(accountAddr, voucher)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Error in receiving voucher %v:\", err.Error())\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\tif lane != 1 {\n\t\terr = fmt.Errorf(\"Not correct lane received expect 1 got %v:\", lane)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\texpected := big.NewInt(0).Add(c.Settings.SearchPrice, offer.GetPrice())\n\tif received.Cmp(expected) < 0 {\n\t\t// Short payment\n\t\t// Refund money\n\t\tif received.Cmp(c.Settings.SearchPrice) <= 0 {\n\t\t\t// No refund\n\t\t} else {\n\t\t\tvar ierr error\n\t\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\t\tif ierr != nil {\n\t\t\t\t// This should never happen\n\t\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t\t}\n\t\t}\n\t\terr = fmt.Errorf(\"Short payment received, expect %v got %v, refund voucher %v\", expected.String(), received.String(), refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\n\t// Payment is fine, verify offer\n\tif offer.Verify(c.OfferSigningPubKey) != nil {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Fail to verify the offer signature, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Verify offer merkle proof\n\tif offer.VerifyMerkleProof() != nil {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Fail to verify the offer merkle proof, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Verify offer expiry\n\tif offer.HasExpired() {\n\t\t// Refund money\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, big.NewInt(0).Sub(received, c.Settings.SearchPrice))\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Offer has expired, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Offer is verified. Respond\n\t// First get the tag\n\ttag := c.OfferMgr.GetTagByCID(offer.GetSubCID())\n\t// Second read the data\n\tdata, err := ioutil.ReadFile(filepath.Join(c.Settings.RetrievalDir, tag))\n\tif err != nil {\n\t\t// Refund money, internal error, refund all\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, received)\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Internal error in finding the content, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\t// Third encoding response\n\tresponse, err := fcrmessages.EncodeDataRetrievalResponse(nonce, tag, data)\n\tif err != nil {\n\t\t// Refund money, internal error, refund all\n\t\tvar ierr error\n\t\trefundVoucher, ierr = c.PaymentMgr.Refund(accountAddr, lane, received)\n\t\tif ierr != nil {\n\t\t\t// This should never happen\n\t\t\tlogging.Error(\"Error in refunding: %v\", ierr.Error())\n\t\t}\n\t\terr = fmt.Errorf(\"Internal error in encoding the response, refund voucher %v\", refundVoucher)\n\t\tlogging.Error(err.Error())\n\t\treturn writer.Write(fcrmessages.CreateFCRACKErrorMsg(nonce, err), c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n\t}\n\tc.OfferMgr.IncrementCIDAccessCount(offer.GetSubCID())\n\n\treturn writer.Write(response, c.MsgSigningKey, c.MsgSigningKeyVer, c.Settings.TCPInactivityTimeout)\n}", "func (_BaseContent *BaseContentTransactor) ProcessRequestPayment(opts *bind.TransactOpts, request_ID *big.Int, payee common.Address, label string, amount *big.Int) (*types.Transaction, error) {\n\treturn _BaseContent.contract.Transact(opts, \"processRequestPayment\", request_ID, payee, label, amount)\n}", "func ProcessStripePayment(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"NOT IMPLEMENTED\"})\n}", "func (s *Server) handleDashboardPaymentView() http.HandlerFunc {\n\tvar o sync.Once\n\tvar tpl *template.Template\n\n\t//steps on the page\n\tsteps := struct {\n\t\tStepDel string\n\t\tStepMarkPaid string\n\t}{\n\t\tStepDel: \"stepDel\",\n\t\tStepMarkPaid: \"stepMarkPaid\",\n\t}\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tctx, logger := GetLogger(s.getCtx(r))\n\t\to.Do(func() {\n\t\t\ttpl = s.loadWebTemplateDashboard(ctx, \"payment-view.html\")\n\t\t})\n\t\tctx, provider, data, errs, ok := s.createTemplateDataDashboard(w, r.WithContext(ctx), tpl, true)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tdata[TplParamActiveNav] = provider.GetURLPayments()\n\t\tdata[TplParamSteps] = steps\n\n\t\t//load the booking\n\t\tnow := data[TplParamCurrentTime].(time.Time)\n\t\tvar paymentUI *paymentUI\n\t\tbookIDStr := r.FormValue(URLParams.BookID)\n\t\tif bookIDStr != \"\" {\n\t\t\tctx, book, ok := s.loadTemplateBook(w, r.WithContext(ctx), tpl, data, errs, bookIDStr, false, false)\n\t\t\tif !ok {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tdata[TplParamFormAction] = book.GetURLPaymentView()\n\n\t\t\t//load the service\n\t\t\tctx, _, ok = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, book.Service.ID, now)\n\t\t\tif !ok {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t//probe for a payment\n\t\t\tctx, payment, err := LoadPaymentByProviderIDAndSecondaryIDAndType(ctx, s.getDB(), provider.ID, book.ID, PaymentTypeBooking)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"load payment\", \"error\", err, \"id\", book.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif payment == nil {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpaymentUI = s.createPaymentUI(payment)\n\t\t} else {\n\t\t\t//load the payment directly\n\t\t\tidStr := r.FormValue(URLParams.PaymentID)\n\t\t\tif idStr == \"\" {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tid := uuid.FromStringOrNil(idStr)\n\t\t\tif id == uuid.Nil {\n\t\t\t\tlogger.Errorw(\"invalid uuid\", \"id\", idStr)\n\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tctx, payment, err := LoadPaymentByID(ctx, s.getDB(), &id)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"load payment\", \"error\", err, \"id\", id)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpaymentUI = s.createPaymentUI(payment)\n\t\t\tdata[TplParamFormAction] = paymentUI.GetURLView()\n\n\t\t\t//probe for a booking\n\t\t\tctx, book, ok := s.loadTemplateBook(w, r.WithContext(ctx), tpl, data, errs, payment.SecondaryID.String(), false, false)\n\t\t\tif ok {\n\t\t\t\tctx, _, _ = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, book.Service.ID, now)\n\t\t\t} else if paymentUI.ServiceID != \"\" {\n\t\t\t\tsvcID := uuid.FromStringOrNil(paymentUI.ServiceID)\n\t\t\t\tif svcID == uuid.Nil {\n\t\t\t\t\tlogger.Errorw(\"invalid uuid\", \"id\", paymentUI.ServiceID)\n\t\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tctx, _, _ = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, &svcID, now)\n\t\t\t}\n\t\t}\n\t\tdata[TplParamPayment] = paymentUI\n\n\t\t//set-up the confirmation\n\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgPaymentMarkPaid)\n\t\tdata[TplParamConfirmSubmitName] = URLParams.Step\n\t\tdata[TplParamConfirmSubmitValue] = steps.StepMarkPaid\n\n\t\t//check the method\n\t\tif r.Method == http.MethodGet {\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\n\t\t//process the step\n\t\tstep := r.FormValue(URLParams.Step)\n\t\tswitch step {\n\t\tcase steps.StepDel:\n\t\t\tctx, err := DeletePayment(ctx, s.getDB(), paymentUI.ID)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"delete payment\", \"error\", err, \"id\", paymentUI.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase steps.StepMarkPaid:\n\t\t\tctx, err := UpdatePaymentDirectCapture(ctx, s.getDB(), paymentUI.ID, &now)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"update payment captured\", \"error\", err, \"id\", paymentUI.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\tdefault:\n\t\t\tlogger.Errorw(\"invalid step\", \"id\", paymentUI.ID, \"step\", step)\n\t\t\ts.SetCookieErr(w, Err)\n\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\t\ts.SetCookieMsg(w, MsgUpdateSuccess)\n\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t}\n}", "func (db DB) handleNewTransaction(log *logrus.Entry, paymentChannel paymentchannel.Channel, order *m.UserOrder, iTx *m.IncomingTransaction, denomAmount *big.Int) (err error) {\n\t//check order amount\n\toa := new(big.Int)\n\toa.SetString(order.ExchangeCurrencyDenominationAmount, 0)\n\n\tcmp := oa.Cmp(denomAmount)\n\tif cmp == -1 || cmp == 1 {\n\t\tif cmp == -1 {\n\t\t\t//order amount < denomAmount\n\t\t\torder.OrderStatus = m.OrderStatusOverPay\n\t\t} else if cmp == 1 {\n\t\t\torder.OrderStatus = m.OrderStatusUnderPay\n\t\t}\n\n\t\t_, err = order.Update(db, boil.Whitelist(m.UserOrderColumns.OrderStatus, m.UserOrderColumns.UpdatedAt))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = paymentChannel.TransferAmount(order, iTx, denomAmount, iTx.SenderAddress, m.TransactionStatusRefund, iTx.BTCSRCOutIndex)\n\t\tif err != nil {\n\t\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error refunding wrong amount\")\n\t\t}\n\t\treturn nil\n\t}\n\n\t//amount payed is exactly the amount bought. we can check/update, if there are coins left\n\tph := new(m.IcoPhase)\n\n\tsqlStr := querying.GetSQLKeyString(`update @ico_phase set @tokens_left=@tokens_left-$1, @updated_at=current_timestamp where id=$2 and @ico_phase_status=$3 and\n\t\t start_time<=current_timestamp and end_time>=current_timestamp and @tokens_left>=$4 returning *`,\n\t\tmap[string]string{\n\t\t\t\"@ico_phase\": m.TableNames.IcoPhase,\n\t\t\t\"@tokens_left\": m.IcoPhaseColumns.TokensLeft,\n\t\t\t\"@updated_at\": m.IcoPhaseColumns.UpdatedAt,\n\t\t})\n\n\terr = queries.Raw(sqlStr, order.TokenAmount, order.IcoPhaseID, m.IcoPhaseStatusActive, order.TokenAmount).Bind(nil, db, ph)\n\tif err != nil {\n\t\t//something is not ok\n\t\t//either left token-amount is to small, or phase is already gone... we will read the data againe and check\n\t\tif err != sql.ErrNoRows {\n\t\t\t// log error\n\t\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error selecting phasedata\")\n\t\t\tiTx.Status = m.TransactionStatusError\n\t\t\tif _, err := iTx.Update(db, boil.Whitelist(m.IncomingTransactionColumns.Status, m.IncomingTransactionColumns.UpdatedAt)); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\n\t\tph, err = m.IcoPhases(qm.Where(\"id=?\", order.IcoPhaseID)).One(db)\n\t\tif err != nil {\n\t\t\tiTx.Status = m.TransactionStatusError\n\t\t\tif _, err := iTx.Update(db, boil.Whitelist(m.IncomingTransactionColumns.Status, m.IncomingTransactionColumns.UpdatedAt)); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\n\t\tif ph.TokensLeft < order.TokenAmount {\n\t\t\torder.OrderStatus = m.OrderStatusNoCoinsLeft\n\t\t}\n\t\tif ph.IcoPhaseStatus != m.IcoPhaseStatusActive {\n\t\t\torder.OrderStatus = m.OrderStatusPhaseExpired\n\t\t}\n\t\tif _, err := order.Update(db, boil.Whitelist(m.UserOrderColumns.OrderStatus, m.UserOrderColumns.UpdatedAt)); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := paymentChannel.TransferAmount(order, iTx, denomAmount, iTx.SenderAddress, m.TransactionStatusRefund, iTx.BTCSRCOutIndex); err != nil {\n\t\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error refunding wrong phase_status or tokenamount\")\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}\n\n\t//everything seems ok -> update the order but first re-check current status\n\terr = order.Reload(db)\n\tif err != nil {\n\t\tiTx.Status = m.TransactionStatusError\n\t\tif _, err := iTx.Update(db, boil.Whitelist(m.IncomingTransactionColumns.Status, m.IncomingTransactionColumns.UpdatedAt)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn err\n\t}\n\n\tif order.OrderStatus != m.OrderStatusPaymentReceived {\n\t\t//order changed meanwhile, eg from second process, we refund and exit\n\t\tif err := paymentChannel.TransferAmount(order, iTx, denomAmount, iTx.SenderAddress, m.TransactionStatusRefund, iTx.BTCSRCOutIndex); err != nil {\n\t\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error refunding wrong order status\")\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n\n\torder.OrderStatus = m.OrderStatusWaitingUserTransaction\n\t_, err = order.Update(db, boil.Whitelist(m.UserOrderColumns.OrderStatus, m.UserOrderColumns.UpdatedAt))\n\tif err != nil {\n\t\tiTx.Status = m.TransactionStatusError\n\t\tif _, err := iTx.Update(db, boil.Whitelist(m.IncomingTransactionColumns.Status, m.IncomingTransactionColumns.UpdatedAt)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn err\n\t}\n\n\t//TODO: move amount to payout-account in payment network\n\n\t//check all user orders and if one is payed, set flag, if not, remove flag\n\tuser, err := m.UserProfiles(qm.Where(\"id=?\", order.UserID)).One(db)\n\tif err != nil {\n\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error selecting user profile-payment-status\")\n\t\treturn nil\n\t}\n\n\tcnt, err := m.UserOrders(qm.Where(\"user_id=? and order_status=?\", order.UserID, m.OrderStatusWaitingUserTransaction)).Count(db)\n\tif cnt > 0 {\n\t\tuser.PaymentState = m.PaymentStateOpen\n\t} else {\n\t\tuser.PaymentState = m.PaymentStateClose\n\t}\n\t_, err = user.Update(db, boil.Whitelist(m.UserProfileColumns.PaymentState, m.UserProfileColumns.UpdatedAt))\n\tif err != nil {\n\t\tlog.WithError(err).WithFields(logrus.Fields{\"order_id\": order.ID, \"transaction_hash\": iTx.TransactionHash}).Error(\"Error updating user profile-payment-status\")\n\t}\n\n\treturn nil\n}", "func (a *App) handleRequest(handler func(storage.Database, http.ResponseWriter, *http.Request)) func(http.ResponseWriter, *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\thandler(a.DB, w, r)\n\t}\n}", "func HandleInsert(w http.ResponseWriter, r *http.Request) {\n\tvar payment models.Payment\n\n\tdecoder := json.NewDecoder(r.Body)\n\tif err := decoder.Decode(&payment); err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(\"Invalid request payload. If you have specified an ID, remove it and let the system generate one.\"))\n\t\treturn\n\t}\n\n\tdefer r.Body.Close()\n\n\t_, err := db.InsertPayment(payment)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(err.Error()))\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusOK)\n}", "func processTxHandler(w http.ResponseWriter, r *http.Request) {\n\tif r.URL.Path != \"/processTx/\" {\n\t\thttp.NotFound(w, r)\n\t\treturn\n\t}\n\n\tif r.Method != \"POST\" { // expecting POST method\n\t\thttp.Error(w, \"Invalid request method.\", 405)\n\t\treturn\n\t}\n\n\tdecoder := json.NewDecoder(r.Body)\n\tvar txIn TxInput\n\n\terr := decoder.Decode(&txIn)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tdefer r.Body.Close()\n\n\t// fmt.Printf(\"\\nTX input:\\n%+v\\n\", txIn)\n\n\ttxResultStr := processTx(&txIn)\n\n\tfmt.Fprintf(w, \"%s\", txResultStr)\n}", "func BobPurchaseDataAPIHandler(w http.ResponseWriter, r *http.Request) {\n\tLog := Logger.NewSessionLogger()\n\n\tLog.Infof(\"start purchase data...\")\n\tvar plog PodLog\n\tplog.Result = LOG_RESULT_FAILED\n\tplog.Operation = LOG_OPERATION_TYPE_BOB_TX\n\tdefer func() {\n\t\terr := insertLogToDB(plog)\n\t\tif err != nil {\n\t\t\tLog.Warnf(\"insert log error! %v\", err)\n\t\t\treturn\n\t\t}\n\t\tnodeRecovery(w, Log)\n\t}()\n\n\trequestData := r.FormValue(\"request_data\")\n\tvar data RequestData\n\terr := json.Unmarshal([]byte(requestData), &data)\n\tif err != nil {\n\t\tLog.Warnf(\"invalid parameter. data=%v, err=%v\", requestData, err)\n\t\tfmt.Fprintf(w, RESPONSE_INCOMPLETE_PARAM)\n\t\treturn\n\t}\n\tLog.Debugf(\"success to parse request data. data=%v\", requestData)\n\n\tif data.MerkleRoot == \"\" || data.AliceIP == \"\" || data.AliceAddr == \"\" || data.BulletinFile == \"\" || data.PubPath == \"\" {\n\t\tLog.Warnf(\"invalid parameter. merkleRoot=%v, AliceIP=%v, AliceAddr=%v, bulletinFile=%v, PubPath=%v\",\n\t\t\tdata.MerkleRoot, data.AliceIP, data.AliceAddr, data.BulletinFile, data.PubPath)\n\t\tfmt.Fprintf(w, RESPONSE_INCOMPLETE_PARAM)\n\t\treturn\n\t}\n\tLog.Debugf(\"read parameters. merkleRoot=%v, AliceIP=%v, AliceAddr=%v, bulletinFile=%v, PubPath=%v\",\n\t\tdata.MerkleRoot, data.AliceIP, data.AliceAddr, data.BulletinFile, data.PubPath)\n\n\tplog.Detail = fmt.Sprintf(\"merkleRoot=%v, AliceIP=%v, AliceAddr=%v, bulletinFile=%v, PubPath=%v\",\n\t\tdata.MerkleRoot, data.AliceIP, data.AliceAddr, data.BulletinFile, data.PubPath)\n\n\tbulletin, err := readBulletinFile(data.BulletinFile, Log)\n\tif err != nil {\n\t\tLog.Warnf(\"failed to read bulletin File. err=%v\", err)\n\t\tfmt.Fprintf(w, RESPONSE_PURCHASE_FAILED)\n\t\treturn\n\t}\n\tplog.Detail = fmt.Sprintf(\"%v, merkle root=%v,\", plog.Detail, bulletin.SigmaMKLRoot)\n\n\tLog.Debugf(\"step0: prepare for transaction...\")\n\tvar params = BobConnParam{data.AliceIP, data.AliceAddr, bulletin.Mode, data.SubMode, data.OT, data.UnitPrice, \"\", bulletin.SigmaMKLRoot}\n\tnode, conn, params, err := preBobConn(params, ETHKey, Log)\n\tif err != nil {\n\t\tLog.Warnf(\"failed to prepare net for transaction. err=%v\", err)\n\t\tfmt.Fprintf(w, RESPONSE_PURCHASE_FAILED)\n\t\treturn\n\t}\n\tdefer func() {\n\t\tif err := node.Close(); err != nil {\n\t\t\tfmt.Errorf(\"failed to close client node: %v\", err)\n\t\t}\n\t\tif err := conn.Close(); err != nil {\n\t\t\tLog.Errorf(\"failed to close connection on client side: %v\", err)\n\t\t}\n\t}()\n\tLog.Debugf(\"[%v]step0: success to establish connecting session with Alice. Alice IP=%v, Alice address=%v\", params.SessionID, params.AliceIPAddr, params.AliceAddr)\n\tplog.Detail = fmt.Sprintf(\"%v, sessionID=%v,\", plog.Detail, params.SessionID)\n\tplog.SessionId = params.SessionID\n\n\tvar tx BobTransaction\n\ttx.SessionID = params.SessionID\n\ttx.Status = TRANSACTION_STATUS_START\n\ttx.Bulletin = bulletin\n\ttx.AliceIP = params.AliceIPAddr\n\ttx.AliceAddr = params.AliceAddr\n\ttx.Mode = params.Mode\n\ttx.SubMode = params.SubMode\n\ttx.OT = params.OT\n\ttx.UnitPrice = params.UnitPrice\n\ttx.BobAddr = fmt.Sprintf(\"%v\", ETHKey.Address.Hex())\n\n\tLog.Debugf(\"[%v]step0: success to prepare for transaction...\", params.SessionID)\n\ttx.Status = TRANSACTION_STATUS_START\n\terr = insertBobTxToDB(tx)\n\tif err != nil {\n\t\tLog.Warnf(\"failed to save transaction to db for Bob. err=%v\", err)\n\t\tfmt.Fprintf(w, fmt.Sprintf(RESPONSE_TRANSACTION_FAILED, \"failed to save transaction to db for Bob.\"))\n\t\treturn\n\t}\n\n\tvar response string\n\tif tx.Mode == TRANSACTION_MODE_PLAIN_POD {\n\t\tswitch tx.SubMode {\n\t\tcase TRANSACTION_SUB_MODE_COMPLAINT:\n\t\t\tif tx.OT {\n\t\t\t\tresponse = BobTxForPOC(node, ETHKey, tx, data.Demands, data.Phantoms, data.BulletinFile, data.PubPath, Log)\n\t\t\t} else {\n\t\t\t\tresponse = BobTxForPC(node, ETHKey, tx, data.Demands, data.BulletinFile, data.PubPath, Log)\n\t\t\t}\n\t\tcase TRANSACTION_SUB_MODE_ATOMIC_SWAP:\n\t\t\tresponse = BobTxForPAS(node, ETHKey, tx, data.Demands, data.BulletinFile, data.PubPath, Log)\n\t\t}\n\t} else if tx.Mode == TRANSACTION_MODE_TABLE_POD {\n\t\tswitch tx.SubMode {\n\t\tcase TRANSACTION_SUB_MODE_COMPLAINT:\n\t\t\tif tx.OT {\n\t\t\t\tresponse = BobTxForTOC(node, ETHKey, tx, data.Demands, data.Phantoms, data.BulletinFile, data.PubPath, Log)\n\t\t\t} else {\n\t\t\t\tresponse = BobTxForTC(node, ETHKey, tx, data.Demands, data.BulletinFile, data.PubPath, Log)\n\t\t\t}\n\t\tcase TRANSACTION_SUB_MODE_ATOMIC_SWAP:\n\t\t\tresponse = BobTxForTAS(node, ETHKey, tx, data.Demands, data.BulletinFile, data.PubPath, Log)\n\t\tcase TRANSACTION_SUB_MODE_VRF:\n\t\t\tif tx.OT {\n\t\t\t\tresponse = BobTxForTOQ(node, ETHKey, tx, data.KeyName, data.KeyValue, data.PhantomKeyValue, data.BulletinFile, data.PubPath, Log)\n\t\t\t} else {\n\t\t\t\tresponse = BobTxForTQ(node, ETHKey, tx, data.KeyName, data.KeyValue, data.BulletinFile, data.PubPath, Log)\n\t\t\t}\n\t\t}\n\t}\n\tvar resp Response\n\terr = json.Unmarshal([]byte(response), &resp)\n\tif err != nil {\n\t\tLog.Warnf(\"failed to parse response. response=%v, err=%v\", response, err)\n\t\tfmt.Fprintf(w, RESPONSE_FAILED_TO_RESPONSE)\n\t\treturn\n\t}\n\tif resp.Code == \"0\" {\n\t\tplog.Result = LOG_RESULT_SUCCESS\n\t}\n\tLog.Debugf(\"[%v]the transaction finish. merkel root=%v, response=%v\", params.SessionID, bulletin.SigmaMKLRoot, response)\n\tfmt.Fprintf(w, response)\n\treturn\n}", "func (p *Politeiawww) handleInvoiceDetails(w http.ResponseWriter, r *http.Request) {\n\tlog.Tracef(\"handleInvoiceDetails\")\n\n\t// Get the invoice details command\n\tvar pd cms.InvoiceDetails\n\t// get version from query string parameters\n\terr := util.ParseGetParams(r, &pd)\n\tif err != nil {\n\t\tRespondWithError(w, r, 0, \"handleInvoiceDetails: ParseGetParams\",\n\t\t\twww.UserError{\n\t\t\t\tErrorCode: www.ErrorStatusInvalidInput,\n\t\t\t})\n\t\treturn\n\t}\n\n\t// Get invoice token from path parameters\n\tpathParams := mux.Vars(r)\n\tpd.Token = pathParams[\"token\"]\n\n\tuser, err := p.sessions.GetSessionUser(w, r)\n\tif err != nil {\n\t\tif !errors.Is(err, sessions.ErrSessionNotFound) {\n\t\t\tRespondWithError(w, r, 0,\n\t\t\t\t\"handleInvoiceDetails: getSessionUser %v\", err)\n\t\t\treturn\n\t\t}\n\t}\n\treply, err := p.processInvoiceDetails(pd, user)\n\tif err != nil {\n\t\tRespondWithError(w, r, 0,\n\t\t\t\"handleInvoiceDetails: processInvoiceDetails %v\", err)\n\t\treturn\n\t}\n\n\t// Reply with the proposal details.\n\tutil.RespondWithJSON(w, http.StatusOK, reply)\n}", "func (p *Politeiawww) handlePayInvoices(w http.ResponseWriter, r *http.Request) {\n\tlog.Tracef(\"handlePayInvoices\")\n\n\tuser, err := p.sessions.GetSessionUser(w, r)\n\tif err != nil {\n\t\tRespondWithError(w, r, 0,\n\t\t\t\"handlePayInvoices: getSessionUser %v\", err)\n\t\treturn\n\t}\n\n\treply, err := p.processPayInvoices(r.Context(), user)\n\tif err != nil {\n\t\tRespondWithError(w, r, 0, \"handlePayInvoices: processPayInvoices %v\",\n\t\t\terr)\n\t\treturn\n\t}\n\n\tutil.RespondWithJSON(w, http.StatusOK, reply)\n}", "func (_obj *DoPayment) Dispatch(ctx context.Context, _val interface{}, req *requestf.RequestPacket, resp *requestf.ResponsePacket, withContext bool) (err error) {\n\t_is := codec.NewReader(tools.Int8ToByte(req.SBuffer))\n\t_os := codec.NewBuffer()\n\tswitch req.SFuncName {\n\tcase \"echoHello\":\n\t\terr := echoHello(ctx, _val, _os, _is, withContext)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\tcase \"health\":\n\t\terr := health(ctx, _val, _os, _is, withContext)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\tcase \"authorise\":\n\t\terr := authorise(ctx, _val, _os, _is, withContext)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\tdefault:\n\t\treturn fmt.Errorf(\"func mismatch\")\n\t}\n\tvar _status map[string]string\n\ts, ok := current.GetResponseStatus(ctx)\n\tif ok && s != nil {\n\t\t_status = s\n\t}\n\tvar _context map[string]string\n\tc, ok := current.GetResponseContext(ctx)\n\tif ok && c != nil {\n\t\t_context = c\n\t}\n\t*resp = requestf.ResponsePacket{\n\t\tIVersion: 1,\n\t\tCPacketType: 0,\n\t\tIRequestId: req.IRequestId,\n\t\tIMessageType: 0,\n\t\tIRet: 0,\n\t\tSBuffer: tools.ByteToInt8(_os.ToBytes()),\n\t\tStatus: _status,\n\t\tSResultDesc: \"\",\n\t\tContext: _context,\n\t}\n\treturn nil\n}", "func (h CreatePaymentRequestHandler) Handle(params paymentrequestop.CreatePaymentRequestParams) middleware.Responder {\n\t// TODO: authorization to create payment request\n\n\treturn h.AuditableAppContextFromRequestWithErrors(params.HTTPRequest,\n\t\tfunc(appCtx appcontext.AppContext) (middleware.Responder, error) {\n\n\t\t\tpayload := params.Body\n\t\t\tif payload == nil {\n\t\t\t\terr := apperror.NewBadDataError(\"Invalid payment request: params Body is nil\")\n\t\t\t\terrPayload := payloads.ClientError(handlers.SQLErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\t\t\t\tappCtx.Logger().Error(err.Error(), zap.Any(\"payload\", errPayload))\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestBadRequest().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tappCtx.Logger().Info(\"primeapi.CreatePaymentRequestHandler info\", zap.String(\"pointOfContact\", params.Body.PointOfContact))\n\n\t\t\tmoveTaskOrderIDString := payload.MoveTaskOrderID.String()\n\t\t\tmtoID, err := uuid.FromString(moveTaskOrderIDString)\n\t\t\tif err != nil {\n\t\t\t\tappCtx.Logger().Error(\"Invalid payment request: params MoveTaskOrderID cannot be converted to a UUID\",\n\t\t\t\t\tzap.String(\"MoveTaskOrderID\", moveTaskOrderIDString), zap.Error(err))\n\t\t\t\t// create a custom verrs for returning a 422\n\t\t\t\tverrs :=\n\t\t\t\t\t&validate.Errors{Errors: map[string][]string{\n\t\t\t\t\t\t\"move_id\": {\"id cannot be converted to UUID\"},\n\t\t\t\t\t},\n\t\t\t\t\t}\n\t\t\t\terrPayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tisFinal := false\n\t\t\tif payload.IsFinal != nil {\n\t\t\t\tisFinal = *payload.IsFinal\n\t\t\t}\n\n\t\t\tpaymentRequest := models.PaymentRequest{\n\t\t\t\tIsFinal: isFinal,\n\t\t\t\tMoveTaskOrderID: mtoID,\n\t\t\t}\n\n\t\t\t// Build up the paymentRequest.PaymentServiceItems using the incoming payload to offload Swagger data coming\n\t\t\t// in from the API. These paymentRequest.PaymentServiceItems will be used as a temp holder to process the incoming API data\n\t\t\tvar verrs *validate.Errors\n\t\t\tpaymentRequest.PaymentServiceItems, verrs, err = h.buildPaymentServiceItems(appCtx, payload)\n\n\t\t\tif err != nil || verrs.HasAny() {\n\n\t\t\t\tappCtx.Logger().Error(\"could not build service items\", zap.Error(err))\n\t\t\t\t// TODO: do not bail out before creating the payment request, we need the failed record\n\t\t\t\t// we should create the failed record and store it as failed with a rejection\n\t\t\t\terrPayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tcreatedPaymentRequest, err := h.PaymentRequestCreator.CreatePaymentRequestCheck(appCtx, &paymentRequest)\n\t\t\tif err != nil {\n\t\t\t\tappCtx.Logger().Error(\"Error creating payment request\", zap.Error(err))\n\t\t\t\tswitch e := err.(type) {\n\t\t\t\tcase apperror.InvalidCreateInputError:\n\t\t\t\t\tverrs := e.ValidationErrors\n\t\t\t\t\tdetail := err.Error()\n\t\t\t\t\tpayload := payloads.ValidationError(detail, h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(payload), err\n\n\t\t\t\tcase apperror.NotFoundError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.NotFoundMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestNotFound().WithPayload(payload), err\n\t\t\t\tcase apperror.ConflictError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.ConflictErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestConflict().WithPayload(payload), err\n\t\t\t\tcase apperror.InvalidInputError:\n\t\t\t\t\tpayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), &validate.Errors{})\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(payload), err\n\t\t\t\tcase apperror.QueryError:\n\t\t\t\t\tif e.Unwrap() != nil {\n\t\t\t\t\t\t// If you can unwrap, log the internal error (usually a pq error) for better debugging\n\t\t\t\t\t\tappCtx.Logger().Error(\"primeapi.CreatePaymentRequestHandler query error\", zap.Error(e.Unwrap()))\n\t\t\t\t\t}\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestInternalServerError().WithPayload(\n\t\t\t\t\t\tpayloads.InternalServerError(nil, h.GetTraceIDFromRequest(params.HTTPRequest))), err\n\n\t\t\t\tcase *apperror.BadDataError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.BadRequestErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestBadRequest().WithPayload(payload), err\n\t\t\t\tdefault:\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestInternalServerError().WithPayload(\n\t\t\t\t\t\tpayloads.InternalServerError(nil, h.GetTraceIDFromRequest(params.HTTPRequest))), err\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturnPayload := payloads.PaymentRequest(createdPaymentRequest)\n\t\t\tappCtx.Logger().Info(\"Successful payment request creation for mto ID\", zap.String(\"moveID\", moveTaskOrderIDString))\n\t\t\treturn paymentrequestop.NewCreatePaymentRequestCreated().WithPayload(returnPayload), nil\n\t\t})\n}", "func ProcessPaymentRequested(ctx fsm.Context, environment ClientDealEnvironment, deal rm.ClientDealState) error {\n\t// If the unseal payment hasn't been made, we need to send funds\n\tif deal.UnsealPrice.GreaterThan(deal.UnsealFundsPaid) {\n\t\tlog.Debugf(\"client: payment needed: unseal price %d > unseal paid %d\",\n\t\t\tdeal.UnsealPrice, deal.UnsealFundsPaid)\n\t\treturn ctx.Trigger(rm.ClientEventSendFunds)\n\t}\n\n\t// If all bytes received have been paid for, we don't need to send funds\n\tif deal.BytesPaidFor >= deal.TotalReceived {\n\t\tlog.Debugf(\"client: no payment needed: bytes paid for %d >= bytes received %d\",\n\t\t\tdeal.BytesPaidFor, deal.TotalReceived)\n\t\treturn nil\n\t}\n\n\t// Not all bytes received have been paid for\n\n\t// If all blocks have been received we need to send a final payment\n\tif deal.AllBlocksReceived {\n\t\tlog.Debugf(\"client: payment needed: all blocks received, bytes paid for %d < bytes received %d\",\n\t\t\tdeal.BytesPaidFor, deal.TotalReceived)\n\t\treturn ctx.Trigger(rm.ClientEventSendFunds)\n\t}\n\n\t// Payments are made in intervals, as bytes are received from the provider.\n\t// If the number of bytes received is at or above the size of the current\n\t// interval, we need to send a payment.\n\tif deal.TotalReceived >= deal.CurrentInterval {\n\t\tlog.Debugf(\"client: payment needed: bytes received %d >= interval %d, bytes paid for %d < bytes received %d\",\n\t\t\tdeal.TotalReceived, deal.CurrentInterval, deal.BytesPaidFor, deal.TotalReceived)\n\t\treturn ctx.Trigger(rm.ClientEventSendFunds)\n\t}\n\n\tlog.Debugf(\"client: no payment needed: received %d < interval %d (paid for %d)\",\n\t\tdeal.TotalReceived, deal.CurrentInterval, deal.BytesPaidFor)\n\treturn nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
handleGeneratetokenized for receive and handle the request from client
func handleDBGeneratetokenized(w http.ResponseWriter, r *http.Request) { defer func() { db.Connection.Close(nil) }() var requestData modelito.RequestTokenized var errorGeneral string var errorGeneralNbr string errorGeneral="" requestData,errorGeneral =obtainParmsGeneratetokenized(r,errorGeneral) ////////////////////////////////////////////////validate parms /// START if errorGeneral=="" { errorGeneral,errorGeneralNbr= ProcessGeneratetokenized(w , requestData) } if errorGeneral!=""{ //send error response if any //prepare an error JSON Response, if any log.Print("CZ STEP Get the ERROR response JSON ready") /// START fieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr) ////////// write the response (ERROR) w.Header().Set("Content-Type", "application/json") w.Write(fieldDataBytesJson) if(err!=nil){ } } }
[ "func generateHandler(w http.ResponseWriter, r *http.Request) {\n\tvar err error\n\n\t// Default length for the body to generate.\n\ttokenLen := 50\n\n\tif r.URL.Query().Get(\"limit\") != \"\" {\n\t\ttokenLen, err = strconv.Atoi(r.URL.Query().Get(\"limit\"))\n\t\tif err != nil {\n\t\t\terrHandler(w, 500, err)\n\t\t}\n\t}\n\n\tout, err := index.Babble(\"\", tokenLen) // Starting seed is left blank for random choice.\n\tif err != nil {\n\t\tif err == ngrams.ErrEmptyIndex {\n\t\t\tm, err := json.Marshal(map[string]interface{}{\n\t\t\t\t\"err\": \"index is empty; please learn ngrams before generating.\",\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\terrHandler(w, 400, err)\n\t\t\t}\n\n\t\t\tw.Write(m)\n\t\t\treturn\n\t\t}\n\n\t\terrHandler(w, 500, err)\n\t}\n\n\tm, err := json.Marshal(map[string]interface{}{\n\t\t\"body\": out,\n\t\t\"limit\": tokenLen,\n\t})\n\tif err != nil {\n\t\terrHandler(w, 500, err)\n\t}\n\n\tw.Write(m)\n\n}", "func TokenizeHandler(request events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// get pan\n\t// tokenize\n\t// store in db\n\t// return token\n\n\treturn events.APIGatewayProxyResponse{\n\t\tBody: \"Tokenize\",\n\t\tStatusCode: 200,\n\t}, nil\n}", "func (o *oauth) tokenHandler(w http.ResponseWriter, r *http.Request) {\n\tw = &rememberingWriter{ResponseWriter: w}\n\n\t// This block is copied from o.server.HandleTokenRequest\n\t// We needed to inspect what's going on a bit.\n\tgt, tgr, verr := o.server.ValidationTokenRequest(r)\n\tif verr != nil {\n\t\tencodeError(w, verr)\n\t\treturn\n\t}\n\tti, verr := o.server.GetAccessToken(gt, tgr)\n\tif verr != nil {\n\t\tencodeError(w, verr)\n\t\treturn\n\t}\n\tdata := o.server.GetTokenData(ti)\n\tbs, err := json.Marshal(data)\n\tif err != nil {\n\t\tencodeError(w, err)\n\t\treturn\n\t}\n\t// (end of copy)\n\n\t// HandleTokenRequest currently returns nil even if the token request\n\t// failed. That menas we can't clearly know if token generation passed or failed.\n\t//\n\t// So we need to find out if an error is written, which we can\n\t// infer by w.WriteHeader call (a 4xx or 5xx status code).\n\tif ww, ok := w.(*rememberingWriter); ok && ww.statusCode > 400 { // wrote error\n\t\ttokenGenerations.Add(1)\n\t\tw.Header().Set(\"X-User-Id\", ti.GetUserID()) // only on non-errors\n\t}\n\n\t// Write our response\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\tw.WriteHeader(http.StatusOK)\n\tw.Write(bs)\n}", "func GenerateToken(s *Server) gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tvar data TokenParameter\n\n\t\tif err := c.BindJSON(&data); err != nil {\n\t\t\tc.JSON(http.StatusBadRequest, gin.H{\"status\": \"JSON Body is missing fields\"})\n\t\t\treturn\n\t\t}\n\n\t\tif err := data.Validate(); err != nil {\n\t\t\tc.JSON(http.StatusBadRequest, gin.H{\"status\": \"JSON Body has invalid data\"})\n\t\t\treturn\n\t\t}\n\n\t\tdeviceId := GetDeviceId(data.Device.Serial)\n\t\ttokenStr := GetTokenString(deviceId)\n\n\t\tif _, err := s.Redis.Do(\"SETEX\", tokenStr, LocalConfig.tokenLifetime, tokenStr); err != nil {\n\t\t\tc.JSON(http.StatusInternalServerError, gin.H{\"status\": \"Internal error\"})\n\t\t\treturn\n\t\t}\n\n\t\tc.JSON(http.StatusOK, gin.H{\n\t\t\t\"deviceid\": deviceId,\n\t\t\t\"token\": tokenStr,\n\t\t\t\"ttl\": LocalConfig.tokenLifetime,\n\t\t})\n\t}\n}", "func GenAuthTokenHandler(c *gin.Context) {\r\n\t// Create a new token object, specifying signing method and the claims\r\n\t// you would like it to contain.\r\n\r\n\ttoken := jwt.NewWithClaims(jwt.SigningMethodHS256, jwt.MapClaims{\r\n\t\t\"foo\": \"bar\",\r\n\t\t\"expire\": func() int64 {\r\n\t\t\tnow := time.Now()\r\n\t\t\tduration, _ := time.ParseDuration(\"14d\")\r\n\t\t\tm1 := now.Add(duration)\r\n\t\t\treturn m1.Unix()\r\n\t\t}(),\r\n\t})\r\n\r\n\t// Sign and get the complete encoded token as a string using the secret\r\n\ttokenString, err := token.SignedString([]byte(utils.AppConfig.Server.SecretKey))\r\n\r\n\tfmt.Println(tokenString, err)\r\n\tc.String(http.StatusOK, tokenString)\r\n}", "func TokenCreateHandler(ctx *gin.Context) {\n\tvar (\n\t\tinput *tokenCreateInput\n\t\tdb *gorm.DB\n\t\tapp *models.App\n\t\ttokenCreateSrv *service.TokenCreate\n\t\treadOnlyI8 int8\n\t\ttokenCreateValue interface{}\n\t\terr error\n\n\t\tcode = 400\n\t\treErrors map[string][]string\n\t\tsuccess bool\n\t\tdata interface{}\n\t)\n\n\tdefer func() {\n\t\tctx.JSON(code, &Response{\n\t\t\tRequestID: ctx.GetInt64(\"requestId\"),\n\t\t\tSuccess: success,\n\t\t\tErrors: reErrors,\n\t\t\tData: data,\n\t\t})\n\t}()\n\n\tinput = ctx.MustGet(\"inputParam\").(*tokenCreateInput)\n\tdb = ctx.MustGet(\"db\").(*gorm.DB)\n\tapp = ctx.MustGet(\"app\").(*models.App)\n\n\tif input.ReadOnly != nil && *input.ReadOnly {\n\t\treadOnlyI8 = 1\n\t}\n\n\ttokenCreateSrv = &service.TokenCreate{\n\t\tBaseService: service.BaseService{\n\t\t\tDB: db,\n\t\t},\n\t\tIP: input.IP,\n\t\tApp: app,\n\t\tPath: *input.Path,\n\t\tSecret: input.Secret,\n\t\tReadOnly: readOnlyI8,\n\t\tExpiredAt: input.ExpiredAt,\n\t\tAvailableTimes: *input.AvailableTimes,\n\t}\n\n\tif err := tokenCreateSrv.Validate(); !reflect.ValueOf(err).IsNil() {\n\t\treErrors = generateErrors(err, \"\")\n\t\treturn\n\t}\n\n\tif tokenCreateValue, err = tokenCreateSrv.Execute(context.Background()); err != nil {\n\t\treErrors = generateErrors(err, \"\")\n\t\treturn\n\t}\n\n\tdata = tokenResp(tokenCreateValue.(*models.Token))\n\tsuccess = true\n\tcode = 200\n}", "func (s *Supervisor) token(q *msg.Request) {\n\tresult := msg.FromRequest(q)\n\t// default result is for the request to fail\n\tresult.Code = 403\n\tresult.Super.Verdict = 403\n\n\t// start response delay timer\n\ttimer := time.NewTimer(1 * time.Second)\n\n\t// start assembly of auditlog entry\n\tresult.Super.Audit = s.auditLog.\n\t\tWithField(`RequestID`, q.ID.String()).\n\t\tWithField(`KexID`, q.Super.Encrypted.KexID).\n\t\tWithField(`IPAddr`, q.RemoteAddr).\n\t\tWithField(`UserName`, `AnonymousCoward`).\n\t\tWithField(`UserID`, `ffffffff-ffff-ffff-ffff-ffffffffffff`).\n\t\tWithField(`Code`, result.Code).\n\t\tWithField(`Verdict`, result.Super.Verdict).\n\t\tWithField(`Section`, q.Section).\n\t\tWithField(`Action`, q.Action).\n\t\tWithField(`Request`, fmt.Sprintf(\"%s::%s\", q.Section, q.Action)).\n\t\tWithField(`Supervisor`, fmt.Sprintf(\"%s::%s=%s\", q.Section, q.Action, q.Super.Task))\n\n\t// tokenRequest/tokenInvalidate are master instance functions\n\tif s.readonly {\n\t\tresult.ReadOnly()\n\t\tresult.Super.Audit.WithField(`Code`, result.Code).Warningln(result.Error)\n\t\tgoto returnImmediate\n\t}\n\n\t// filter requests with invalid task\n\tswitch q.Super.Task {\n\tcase msg.TaskRequest:\n\tcase msg.TaskInvalidateGlobal:\n\tcase msg.TaskInvalidateAccount:\n\tcase msg.TaskInvalidate:\n\tdefault:\n\t\tresult.UnknownTask(q)\n\t\tresult.Super.Audit.WithField(`Code`, result.Code).Warningln(result.Error)\n\t\tgoto returnImmediate\n\t}\n\n\t// select correct taskhandler\n\tswitch q.Super.Task {\n\tcase msg.TaskRequest:\n\t\ts.tokenRequest(q, &result)\n\tcase msg.TaskInvalidateGlobal:\n\t\ts.tokenInvalidateGlobal(q, &result)\n\tcase msg.TaskInvalidateAccount:\n\t\ts.tokenInvalidateAccount(q, &result)\n\tcase msg.TaskInvalidate:\n\t\ts.tokenInvalidate(q, &result)\n\t}\n\n\t// wait for delay timer to trigger\n\t<-timer.C\n\nreturnImmediate:\n\t// cleanup delay timer\n\tif timer.Stop() {\n\t\t<-timer.C\n\t}\n\tq.Reply <- result\n}", "func GenerateToken(res http.ResponseWriter, req *http.Request) {\n\tlog.Printf(\"GenerateToken Received request: %v\", req)\n\t// only valid with PUT\n\tif req.Method != http.MethodPut {\n\t\treturn\n\t}\n\t// check that header contains valid ttl\n\trequestedTTL := req.Header.Get(tokenTTLHeader)\n\tvalidTTL, err := extractValidTTL(requestedTTL)\n\tif err != nil {\n\t\tlog.Printf(\"Something went wrong with ttl validation: %v with requested TTL: %v\", err.Error(), requestedTTL)\n\t\tserver.ReturnBadRequestResponse(res)\n\t\treturn\n\t}\n\n\tkey := make([]byte, 32)\n\t_, err = rand.Read(key)\n\tif err != nil {\n\t\tserver.FormatAndReturnTextResponse(res, \"Something went wrong with token creation\")\n\t\treturn\n\t}\n\n\ttokenValue := base64.StdEncoding.EncodeToString(key)\n\ttoken := v2Token{\n\t\tValue: tokenValue,\n\t\tTTL: validTTL,\n\t\tCreatedAt: time.Now(),\n\t}\n\tgeneratedTokens[token.Value] = token\n\tres.Header().Set(tokenTTLHeader, strconv.Itoa(token.TTL))\n\tserver.FormatAndReturnTextResponse(res, token.Value)\n}", "func (r *oauthProxy) tokenHandler(w http.ResponseWriter, req *http.Request) {\n\tuser, err := r.getIdentity(req)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Write(user.token.Payload)\n}", "func (th *tmplHandler) getTokenHandler(w http.ResponseWriter, r *http.Request) {\n\tw.WriteHeader(http.StatusOK)\n\tif err := th.ExecuteTemplate(w, \"token_form.tmpl\", nil); err != nil {\n\t\tlog.Warnf(\"could not execute token tmpl for get request: %s\", err)\n\t\trenderError(w, th.Lookup(\"errors.tmpl\"), responseError{})\n\t}\n}", "func HandleMytokenFromTransferCode(ctx *fiber.Ctx) *model.Response {\n\trlog := logger.GetRequestLogger(ctx)\n\trlog.Debug(\"Handle mytoken from transfercode\")\n\treq := response.NewExchangeTransferCodeRequest()\n\tif err := errors.WithStack(json.Unmarshal(ctx.Body(), &req)); err != nil {\n\t\treturn model.ErrorToBadRequestErrorResponse(err)\n\t}\n\trlog.Trace(\"Parsed request\")\n\tvar errorRes *model.Response = nil\n\tvar tokenStr string\n\tif err := db.Transact(\n\t\trlog, func(tx *sqlx.Tx) error {\n\t\t\tstatus, err := transfercoderepo.CheckTransferCode(rlog, tx, req.TransferCode)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif !status.Found {\n\t\t\t\terrorRes = &model.Response{\n\t\t\t\t\tStatus: fiber.StatusUnauthorized,\n\t\t\t\t\tResponse: api.ErrorBadTransferCode,\n\t\t\t\t}\n\t\t\t\treturn errors.New(errResPlaceholder)\n\t\t\t}\n\t\t\tif status.Expired {\n\t\t\t\terrorRes = &model.Response{\n\t\t\t\t\tStatus: fiber.StatusUnauthorized,\n\t\t\t\t\tResponse: api.ErrorTransferCodeExpired,\n\t\t\t\t}\n\t\t\t\treturn errors.New(errResPlaceholder)\n\t\t\t}\n\t\t\ttokenStr, err = transfercoderepo.PopTokenForTransferCode(\n\t\t\t\trlog, tx, req.TransferCode, *ctxutils.ClientMetaData(ctx),\n\t\t\t)\n\t\t\treturn err\n\t\t},\n\t); err != nil {\n\t\tif errorRes != nil {\n\t\t\treturn errorRes\n\t\t}\n\t\trlog.Errorf(\"%s\", errorfmt.Full(err))\n\t\treturn model.ErrorToInternalServerErrorResponse(err)\n\t}\n\n\ttoken, err := universalmytoken.Parse(rlog, tokenStr)\n\tif err != nil {\n\t\trlog.Errorf(\"%s\", errorfmt.Full(err))\n\t\treturn model.ErrorToBadRequestErrorResponse(err)\n\t}\n\tmt, err := mytoken.ParseJWT(token.JWT)\n\tif err != nil {\n\t\trlog.Errorf(\"%s\", errorfmt.Full(err))\n\t\treturn model.ErrorToInternalServerErrorResponse(err)\n\t}\n\treturn &model.Response{\n\t\tStatus: fiber.StatusOK,\n\t\tResponse: response.MytokenResponse{\n\t\t\tMytokenResponse: api.MytokenResponse{\n\t\t\t\tMytoken: token.OriginalToken,\n\t\t\t\tExpiresIn: mt.ExpiresIn(),\n\t\t\t\tCapabilities: mt.Capabilities,\n\t\t\t\tMOMID: mt.ID.Hash(),\n\t\t\t},\n\t\t\tMytokenType: token.OriginalTokenType,\n\t\t\tRestrictions: mt.Restrictions,\n\t\t},\n\t}\n\n}", "func (o *oauth) createTokenHandler(auth authable) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tuserId, err := auth.findUserId(extractCookie(r).Value)\n\t\tif err != nil {\n\t\t\t// user not found, return\n\t\t\tw.WriteHeader(http.StatusForbidden)\n\t\t\treturn\n\t\t}\n\n\t\trecords, err := o.clientStore.GetByUserID(userId)\n\t\tif err != nil && !strings.Contains(err.Error(), \"not found\") {\n\t\t\tinternalError(w, err, \"oauth\")\n\t\t\treturn\n\t\t}\n\t\tif len(records) == 0 { // nothing found, so fake one\n\t\t\trecords = append(records, &models.Client{})\n\t\t}\n\n\t\tclients := make([]*models.Client, len(records))\n\t\tfor i := range records {\n\t\t\terr = o.clientStore.DeleteByID(records[i].GetID())\n\t\t\tif err != nil && !strings.Contains(err.Error(), \"not found\") {\n\t\t\t\tinternalError(w, err, \"oauth\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tclients[i] = &models.Client{\n\t\t\t\tID: generateID()[:12],\n\t\t\t\tSecret: generateID(),\n\t\t\t\tDomain: Domain,\n\t\t\t\tUserID: userId,\n\t\t\t}\n\n\t\t\t// Write client into oauth clients db.\n\t\t\tif err := o.clientStore.Set(clients[i].GetID(), clients[i]); err != nil {\n\t\t\t\tinternalError(w, err, \"oauth\")\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t// metrics\n\t\tclientGenerations.Add(1)\n\n\t\t// render back new client info\n\t\ttype response struct {\n\t\t\tClients []*models.Client `json:\"clients\"`\n\t\t}\n\t\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\t\tif err := json.NewEncoder(w).Encode(&response{clients}); err != nil {\n\t\t\tinternalError(w, err, \"oauth\")\n\t\t\treturn\n\t\t}\n\t}\n}", "func (s *Server) handleToken(w http.ResponseWriter, req *http.Request) error {\n\tsession, err := s.cookieStore.Get(req, UserSessionName)\n\tif err != nil {\n\t\treturn err\n\t}\n\tw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\temail, ok := session.Values[\"email\"]\n\tfill := &tokenFill{}\n\tif ok {\n\t\temailStr, ok := email.(string)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"unexpected session value type %T\", email)\n\t\t}\n\t\tfill.Email = emailStr\n\t\tif s.opts.UseJWT {\n\t\t\ttoken, err := s.GetJWT(emailStr)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tfill.Token = token\n\t\t}\n\t}\n\treturn tokenTmpl.Execute(w, fill)\n}", "func (c *Client) Generate(msg []byte, t *dns.TSIG) ([]byte, error) {\n\tif dns.CanonicalName(t.Algorithm) != tsig.GSS {\n\t\treturn nil, dns.ErrKeyAlg\n\t}\n\n\tc.m.RLock()\n\tdefer c.m.RUnlock()\n\n\tctx, ok := c.ctx[t.Hdr.Name]\n\tif !ok {\n\t\treturn nil, dns.ErrSecret\n\t}\n\n\ttoken := gssapi.MICToken{\n\t\tFlags: gssapi.MICTokenFlagAcceptorSubkey,\n\t\tSndSeqNum: ctx.seq,\n\t\tPayload: msg,\n\t}\n\n\tif err := token.SetChecksum(ctx.key, keyusage.GSSAPI_INITIATOR_SIGN); err != nil {\n\t\treturn nil, err\n\t}\n\n\tb, err := token.Marshal()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx.seq++\n\n\treturn b, nil\n}", "func (h *LolVideoHandler) HandleDatagenRequest() func(http.ResponseWriter, *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\tdatagenRequest := &models.DatagenRequest{}\n\t\tif err := vsjson.DecodeLimit(r.Body, 2<<20, datagenRequest); err != nil {\n\t\t\tlog.Println(err)\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\tif err := datagenRequest.Valid(coordinator.AwsRegionMap); err != nil {\n\t\t\tlog.Println(err)\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\tmessageJson, err := json.Marshal(datagenRequest)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Unable to marshal datagenRequest\", err)\n\t\t\tmessageJson = []byte{}\n\t\t}\n\t\tlog.Println(\"Executing valid datagen request\", string(messageJson))\n\n\t\treplayRequest := datagenRequest.ReplayRequest\n\t\tdisplayStr := InitConnection(replayRequest.WorkerInstance, string(messageJson))\n\t\tec2Client := coordinator.GetEc2Client(aws.MustEnvCreds(), replayRequest.Region)\n\n\t\tcommands := []*coordinator.VNCCommand{}\n\n\t\tcommands = append(commands, coordinator.CommandStartEloBuddy()...)\n\t\t// Datagen requests should terminate when the game ends.\n\t\t// This time represents a timeout in case the client hangs and never ends.\n\t\t// Extend it out a bit more generously\n\t\treplayRequest.GameLengthSeconds += 300\n\t\tcommands = append(commands, coordinator.CommandRunReplay(replayRequest, 0, false)...)\n\t\tif datagenRequest.IncreasedPlaybackSpeed {\n\t\t\tcommands = append(commands, coordinator.CommandSpeedUpReplay()...)\n\t\t}\n\n\t\t// Run all the vnc commands. This can take several minutes\n\t\t// Run it in a thread to prevent timeout\n\t\tgo func() {\n\t\t\tcoordinator.RunTaskOnWorker(displayStr, replayRequest.WorkerInstance, coordinator.VncPasswordFile, ec2Client, commands)\n\t\t\tCleanup(replayRequest.WorkerInstance, displayStr)\n\t\t}()\n\t\tw.WriteHeader(http.StatusOK)\n\t}\n}", "func generateHandler(requestChannel chan<- *keystore.Request, handler func(http.ResponseWriter, *http.Request, chan<- *keystore.Request)) http.Handler {\n\n\t// Handler that passes the requests along with the request and response\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { handler(w, r, requestChannel) })\n}", "func (p *pbft) handleClientRequest(content []byte) {\n\tfmt.Println(\"The primary node has received the request from the client.\")\n\t//The Request structure is parsed using JSON\n\tr := new(Request)\n\terr := json.Unmarshal(content, r)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\t//to add infoID\n\tp.sequenceIDAdd()\n\t//to get the digest\n\tdigest := getDigest(*r)\n\tfmt.Println(\"The request has been stored into the temporary message pool.\")\n\t//to store into the temp message pool\n\tp.messagePool[digest] = *r\n\t//to sign the digest by the primary node\n\tdigestByte, _ := hex.DecodeString(digest)\n\tsignInfo := p.RsaSignWithSha256(digestByte, p.node.rsaPrivKey)\n\t//setup PrePrepare message and send to other nodes\n\tpp := PrePrepare{*r, digest, p.sequenceID, signInfo}\n\tb, err := json.Marshal(pp)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\tfmt.Println(\"sending PrePrepare messsage to all the other nodes...\")\n\t//to send PrePrepare message to other nodes\n\tp.broadcast(cPrePrepare, b)\n\tfmt.Println(\"PrePrepare is done.\")\n}", "func (s *Server) handleCustomerGetToken(writer http.ResponseWriter, request *http.Request) {\n\tvar item *types.Auth\n\n\terr := json.NewDecoder(request.Body).Decode(&item)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\thttp.Error(writer, http.StatusText(http.StatusBadRequest), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\ttoken, err := s.customersSvc.Token(request.Context(), item.Login, item.Password)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\thttp.Error(writer, http.StatusText(http.StatusInternalServerError), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\trespondJSON(writer, &types.Token{Token: token})\n}", "func HandleMytokenFromMytoken(ctx *fiber.Ctx) *model.Response {\n\trlog := logger.GetRequestLogger(ctx)\n\trlog.Debug(\"Handle mytoken from mytoken\")\n\treq := response.NewMytokenRequest()\n\tif err := errors.WithStack(json.Unmarshal(ctx.Body(), &req)); err != nil {\n\t\treturn model.ErrorToBadRequestErrorResponse(err)\n\t}\n\tusedRestriction, mt, errRes := HandleMytokenFromMytokenReqChecks(rlog, req, ctx.IP(), ctx)\n\tif errRes != nil {\n\t\treturn errRes\n\t}\n\treturn HandleMytokenFromMytokenReq(rlog, mt, req, ctxutils.ClientMetaData(ctx), usedRestriction)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Bit returns a uint32 with vth bit set to 1.
func Bit(v int) uint32 { return uint32(1) << uint32(v) }
[ "func (d bitvector) get(bit uint32) uint {\n\n\tshift := bit % 32\n\tbb := d[bit/32]\n\tbb &= (1 << shift)\n\n\treturn uint(bb >> shift)\n}", "func (x *Int) Bit(i int) uint {}", "func bitOf(vec Flag, val Value) (uint, uint32, error) {\n\tif vec > Inheritable || val > Value(words*32) {\n\t\treturn 0, 0, ErrBadValue\n\t}\n\tu := uint(val)\n\treturn u / 32, uint32(1) << (u % 32), nil\n}", "func IntBit(x *big.Int, i int) uint", "func MSB32(x uint32) uint32", "func Val(value byte, bit byte) byte {\n\treturn (value >> bit) & 1\n}", "func (b Bits) Bit(n int) int {\n\tif n < 0 || n >= b.Num {\n\t\tpanic(\"bit number out of range\")\n\t}\n\treturn int(b.Bits[n>>6] >> uint(n&63) & 1)\n}", "func OnBit(num int, nth int) int {\n\treturn num | (1 << uint(nth))\n}", "func OnBit(num, nth int) int {\n\treturn num | (1 << uint(nth))\n}", "func Uint32() uint32", "func (bitmap *bitmap) Bit(index int) int {\n\tif index >= bitmap.Size {\n\t\tpanic(\"index out of range\")\n\t}\n\n\tdiv, mod := index/8, index%8\n\treturn int((uint(bitmap.data[div]) & (1 << uint(7-mod))) >> uint(7-mod))\n}", "func Bitno(b uint64) int", "func setBit(bitmap uint32, index uint) uint32 {\n\treturn (bitmap | (1 << index))\n}", "func GetBit(h []byte, i uint) int {\n\n\tsByte := int(h[i>>3])\n\treturn (sByte >> (i & 7)) & 1\n}", "func getBitValue(bitHolder uint64, bitPos int) int {\n\tsetBit := bitHolder & (1 << bitPos)\n\tsetBit = setBit >> bitPos\n\treturn int(setBit)\n}", "func (d bitvector) set(bit uint32) {\n\td[bit/32] |= (1 << (bit % 32))\n}", "func getBit(index index, position uint8) uint8 {\n\treturn uint8((uint64(index) >> position) & 1)\n}", "func refBit(val int, i uint) int {\n\treturn (val >> i) & 1\n}", "func bit(cpu *CPU, r, b byte) {\n\tbit := (r>>b)&1 == 0\n\tcpu.SetZero(bit)\n\tcpu.SetNegative(false)\n\tcpu.SetHalfCarry(true)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SetPostviewImageSize sets the Post View image size for the camera: The possible options are: "2M" a smaller preview, usually 2Megpixels in size, sometimes not camera dependant "Original" the size of the image taken
func (c *Camera) SetPostviewImageSize(size PostViewSize) (err error) { _, err = c.newRequest(endpoints.Camera, "setPostviewImageSize", size).Do() return }
[ "func (c *Camera) GetPostviewImageSize() (size string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getPostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) > 0 {\n\t\terr = json.Unmarshal(resp.Result[0], &size)\n\t}\n\n\treturn\n}", "func (c *Camera) SetImageSize(width int, height int) (err error) {\n\tc.imageWidth = width\n\tc.imageHeight = height\n\n\terr = c.Lens.setAspectRatio(float64(width) / float64(height))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tc.output = image.NewRGBA(image.Rect(0, 0, c.imageWidth, c.imageHeight))\n\treturn\n}", "func (c *Camera) GetSupportedPostviewImageSize() (sizes []string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getSupportedPostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) > 0 {\n\t\terr = json.Unmarshal(resp.Result[0], &sizes)\n\t}\n\n\treturn\n}", "func (canvas *Canvas) SetSize(width, height Unit) {\n\tcanvas.page.MediaBox = Rectangle{Point{0, 0}, Point{width, height}}\n}", "func (c *Camera) GetAvailablePostviewImageSize() (current string, available []string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getAvailablePostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) >= 1 {\n\t\t// Current size\n\t\tif err := json.Unmarshal(resp.Result[0], &current); err != nil {\n\t\t\treturn current, available, err\n\t\t}\n\n\t\t// Available sizes\n\t\tif err := json.Unmarshal(resp.Result[1], &available); err != nil {\n\t\t\treturn current, available, err\n\t\t}\n\t}\n\n\treturn\n}", "func (c *Client) RenterSetStreamCacheSizePost(cacheSize uint64) (err error) {\n\tvalues := url.Values{}\n\tvalues.Set(\"streamcachesize\", strconv.FormatUint(cacheSize, 10))\n\terr = c.post(\"/renter\", values.Encode(), nil)\n\treturn\n}", "func (r *FakeImageService) SetFakeImageSize(size uint64) {\n\tr.Lock()\n\tdefer r.Unlock()\n\n\tr.FakeImageSize = size\n}", "func qr_decoder_set_image_size(p _QrDecoderHandle, width, height, depth, channel int) _QrDecoderHandle {\n\tv := C.qr_decoder_set_image_size(C.QrDecoderHandle(p),\n\t\tC.int(width), C.int(height), C.int(depth), C.int(channel),\n\t)\n\treturn _QrDecoderHandle(v)\n}", "func (b *builder) SetSize(width, height pic.Twiplet, dpi int32) {\n\tb.width = width.Pixels(dpi)\n\tb.height = height.Pixels(dpi)\n\tb.dpi = float64(dpi)\n}", "func (self *RenderWindow) SetSize(width uint, height uint) void {\n return C.sfRenderWindow_SetSize(self.Cref, width, height)\n}", "func (w *WebGLRenderTarget) SetSize(width, height float64) *WebGLRenderTarget {\n\tw.p.Call(\"setSize\", width, height)\n\treturn w\n}", "func (r *ImageRef) SetPageHeight(height int) error {\n\tout, err := vipsCopyImage(r.image)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvipsSetPageHeight(out, height)\n\n\tr.setImage(out)\n\treturn nil\n}", "func (kf *KalmanFilter) SetStatePost(statePost Mat) {\n\tC.KalmanFilter_SetStatePost(kf.p, statePost.p)\n}", "func (b *GoGLBackendOffscreen) SetSize(w, h int) {\n\tb.GoGLBackend.SetBounds(0, 0, w, h)\n\tb.offscrImg.w = b.offscrBuf.w\n\tb.offscrImg.h = b.offscrBuf.h\n}", "func (v *Video) SetSize(width int, height int) {\n\tv.width = width\n\tv.height = height\n\tv.filters = append(v.filters, fmt.Sprintf(\"scale=%d:%d\", width, height))\n}", "func (m *PlannerTaskDetails) SetPreviewType(value *PlannerPreviewType)() {\n err := m.GetBackingStore().Set(\"previewType\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *ServiceHealthIssuePost) SetPostType(value *PostType)() {\n err := m.GetBackingStore().Set(\"postType\", value)\n if err != nil {\n panic(err)\n }\n}", "func (sf *TWindow) SetSize(sizeX size.ASizeX, sizeY size.ASizeY) {\n\tif sizeX < 0 {\n\t\tsizeX = 0\n\t}\n\tif sizeY < 0 {\n\t\tsizeY = 0\n\t}\n\tsf.chSize <- size.NewSize(sizeX, sizeY)\n}", "func (pu *PostUpdate) SetViewCount(i int) *PostUpdate {\n\tpu.mutation.ResetViewCount()\n\tpu.mutation.SetViewCount(i)\n\treturn pu\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetPostviewImageSize obtains the current Post View Image size from the camera: The possible options are: "2M" a smaller preview, usually 2Megpixels in size, sometimes not camera dependant "Original" the size of the image taken
func (c *Camera) GetPostviewImageSize() (size string, err error) { resp, err := c.newRequest(endpoints.Camera, "getPostviewImageSize").Do() if err != nil { return } if len(resp.Result) > 0 { err = json.Unmarshal(resp.Result[0], &size) } return }
[ "func (c *Camera) GetSupportedPostviewImageSize() (sizes []string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getSupportedPostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) > 0 {\n\t\terr = json.Unmarshal(resp.Result[0], &sizes)\n\t}\n\n\treturn\n}", "func (c *Camera) GetAvailablePostviewImageSize() (current string, available []string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getAvailablePostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) >= 1 {\n\t\t// Current size\n\t\tif err := json.Unmarshal(resp.Result[0], &current); err != nil {\n\t\t\treturn current, available, err\n\t\t}\n\n\t\t// Available sizes\n\t\tif err := json.Unmarshal(resp.Result[1], &available); err != nil {\n\t\t\treturn current, available, err\n\t\t}\n\t}\n\n\treturn\n}", "func (c *Camera) SetPostviewImageSize(size PostViewSize) (err error) {\n\t_, err = c.newRequest(endpoints.Camera, \"setPostviewImageSize\", size).Do()\n\treturn\n}", "func (f *XtermImage) GetSize() (width, height int) {\n\twidth = f.timage.Bounds().Max.X\n\theight = f.timage.Bounds().Max.Y\n\n\treturn\n}", "func getOriginalSizeUrl(flickrOauth FlickrOAuth, photo Photo) (string, string) {\n\n\tif photo.Media == \"photo\" {\n\t\treturn photo.OriginalUrl, \"\"\n\t}\n\n\textras := map[string]string{\"photo_id\": photo.Id}\n\n\tvar err error\n\tvar body []byte\n\n\tbody, err = makeGetRequest(func() string { return generateOAuthUrl(apiBaseUrl, \"flickr.photos.getSizes\", flickrOauth, extras) })\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tresponse := PhotoSizeResponse{}\n\terr = xml.Unmarshal(body, &response)\n\tif err != nil {\n\t\tlogMessage(\"Could not unmarshal body, check logs for body detail.\", true)\n\t\tlogMessage(string(body), false)\n\t\treturn \"\", \"\"\n\t}\n\n\tphotoUrl := \"\"\n\tvideoUrl := \"\"\n\tfor _, v := range response.SizesContainer.Sizes {\n\t\tif v.Label == \"Original\" {\n\t\t\tphotoUrl = v.Url\n\t\t}\n\n\t\tif v.Label == \"Video Original\" {\n\t\t\tvideoUrl = v.Url\n\t\t}\n\t}\n\n\treturn photoUrl, videoUrl\n}", "func (canvas *Canvas) Size() (width, height Unit) {\n\tmbox := canvas.page.MediaBox\n\treturn mbox.Dx(), mbox.Dy()\n}", "func (m *Manager) GetImageSize(hash string) (int64, error) {\n\tpath := filepath.Join(m.Options.Directory, hash)\n\tif _, err := os.Stat(path); err != nil {\n\t\treturn 0, fmt.Errorf(\"failed to locate image path: %v\", err)\n\t}\n\n\t// FIXME need a real way to do this\n\treturn 0, nil\n}", "func (image *JPGImage) GetFileSize() uint64 {\n\treturn uint64(len(image.data))\n}", "func (is ImageSurface) Size() Point {\n\treturn Point{float64(is.width), float64(is.height)}\n}", "func (tc *taskContext) getPostTimeout() time.Duration {\n\ttc.RLock()\n\tdefer tc.RUnlock()\n\n\tif tc.taskConfig != nil && tc.taskConfig.Project != nil && tc.taskConfig.Project.PostTimeoutSecs != 0 {\n\t\treturn time.Duration(tc.taskConfig.Project.PostTimeoutSecs) * time.Second\n\t}\n\treturn defaultPostTimeout\n}", "func (hmd *Hmd) GetFovTextureSize(eye EyeType, fov FovPort, pixelsPerDisplayPixel float32) Sizei {\n\tvar cFov C.ovrFovPort\n\tcFov.DownTan = C.float(fov.DownTan)\n\tcFov.LeftTan = C.float(fov.LeftTan)\n\tcFov.RightTan = C.float(fov.RightTan)\n\tcFov.UpTan = C.float(fov.UpTan)\n\treturn sizei(C.ovrHmd_GetFovTextureSize(hmd.cptr(), C.ovrEyeType(eye), cFov, C.float(pixelsPerDisplayPixel)))\n}", "func qr_decoder_set_image_size(p _QrDecoderHandle, width, height, depth, channel int) _QrDecoderHandle {\n\tv := C.qr_decoder_set_image_size(C.QrDecoderHandle(p),\n\t\tC.int(width), C.int(height), C.int(depth), C.int(channel),\n\t)\n\treturn _QrDecoderHandle(v)\n}", "func (img *Image) Size() (int, int) { return img.img.Size() }", "func GetPageSize(reqSize uint32) int {\n\t// default page size\n\tvar pageSize = 256\n\n\tif reqSize != 0 {\n\t\tpageSize = int(reqSize)\n\n\t\t// max page size\n\t\tif pageSize > 2048 {\n\t\t\tpageSize = 2048\n\t\t}\n\t}\n\n\treturn pageSize\n}", "func (me *Image) Size() util.Size {\n\tvar s util.Size\n\ts.Width = me.key.width\n\ts.Height = me.key.height\n\treturn s\n}", "func (a *PhonebookAccess1) GetFixedImageSize() (bool, error) {\n\tv, err := a.GetProperty(\"FixedImageSize\")\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn v.Value().(bool), nil\n}", "func GetVMSize(vm *compute.VirtualMachine) (Vmsize compute.VirtualMachineSizeTypes) {\n\n\tVmsize = vm.VirtualMachineProperties.HardwareProfile.VMSize\n\treturn\n\n}", "func (p *Photo) GetSizes() (value []PhotoSizeClass) {\n\tif p == nil {\n\t\treturn\n\t}\n\treturn p.Sizes\n}", "func (w *Widget) GetSizeRequest() (width, height int) {\n\tfields := w.Candy().Guify(\"gtk_widget_get_size_request\", w, 0, 0).Fields()\n\twidth = fields[0].MustInt()\n\theight = fields[1].MustInt()\n\treturn\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetSupportedPostviewImageSize obtains the supported Post View Image sizes from the camera
func (c *Camera) GetSupportedPostviewImageSize() (sizes []string, err error) { resp, err := c.newRequest(endpoints.Camera, "getSupportedPostviewImageSize").Do() if err != nil { return } if len(resp.Result) > 0 { err = json.Unmarshal(resp.Result[0], &sizes) } return }
[ "func (c *Camera) GetAvailablePostviewImageSize() (current string, available []string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getAvailablePostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) >= 1 {\n\t\t// Current size\n\t\tif err := json.Unmarshal(resp.Result[0], &current); err != nil {\n\t\t\treturn current, available, err\n\t\t}\n\n\t\t// Available sizes\n\t\tif err := json.Unmarshal(resp.Result[1], &available); err != nil {\n\t\t\treturn current, available, err\n\t\t}\n\t}\n\n\treturn\n}", "func (c *Camera) GetPostviewImageSize() (size string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getPostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) > 0 {\n\t\terr = json.Unmarshal(resp.Result[0], &size)\n\t}\n\n\treturn\n}", "func (c *Camera) SetPostviewImageSize(size PostViewSize) (err error) {\n\t_, err = c.newRequest(endpoints.Camera, \"setPostviewImageSize\", size).Do()\n\treturn\n}", "func (p *Photo) GetVideoSizes() (value []VideoSizeClass, ok bool) {\n\tif p == nil {\n\t\treturn\n\t}\n\tif !p.Flags.Has(1) {\n\t\treturn value, false\n\t}\n\treturn p.VideoSizes, true\n}", "func ImageSize(\n\tsize imager.ImageSize,\n\tsupportedSizes []imager.ImageSize,\n) (err error) {\n\tfor _, s := range supportedSizes {\n\t\tif s == size {\n\t\t\treturn nil\n\t\t}\n\t}\n\n\treturn fmt.Errorf(\"expected one of %v\", supportedSizes)\n}", "func (a *PhonebookAccess1) GetFixedImageSize() (bool, error) {\n\tv, err := a.GetProperty(\"FixedImageSize\")\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn v.Value().(bool), nil\n}", "func (p *Photo) GetSizes() (value []PhotoSizeClass) {\n\tif p == nil {\n\t\treturn\n\t}\n\treturn p.Sizes\n}", "func (f *framebuffer) viewportSize() (int, int) {\n\t// On some environments, viewport size must be within the framebuffer size.\n\t// e.g. Edge (#71), Chrome on GPD Pocket (#420), macOS Mojave (#691).\n\t// Use the same size of the framebuffer here.\n\treturn f.width, f.height\n}", "func qr_decoder_set_image_size(p _QrDecoderHandle, width, height, depth, channel int) _QrDecoderHandle {\n\tv := C.qr_decoder_set_image_size(C.QrDecoderHandle(p),\n\t\tC.int(width), C.int(height), C.int(depth), C.int(channel),\n\t)\n\treturn _QrDecoderHandle(v)\n}", "func (t *TargetBuilder) MaxImgSizes() []int {\n\tsz0 := t.bspPkg.FlashMap.Areas[flash.FLASH_AREA_NAME_IMAGE_0].Size\n\tsz1 := t.bspPkg.FlashMap.Areas[flash.FLASH_AREA_NAME_IMAGE_1].Size\n\ttrailerSz := t.bootTrailerSize()\n\n\treturn []int{\n\t\tsz0 - trailerSz,\n\t\tsz1 - trailerSz,\n\t}\n}", "func AvpictureGetSize(pf PixelFormat, w, h int) int {\n\treturn int(C.avpicture_get_size((C.enum_AVPixelFormat)(pf), C.int(w), C.int(h)))\n}", "func (m *wasiSnapshotPreview1Impl) environSizesGet() (r0 wasiSize, r1 wasiSize, err wasiErrno) {\n\tsize := 0\n\tfor _, s := range m.env {\n\t\tsize += len(s) + 1\n\t}\n\treturn wasiSize(len(m.env)), wasiSize(size), wasiErrnoSuccess\n}", "func PossibleImageSizeValues() []ImageSize {\n\treturn []ImageSize{\n\t\tImageSize512x512,\n\t\tImageSize1024x1024,\n\t\tImageSize256x256,\n\t}\n}", "func (img *Image) Size() (int, int) { return img.img.Size() }", "func (g *Gui) getWindowSize() (float32, float32) {\n\tw, h := g.cfg.window.w, g.cfg.window.h\n\t// Maintain the image aspect ratio in case the image width and height is greater than the predefined window.\n\tr := getRatio(w, h)\n\tif w > maxScreenX && h > maxScreenY {\n\t\tw = w * r\n\t\th = h * r\n\t}\n\treturn w, h\n}", "func (is ImageSurface) Size() Point {\n\treturn Point{float64(is.width), float64(is.height)}\n}", "func (is ImageSize) Size() (width int, height int) {\n\tconst tokensWidthHeightCount = 2\n\n\tsizeTokens := strings.Split(string(is), \"x\")\n\tif len(sizeTokens) != tokensWidthHeightCount {\n\t\treturn 0, 0\n\t}\n\n\tvar err error\n\twidth, err = strconv.Atoi(sizeTokens[0])\n\tswitch {\n\tcase err != nil:\n\t\tfallthrough\n\tcase width <= 0:\n\t\treturn 0, 0\n\t}\n\n\theight, err = strconv.Atoi(sizeTokens[1])\n\tswitch {\n\tcase err != nil:\n\t\tfallthrough\n\tcase height <= 0:\n\t\treturn 0, 0\n\t}\n\n\treturn width, height\n}", "func (f *XtermImage) GetSize() (width, height int) {\n\twidth = f.timage.Bounds().Max.X\n\theight = f.timage.Bounds().Max.Y\n\n\treturn\n}", "func GetBlobSizesR(state kv.KVStoreReader, blobHash hashing.HashValue) *collections.ImmutableMap {\n\treturn collections.NewMapReadOnly(state, sizesMapName(blobHash))\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
GetAvailablePostviewImageSize obtains the current and available Post View Image sizes from the camera
func (c *Camera) GetAvailablePostviewImageSize() (current string, available []string, err error) { resp, err := c.newRequest(endpoints.Camera, "getAvailablePostviewImageSize").Do() if err != nil { return } if len(resp.Result) >= 1 { // Current size if err := json.Unmarshal(resp.Result[0], &current); err != nil { return current, available, err } // Available sizes if err := json.Unmarshal(resp.Result[1], &available); err != nil { return current, available, err } } return }
[ "func (c *Camera) GetSupportedPostviewImageSize() (sizes []string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getSupportedPostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) > 0 {\n\t\terr = json.Unmarshal(resp.Result[0], &sizes)\n\t}\n\n\treturn\n}", "func (c *Camera) GetPostviewImageSize() (size string, err error) {\n\tresp, err := c.newRequest(endpoints.Camera, \"getPostviewImageSize\").Do()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(resp.Result) > 0 {\n\t\terr = json.Unmarshal(resp.Result[0], &size)\n\t}\n\n\treturn\n}", "func (c *Camera) SetPostviewImageSize(size PostViewSize) (err error) {\n\t_, err = c.newRequest(endpoints.Camera, \"setPostviewImageSize\", size).Do()\n\treturn\n}", "func (a *PhonebookAccess1) GetFixedImageSize() (bool, error) {\n\tv, err := a.GetProperty(\"FixedImageSize\")\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn v.Value().(bool), nil\n}", "func AvpictureGetSize(pf PixelFormat, w, h int) int {\n\treturn int(C.avpicture_get_size((C.enum_AVPixelFormat)(pf), C.int(w), C.int(h)))\n}", "func (m *wasiSnapshotPreview1Impl) environSizesGet() (r0 wasiSize, r1 wasiSize, err wasiErrno) {\n\tsize := 0\n\tfor _, s := range m.env {\n\t\tsize += len(s) + 1\n\t}\n\treturn wasiSize(len(m.env)), wasiSize(size), wasiErrnoSuccess\n}", "func (p *PdfiumImplementation) FPDFBitmap_GetHeight(request *requests.FPDFBitmap_GetHeight) (*responses.FPDFBitmap_GetHeight, error) {\n\tp.Lock()\n\tdefer p.Unlock()\n\n\tbitmapHandle, err := p.getBitmapHandle(request.Bitmap)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\theight := C.FPDFBitmap_GetHeight(bitmapHandle.handle)\n\treturn &responses.FPDFBitmap_GetHeight{\n\t\tHeight: int(height),\n\t}, nil\n}", "func (img *Image) Size() (int, int) { return img.img.Size() }", "func (f *XtermImage) GetSize() (width, height int) {\n\twidth = f.timage.Bounds().Max.X\n\theight = f.timage.Bounds().Max.Y\n\n\treturn\n}", "func GetBlobSizesR(state kv.KVStoreReader, blobHash hashing.HashValue) *collections.ImmutableMap {\n\treturn collections.NewMapReadOnly(state, sizesMapName(blobHash))\n}", "func (f *framebuffer) viewportSize() (int, int) {\n\t// On some environments, viewport size must be within the framebuffer size.\n\t// e.g. Edge (#71), Chrome on GPD Pocket (#420), macOS Mojave (#691).\n\t// Use the same size of the framebuffer here.\n\treturn f.width, f.height\n}", "func (g *Gui) getWindowSize() (float32, float32) {\n\tw, h := g.cfg.window.w, g.cfg.window.h\n\t// Maintain the image aspect ratio in case the image width and height is greater than the predefined window.\n\tr := getRatio(w, h)\n\tif w > maxScreenX && h > maxScreenY {\n\t\tw = w * r\n\t\th = h * r\n\t}\n\treturn w, h\n}", "func UlBitmapGetSize(bitmap ULBitmap) uint {\n\tcbitmap, _ := *(*C.ULBitmap)(unsafe.Pointer(&bitmap)), cgoAllocsUnknown\n\t__ret := C.ulBitmapGetSize(cbitmap)\n\t__v := (uint)(__ret)\n\treturn __v\n}", "func (is ImageSurface) Size() Point {\n\treturn Point{float64(is.width), float64(is.height)}\n}", "func (ly *Layout) AvailSize() mat32.Vec2 {\n\tspc := ly.BoxSpace()\n\tavail := ly.LayState.Alloc.Size.SubScalar(spc.Right) // spc is for right size space\n\tparni, _ := KiToNode2D(ly.Par)\n\tif parni != nil {\n\t\tvp := parni.AsViewport2D()\n\t\tif vp != nil {\n\t\t\tif vp.ViewportSafe() == nil {\n\t\t\t\t// SidesTODO: might not be right\n\t\t\t\tavail = mat32.NewVec2FmPoint(ly.VpBBox.Size()).SubScalar(spc.Right)\n\t\t\t\t// fmt.Printf(\"non-nil par ly: %v vp: %v %v\\n\", ly.Path(), vp.Path(), avail)\n\t\t\t}\n\t\t}\n\t}\n\treturn avail\n}", "func qr_decoder_set_image_size(p _QrDecoderHandle, width, height, depth, channel int) _QrDecoderHandle {\n\tv := C.qr_decoder_set_image_size(C.QrDecoderHandle(p),\n\t\tC.int(width), C.int(height), C.int(depth), C.int(channel),\n\t)\n\treturn _QrDecoderHandle(v)\n}", "func (r *ImageRef) GetPageHeight() int {\n\treturn vipsGetPageHeight(r.image)\n}", "func (p *Photo) GetVideoSizes() (value []VideoSizeClass, ok bool) {\n\tif p == nil {\n\t\treturn\n\t}\n\tif !p.Flags.Has(1) {\n\t\treturn value, false\n\t}\n\treturn p.VideoSizes, true\n}", "func GetDimensions(imageData io.Reader) (int, int, error) {\n\tcfg, _, err := image.DecodeConfig(imageData)\n\tif seeker, ok := imageData.(io.ReadSeeker); ok {\n\t\tdefer seeker.Seek(0, 0)\n\t}\n\treturn cfg.Width, cfg.Height, err\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
HandleRequest handles incoming request
func HandleRequest(ctx context.Context, request events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) { _, _ = pretty.Println("parsed:", request.Body) return events.APIGatewayProxyResponse{Body: "response is working", StatusCode: 200}, nil }
[ "func handleRequest(req api.Request, resp api.Response) (next bool, reqCtx uint32) {\n\t// Serve a response that shows the invoked request URI.\n\tresp.Headers().Set(\"Content-Type\", \"text/plain\")\n\tresp.Body().WriteString(\"hello \" + req.GetURI())\n\treturn // skip any downstream middleware, as we wrote a response.\n}", "func HandleRequest(c *gin.Context, isView bool) {\n\tvar err error\n\n\tid := c.Param(\"uniuri\")\n\tre := models.Resource{}\n\n\tif err = re.Get(id); err != nil || re.Key == \"\" {\n\t\tlogger.InfoC(c, \"server\", \"Not found\", id)\n\t\tc.AbortWithStatus(http.StatusNotFound)\n\t\treturn\n\t}\n\tre.LogFetched(c)\n\tif conf.C.AlwaysDownload {\n\t\tc.Header(\"Content-Type\", \"application/octet-stream\")\n\t\tc.Header(\"Content-Disposition\", \"attachment; filename=\\\"\"+re.Name+\"\\\"\")\n\t} else {\n\t\tc.Header(\"Content-Disposition\", \"filename=\\\"\"+re.Name+\"\\\"\")\n\t}\n\tfile := path.Join(conf.C.UploadDir, re.Key)\n\tif _, err := os.Stat(file); err != nil {\n\t\tlogger.ErrC(c, \"server\", fmt.Sprintf(\"Couldn't open %s\", re.Key), err)\n\t\tc.AbortWithStatus(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tc.File(file)\n\n\tif isView && re.Once {\n\t\tre.Delete()\n\t\tre.LogDeleted(c)\n\t}\n}", "func (c *Core) handleRequest(fctx *fasthttp.RequestCtx) {\n\tctx := c.assignCtx(fctx)\n\tdefer c.releaseCtx(ctx)\n\tif ctx.methodINT == -1 {\n\t\tctx.Status(StatusBadRequest).SendString(\"Invalid http method\")\n\t\treturn\n\t}\n\n\tstart := time.Now()\n\t// Delegate next to handle the request\n\t// Find match in stack\n\tmatch, err := c.next(ctx)\n\tif err != nil {\n\t\t_ = ctx.SendStatus(StatusInternalServerError)\n\t}\n\t// Generate ETag if enabled\n\tif match && c.ETag {\n\t\tsetETag(ctx, false)\n\t}\n\tif c.Debug {\n\t\td := time.Since(start)\n\t\t// d := time.Now().Sub(start).String()\n\t\tLog.D(\"%s %s %d %s\\n\", ctx.method, ctx.path, ctx.Response.StatusCode(), d)\n\t}\n}", "func (srv *server) handleRequest(clt *Client, msg *Message) {\n\treplyPayload, returnedErr := srv.impl.OnRequest(\n\t\tcontext.Background(),\n\t\tclt,\n\t\tmsg,\n\t)\n\tswitch returnedErr.(type) {\n\tcase nil:\n\t\tsrv.fulfillMsg(clt, msg, replyPayload)\n\tcase ReqErr:\n\t\tsrv.failMsg(clt, msg, returnedErr)\n\tcase *ReqErr:\n\t\tsrv.failMsg(clt, msg, returnedErr)\n\tdefault:\n\t\tsrv.errorLog.Printf(\"Internal error during request handling: %s\", returnedErr)\n\t\tsrv.failMsg(clt, msg, returnedErr)\n\t}\n}", "func (res *Resource) HandleRequest(w http.ResponseWriter, r *http.Request) {\n\tif r.Method != http.MethodGet && r.Method != http.MethodHead {\n\t\thandleMethodNotAllowed(w, r)\n\t\treturn\n\t}\n\n\treader := bytes.NewReader(res.Content)\n\thttp.ServeContent(w, r, res.FileName, res.ModTime, reader)\n}", "func HandleRequest(request events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// Get request parameters\n\tprojectID := request.PathParameters[\"pid\"]\n\tcookie := auth.ExtractCookie(request.Headers[\"Cookie\"])\n\n\t// Perform the action\n\turl, err := actionFunc(projectID, cookie, auth.VerifyCookie, dao.Dynamo)\n\n\t// Handle the output\n\tresponse := &getDownloadResponse{URL: url}\n\treturn http.GatewayResponse(response, \"\", err), nil\n}", "func handleRequest(request *http.Request, t http.RoundTripper) (rsp *http.Response) {\n\tvar err error\n\n\tif rsp, err = t.RoundTrip(request); err != nil {\n\t\tlog.Println(\"Request failed:\", err)\n\t}\n\n\treturn\n}", "func (app *App) handleRequest(handler RequestHandlerFunction) http.HandlerFunc {\r\n\treturn func(w http.ResponseWriter, r *http.Request) {\r\n\t\thandler(app.DB, w, r)\r\n\t}\r\n}", "func (h HTTPHandlerFunc) HandleRequest(c context.Context, fc *fasthttp.RequestCtx) error {\n\treturn h(c, fc)\n}", "func (s *Server) HandleRequest(w dns.ResponseWriter, r *dns.Msg) {\n\tresp := &dns.Msg{}\n\tresp.SetReply(r)\n\n\tfor _, q := range r.Question {\n\t\tans := s.handleQuestion(q)\n\t\tif ans != nil {\n\t\t\tresp.Answer = append(resp.Answer, ans...)\n\t\t}\n\t}\n\n\terr := w.WriteMsg(resp)\n\tif err != nil {\n\t\ts.logger.Println(\"ERROR : \" + err.Error())\n\t}\n\tw.Close()\n\n}", "func (pi *PackageIndexer) HandleRequest(req Request) string {\n\t// bad request made \n\tif req.err != \"\" {\n\t\treturn ERROR \n\t}\n\t// set the name of the package \t\n\tpack := Package{name: req.pack}\t\t\t\t\t\t\t\t\n\t// add package dependencies \n\tfor _, name := range req.dep {\t\t\t\t\t\t\t\t \n\t\tpack.deps = append(pack.deps, &Package{name: name})\n\t}\n\t// check command type \n switch req.comm {\t\t\t\t\t\t\t\t\t\t\t\n case \"INDEX\":\n return pi.Index(&pack)\n case \"REMOVE\":\n return pi.Remove(&pack)\n case \"QUERY\":\n return pi.Query(pack.name)\n }\n\n // otherwise, error with request \n return ERROR \t\t\t\t\t\t\t\t\t\t\t\t\n}", "func (c *BFTChain) HandleRequest(sender uint64, req []byte) {\n\tc.Logger.Debugf(\"HandleRequest from %d\", sender)\n\tif _, err := c.verifier.VerifyRequest(req); err != nil {\n\t\tc.Logger.Warnf(\"Got bad request from %d: %v\", sender, err)\n\t\treturn\n\t}\n\tc.consensus.SubmitRequest(req)\n}", "func (r *route) handleRequest(w http.ResponseWriter, req *http.Request) {\n pathParams := r.parsePatternParams(req.URL.Path)\n if req.URL.RawQuery != \"\" && pathParams != \"\" {\n req.URL.RawQuery += \"&\"\n }\n req.URL.RawQuery += pathParams\n r.handler.ServeHTTP(w,req)\n}", "func (c *Client) HandleRequest(req *http.Request) (res *http.Response, err error) {\n\treq.URL.Path = \"/api/v\" + c.APIVersion + req.URL.Path\n\n\t// Fill out Host and Scheme if it is empty\n\tif req.URL.Host == \"\" {\n\t\treq.URL.Host = c.URLHost\n\t}\n\tif req.URL.Scheme == \"\" {\n\t\treq.URL.Scheme = c.URLScheme\n\t}\n\tif req.Header.Get(\"User-Agent\") == \"\" {\n\t\treq.Header.Set(\"User-Agent\", c.UserAgent)\n\t}\n\tif req.Header.Get(\"Authorization\") == \"\" {\n\t\treq.Header.Set(\"Authorization\", \"Bot \"+c.Token)\n\t}\n\n\tres, err = c.HTTP.Do(req)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif res.StatusCode == http.StatusUnauthorized {\n\t\terr = errors.New(\"Invalid token passed\")\n\t\treturn\n\t}\n\n\treturn\n}", "func handleRequest(ctx context.Context, event events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// decode the event parameter\n\tvar data EventData\n\tif err := json.Unmarshal([]byte(event.Body), &data); err != nil {\n\t\treturn events.APIGatewayProxyResponse{StatusCode: 500}, err\n\t}\n\n\t// prepare the response string\n\tcurrentTime := time.Now()\n\tcurrentTimeStr := currentTime.Format(\"2006-01-02 15:04:05\")\n\tresponseStr := fmt.Sprintf(\"Hello from AWS Lambda, %s! Its %s\", data.Name, currentTimeStr)\n\n\t// return the response\n\treturn events.APIGatewayProxyResponse{Body: responseStr, StatusCode: 200}, nil\n}", "func HandleRequest(ctx context.Context) error {\n\tfmt.Println(\"Hello Go from Lambda!\")\n\treturn nil\n}", "func (q *Queue) HandleRequest(ctx context.Context, r *Request) error {\n\tconsumer := q.Consumer()\n\n\treturn consumer.Handle(r)\n}", "func HandleRequest(client github.Client, event *github.GenericRequestEvent) error {\n\treturn plugins.HandleRequest(client, event)\n}", "func (api *Api) handleRequest(handler RequestHandlerFunction) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\thandler(api.DB, w, r)\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
New creates and returns (but does not start) a new KeyValueServer.
func New(store kvstore.KVStore) KeyValueServer { // TODO: implement this! var server keyValueServer server.clientNum = 0 server.listener = nil server.readChan = make(chan []byte) server.channelMap = make(map[net.Conn]chan []byte) // 使用接口时,返回接口类型变量, 参考 book p113 return &server }
[ "func New() KeyValueServer {\n\treturn &keyValueServer{\n\t\tnil,\n\t\tmake([]*client, 0),\n\t\tmake(chan *request),\n\t\tmake(chan *request),\n\t\tmake(chan []byte),\n\t\tmake(chan net.Conn),\n\t\tmake(chan int),\n\t\tmake(chan int),\n\t\tmake(chan *client),\n\t\tmake(chan int),\n\t\tmake(chan int)}\n}", "func New() KeyValueServer {\n\t// TODO: implement this!\n\treturn keyValueServer{protocol:\"tcp\",count:0 };\n}", "func New() KeyValueServer {\n\tinit_db()\n\tkvs := &keyValueServer{\n\t\tclienter: make(map[int]*Clienter),\n\t\tconnectedClients: -1,\n\t\tconnChannel: make(chan net.Conn),\n\t\tclose_signal: make(chan bool),\n\t\tcnt_signal_in: make(chan bool),\n\t\tcnt_signal_out: make(chan int),\n\t\tdelete: make(chan *Clienter),\n\t\treq: make(chan *Request),\n\t}\n\treturn kvs\n}", "func NewServer(c *config.Config) (*ToyKVServer, error) {\n\ttk := &ToyKVServer{\n\t\tlogger: log.New(os.Stderr, \"[toykv] \", log.LstdFlags|log.Lshortfile),\n\t}\n\n\tvar err error\n\ttk.store = store.NewStore(c.RaftDir, c.RaftPort)\n\n\tbootstrap := (c.Join == \"\")\n\terr = tk.store.Init(bootstrap, c.NodeID)\n\tif err != nil {\n\t\ttk.logger.Print(err.Error())\n\t\treturn nil, err\n\t}\n\n\tif !bootstrap {\n\t\t// send join request to node already exists\n\t\trc := goredis.NewClient(c.Join, \"\")\n\t\ttk.logger.Printf(\"join request send to %s\\n\", c.Join)\n\t\t_, err := rc.Do(\"join\", c.RaftPort, c.NodeID)\n\t\tif err != nil {\n\t\t\ttk.logger.Println(err)\n\t\t}\n\t\trc.Close()\n\t}\n\n\ttk.listener, err = net.Listen(\"tcp\", c.ListenAddress)\n\tif err != nil {\n\t\ttk.logger.Print(err.Error())\n\t\treturn nil, err\n\t}\n\n\ttk.logger.Printf(\"toykv listen %s successfully\", c.ListenAddress)\n\treturn tk, nil\n}", "func NewServer() *Server {}", "func New() *KV {\n\treturn &KV{data: map[string]string{}}\n}", "func New(level int, rootdir string, flushThresh store.KVLen, blockSize store.KVLen) (*Server, error) {\n\n\tif err := os.MkdirAll(rootdir, 0777); err != nil {\n\t\treturn nil, fmt.Errorf(\"Create root directory failed | rootdir=%v | err=[%w]\", rootdir, err)\n\t}\n\n\treturn &Server{\n\t\tlogger: logging.New(level),\n\n\t\trootdir: rootdir,\n\t\tflushThresh: flushThresh,\n\t\tblockSize: blockSize,\n\n\t\tinitDone: false,\n\t}, nil\n}", "func New(m int, k int) *Server {\n\treturn &Server{bf: bloom.New(uint(m), uint(k))}\n}", "func New(path string) (*KV, error) {\n\tb, err := bolt.Open(path, 0644, nil)\n\treturn &KV{db: b}, err\n}", "func New() *Server {\n\treturn &Server{}\n}", "func New(addr string, password string) *KVStore {\n\tconst maxRetries = 5\n\n\tclient := redis.NewClient(&redis.Options{\n\t\tAddr: addr,\n\t\tMaxRetries: maxRetries,\n\t\tPassword: password,\n\t})\n\n\treturn &KVStore{client: client}\n}", "func (t *OpenconfigSystem_System_Ntp_Servers) NewServer(Address string) (*OpenconfigSystem_System_Ntp_Servers_Server, error){\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Server == nil {\n\t\tt.Server = make(map[string]*OpenconfigSystem_System_Ntp_Servers_Server)\n\t}\n\n\tkey := Address\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Server[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Server\", key)\n\t}\n\n\tt.Server[key] = &OpenconfigSystem_System_Ntp_Servers_Server{\n\t\tAddress: &Address,\n\t}\n\n\treturn t.Server[key], nil\n}", "func New() (*server, error) {\n\tenv, err := env.New()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &server{\n\t\tenv: env,\n\t}, nil\n}", "func (t *OpenconfigSystem_System_Dns_Servers) NewServer(Address string) (*OpenconfigSystem_System_Dns_Servers_Server, error){\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Server == nil {\n\t\tt.Server = make(map[string]*OpenconfigSystem_System_Dns_Servers_Server)\n\t}\n\n\tkey := Address\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Server[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Server\", key)\n\t}\n\n\tt.Server[key] = &OpenconfigSystem_System_Dns_Servers_Server{\n\t\tAddress: &Address,\n\t}\n\n\treturn t.Server[key], nil\n}", "func New(pipeName string, hnd daemon.Handler) *Server {\n\treturn nil\n}", "func New(bikeAccessor bike.Accessor) *Server {\n\treturn &Server{bikeAccessor: bikeAccessor}\n}", "func New(name, group, address string) *Server {\n\ts := &Server{\n\t\tname: name,\n\t\tgroup: group,\n\t\taddress: address,\n\t}\n\n\treturn s\n}", "func (t *OpenconfigOfficeAp_System_Ntp_Servers) NewServer(Address string) (*OpenconfigOfficeAp_System_Ntp_Servers_Server, error) {\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Server == nil {\n\t\tt.Server = make(map[string]*OpenconfigOfficeAp_System_Ntp_Servers_Server)\n\t}\n\n\tkey := Address\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Server[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Server\", key)\n\t}\n\n\tt.Server[key] = &OpenconfigOfficeAp_System_Ntp_Servers_Server{\n\t\tAddress: &Address,\n\t}\n\n\treturn t.Server[key], nil\n}", "func (t *OpenconfigOfficeAp_System_Dns_Servers) NewServer(Address string) (*OpenconfigOfficeAp_System_Dns_Servers_Server, error) {\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Server == nil {\n\t\tt.Server = make(map[string]*OpenconfigOfficeAp_System_Dns_Servers_Server)\n\t}\n\n\tkey := Address\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Server[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Server\", key)\n\t}\n\n\tt.Server[key] = &OpenconfigOfficeAp_System_Dns_Servers_Server{\n\t\tAddress: &Address,\n\t}\n\n\treturn t.Server[key], nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
InterceptRequest creates new request interceptor
func InterceptRequest(f func(http.Header)) *RequestInterceptor { return &RequestInterceptor{Intercept: f} }
[ "func (r *Request) Intercept(interceptor Intercept) *Request {\n\tr.interceptor = interceptor\n\treturn r\n}", "func (a *APITest) Intercept(interceptor Intercept) *APITest {\n\ta.request.interceptor = interceptor\n\treturn a\n}", "func AddRequestInterceptor(handler RequestInterceptor) {\n\tinterceptors = append(interceptors, handler)\n}", "func (this Interceptor) Intercept(url string, exec rack.Middleware) error {\n\tif this[url] != nil {\n\t\treturn PreExistingInterceptorError{url}\n\t}\n\tthis[url] = exec\n\treturn nil\n}", "func (c *UrlReplaceHandler) Intercept(pipeline Pipeline, middlewareIndex int, req *http.Request) (*http.Response, error) {\n\treqOption, ok := req.Context().Value(urlReplaceOptionKey).(urlReplaceOptionsInt)\n\tif !ok {\n\t\treqOption = &c.options\n\t}\n\n\tobsOptions := GetObservabilityOptionsFromRequest(req)\n\tctx := req.Context()\n\tvar span trace.Span\n\tif obsOptions != nil {\n\t\tctx, span = otel.GetTracerProvider().Tracer(obsOptions.GetTracerInstrumentationName()).Start(ctx, \"UrlReplaceHandler_Intercept\")\n\t\tspan.SetAttributes(attribute.Bool(\"com.microsoft.kiota.handler.url_replacer.enable\", true))\n\t\tdefer span.End()\n\t\treq = req.WithContext(ctx)\n\t}\n\n\tif !reqOption.IsEnabled() || len(reqOption.GetReplacementPairs()) == 0 {\n\t\treturn pipeline.Next(req, middlewareIndex)\n\t}\n\n\treq.URL.Path = ReplacePathTokens(req.URL.Path, reqOption.GetReplacementPairs())\n\n\tif span != nil {\n\t\tspan.SetAttributes(attribute.String(\"http.request_url\", req.RequestURI))\n\t}\n\n\treturn pipeline.Next(req, middlewareIndex)\n}", "func NewMockRequestInterceptor(t interface {\n\tmock.TestingT\n\tCleanup(func())\n}) *MockRequestInterceptor {\n\tmock := &MockRequestInterceptor{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func ApplyInterceptor(ctx context.Context, req *Request) (*Request, error) {\n\tif registeredInterceptor == nil {\n\t\treturn req, nil\n\t}\n\treturn registeredInterceptor.Apply(ctx, req)\n}", "func InjectTrace(ctx context.Context, incomingReq *restful.Request,\n\toutgoingReq *http.Request) (*http.Request, opentracing.Span, context.Context) {\n\tspan, newCtx := StartSpanFromContext(ctx, \"outgoing request\")\n\tif span != nil {\n\t\text.HTTPUrl.Set(span, outgoingReq.Host+outgoingReq.RequestURI)\n\t\text.HTTPMethod.Set(span, outgoingReq.Method)\n\t\t_ = span.Tracer().Inject(\n\t\t\tspan.Context(),\n\t\t\topentracing.HTTPHeaders,\n\t\t\topentracing.HTTPHeadersCarrier(outgoingReq.Header))\n\n\t\tfor _, header := range forwardHeaders {\n\t\t\tif value := incomingReq.Request.Header.Get(header); value != \"\" {\n\t\t\t\toutgoingReq.Header.Set(header, value)\n\t\t\t}\n\t\t}\n\t} else {\n\t\treturn outgoingReq, nil, nil\n\t}\n\n\tif logrus.GetLevel() >= logrus.DebugLevel {\n\t\theader := make(map[string]string)\n\n\t\tfor key, val := range outgoingReq.Header {\n\t\t\tkey = strings.ToLower(key)\n\t\t\tif !strings.Contains(key, \"auth\") {\n\t\t\t\theader[key] = val[0]\n\t\t\t}\n\t\t}\n\n\t\tlogrus.Debug(\"outgoing header : \", header)\n\t}\n\n\tif abTraceID := incomingReq.Request.Header.Get(event.TraceIDKey); abTraceID != \"\" {\n\t\toutgoingReq.Header.Set(event.TraceIDKey, abTraceID)\n\t}\n\n\treturn outgoingReq, span, newCtx\n}", "func NewInterceptReader(reader resp.AllReadCloser) *InterceptReader {\n\treturn NewInterceptReaderWithBuffer(reader, make([]byte, reader.Len()))\n}", "func Interceptor(opts ...Option) gin.HandlerFunc {\n\tset := newOptionSet(opts...)\n\n\treturn func(ctx *gin.Context) {\n\t\tctx.Set(rkgininter.RpcEntryNameKey, set.EntryName)\n\n\t\trequestId := rkcommon.GenerateRequestId()\n\t\tctx.Header(rkginctx.RequestIdKey, requestId)\n\n\t\tevent := rkginctx.GetEvent(ctx)\n\t\tevent.SetRequestId(requestId)\n\t\tevent.SetEventId(requestId)\n\n\t\tctx.Header(set.AppNameKey, rkentry.GlobalAppCtx.GetAppInfoEntry().AppName)\n\t\tctx.Header(set.AppVersionKey, rkentry.GlobalAppCtx.GetAppInfoEntry().Version)\n\n\t\tnow := time.Now()\n\t\tctx.Header(set.AppUnixTimeKey, now.Format(time.RFC3339Nano))\n\t\tctx.Header(set.ReceivedTimeKey, now.Format(time.RFC3339Nano))\n\n\t\tctx.Next()\n\t}\n}", "func RequestLimitInterceptor(requestLimit int) grpc.UnaryServerInterceptor {\n\tsem := semaphore.NewWeighted(int64(requestLimit))\n\treturn func(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {\n\t\tif !sem.TryAcquire(1) {\n\t\t\treturn nil, status.Error(codes.Unavailable, \"Too many pending requests. Please retry later.\")\n\t\t}\n\t\tdefer sem.Release(1)\n\t\treturn handler(ctx, req)\n\t}\n}", "func AddToRequest(app *App) server.PreHandlerFunc {\n\treturn func(req *http.Request) *http.Request {\n\t\tnewCtx := With(req.Context(), app)\n\t\treturn req.Clone(newCtx)\n\t}\n}", "func (bas *BaseService) OnRequest(ctx context.Context, args Args) {}", "func RegisterInterceptor(newRI RequestInterceptor) (restore func()) {\n\toldRI := registeredInterceptor\n\tregisteredInterceptor = newRI\n\treturn func() {\n\t\tregisteredInterceptor = oldRI\n\t}\n}", "func InjectRequestLogger(l Logger) Middleware {\n\treturn func(inner HandlerFunc) HandlerFunc {\n\t\treturn func(rw http.ResponseWriter, req *http.Request) error {\n\t\t\tif l != nil {\n\t\t\t\treq = Inject(req, ctxKeyLogger, l)\n\t\t\t}\n\n\t\t\treturn inner(rw, req)\n\t\t}\n\t}\n}", "func NewInboundRequest(\n baseRequest *http.Request,\n pathParams PathParams,\n) *InboundRequest {\n req := &InboundRequest{\n Request: *baseRequest,\n PathParams: pathParams,\n }\n return req\n}", "func incrInterceptedRequestStatDelta() {\n\tStatMu.Mutex.Lock()\n\n\t// increment the requests counter\n\t*(StatMu.InstanceStat.InterceptedRequests) = *(StatMu.InstanceStat.InterceptedRequests) + uint64(1)\n\tStatMu.Mutex.Unlock()\n\n}", "func InterceptWithReqModifier(method string, modifier SafeLoggingModifier) InterceptOption {\n\treturn func(cfg *interceptConfig) {\n\t\tcfg.reqModifiers[method] = modifier\n\t}\n}", "func (h standAloneHandler) requestToProxyRequest(r *http.Request) (context.Context, *APIGatewayProxyRequest) {\n\tctx := context.Background()\n\treq := APIGatewayProxyRequest{\n\t\tPath: r.URL.Path,\n\t\tHTTPMethod: r.Method,\n\t\tRequestContext: events.APIGatewayProxyRequestContext{\n\t\t\tHTTPMethod: r.Method,\n\t\t},\n\t}\n\n\t// transfer the headers over to the event\n\treq.Headers = map[string]string{}\n\tfor k, v := range r.Header {\n\t\treq.Headers[k] = strings.Join(v, \"; \")\n\t}\n\n\t// Querystring params\n\tparams := r.URL.Query()\n\tparamsMap := map[string]string{}\n\tfor k, _ := range params {\n\t\tparamsMap[k] = params.Get(k)\n\t}\n\treq.QueryStringParameters = paramsMap\n\n\t// Path params (just the proxy+ path ... but it does not have the preceding slash)\n\treq.PathParameters = map[string]string{\n\t\t\"proxy\": r.URL.Path[1:len(r.URL.Path)],\n\t}\n\treq.Resource = \"/{proxy+}\"\n\treq.RequestContext.ResourcePath = \"/{proxy+}\"\n\n\t// Identity info: user agent, IP, etc.\n\treq.RequestContext.Identity.UserAgent = r.Header.Get(\"User-Agent\")\n\treq.RequestContext.Identity.SourceIP = r.Header.Get(\"X-Forwarded-For\")\n\tif req.RequestContext.Identity.SourceIP == \"\" {\n\t\tip, _, err := net.SplitHostPort(r.RemoteAddr)\n\t\tif err == nil {\n\t\t\treq.RequestContext.Identity.SourceIP = net.ParseIP(ip).String()\n\t\t}\n\t}\n\n\t// Stage will be \"local\" for now? I'm not sure what makes sense here. Local gateway. Local. Debug. ¯\\_(ツ)_/¯\n\treq.RequestContext.Stage = \"local\"\n\t// TODO: Stage variables would need to be pulled from the aegis.yaml ...\n\t// so now the config file has to be next to the app... otherwise some defaults will be set like \"local\"\n\t// and no stage variables i suppose.\n\t// evt.StageVariables =\n\n\t// The request id will simply be a timestamp to help keep it unique, but also allowing it to be easily sorted\n\treq.RequestContext.RequestID = strconv.FormatInt(time.Now().UnixNano(), 10)\n\n\t// pass along the body\n\tbodyData, err := ioutil.ReadAll(r.Body)\n\tif err == nil {\n\t\treq.Body = string(bodyData)\n\t}\n\n\treturn ctx, &req\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
InterceptResponse creates new response interceptor
func InterceptResponse(f ResponseInterceptFunc) *ResponseInterceptor { return &ResponseInterceptor{Intercept: f} }
[ "func (m *middleware) InterceptResponse(ctx context.Context, next gqlgen.ResponseHandler) *gqlgen.Response {\n\tresp := next(ctx)\n\n\toperations, ok := FromCtx(ctx)\n\tif !ok {\n\t\treturn resp\n\t}\n\n\tlocations := make([]string, 0)\n\tfor _, operation := range *operations {\n\t\toperationURL := fmt.Sprintf(\"%s/%s/%s\", m.directorURL, operation.ResourceType, operation.ResourceID)\n\t\tlocations = append(locations, operationURL)\n\t}\n\n\tif len(locations) > 0 {\n\t\treqCtx := gqlgen.GetOperationContext(ctx)\n\t\tgqlgen.RegisterExtension(ctx, LocationsParam, locations)\n\t\tresp.Extensions = gqlgen.GetExtensions(ctx)\n\n\t\tjsonPropsToDelete := make([]string, 0)\n\t\tfor _, gqlOperation := range reqCtx.Doc.Operations {\n\t\t\tfor _, gqlSelection := range gqlOperation.SelectionSet {\n\t\t\t\tgqlField, ok := gqlSelection.(*ast.Field)\n\t\t\t\tif !ok {\n\t\t\t\t\tlog.C(ctx).Errorf(\"Unable to prepare final response: gql field has unexpected type %T instead of *ast.Field\", gqlSelection)\n\t\t\t\t\treturn gqlgen.ErrorResponse(ctx, \"unable to prepare final response\")\n\t\t\t\t}\n\n\t\t\t\tmutationAlias := gqlField.Alias\n\t\t\t\tfor _, gqlArgument := range gqlField.Arguments {\n\t\t\t\t\tif gqlArgument.Name == ModeParam && gqlArgument.Value.Raw == string(graphql.OperationModeAsync) {\n\t\t\t\t\t\tjsonPropsToDelete = append(jsonPropsToDelete, mutationAlias)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tnewData, err := cleanupFields(resp, jsonPropsToDelete)\n\t\tif err != nil {\n\t\t\tlog.C(ctx).WithError(err).Errorf(\"Unable to process and delete unnecessary bytes from response body: %v\", err)\n\t\t\treturn gqlgen.ErrorResponse(ctx, \"failed to prepare response body\")\n\t\t}\n\n\t\tresp.Data = newData\n\t}\n\n\treturn resp\n}", "func (a *APITest) Intercept(interceptor Intercept) *APITest {\n\ta.request.interceptor = interceptor\n\treturn a\n}", "func (k <%= pluginInternal %>) OnResponse(ctx context.Context, p spec.APIProxy, c controller.Controller, r *http.Request, resp *http.Response, span opentracing.Span) error {\n\treturn nil\n}", "func (r *Request) Intercept(interceptor Intercept) *Request {\n\tr.interceptor = interceptor\n\treturn r\n}", "func (p *Proxy) onResponse(resp *http.Response, ctx *goproxy.ProxyCtx) *http.Response {\n\tfor _, h := range mubeng.HopHeaders {\n\t\tresp.Header.Del(h)\n\t}\n\n\treturn resp\n}", "func NewInterceptReader(reader resp.AllReadCloser) *InterceptReader {\n\treturn NewInterceptReaderWithBuffer(reader, make([]byte, reader.Len()))\n}", "func (c *UrlReplaceHandler) Intercept(pipeline Pipeline, middlewareIndex int, req *http.Request) (*http.Response, error) {\n\treqOption, ok := req.Context().Value(urlReplaceOptionKey).(urlReplaceOptionsInt)\n\tif !ok {\n\t\treqOption = &c.options\n\t}\n\n\tobsOptions := GetObservabilityOptionsFromRequest(req)\n\tctx := req.Context()\n\tvar span trace.Span\n\tif obsOptions != nil {\n\t\tctx, span = otel.GetTracerProvider().Tracer(obsOptions.GetTracerInstrumentationName()).Start(ctx, \"UrlReplaceHandler_Intercept\")\n\t\tspan.SetAttributes(attribute.Bool(\"com.microsoft.kiota.handler.url_replacer.enable\", true))\n\t\tdefer span.End()\n\t\treq = req.WithContext(ctx)\n\t}\n\n\tif !reqOption.IsEnabled() || len(reqOption.GetReplacementPairs()) == 0 {\n\t\treturn pipeline.Next(req, middlewareIndex)\n\t}\n\n\treq.URL.Path = ReplacePathTokens(req.URL.Path, reqOption.GetReplacementPairs())\n\n\tif span != nil {\n\t\tspan.SetAttributes(attribute.String(\"http.request_url\", req.RequestURI))\n\t}\n\n\treturn pipeline.Next(req, middlewareIndex)\n}", "func NewWriterInterceptor(w http.ResponseWriter, req *http.Request, fn ResModifierFunc) *WriterInterceptor {\n\tres := &http.Response{\n\t\tRequest: req,\n\t\tStatusCode: 200,\n\t\tStatus: \"200 OK\",\n\t\tProto: \"HTTP/1.1\",\n\t\tProtoMajor: 1,\n\t\tProtoMinor: 1,\n\t\tHeader: make(http.Header),\n\t\tBody: ioutil.NopCloser(bytes.NewReader([]byte{})),\n\t}\n\treturn &WriterInterceptor{mutex: &sync.Mutex{}, writer: w, modifier: fn, response: res}\n}", "func (f *genericFilter) OnResponse(_ context.Context, result protocol.Result, _ protocol.Invoker,\n\t_ protocol.Invocation) protocol.Result {\n\treturn result\n}", "func InterceptWithRespModifier(method string, modifier SafeLoggingModifier) InterceptOption {\n\treturn func(cfg *interceptConfig) {\n\t\tcfg.respModifiers[method] = modifier\n\t}\n}", "func (tunnel *TunnelHandler) OnResponse(filters ...Filter) *RespFilterGroup {\n\treturn &RespFilterGroup{ctx: tunnel.Ctx, filters: filters}\n}", "func (this Interceptor) Intercept(url string, exec rack.Middleware) error {\n\tif this[url] != nil {\n\t\treturn PreExistingInterceptorError{url}\n\t}\n\tthis[url] = exec\n\treturn nil\n}", "func NewResponseModifier(req *http.Request, res *http.Response) *ResponseModifier {\n\treturn &ResponseModifier{Request: req, Response: res, Header: res.Header}\n}", "func (a *ContinueRequestArgs) SetInterceptResponse(interceptResponse bool) *ContinueRequestArgs {\n\ta.InterceptResponse = &interceptResponse\n\treturn a\n}", "func Response(fn ResModifierFunc) func(http.Handler) http.Handler {\n\treturn func(h http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tif r.Method == \"OPTIONS\" || r.Method == \"HEAD\" {\n\t\t\t\th.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twriter := NewWriterInterceptor(w, r, fn)\n\t\t\tdefer h.ServeHTTP(writer, r)\n\n\t\t\tnotifier, ok := w.(http.CloseNotifier)\n\t\t\tif !ok {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnotify := notifier.CloseNotify()\n\t\t\tgo func() {\n\t\t\t\t<-notify\n\t\t\t\twriter.Close()\n\t\t\t}()\n\t\t})\n\t}\n}", "func (f *Filter) ModifyResponse(ctx *martian.Context, res *http.Response) error {\n\tif f.matches(res.Request) && f.resmod != nil {\n\t\treturn f.resmod.ModifyResponse(ctx, res)\n\t}\n\n\treturn nil\n}", "func (w *response) Hijack() { w.hijacked = true }", "func Interceptor(opts ...Option) gin.HandlerFunc {\n\tset := newOptionSet(opts...)\n\n\treturn func(ctx *gin.Context) {\n\t\tctx.Set(rkgininter.RpcEntryNameKey, set.EntryName)\n\n\t\trequestId := rkcommon.GenerateRequestId()\n\t\tctx.Header(rkginctx.RequestIdKey, requestId)\n\n\t\tevent := rkginctx.GetEvent(ctx)\n\t\tevent.SetRequestId(requestId)\n\t\tevent.SetEventId(requestId)\n\n\t\tctx.Header(set.AppNameKey, rkentry.GlobalAppCtx.GetAppInfoEntry().AppName)\n\t\tctx.Header(set.AppVersionKey, rkentry.GlobalAppCtx.GetAppInfoEntry().Version)\n\n\t\tnow := time.Now()\n\t\tctx.Header(set.AppUnixTimeKey, now.Format(time.RFC3339Nano))\n\t\tctx.Header(set.ReceivedTimeKey, now.Format(time.RFC3339Nano))\n\n\t\tctx.Next()\n\t}\n}", "func (forward *ForwardHandler) OnResponse(filters ...Filter) *RespFilterGroup {\n\treturn &RespFilterGroup{ctx: forward.Ctx, filters: filters}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
ZK assisted RingBuilder implementation. The ring type or the parameters might be designated via the zookeeper.
func (c *ZKCluster) BuildRing(shards []Shard) *HashRing { switch strings.ToLower(c.info.Options.RingType) { case "consistent": nreplica, err := strconv.Atoi(c.info.Options.RingParams) if err != nil { nreplica = len(shards) } ring := &ConsistentRing{ Nreplica: nreplica, } return ring.BuildRing(shards) default: return nil } }
[ "func parseRing(hostname, algo string, replicas int) *hashing.JSONRingType {\n\tif flag.NArg() < 1 {\n\t\tlog.Printf(\"You must have at least 1 node in your hash ring\")\n\t\tusage()\n\t\tos.Exit(1)\n\t}\n\tring := new(hashing.JSONRingType)\n\tring.Name = hostname\n\tring.Algo = algo\n\tring.Replicas = replicas\n\tfor _, v := range flag.Args() {\n\t\tn, err := hashing.NewNodeParser(v)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Error parsing hashring: %s\", err.Error())\n\t\t}\n\t\tring.Nodes = append(ring.Nodes, n)\n\t}\n\n\treturn ring\n}", "func (b *Builder) Build() (*ZkPlus, error) {\n\tprefix := b.pathPrefix\n\tif len(prefix) == 0 {\n\t\tprefix = \"\"\n\t} else if prefix[0] != '/' {\n\t\treturn nil, errInvalidPathPrefix\n\t} else if prefix[len(prefix)-1] == '/' {\n\t\treturn nil, errInvalidPathSuffix\n\t}\n\tb.logger.Log(logkey.ZkPrefix, prefix, \"new with prefix\")\n\n\tret := &ZkPlus{\n\t\tpathPrefix: prefix,\n\t\tlogger: b.logger,\n\t\tcreateRoot: b.createRoot,\n\n\t\tzkConnector: b.zkConnector,\n\t\texposedChan: make(chan zk.Event),\n\t\tshouldQuit: make(chan chan struct{}),\n\t\taskForConn: make(chan chan zktest.ZkConnSupported),\n\t}\n\tgo ret.eventLoop()\n\treturn ret, nil\n}", "func NewRing(ctx context.Context, keyConfig *schema.EncryptionKeys) (*Ring, error) {\n\textsvc, err := NewKey(ctx, keyConfig.ExternalServiceKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Ring{\n\t\tExternalServiceKey: extsvc,\n\t}, nil\n}", "func Builder(bidderName openrtb_ext.BidderName, config config.Adapter, server config.Server) (adapters.Bidder, error) {\n\tendpoint := config.Endpoint\n\tif endpoint == \"\" {\n\t\tendpoint = Endpoint // Hardcoded default\n\t}\n\n\tbidder := &TelariaAdapter{\n\t\tURI: endpoint,\n\t}\n\treturn bidder, nil\n}", "func Create(conf *Config, trans Transport) (*Ring, error) {\n\t// Initialize the hash bits\n\tconf.hashBits = conf.HashFunc().Size() * 8\n\n\t// Create and initialize a ring\n\tring := &Ring{}\n\tring.init(conf, trans)\n\tring.setLocalSuccessors()\n\tring.setLocalPredecessors()\n\tring.schedule()\n\treturn ring, nil\n}", "func Builder(_ openrtb_ext.BidderName, config config.Adapter, _ config.Server) (adapters.Bidder, error) {\n\tbidder := &adapter{\n\t\tendpoint: config.Endpoint,\n\t}\n\treturn bidder, nil\n}", "func Builder(bidderName openrtb_ext.BidderName, config config.Adapter, server config.Server) (adapters.Bidder, error) {\n\tendpointTemplate, err := template.New(\"endpointTemplate\").Parse(config.Endpoint)\n\tif err != nil {\n\t\treturn nil, errors.New(\"Unable to parse endpoint template\")\n\t}\n\n\twhiteSpace := regexp.MustCompile(`\\s+`)\n\n\tbidder := &AdOceanAdapter{\n\t\tendpointTemplate: endpointTemplate,\n\t\tmeasurementCode: whiteSpace.ReplaceAllString(measurementCode, \" \"),\n\t}\n\treturn bidder, nil\n}", "func createChordRing() {\n\tsuccessor.Address = selfaddr\n\tsuccessor.Chordid = chordid\n\tpredecessor.Address = successor.Address\n\tpredecessor.Chordid = successor.Chordid\n\tfingertable[1] = NodeInfo{chordid, selfaddr}\n\tfmt.Println(\"Created ChordRing\")\n}", "func InitRing(myIp string, myId int) {\n\tfmt.Println(\"\\n------------------------------------------------------------------------------\")\n\tfmt.Println(\"Test 1: creating/join chord ring ...\")\n\tstart1 := time.Now()\n\n\t// scan for ring\n\tfmt.Println(\"\\nScanning for ring ...\")\n\tipInRing, _ := chord.CheckRing()\n\tringSize := len(ipInRing)\n\tfmt.Println(\"Ring scan completed!\\nNodes in ring: \", ipInRing, \"\\nRing size: \", ringSize)\n\n\t// init node\n\tfmt.Println(\"\\nCreating node ...\")\n\tchord.ChordNode = &chord.Node{\n\t\tIdentifier: myId,\n\t\tIP: myIp,\n\t}\n\tfmt.Println(\"\\nActivating node ...\")\n\tgo node_listen(myIp)\n\n\t// create / join ring\n\tif ringSize == 0 {\n\t\t// Ring does NOT exists => CREATE ring\n\t\tfmt.Println(\"\\nRing does NOT exists!\\nCreating new ring at \", myIp)\n\t\tchord.ChordNode.CreateNodeAndJoin(nil)\n\t\tfmt.Println(\"New ring successfully created!\")\n\t} else {\n\t\t// Ring EXISTS => JOIN ring\n\t\tfmt.Println(\"\\nRing does EXISTS!\")\n\t\tremoteIp := ipInRing[0]\n\t\tremoteId := chord.Hash(remoteIp)\n\t\tremoteNode := &chord.RemoteNode{\n\t\t\tIdentifier: remoteId,\n\t\t\tIP: remoteIp,\n\t\t}\n\n\t\tchord.ChordNode.IP = myIp\n\t\tchord.ChordNode.Identifier = myId\n\n\t\tfmt.Println(\"Joining ring via \", remoteId, \"(\", remoteIp, \")\")\n\t\tchord.ChordNode.CreateNodeAndJoin(remoteNode)\n\t\tfmt.Println(\"Node \", myId, \" successfully joined ring!\")\n\t}\n\n\tend1 := time.Now()\n\tduration1 := end1.Sub(start1)\n\tfmt.Println(\"Test 1 COMPLETED!!!\\nDuration \", duration1)\n\tfmt.Println(\"------------------------------------------------------------------------------\\n \")\n\tchord.ChordNode.PrintNode()\n}", "func Builder(bidderName openrtb_ext.BidderName, config config.Adapter, server config.Server) (adapters.Bidder, error) {\n\ttemplate, err := template.New(\"endpointTemplate\").Parse(config.Endpoint)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to parse endpoint url template: %v\", err)\n\t}\n\n\tbidder := &SmartRTBAdapter{\n\t\tEndpointTemplate: template,\n\t}\n\treturn bidder, nil\n}", "func Builder(bidderName openrtb_ext.BidderName, config config.Adapter, server config.Server) (adapters.Bidder, error) {\n\tbidder := &adapter{\n\t\tURI: config.Endpoint,\n\t}\n\treturn bidder, nil\n}", "func (ls *libstore) BuildRing() {\n\tls.mux.Lock()\n\tdefer ls.mux.Unlock()\n\n\tfor _, node := range ls.storageServers {\n\t\tfor _, id := range node.VirtualIDs {\n\t\t\tls.virtualIDRing[id] = node.HostPort\n\t\t}\n\t}\n}", "func Builder(bidderName openrtb_ext.BidderName, config config.Adapter, _ config.Server) (adapters.Bidder, error) {\n\tbidder := &Adapter{\n\t\turl: config.Endpoint,\n\t}\n\treturn bidder, nil\n}", "func (c *Channel) Ring(key *ari.Key) error {\n\treturn c.client.post(\"/channels/\"+key.ID+\"/ring\", nil, nil)\n}", "func Builder(bidderName openrtb_ext.BidderName, config config.Adapter, server config.Server) (adapters.Bidder, error) {\n\tbidder := &ConsumableAdapter{\n\t\tclock: realInstant{},\n\t\tendpoint: config.Endpoint,\n\t}\n\treturn bidder, nil\n}", "func (a *app) initBuilder() (di.Builder, error) {\n\treturn di.NewBuilder(\n\t\tdi.Provide(a.provideRootContext, di.Unshared()),\n\t\tdi.Provide(\n\t\t\ta.provideRootCmd,\n\t\t\tdi.Constraint(0, di.Optional(true), withPersistentPreRunner()),\n\t\t\tdi.Constraint(1, di.Optional(true), withPersistentFlags()),\n\t\t\tdi.Constraint(2, di.Optional(true), withCliCommand()),\n\t\t\tasRootCommand(),\n\t\t),\n\t\tdi.Provide(a.provideVersionCmd, AsCliCommand()),\n\t)\n}", "func (f *ring2Factory) NewBuilder() *resource.Builder {\n\tclientMapperFunc := resource.ClientMapperFunc(f.objectMappingFactory.ClientForMapping)\n\tmapper, typer := f.objectMappingFactory.Object()\n\n\tunstructuredClientMapperFunc := resource.ClientMapperFunc(f.objectMappingFactory.UnstructuredClientForMapping)\n\n\tcategoryExpander := f.objectMappingFactory.CategoryExpander()\n\n\treturn resource.NewBuilder(\n\t\t&resource.Mapper{\n\t\t\tRESTMapper: mapper,\n\t\t\tObjectTyper: typer,\n\t\t\tClientMapper: clientMapperFunc,\n\t\t\tDecoder: InternalVersionDecoder(),\n\t\t},\n\t\t&resource.Mapper{\n\t\t\tRESTMapper: mapper,\n\t\t\tObjectTyper: typer,\n\t\t\tClientMapper: unstructuredClientMapperFunc,\n\t\t\tDecoder: unstructured.UnstructuredJSONScheme,\n\t\t},\n\t\tcategoryExpander,\n\t)\n}", "func Builder(bidderName openrtb_ext.BidderName, config config.Adapter, server config.Server) (adapters.Bidder, error) {\n\tbidder := &MgidAdapter{\n\t\tendpoint: config.Endpoint,\n\t}\n\treturn bidder, nil\n}", "func Builder(bidderName openrtb_ext.BidderName, config config.Adapter) (adapters.Bidder, error) {\n\tbidder := &DaxAdapter{\n\t\tendpoint: config.Endpoint,\n\t}\n\treturn bidder, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
The watcher for status updates via the zookeeper.
func (c *ZKCluster) watchStatusUpdates() { status_node := ZK_ROOT + "/" + c.info.Name + "/status" for { exists, stat, event, err := c.zc.conn.ExistsW(status_node) if err == nil && exists && stat.Version > c.version { if statusbytes, stat, err := c.zc.conn.Get(status_node); err == nil { var status map[string]ShardStatus if err := json.Unmarshal(statusbytes, &status); err == nil { for k, v := range status { if c.status[k] != v { c.status[k] = v c.updates <- v } } c.version = stat.Version } } } select { case <- event: case <- c.checkerdone: return } } }
[ "func (s *ConfigService) Watch(d time.Duration)", "func (s *Server) Watch(in *grpc_health_v1.HealthCheckRequest, server grpc_health_v1.Health_WatchServer) error {\n\tresp := &grpc_health_v1.HealthCheckResponse{Status: grpc_health_v1.HealthCheckResponse_SERVING}\n\treturn server.Send(resp)\n}", "func (s *server) Watch(request *healthgrpc.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error {\n\tresponse, err := s.Check(context.Background(), request)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := stream.Send(response); err != nil {\n\t\treturn err\n\t}\n\n\tc := make(chan healthgrpc.HealthCheckResponse_ServingStatus)\n\tticker := time.NewTicker(1 * time.Second)\n\tdefer ticker.Stop()\n\n\tgo func() {\n\t\tresponseStatus := response.Status\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tupdatedResponse := s.serviceMap[request.Service].Check()\n\t\t\t\tif updatedResponse != responseStatus {\n\t\t\t\t\tc <- updatedResponse\n\t\t\t\t\tresponseStatus = updatedResponse\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tfor {\n\t\tselect {\n\t\tcase <-stream.Context().Done():\n\t\t\treturn nil\n\t\tcase s := <-c:\n\t\t\tif err := stream.SendMsg(&healthgrpc.HealthCheckResponse{Status: s}); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n}", "func (t *Tuner) WatchStatus(handler func(Status)) watch.Watch {\n\treturn t.status.Watch(handler)\n}", "func (s *InnerService) Watch(bcsTLSConfig bcsoptions.TLS) error {\n\tglog.Infof(\"start to watch service[%s] from ZK\", s.name)\n\n\tif s.eventChan == nil {\n\t\treturn fmt.Errorf(\"event channel not initialized\")\n\t}\n\n\tfor data := range s.eventChan {\n\t\tglog.Infof(\"received ZK event, Server: %+v\", data.Server)\n\t\tif data.Err != nil {\n\t\t\tglog.Errorf(\"%s service discover failed, %+v\", s.name, data.Err)\n\t\t\tcontinue\n\t\t}\n\t\ts.update(data.Server, bcsTLSConfig)\n\t}\n\n\treturn nil\n}", "func createStatusReconciler(adapter *Adapter) chan statusReconcilerChannelEntry {\n\n\tsenderChannel := make(chan statusReconcilerChannelEntry)\n\n\tgo func() {\n\n\t\t// This map is the single source of truth re: what odo expects the cluster namespace to look like; when\n\t\t// new events are received that contain pod data that differs from this, the user should be informed of the delta\n\t\t// (and this 'truth' should be updated.)\n\t\t//\n\t\t// Map key is pod UID\n\t\tmostRecentPodStatus := map[string]*KubernetesPodStatus{}\n\n\t\tfor {\n\n\t\t\tentry := <-senderChannel\n\n\t\t\tif entry.isWatchThreadRestarted {\n\t\t\t\t// On network disconnect, clear the status map\n\t\t\t\tmostRecentPodStatus = map[string]*KubernetesPodStatus{}\n\t\t\t}\n\n\t\t\tif entry.err != nil {\n\t\t\t\tadapter.Logger().ReportError(entry.err, machineoutput.TimestampNow())\n\t\t\t\tklog.V(4).Infof(\"Error received on status reconciler channel %v\", entry.err)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif entry.pods == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Map key is pod UID (we don't use the map value)\n\t\t\tentryPodUIDs := map[string]string{}\n\t\t\tfor _, pod := range entry.pods {\n\t\t\t\tentryPodUIDs[string(pod.UID)] = string(pod.UID)\n\t\t\t}\n\n\t\t\tchangeDetected := false\n\n\t\t\t// This section of the algorithm only works if the entry was from a podlist (which contains the full list\n\t\t\t// of all pods that exist in the namespace), rather than the watch (which contains only one pod in\n\t\t\t// the namespace.)\n\t\t\tif entry.isCompleteListOfPods {\n\t\t\t\t// Detect if there exists a UID in mostRecentPodStatus that is not in entry; if so, one or more previous\n\t\t\t\t// pods have disappeared, so set changeDetected to true.\n\t\t\t\tfor mostRecentPodUID := range mostRecentPodStatus {\n\t\t\t\t\tif _, exists := entryPodUIDs[mostRecentPodUID]; !exists {\n\t\t\t\t\t\tklog.V(4).Infof(\"Status change detected: Could not find previous pod %s in most recent pod status\", mostRecentPodUID)\n\t\t\t\t\t\tdelete(mostRecentPodStatus, mostRecentPodUID)\n\t\t\t\t\t\tchangeDetected = true\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif !changeDetected {\n\n\t\t\t\t// For each pod we received a status for, determine if it is a change, and if so, update mostRecentPodStatus\n\t\t\t\tfor _, pod := range entry.pods {\n\t\t\t\t\tpodVal := CreateKubernetesPodStatusFromPod(*pod)\n\n\t\t\t\t\tif entry.isDeleteEventFromWatch {\n\t\t\t\t\t\tdelete(mostRecentPodStatus, string(pod.UID))\n\t\t\t\t\t\tklog.V(4).Infof(\"Removing deleted pod %s\", pod.UID)\n\t\t\t\t\t\tchangeDetected = true\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\n\t\t\t\t\t// If a pod exists in the new pod status, that we have not seen before, then a change is detected.\n\t\t\t\t\tprevValue, exists := mostRecentPodStatus[string(pod.UID)]\n\t\t\t\t\tif !exists {\n\t\t\t\t\t\tmostRecentPodStatus[string(pod.UID)] = &podVal\n\t\t\t\t\t\tklog.V(4).Infof(\"Adding new pod to most recent pod status %s\", pod.UID)\n\t\t\t\t\t\tchangeDetected = true\n\n\t\t\t\t\t} else {\n\t\t\t\t\t\t// If the pod exists in both the old and new status, then do a deep comparison\n\t\t\t\t\t\tareEqual := areEqual(&podVal, prevValue)\n\t\t\t\t\t\tif areEqual != \"\" {\n\t\t\t\t\t\t\tmostRecentPodStatus[string(pod.UID)] = &podVal\n\t\t\t\t\t\t\tklog.V(4).Infof(\"Pod value %s has changed: %s\", pod.UID, areEqual)\n\t\t\t\t\t\t\tchangeDetected = true\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// On change: output all pods (our full knowledge of the odo-managed components in the namespace) as a single JSON event\n\t\t\tif changeDetected {\n\n\t\t\t\tpodStatuses := []machineoutput.KubernetesPodStatusEntry{}\n\n\t\t\t\tfor _, v := range mostRecentPodStatus {\n\n\t\t\t\t\tstartTime := \"\"\n\t\t\t\t\tif v.StartTime != nil {\n\t\t\t\t\t\tstartTime = machineoutput.FormatTime(*v.StartTime)\n\t\t\t\t\t}\n\n\t\t\t\t\tpodStatuses = append(podStatuses, machineoutput.KubernetesPodStatusEntry{\n\t\t\t\t\t\tName: v.Name,\n\t\t\t\t\t\tContainers: v.Containers,\n\t\t\t\t\t\tInitContainers: v.InitContainers,\n\t\t\t\t\t\tLabels: v.Labels,\n\t\t\t\t\t\tPhase: v.Phase,\n\t\t\t\t\t\tUID: v.UID,\n\t\t\t\t\t\tStartTime: startTime,\n\t\t\t\t\t})\n\t\t\t\t}\n\n\t\t\t\tadapter.Logger().KubernetesPodStatus(podStatuses, machineoutput.TimestampNow())\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn senderChannel\n}", "func (s *healthServer) Watch(\n\tin *grpc_health_v1.HealthCheckRequest,\n\tw grpc_health_v1.Health_WatchServer,\n) error {\n\treturn status.Errorf(codes.Unimplemented, \"(c *healthServer) Watch() is not implemented\")\n}", "func (s *stateManager) Watch(watcher *AllocationWatcher) func() {\n\tstopChan := make(chan interface{})\n\ts.stopChan = append(s.stopChan, stopChan)\n\tctx := context.Background()\n\n\tkey := fmt.Sprintf(\"%s/allocations\", etcdPrefix)\n\twatchChan := s.cli.Watch(ctx, key, clientv3.WithPrefix(), clientv3.WithPrevKV())\n\n\tstopFunc := func() {\n\t\tstopChan <- true\n\t}\n\n\t// Start a new thread and watch for changes in etcd\n\tgo s.watchChannel(watchChan, stopChan, watcher)\n\n\treturn stopFunc\n}", "func (h *HealthImpl) Watch(in *grpc_health_v1.HealthCheckRequest, stream grpc_health_v1.Health_WatchServer) error {\n\treturn nil\n}", "func (m *manager) Watch(addr string) {\n\tm.watch.Watch(addr)\n}", "func (l *EtcdLeader) Watch(ctx context.Context) {\n\tl.wrapper.WatchLeader(ctx, l.pariticipant, l.revision)\n}", "func LeaderWatcher() {\n\t//initial check\n\tresp, err := kapi.Get(context.Background(), \"/leader\", nil)\n\tif err != nil {\n\t\tclierr := err.(client.Error)\n\t\tlog.Println(clierr.Code)\n\t\tSetLeader()\n\t} else {\n\t\tLeader = resp.Node.Value\n\t}\n\n\t//keep watching for changes\n\twatcher := kapi.Watcher(\"/leader\", nil)\n\tfor {\n\t\tresp, err := watcher.Next(context.Background())\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t}\n\n\t\tif resp.Action == \"expire\" {\n\t\t\tSetLeader()\n\t\t} else {\n\t\t\tLeader = resp.Node.Value\n\t\t\tlog.Printf(\"Current Leader: %s\\n\", Leader)\n\t\t}\n\t}\n}", "func (k *Krypton) WatchDeployStatus(appname, entrypoint, nodename string) etcdclient.Watcher {\n\tif appname == \"\" {\n\t\tentrypoint = \"\"\n\t}\n\tif entrypoint == \"\" {\n\t\tnodename = \"\"\n\t}\n\tkey := filepath.Join(containerDeployPrefix, appname, entrypoint, nodename)\n\treturn k.etcd.Watcher(key, &etcdclient.WatcherOptions{Recursive: true})\n}", "func watcher(configModel model.Config) {\n\t// Set the client variable\n\tconfig.Client = configModel.Client.Name\n\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer watcher.Close()\n\n\tdone := make(chan struct{})\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event := <-watcher.Events:\n\t\t\t\tif event.Op&fsnotify.Write == fsnotify.Write {\n\t\t\t\t\tlogs.INFO.Println(\"Modified file -> \", event.Name)\n\t\t\t\t\t// When the file name has not been defined, it is time to\n\t\t\t\t\t// use the SetFile() method to add a new file to read.\n\t\t\t\t\tif filename == \"\" {\n\t\t\t\t\t\tstore.SetFile(event.Name)\n\t\t\t\t\t\tfilename = event.Name\n\t\t\t\t\t}\n\t\t\t\t\tif filename != \"\" && filename != event.Name {\n\t\t\t\t\t\tlogs.INFO.Println(\"Reset seek\")\n\t\t\t\t\t\tseek = 0\n\t\t\t\t\t}\n\t\t\t\t\treadLines(event.Name)\n\t\t\t\t}\n\t\t\tcase err := <-watcher.Errors:\n\t\t\t\tlogs.CRITICAL.Println(\"Error on watcher: \", err)\n\t\t\t}\n\t\t}\n\t}()\n\terr = watcher.Add(configModel.Pathlog.Name)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\t<-done\n}", "func (c *PumpsClient) watchStatus(revision int64) {\n\trch := c.EtcdRegistry.WatchNode(c.ctx, c.nodePath, revision)\n\n\tfor {\n\t\tselect {\n\t\tcase <-c.ctx.Done():\n\t\t\tlog.Info(\"watch status finished\", zap.String(\"category\", \"pumps client\"))\n\t\t\treturn\n\t\tcase wresp := <-rch:\n\t\t\tif wresp.Err() != nil {\n\t\t\t\t// meet error, watch from the latest revision.\n\t\t\t\t// pump will update the key periodly, it's ok for we to lost some event here\n\t\t\t\tlog.Warn(\"watch status meet error\", zap.String(\"category\", \"pumps client\"), zap.Error(wresp.Err()))\n\t\t\t\trch = c.EtcdRegistry.WatchNode(c.ctx, c.nodePath, 0)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tfor _, ev := range wresp.Events {\n\t\t\t\tstatus := &node.Status{}\n\t\t\t\terr := json.Unmarshal(ev.Kv.Value, &status)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(\"unmarshal pump status failed\", zap.String(\"category\", \"pumps client\"), zap.ByteString(\"value\", ev.Kv.Value), zap.Error(err))\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tswitch ev.Type {\n\t\t\t\tcase mvccpb.PUT:\n\t\t\t\t\tif !c.exist(status.NodeID) {\n\t\t\t\t\t\tlog.Info(\"find a new pump\", zap.String(\"category\", \"pumps client\"), zap.String(\"NodeID\", status.NodeID))\n\t\t\t\t\t\tc.addPump(NewPumpStatus(status, c.Security), true)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\n\t\t\t\t\tpump, availableChanged, available := c.updatePump(status)\n\t\t\t\t\tif availableChanged {\n\t\t\t\t\t\tlog.Info(\"pump's state is changed\", zap.String(\"category\", \"pumps client\"), zap.String(\"NodeID\", pump.Status.NodeID), zap.String(\"state\", status.State))\n\t\t\t\t\t\tc.setPumpAvailable(pump, available)\n\t\t\t\t\t}\n\n\t\t\t\tcase mvccpb.DELETE:\n\t\t\t\t\t// now will not delete pump node in fact, just for compatibility.\n\t\t\t\t\tnodeID := node.AnalyzeNodeID(string(ev.Kv.Key))\n\t\t\t\t\tlog.Info(\"remove pump\", zap.String(\"category\", \"pumps client\"), zap.String(\"NodeID\", nodeID))\n\t\t\t\t\tc.removePump(nodeID)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func Watch(c cookoo.Context, p *cookoo.Params) (interface{}, cookoo.Interrupt) {\n\tdir := cookoo.GetString(\"dir\", \".\", p)\n\troute := cookoo.GetString(\"update\", \"@update\", p)\n\n\tr, ok := c.Has(\"router\")\n\tif !ok {\n\t\treturn time.Now(), fmt.Errorf(\"Could not find 'router' in context.\")\n\t}\n\n\trouter := r.(*cookoo.Router)\n\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer watcher.Close()\n\twatcher.Add(dir)\n\n\tfmt.Printf(\"[INFO] Watching %s for changes to .codl files.\\n\", dir)\n\n\t// Watch for updates to files.\n\tfor {\n\t\tselect {\n\t\tcase good := <-watcher.Events:\n\n\t\t\t// Look for create, write, and rename events.\n\t\t\tswitch good.Op {\n\t\t\t//case fsnotify.Create, fsnotify.Write, fsnotify.Rename:\n\t\t\tcase fsnotify.Write, fsnotify.Create:\n\t\t\t\tif path.Ext(good.Name) != \".codl\" {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tfmt.Printf(\"[INFO] %s has changed. Updating. (%s)\\n\", good.Name, good.String())\n\t\t\t\tc.Put(\"files\", []string{good.Name})\n\t\t\t\terr := router.HandleRequest(route, c, false)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Printf(\"[ERROR] %s\\n\", err)\n\t\t\t\t\t//return time.Now(), err\n\t\t\t\t}\n\t\t\t\tc.Put(\"lastUpdated\", time.Now())\n\n\t\t\t// Log but otherwise ignore Remove.\n\t\t\tcase fsnotify.Remove:\n\t\t\t\tfmt.Printf(\"[INFO] %s has been removed.\\n\", good.Name)\n\t\t\t}\n\t\tcase bad := <-watcher.Errors:\n\t\t\tc.Logf(\"warn\", \"Error watching: %s\", bad.Error())\n\t\t}\n\t}\n}", "func (api *versionAPI) Watch(handler VersionHandler) error {\n\tapi.ct.startWorkerPool(\"Version\")\n\treturn api.ct.WatchVersion(handler)\n}", "func (s *ServiceStorage) WatchStatus(ctx context.Context, service chan *types.Service) error {\n\n\tlog.V(logLevel).Debug(\"storage:etcd:service:> watch service\")\n\n\tconst filter = `\\b\\/` + serviceStorage + `\\/(.+):(.+)/status\\b`\n\tclient, destroy, err := getClient(ctx)\n\tif err != nil {\n\t\tlog.V(logLevel).Errorf(\"storage:etcd:service:> watch service err: %s\", err.Error())\n\t\treturn err\n\t}\n\tdefer destroy()\n\n\tr, _ := regexp.Compile(filter)\n\tkey := keyCreate(serviceStorage)\n\tcb := func(action, key string, _ []byte) {\n\t\tkeys := r.FindStringSubmatch(key)\n\t\tif len(keys) < 3 {\n\t\t\treturn\n\t\t}\n\n\t\tif action == \"delete\" {\n\t\t\treturn\n\t\t}\n\n\t\tif d, err := s.Get(ctx, keys[1], keys[2]); err == nil {\n\t\t\tservice <- d\n\t\t}\n\t}\n\n\tif err := client.Watch(ctx, key, filter, cb); err != nil {\n\t\tlog.V(logLevel).Errorf(\"storage:etcd:service:> watch service err: %s\", err.Error())\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (e *etcdCacheEntry) Watch(c *EtcdConfig) {\n e.Lock()\n defer e.Unlock()\n e.startWatching(c)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
doFetch do actual http call and will use given ctx as http request context.
func doFetch(ctx context.Context, url string, wg *sync.WaitGroup, result chan<- fetchResult) { start := time.Now() defer wg.Done() var ( err error req *http.Request res *http.Response body []byte ) resChan := fetchResult{url: url, err: err} req, err = http.NewRequest("GET", url, nil) if err != nil { resChan.totalTime = time.Since(start).Seconds() resChan.err = err result <- resChan return } res, err = httpClient().Do(req.WithContext(ctx)) if err != nil { if ctx.Err() != nil { err = errTimeout } resChan.totalTime = time.Since(start).Seconds() resChan.err = err result <- resChan return } defer res.Body.Close() body, err = ioutil.ReadAll(res.Body) if err != nil { resChan.totalTime = time.Since(start).Seconds() resChan.err = err result <- resChan return } var r interface{} err = json.Unmarshal(body, &r) resChan.totalTime = time.Since(start).Seconds() resChan.err = err resChan.content = fmt.Sprintf("%v", r) result <- resChan }
[ "func (tcl *HTTPClient) Do(ctx context.Context, req *http.Request) (*http.Response, error) {\n\turl := \"\"\n\tif req.URL != nil {\n\t\turl = req.URL.String()\n\t}\n\n\tstart := time.Now()\n\trsp, err := tcl.cl.Do(ctx, req)\n\tif err != nil {\n\t\tObserveCount(tcl.provider, url, req.Method, 0)\n\t\tObserveLatency(tcl.provider, url, req.Method, 0, time.Since(start))\n\t\treturn rsp, err\n\t}\n\n\tObserveCount(tcl.provider, url, req.Method, rsp.StatusCode)\n\tObserveLatency(tcl.provider, url, req.Method, rsp.StatusCode, time.Since(start))\n\n\treturn rsp, err\n}", "func DoCtx(ctx context.Context, req *http.Request, resp interface{}) (*http.Response, error) {\n\tr := req.Clone(ctx)\n\n\treturn Do(r, resp)\n}", "func do(ctx context.Context, client *http.Client, req *http.Request) (*http.Response, error) {\n\tif client == nil {\n\t\tclient = http.DefaultClient\n\t}\n\tresp, err := client.Do(req.WithContext(ctx))\n\t// If we got an error, and the context has been canceled,\n\t// the context's error is probably more useful.\n\tif err != nil {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\terr = ctx.Err()\n\t\tdefault:\n\t\t}\n\t}\n\treturn resp, err\n}", "func (this *DefaultExtender) Fetch(ctx *URLContext, userAgent string, headRequest bool) (*http.Response, error) {\n\tvar reqType string\n\n\t// Prepare the request with the right user agent\n\tif headRequest {\n\t\treqType = \"HEAD\"\n\t} else {\n\t\treqType = \"GET\"\n\t}\n\treq, e := http.NewRequest(reqType, ctx.url.String(), nil)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treq.Header.Set(\"User-Agent\", userAgent)\n\treturn HttpClient.Do(req)\n}", "func (g *Goproxy) serveFetch(\n\trw http.ResponseWriter,\n\treq *http.Request,\n\tname string,\n\ttempDir string,\n) {\n\tf, err := newFetch(g, name, tempDir)\n\tif err != nil {\n\t\tresponseNotFound(rw, req, 86400, err)\n\t\treturn\n\t}\n\n\tvar isDownload bool\n\tswitch f.ops {\n\tcase fetchOpsDownloadInfo, fetchOpsDownloadMod, fetchOpsDownloadZip:\n\t\tisDownload = true\n\t}\n\n\tnoFetch, _ := strconv.ParseBool(req.Header.Get(\"Disable-Module-Fetch\"))\n\tif noFetch {\n\t\tvar cacheControlMaxAge int\n\t\tif isDownload {\n\t\t\tcacheControlMaxAge = 604800\n\t\t} else {\n\t\t\tcacheControlMaxAge = 60\n\t\t}\n\n\t\tg.serveCache(\n\t\t\trw,\n\t\t\treq,\n\t\t\tf.name,\n\t\t\tf.contentType,\n\t\t\tcacheControlMaxAge,\n\t\t\tfunc() {\n\t\t\t\tresponseNotFound(\n\t\t\t\t\trw,\n\t\t\t\t\treq,\n\t\t\t\t\t60,\n\t\t\t\t\t\"temporarily unavailable\",\n\t\t\t\t)\n\t\t\t},\n\t\t)\n\n\t\treturn\n\t}\n\n\tif isDownload {\n\t\tg.serveCache(rw, req, f.name, f.contentType, 604800, func() {\n\t\t\tg.serveFetchDownload(rw, req, f)\n\t\t})\n\t\treturn\n\t}\n\n\tfr, err := f.do(req.Context())\n\tif err != nil {\n\t\tg.serveCache(rw, req, f.name, f.contentType, 60, func() {\n\t\t\tg.logErrorf(\n\t\t\t\t\"failed to %s module version: %s: %v\",\n\t\t\t\tf.ops,\n\t\t\t\tf.name,\n\t\t\t\terr,\n\t\t\t)\n\t\t\tresponseError(rw, req, err, true)\n\t\t})\n\t\treturn\n\t}\n\n\tcontent, err := fr.Open()\n\tif err != nil {\n\t\tg.logErrorf(\"failed to open fetch result: %s: %v\", f.name, err)\n\t\tresponseInternalServerError(rw, req)\n\t\treturn\n\t}\n\tdefer content.Close()\n\n\tif err := g.putCache(req.Context(), f.name, content); err != nil {\n\t\tg.logErrorf(\"failed to cache module file: %s: %v\", f.name, err)\n\t\tresponseInternalServerError(rw, req)\n\t\treturn\n\t} else if _, err := content.Seek(0, io.SeekStart); err != nil {\n\t\tg.logErrorf(\n\t\t\t\"failed to seek fetch result content: %s: %v\",\n\t\t\tf.name,\n\t\t\terr,\n\t\t)\n\t\tresponseInternalServerError(rw, req)\n\t\treturn\n\t}\n\n\tresponseSuccess(rw, req, content, f.contentType, 60)\n}", "func (c *Client) Fetch(ctx context.Context, p *FetchPayload) (err error) {\n\t_, err = c.FetchEndpoint(ctx, p)\n\treturn\n}", "func (c *client) do(ctx context.Context, req *http.Request, result interface{}) error {\n\treq = req.WithContext(ctx)\n\tresp, err := httpClient.Do(req)\n\tif err != nil {\n\t\t// Request failed\n\t\treturn errors.WithStack(err)\n\t}\n\n\t// Read content\n\tdefer resp.Body.Close()\n\tbody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\n\t// Check status\n\tstatusCode := resp.StatusCode\n\tif statusCode != 200 {\n\t\tif err := provisioner.ParseResponseError(resp, body); err != nil {\n\t\t\treturn errors.WithStack(err)\n\t\t}\n\t\treturn errors.WithStack(errors.Newf(\"Invalid status %d\", statusCode))\n\t}\n\n\t// Got a success status\n\tif result != nil {\n\t\tif err := json.Unmarshal(body, result); err != nil {\n\t\t\tmethod := resp.Request.Method\n\t\t\turl := resp.Request.URL.String()\n\t\t\treturn errors.Wrapf(err, \"Failed decoding response data from %s request to %s: %v\", method, url, err)\n\t\t}\n\t}\n\treturn nil\n}", "func HttpDo(ctx context.Context, request *http.Request, f func(*http.Response, error) error) error {\n\ttr := &http.Transport{}\n\tclient := &http.Client{Transport: tr}\n\tc := make(chan error, 1)\n\n\tgo func() {\n\t\tc <- f(client.Do(request))\n\t}()\n\n\tselect {\n\tcase <-ctx.Done():\n\t\ttr.CancelRequest(request)\n\t\t<-c\n\t\treturn ctx.Err()\n\tcase err := <-c:\n\t\treturn err\n\t}\n}", "func DoReqWithCtx(req *http.Request, f func(*http.Response, error) error) error {\n\tctx := req.Context()\n\tc := make(chan error, 1)\n\tgo func() { c <- f(http.DefaultClient.Do(req)) }()\n\tselect {\n\tcase <-ctx.Done():\n\t\t<-c\n\t\treturn ctx.Err()\n\n\tcase err := <-c:\n\t\treturn err\n\t}\n}", "func (c Client) Do(ctx context.Context, r storage.Runner) (storage.Fetcher, error) {\n\tif err := ctx.Err(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"Context already closed\")\n\t}\n\tres := make(chan storage.Fetcher, 1)\n\tch := run(res, r)\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn nil, errors.Wrap(ctx.Err(), \"Context closed\")\n\tcase result := <-res:\n\t\treturn result, nil\n\tcase err := <-ch:\n\t\treturn nil, errors.Wrap(err, \"Transaction failed\")\n\t}\n}", "func (c Client) Fetch() (*FetchContextResponse, error) {\n\treturn c.FetchWithContext(context.Background())\n}", "func (c *Client) FetchData(ctx context.Context, url string) ([]byte, error) {\n\n\t// Implement semaphores to ensure maximum concurrency threshold.\n\tc.semaphore <- struct{}{}\n\tdefer func() { <-c.semaphore }()\n\n\t// If there is an in-flight request for a unique URL, send response\n\t// from the in-flight request. Else, create the in-flight request.\n\tresponseRaw, err, shared := c.RequestGroup.Do(url, func() (interface{}, error) {\n\t\treturn c.fetchResponse(ctx)\n\t})\n\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\n\tlog.Infof(\"in-flight status : %t\", shared)\n\n\t//time.Sleep(time.Second * 4)\n\n\tresponse := responseRaw.([]byte)\n\n\treturn response, err\n}", "func (ri *RestInvoker) ContextDo(ctx context.Context, req *rest.Request, options ...InvocationOption) (*rest.Response, error) {\n\tif string(req.GetRequest().URL.Scheme) != \"cse\" {\n\t\treturn nil, fmt.Errorf(\"scheme invalid: %s, only support cse://\", req.GetRequest().URL.Scheme)\n\t}\n\n\topts := getOpts(req.GetRequest().Host, options...)\n\topts.Protocol = common.ProtocolRest\n\n\tresp := rest.NewResponse()\n\n\tinv := invocation.New(ctx)\n\twrapInvocationWithOpts(inv, opts)\n\tinv.MicroServiceName = req.GetRequest().Host\n\t// TODO load from openAPI schema\n\t// inv.SchemaID = schemaID\n\t// inv.OperationID = operationID\n\tinv.Args = req\n\tinv.Reply = resp\n\tinv.URLPathFormat = req.Req.URL.Path\n\n\tinv.SetMetadata(common.RestMethod, req.GetMethod())\n\n\terr := ri.invoke(inv)\n\treturn resp, err\n}", "func (tc *TracingClient) Do(req *http.Request) (*http.Response, error) {\n\tvar parentContext opentracing.SpanContext\n\n\tif span := opentracing.SpanFromContext(req.Context()); span != nil {\n\t\tparentContext = span.Context()\n\t}\n\n\tappSpan := tc.tracer.StartSpan(\"HTTP Client\", opentracing.ChildOf(parentContext))\n\n\text.HTTPMethod.Set(appSpan, req.Method)\n\text.HTTPUrl.Set(appSpan, req.URL.String())\n\n\tres, err := tc.Client.Do(\n\t\treq.WithContext(opentracing.ContextWithSpan(req.Context(), appSpan)),\n\t)\n\tif err != nil {\n\t\text.Error.Set(appSpan, true)\n\t\tappSpan.SetTag(\"error\", err.Error())\n\t\tappSpan.Finish()\n\t\treturn res, err\n\t}\n\n\text.HTTPStatusCode.Set(appSpan, uint16(res.StatusCode))\n\tif res.StatusCode > 399 {\n\t\text.Error.Set(appSpan, true)\n\t}\n\tif req.Method == \"HEAD\" {\n\t\tappSpan.Finish()\n\t} else {\n\t\tres.Body = &spanCloser{\n\t\t\tReadCloser: res.Body,\n\t\t\tsp: appSpan,\n\t\t\ttraceEnabled: tc.httpTrace,\n\t\t}\n\t}\n\treturn res, nil\n}", "func DoHttpRequest(cfg *config.Configuration, req *http.Request, useCreds bool) (*http.Response, error) {\n\tvar creds auth.Creds\n\tif useCreds {\n\t\tc, err := auth.GetCreds(cfg, req)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tcreds = c\n\t}\n\n\treturn doHttpRequest(cfg, req, creds)\n}", "func (fs FetchService) Fetch(req interface{}) (interface{}, error) {\n\t res, err := fs.Response(req)\n\t if err != nil {\n\t\t \treturn nil, err\n\t\t }\n\t return res, nil\n\n}", "func DoRequest(ctx context.Context, req *http.Request) (*http.Response, error) {\n\tclient := ContextClient(ctx)\n\n\tif req.UserAgent() == \"\" {\n\t\treq.Header.Set(\"User-Agent\", ContextUserAgent(ctx))\n\t}\n\n\tresp, err := client.Do(req.WithContext(ctx))\n\t// If we got an error, and the context has been canceled,\n\t// the context's error is probably more useful.\n\tif err != nil {\n\t\t<-ctx.Done()\n\t\terr = ctx.Err()\n\t}\n\n\treturn resp, err\n}", "func (ri *RestInvoker) ContextDo(ctx context.Context, req *rest.Request, options ...InvocationOption) (*rest.Response, error) {\n\topts := getOpts(string(req.GetRequest().Host()), options...)\n\topts.Protocol = common.ProtocolRest\n\tif len(opts.Filters) == 0 {\n\t\topts.Filters = ri.opts.Filters\n\t}\n\tif string(req.GetRequest().URI().Scheme()) != \"cse\" {\n\t\treturn nil, fmt.Errorf(\"Scheme invalid: %s, only support cse://\", req.GetRequest().URI().Scheme())\n\t}\n\tif req.GetHeader(\"Content-Type\") == \"\" {\n\t\treq.SetHeader(\"Content-Type\", \"application/json\")\n\t}\n\tnewReq := req.Copy()\n\tdefer newReq.Close()\n\tresp := rest.NewResponse()\n\tnewReq.SetHeader(common.HeaderSourceName, config.SelfServiceName)\n\tinv := invocation.CreateInvocation()\n\twrapInvocationWithOpts(inv, opts)\n\tinv.AppID = config.GlobalDefinition.AppID\n\tinv.MicroServiceName = string(req.GetRequest().Host())\n\tinv.Args = newReq\n\tinv.Reply = resp\n\tinv.Ctx = ctx\n\tinv.URLPathFormat = req.Req.URI().String()\n\tinv.MethodType = req.GetMethod()\n\tc, err := handler.GetChain(common.Consumer, ri.opts.ChainName)\n\tif err != nil {\n\t\tlager.Logger.Errorf(err, \"Handler chain init err.\")\n\t\treturn nil, err\n\t}\n\tc.Next(inv, func(ir *invocation.InvocationResponse) error {\n\t\terr = ir.Err\n\t\treturn err\n\t})\n\treturn resp, err\n}", "func Do(ctx context.Context, client *http.Client, req *http.Request) (*http.Response, error) {\n\tif client == nil {\n\t\tclient = http.DefaultClient\n\t}\n\tresp, err := client.Do(req.WithContext(ctx))\n\t// If we got an error, and the context has been canceled,\n\t// the context's error is probably more useful.\n\tif err != nil {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\terr = ctx.Err()\n\t\tdefault:\n\t\t}\n\t}\n\treturn resp, err\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
NewBytesBuffer create a bytes buffer
func NewBytesBuffer(p []byte) *BytesBuffer { return &BytesBuffer{reader: bytes.NewReader(p)} }
[ "func NewBufferBytes(data []byte) *Buffer {\n\treturn &Buffer{refCount: 0, buf: data, length: len(data)}\n}", "func NewBuffer(buffer [BUFFER_SIZE]byte) *BytePacketBuffer {\n\treturn &BytePacketBuffer{\n\t\tbuf: &buffer,\n\t\tpos: 0,\n\t}\n}", "func newBuffer() *bytes.Buffer {\n\treturn bytes.NewBuffer(make([]byte, bytes.MinRead))\n}", "func newBuffer() Buffer {\n\treturn &buffer{\n\t\tbytes: make([]byte, 0, 64),\n\t}\n}", "func newBuffer(b []byte) *buffer {\n\treturn &buffer{proto.NewBuffer(b), 0}\n}", "func NewBuffer(inp []byte) *ByteBuffer {\n\tif inp == nil {\n\t\tinp = make([]byte, 0, 512)\n\t}\n\treturn &ByteBuffer{Buffer: bytes.NewBuffer(inp)}\n}", "func newBuffer(buf []byte) *Buffer {\n\treturn &Buffer{data: buf}\n}", "func newBuffer(size int) *bytes.Buffer {\n\tb := new(bytes.Buffer)\n\tb.Grow(size)\n\treturn b\n}", "func (b *defaultByteBuffer) NewBuffer() ByteBuffer {\n\treturn NewWriterBuffer(256)\n}", "func newBuffer(data []byte) *buffer {\n\treturn &buffer{\n\t\toffset: 0,\n\t\tdata: data,\n\t}\n}", "func NewBuffer() *Buffer {\n\treturn NewBufferWithSize(initialSize)\n}", "func New(b []byte) *Buffer {\n\treturn &Buffer{b: b}\n}", "func NewBuffer() Buffer {\n\treturn &buffer{}\n}", "func NewBuffer() *Buffer {\n\treturn &Buffer{}\n}", "func NewBuffer(sz int) *Buffer {\n\treturn &Buffer{\n\t\tbuf: Calloc(sz),\n\t\toffset: 0,\n\t}\n}", "func GetBytesBuffer(size int) *bytes.Buffer {\n\tswitch {\n\n\tcase size > 0 && size <= 256:\n\t\treturn GetBytesBuffer256()\n\n\tcase size > 256 && size <= 512:\n\t\treturn GetBytesBuffer512()\n\n\tcase size > 512 && size <= 1024:\n\t\treturn GetBytesBuffer1K()\n\n\tcase size > 1024 && size <= 2048:\n\t\treturn GetBytesBuffer2K()\n\n\tcase size > 2048 && size <= 4096:\n\t\treturn GetBytesBuffer4K()\n\n\tcase size > 4096 && size <= 8192:\n\t\treturn GetBytesBuffer8K()\n\n\tcase size > 8192 && size <= 16384:\n\t\treturn GetBytesBuffer16K()\n\n\tcase size > 16384 && size <= 32768:\n\t\treturn GetBytesBuffer32K()\n\n\tcase size > 32768 && size <= 65536:\n\t\treturn GetBytesBuffer64K()\n\n\tcase size > 65536 && size <= 131072:\n\t\treturn GetBytesBuffer128K()\n\n\tcase size > 131072 && size <= 262144:\n\t\treturn GetBytesBuffer256K()\n\n\tcase size > 262144 && size <= 524288:\n\t\treturn GetBytesBuffer512K()\n\n\tcase size > 524288 && size <= 1048576:\n\t\treturn GetBytesBuffer1M()\n\n\tcase size > 1048576 && size <= 2097152:\n\t\treturn GetBytesBuffer2M()\n\n\tcase size > 2097152 && size <= 4194304:\n\t\treturn GetBytesBuffer4M()\n\n\tcase size > 4194304 && size <= 8388608:\n\t\treturn GetBytesBuffer8M()\n\n\tcase size > 8388608 && size <= 16777216:\n\t\treturn GetBytesBuffer16M()\n\n\tdefault:\n\t\treturn bytes.NewBuffer(make([]byte, size))\n\t}\n}", "func NewBuffer() *Buffer { return globalPool.NewBuffer() }", "func NewByteSliceBuffer(size uint64) *ByteSliceBuffer {\n\treturn &ByteSliceBuffer{\n\t\tBuffer: New(size, 0),\n\t\tdata: make([][]byte, size),\n\t}\n}", "func NewAttachedBytes(buffer []byte, offset int, size int) *Buffer {\n result := NewEmptyBuffer()\n result.AttachBytes(buffer, offset, size)\n return result\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Read from the byte buffer
func (bb *BytesBuffer) Read(p []byte) (n int, err error) { return bb.reader.Read(p) }
[ "func Read(b []byte) { Reader.Read(b) }", "func (cb *Buffer) Read(buf []byte) (int, error) {\n\tif buf == nil || len(buf) == 0 {\n\t\treturn 0, fmt.Errorf(\"Target buffer is null or empty\")\n\t}\n\n\ttoRead := min(len(buf), cb.ReadAvailability())\n\n\tlBytes := min(cb.rpos, toRead)\n\tcopy(buf[toRead-lBytes:toRead], cb.buffer[cb.rpos-lBytes:cb.rpos])\n\n\tif toRead > lBytes {\n\t\trBytes := toRead - lBytes\n\t\tcopy(buf[:rBytes], cb.buffer[len(cb.buffer)-rBytes:len(cb.buffer)])\n\t\tcb.rpos = len(cb.buffer) - rBytes\n\t} else {\n\t\tcb.rpos -= lBytes\n\t}\n\n\tcb.full = false\n\treturn toRead, nil\n}", "func ReadBytes(buffer []byte, offset int, size int) []byte {\n return buffer[offset:offset + size]\n}", "func (r *binaryReader) readBuf(len int) ([]byte, error) {\n\tb := r.buf[:len]\n\tn, err := r.Read(b)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif n != len {\n\t\treturn nil, errors.New(\"TODO failed to read enough bytes\")\n\t}\n\treturn b, nil\n}", "func ReadByte(buffer []byte, offset int) byte {\n return buffer[offset]\n}", "func (bs *ByteStream) Read(b []byte) (n int, err error) {\n\tif len(b) == 0 {\n\t\treturn 0, nil\n\t}\n\tn, err = bs.r.Read(b)\n\tbs.pos += n\n\treturn\n}", "func (b *buffer) readbyte() (c byte) {\n\tc = b.buf[b.off]\n\tif b.size == 1 {\n\t\tb.off = 0\n\t\tb.size = 0\n\t} else {\n\t\tb.off++\n\t\tb.size--\n\t\tif b.off == len(b.buf) {\n\t\t\tb.off = 0\n\t\t}\n\t}\n\treturn c\n}", "func (o *BytesItem) Read(obuf []byte, off uint64, rcount uint32) (uint32, error) {\n\tif off > 0 {\n\t\treturn 0, warp9.WarpErrorEOF\n\t}\n\tbuf := o.bytes.Bytes()\n\tn := copy(obuf, buf)\n\tif n != len(buf) {\n\t\treturn 0, warp9.ErrorCode(warp9.Ebufsmall)\n\t}\n\treturn uint32(n), nil\n}", "func (jbobject *JavaNioCharBuffer) Read(a JavaNioCharBufferInterface) (int, error) {\n\tconv_a := javabind.NewGoToJavaCallable()\n\tif err := conv_a.Convert(a); err != nil {\n\t\tpanic(err)\n\t}\n\tjret, err := jbobject.CallMethod(javabind.GetEnv(), \"read\", javabind.Int, conv_a.Value().Cast(\"java/nio/CharBuffer\"))\n\tif err != nil {\n\t\tvar zero int\n\t\treturn zero, err\n\t}\n\tconv_a.CleanUp()\n\treturn jret.(int), nil\n}", "func (r *Reader) ReadBytes(length int) []byte {\n\tif len(r.buffer) <= r.index+length-1 {\n\t\tlog.Panic(\"Error reading []byte: buffer is too small!\")\n\t}\n\n\tvar data = r.buffer[r.index : r.index+length]\n\tr.index += length\n\n\treturn data\n}", "func (t *TrackRemote) Read(b []byte) (n int, err error) { return }", "func (e *ObservableEditableBuffer) Read(q0 int, r []rune) (int, error) {\n\treturn e.f.Read(q0, r)\n}", "func (b *BinaryReader) readToBuffer(numBytes int, off int64) (int, error) {\n\tif numBytes > bufferSizeInBytes {\n\t\treturn 0, fmt.Errorf(\"Expected (0-%d) however found: %d\", bufferSizeInBytes, numBytes)\n\t}\n\n\treturn b.from.ReadAt(b.buffer[:numBytes], off)\n}", "func readBytesBuf(buf []byte, r io.Reader) ([]byte, error) {\n\t_, err := r.Read(buf)\n\treturn buf, err\n}", "func (d *videoDecryptor) Read(buf []byte) (int, error) {\n\tn, err := d.Reader.Read(buf)\n\tif err != nil {\n\t\treturn n, err\n\t}\n\txorBuff(n, d.Offset, buf, d.Key1, d.Key2)\n\td.Offset += n\n\treturn n, err\n}", "func (m *MockNetpollReader) ReadByte() (b byte, err error) {\n\treturn\n}", "func (v *VirtioSocketConnection) Read(b []byte) (n int, err error) { return v.rawConn.Read(b) }", "func (con *Connection) Read(b []byte) (int, error) {\n\tif con.getDecrypter() != nil {\n\t\treturn con.DecryptedRead(b)\n\t}\n\n\treturn con.connection.Read(b)\n}", "func (d *decoder) readBytes(n int) ([]byte, error) {\n\tif d.len < d.pos+n {\n\t\treturn nil, ErrUnexpectedFileEnd\n\t}\n\tn, d.pos = d.pos, d.pos+n\n\treturn d.buf[n:d.pos], nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Close the bytes buffer
func (bb *BytesBuffer) Close() error { return nil }
[ "func (b *Buffer) Close() error { return nil }", "func (b *Bytes) Close() error {\n\tif b.p != nil {\n\t\tPut(b.p)\n\t\tb.p = nil\n\t}\n\treturn nil\n}", "func (buf *Buffer) Close() error {\n\tbuf.Closed = true\n\treturn nil\n}", "func (buf *logBuffer) close() {\n\tbuf.Flush()\n\tbuf.file.Close()\n\treturn\n}", "func (b *Buffer) Close() error {\n\tb.Unmap()\n\treturn os.Remove(b.filename)\n}", "func (b *Buffer) close() error {\n\tif b.file == nil {\n\t\treturn nil\n\t}\n\n\tpath := b.file.Name()\n\n\tb.log(2, \"renaming %q\", path)\n\terr := os.Rename(path, path+\".closed\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif b.BufferSize != 0 {\n\t\tb.log(2, \"flushing %q\", path)\n\t\terr = b.buf.Flush()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tb.log(2, \"closing %q\", path)\n\treturn b.file.Close()\n}", "func (r *bytesReader) Close() error {\n\t_, err := r.Seek(0, io.SeekStart)\n\treturn err\n}", "func (b *WriteBuffer) Close() {\n\tb.mu.Lock()\n\tdefer b.mu.Unlock()\n\n\tif b.alloc != nil {\n\t\tfor _, s := range b.inner.Slices {\n\t\t\tb.alloc.releaseChunk(s)\n\t\t}\n\n\t\tb.alloc = nil\n\t}\n\n\tb.inner.invalidate()\n}", "func (bbw *Writer) Close() ([]byte, error) {\n\tif bbw.clsdPos >= 0 {\n\t\treturn bbw.buf[:bbw.clsdPos], nil\n\t}\n\tif len(bbw.buf)-bbw.offs < 4 {\n\t\tbbw.clsdPos = bbw.offs\n\t\tbbw.buf = bbw.buf[:bbw.clsdPos]\n\t\treturn bbw.buf, nil\n\t}\n\tbinary.BigEndian.PutUint32(bbw.buf[bbw.offs:], uint32(0xFFFFFFFF))\n\tbbw.clsdPos = bbw.offs\n\tbbw.offs = len(bbw.buf)\n\treturn bbw.buf[:bbw.clsdPos], nil\n}", "func (rr *Reader) Close() {\n\tif rr.Err == nil && len(rr.Bytes()) != 0 {\n\t\trr.Err = errors.New(\"excess bytes in buffer\")\n\t}\n}", "func (b *BufferedReadCloser) Close() error {\n\tb.b = nil\n\treturn nil\n}", "func (r *bodyReader) Close() error {\n\tswitch r.contentEncoding {\n\tcase \"\":\n\t\treturn nil\n\tcase \"gzip\":\n\t\treturn r.r.Close()\n\tdefault:\n\t\tpanic(\"Unreachable\")\n\t}\n}", "func (cw *CappedWriter) Close() error { return nil }", "func Bpf_object__close(object *Struct_bpf_object) {\n\t C.bpf_object__close()\n}", "func (ch *Channel) Close() {}", "func (e *encoder) Close() error {\n\t// If there's anything left in the buffer, flush it out\n\tif e.err == nil && e.nbuf > 0 {\n\t\te.enc.Encode(e.out[0:], e.buf[0:e.nbuf])\n\t\tencodedLen := e.enc.EncodedLen(e.nbuf)\n\t\te.nbuf = 0\n\t\t_, e.err = e.w.Write(e.out[0:encodedLen])\n\t}\n\treturn e.err\n}", "func (c *UDPChannel) Close() {\n\n}", "func (s *Basememcached_protocolListener) ExitBytes(ctx *BytesContext) {}", "func (r *body) Close() error { return nil }" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
ToEd25519 converts the public key p into a ed25519 key. (x, y) = (sqrt(486664)u/v, (u1)/(u+1))
func (p PublicKey) ToEd25519() (ed25519.PublicKey, error) { a, err := convertMont(p) if err != nil { return nil, err } return a.Bytes(), nil }
[ "func Ed25519PublicKeyToCurve25519(pk ed25519.PublicKey) []byte {\n\t// ed25519.PublicKey is a little endian representation of the y-coordinate,\n\t// with the most significant bit set based on the sign of the x-coordinate.\n\tbigEndianY := make([]byte, ed25519.PublicKeySize)\n\tfor i, b := range pk {\n\t\tbigEndianY[ed25519.PublicKeySize-i-1] = b\n\t}\n\tbigEndianY[0] &= 0b0111_1111\n\n\t// The Montgomery u-coordinate is derived through the bilinear map\n\t//\n\t// u = (1 + y) / (1 - y)\n\t//\n\t// See https://blog.filippo.io/using-ed25519-keys-for-encryption.\n\ty := new(big.Int).SetBytes(bigEndianY)\n\tdenom := big.NewInt(1)\n\tdenom.ModInverse(denom.Sub(denom, y), curve25519P) // 1 / (1 - y)\n\tu := y.Mul(y.Add(y, big.NewInt(1)), denom)\n\tu.Mod(u, curve25519P)\n\n\tout := make([]byte, curve25519.PointSize)\n\tuBytes := u.Bytes()\n\tfor i, b := range uBytes {\n\t\tout[len(uBytes)-i-1] = b\n\t}\n\n\treturn out\n}", "func Ed25519PublicKey(pk crypto.PublicKey) PublicKey {\n\treturn PublicKey{\n\t\tAlgorithm: SignatureAlgoEd25519,\n\t\tKey: pk[:],\n\t}\n}", "func PublicEd25519toCurve25519(pub []byte) ([]byte, error) {\n\tif len(pub) == 0 {\n\t\treturn nil, errors.New(\"public key is nil\")\n\t}\n\n\tif len(pub) != ed25519.PublicKeySize {\n\t\treturn nil, fmt.Errorf(\"%d-byte key size is invalid\", len(pub))\n\t}\n\n\tpkOut := new([Curve25519KeySize]byte)\n\tpKIn := new([Curve25519KeySize]byte)\n\tcopy(pKIn[:], pub)\n\n\tsuccess := extra25519.PublicKeyToCurve25519(pkOut, pKIn)\n\tif !success {\n\t\treturn nil, errors.New(\"error converting public key\")\n\t}\n\n\treturn pkOut[:], nil\n}", "func Ed25519PrivateKeyToCurve25519(pk ed25519.PrivateKey) []byte {\n\th := sha512.New()\n\th.Write(pk.Seed())\n\tout := h.Sum(nil)\n\n\t// No idea why, copy-pasted from libsodium\n\tout[0] &= 248\n\tout[31] &= 127\n\tout[31] |= 64\n\n\treturn out[:curve25519.ScalarSize]\n}", "func parseED25519Key(key ssh.PublicKey) (ed25519.PublicKey, error) {\n\tvar sshWire struct {\n\t\tName string\n\t\tKeyBytes []byte\n\t}\n\tif err := ssh.Unmarshal(key.Marshal(), &sshWire); err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to unmarshal key %v: %v\", key.Type(), err)\n\t}\n\treturn ed25519.PublicKey(sshWire.KeyBytes), nil\n}", "func NewEdX25519PublicKey(b *[ed25519.PublicKeySize]byte) *EdX25519PublicKey {\n\treturn &EdX25519PublicKey{\n\t\tid: MustID(edx25519KeyHRP, b[:]),\n\t\tpublicKey: b,\n\t}\n}", "func CURVE25519_convert_X_to_Ed(x []byte) ([]byte, error) {\n\tret := make([]byte, 32)\n\tx25519 := (*C.uchar)(unsafe.Pointer(&x[0]))\n\ted25519 := (*C.uchar)(unsafe.Pointer(&ret[0]))\n\tif C.CURVE25519_convert_X_to_Ed(ed25519, x25519) == 1 {\n\t\treturn ret, nil\n\t}\n\treturn nil, errors.New(\"Invalid x25519 point to convert!\")\n}", "func PrivKeyToEd25519PrivateKey(privKey crypto.PrivKey) *ed25519.PrivateKey {\n\tkey, err := crypto.PrivKeyToStdKey(privKey)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t\treturn nil\n\t}\n\tif pk, ok := key.(*ed25519.PrivateKey); ok {\n\t\treturn pk\n\t}\n\treturn nil\n}", "func (pk *ECPublicKey) toECDSA() *ecdsa.PublicKey {\n\tecdsaPub := new(ecdsa.PublicKey)\n\tecdsaPub.Curve = pk.Curve\n\tecdsaPub.X = pk.X\n\tecdsaPub.Y = pk.Y\n\n\treturn ecdsaPub\n}", "func (k EdDSAPublic) ToECDH() ECDHPublic {\n\tout := make([]byte, ECDHKeyLength)\n\trv := C.crypto_sign_ed25519_pk_to_curve25519(g2cbt(out), g2cbt(k))\n\tif rv != 0 {\n\t\tpanic(\"crypto_sign_ed25519_sk_to_curve25519 returned non-zero\")\n\t}\n\treturn out\n}", "func PrivateKeyToCurve25519(privateKey []byte) (curvePrivate []byte) {\n\th := sha512.New()\n\th.Write(privateKey)\n\tdigest := h.Sum(nil)\n\n\t// key clamping\n\tdigest[0] &= 248\n\tdigest[31] &= 127\n\tdigest[31] |= 64\n\n\treturn digest[:32]\n}", "func (privKey *YubiHsmPrivateKey) exportEd25519Pubkey(keyData []byte) error {\n\tif len(keyData) != YubiEd25519PubKeySize {\n\t\treturn errors.New(\"Invalid ed25519 public key data size\")\n\t}\n\n\tprivKey.pubKeyBytes = make([]byte, YubiEd25519PubKeySize)\n\tcopy(privKey.pubKeyBytes[:], keyData[:])\n\n\treturn nil\n}", "func NewEdX25519PublicKeyFromID(id ID) (*EdX25519PublicKey, error) {\n\tif id == \"\" {\n\t\treturn nil, errors.Errorf(\"empty id\")\n\t}\n\thrp, b, err := id.Decode()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif hrp != edx25519KeyHRP {\n\t\treturn nil, errors.Errorf(\"invalid key type for edx25519\")\n\t}\n\tif len(b) != ed25519.PublicKeySize {\n\t\treturn nil, errors.Errorf(\"invalid ed25519 public key bytes\")\n\t}\n\treturn &EdX25519PublicKey{\n\t\tid: id,\n\t\tpublicKey: Bytes32(b),\n\t}, nil\n}", "func GenerateEd25519Key(src io.Reader) (PrivKey, PubKey, error) {\n\tpub, priv, err := ed25519.GenerateKey(src)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn &Ed25519PrivateKey{\n\t\t\tk: priv,\n\t\t},\n\t\t&Ed25519PublicKey{\n\t\t\tk: pub,\n\t\t},\n\t\tnil\n}", "func TestExtractEd25519PublicKey(t *testing.T) {\n\tpub, priv, err := ed25519.GenerateKey(rand.Reader)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif g, e := extractEd25519Pubkey(priv), pub; !bytes.Equal(g[:], e[:]) {\n\t\tt.Errorf(\"assumed public key would be in private key: %x != %x\", g, e)\n\t}\n}", "func SecretEd25519toCurve25519(priv []byte) ([]byte, error) {\n\tif len(priv) == 0 {\n\t\treturn nil, errors.New(\"private key is nil\")\n\t}\n\n\tsKIn := new([ed25519.PrivateKeySize]byte)\n\tcopy(sKIn[:], priv)\n\n\tsKOut := new([Curve25519KeySize]byte)\n\textra25519.PrivateKeyToCurve25519(sKOut, sKIn)\n\n\treturn sKOut[:], nil\n}", "func (v *Point) ToPublicKey() *ecdsa.PublicKey {\n\tv.toAffine()\n\n\tpk := secp256k1.NewPublicKey(&v.p.X, &v.p.Y)\n\n\treturn pk.ToECDSA()\n}", "func CreatePublicKeyX25519FromBase64(publicKeyBase64 string) (*X25519.PublicKey, error) {\n publicKeyBytes, err := base64.StdEncoding.DecodeString(publicKeyBase64)\n if err != nil {\n return nil, err\n }\n return X25519.NewPublicKey(publicKeyBytes), nil\n}", "func NewEdX25519KeyFromPrivateKey(privateKey *[ed25519.PrivateKeySize]byte) *EdX25519Key {\n\t// Derive public key from private key\n\tedpk := ed25519.PrivateKey(privateKey[:])\n\tpublicKey := edpk.Public().(ed25519.PublicKey)\n\tif len(publicKey) != ed25519.PublicKeySize {\n\t\tpanic(errors.Errorf(\"invalid public key bytes (len=%d)\", len(publicKey)))\n\t}\n\n\tvar privateKeyBytes [ed25519.PrivateKeySize]byte\n\tcopy(privateKeyBytes[:], privateKey[:ed25519.PrivateKeySize])\n\n\tvar publicKeyBytes [ed25519.PublicKeySize]byte\n\tcopy(publicKeyBytes[:], publicKey[:ed25519.PublicKeySize])\n\n\treturn &EdX25519Key{\n\t\tprivateKey: &privateKeyBytes,\n\t\tpublicKey: NewEdX25519PublicKey(&publicKeyBytes),\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Sign signs the given message with the private key p and returns a signature. It implements the XEdDSA sign method defined in XEdDSA performs two passes over messages to be signed and therefore cannot handle prehashed messages. Thus opts.HashFunc() must return zero to indicate the message hasn't been hashed. This can be achieved by passing crypto.Hash(0) as the value for opts.
func (p PrivateKey) Sign(rand io.Reader, message []byte, opts crypto.SignerOpts) (signature []byte, err error) { if opts.HashFunc() != crypto.Hash(0) { return nil, errors.New("x25519: cannot sign hashed message") } return Sign(rand, p, message) }
[ "func (priv PrivateKey) Sign(rand io.Reader, message []byte, opts crypto.SignerOpts) (signature []byte, err error) {\n\tif opts.HashFunc() != crypto.Hash(0) {\n\t\treturn nil, errors.New(\"sign: cannot sign hashed message\")\n\t}\n\n\tout := Sign(message, priv)\n\treturn out[:], nil\n}", "func Sign(message, secretKey []byte) ([]byte, error) {\n\treturn defaultPH.cryptoSign(message, secretKey)\n}", "func (sk PrivateKey) Sign(message []byte, hasher Hasher) ([]byte, error) {\n\treturn sk.privateKey.Sign(message, hasher)\n}", "func (pk *PrivateKey) Sign(message []byte) *Signature {\n var signature Signature\n copy(signature[:], ed25519.Sign(pk[:], message)[:])\n return &signature\n}", "func (pk PrivateKey) Sign(message []byte) Signature {\n\tvar signature Signature\n\tcopy(signature[:], ed25519.Sign(pk[:], message)[:])\n\treturn signature\n}", "func (priv *PrivateKey) Sign(rand io.Reader, msg []byte, opts crypto.SignerOpts) ([]byte, error)", "func (k *KeyPairEd25519) Sign(message []byte) ([]byte, error) {\n\tres, err := k.privateKey.Sign(nil, message, crypto.Hash(0))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"calling sign: %v\", err)\n\t}\n\treturn res, nil\n}", "func (p *KeyPair) Sign(message []byte) ([]byte, error) {\n\tprivateKey := p.ToEcdsa()\n\thash := sha256.Sum256(message)\n\tr, s, err := ecdsa.Sign(rand.Reader, privateKey, hash[:])\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tparams := privateKey.Curve.Params()\n\tcurveOrderByteSize := params.P.BitLen() / 8\n\trBytes, sBytes := r.Bytes(), s.Bytes()\n\tsignature := make([]byte, curveOrderByteSize*2)\n\tcopy(signature[curveOrderByteSize-len(rBytes):], rBytes)\n\tcopy(signature[curveOrderByteSize*2-len(sBytes):], sBytes)\n\n\treturn signature, nil\n}", "func (k *onChainPrivateKey) Sign(msg []byte) (signature []byte, err error) {\n\tsig, err := crypto.Sign(onChainHash(msg), (*ecdsa.PrivateKey)(k))\n\treturn sig, err\n}", "func (_Ethdkg *EthdkgCallerSession) Sign(message []byte, privK *big.Int) ([2]*big.Int, error) {\n\treturn _Ethdkg.Contract.Sign(&_Ethdkg.CallOpts, message, privK)\n}", "func (l LocalIdentity) Sign(message []byte) ed25519.Signature {\n\treturn l.privateKey.Sign(message)\n}", "func (pk *PrivKeyLedgerEd25519) Sign(msg []byte) crypto.Signature {\n\t// oh, I wish there was better error handling\n\tdev, err := getLedger()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tpub, sig, err := signLedger(dev, msg)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// if we have no pubkey yet, store it for future queries\n\tif pk.CachedPubKey.Empty() {\n\t\tpk.CachedPubKey = pub\n\t} else if !pk.CachedPubKey.Equals(pub) {\n\t\tpanic(\"signed with a different key than stored\")\n\t}\n\treturn sig\n}", "func Sign(message string) (string, error) {\n\n\t// TODO check length on string\n\t// Sign\n\tvar h hash.Hash\n\th = sha256.New()\n\n\tio.WriteString(h, message)\n\tsignhash := h.Sum(nil)\n\n\trsaKey, err := loadPrivateKeyFromFile()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\trsaSignature, err := rsa.SignPKCS1v15(rand.Reader, rsaKey, crypto.SHA256, signhash)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn \"\", err\n\t}\n\n\tsEnc := base64.StdEncoding.EncodeToString(rsaSignature)\n\treturn sEnc, nil\n}", "func Sign(opriv, rpriv *btcec.PrivateKey, m []byte) []byte {\n\tR := rpriv.PubKey()\n\tk := rpriv.D\n\tv := opriv.D\n\n\t// h(R,m) * v\n\thv := new(big.Int).Mul(hash(R, m), v)\n\n\t// k - h(R,m) * v\n\ts := new(big.Int).Sub(k, hv)\n\n\t// s mod N\n\ts = new(big.Int).Mod(s, btcec.S256().N)\n\n\treturn s.Bytes()\n}", "func SignPSS(rand io.Reader, priv *rsa.PrivateKey, hash crypto.Hash, hashed []byte, opts *rsa.PSSOptions,) ([]byte, error)", "func (_BondedECDSAKeep *BondedECDSAKeepTransactor) Sign(opts *bind.TransactOpts, _digest [32]byte) (*types.Transaction, error) {\n\treturn _BondedECDSAKeep.contract.Transact(opts, \"sign\", _digest)\n}", "func Sign(suite suites.Suite, x kyber.Scalar, msg []byte) ([]byte, error) {\n\tHM := hashToPoint(suite, msg)\n\txHM := HM.Mul(x, HM)\n\ts, err := xHM.MarshalBinary()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn s, nil\n}", "func Sign(k string, p *Packet) string {\n\treturn GenerateSignature(p.Header.MessageID, k, p.Header.Timestamp.String())\n}", "func Sign(secretKey SecretKey, message string) (signedMessage SignedMessage, err error) {\n\tecdsaKey := ecdsa.PrivateKey(secretKey)\n\tr, s, err := ecdsa.Sign(rand.Reader, &ecdsaKey, []byte(message))\n\tsignedMessage.Signature.R = r\n\tsignedMessage.Signature.S = s\n\tsignedMessage.Message = message\n\treturn\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Verify reports whether sig is a valid signature of message by publicKey. It will panic if len(publicKey) is not PublicKeySize. It implements the XEdDSA verify method defined in xeddsa_verify(u, M, (R || s)): if u >= p or R.y >= 2|p| or s >= 2|q|: return false A = convert_mont(u) if not on_curve(A): return false h = hash(R || A || M) (mod q) Rcheck = sB hA if bytes_equal(R, Rcheck): return true return false
func Verify(publicKey PublicKey, message, sig []byte) bool { // The following code should be equivalent to: // // pub, err := publicKey.ToEd25519() // if err != nil { // return false // } // return ed25519.Verify(pub, message, sig) if l := len(publicKey); l != PublicKeySize { panic("x25519: bad public key length: " + strconv.Itoa(l)) } if len(sig) != SignatureSize || sig[63]&0xE0 != 0 { return false } a, err := convertMont(publicKey) if err != nil { return false } hh := sha512.New() hh.Write(sig[:32]) hh.Write(a.Bytes()) hh.Write(message) hDigest := make([]byte, 0, sha512.Size) hDigest = hh.Sum(hDigest) h, err := edwards25519.NewScalar().SetUniformBytes(hDigest) if err != nil { return false } s, err := edwards25519.NewScalar().SetCanonicalBytes(sig[32:]) if err != nil { return false } minusA := (&edwards25519.Point{}).Negate(a) r := (&edwards25519.Point{}).VarTimeDoubleScalarBaseMult(h, minusA, s) return subtle.ConstantTimeCompare(sig[:32], r.Bytes()) == 1 }
[ "func Verify(sig []byte, publicKey PublicKey) bool {\n\tif l := len(publicKey); l != PublicKeySize {\n\t\tpanic(\"sign: bad public key length: \" + strconv.Itoa(l))\n\t}\n\n\tif len(sig) < SignatureSize || sig[63]&224 != 0 {\n\t\treturn false\n\t}\n\tmsg := sig[SignatureSize:]\n\tsig = sig[:SignatureSize]\n\n\treturn ed25519.Verify(ed25519.PublicKey(publicKey), msg, sig)\n}", "func Verify(pub *PublicKey, hash []byte, r, s *big.Int) bool", "func (sig *Signature) VerifySignature(publicKey interface{}, encoding string) bool {\n\tif sig.Data == nil {\n\t\tlog.Warn(\"sig does not contain signature data\", \"sig\", sig)\n\t\treturn false\n\t}\n\tif publicKey == nil {\n\t\tlog.Warn(\"PublicKey is nil\")\n\t\treturn false\n\t}\n\tencoding += sig.GetSignatureMetaData().String()\n\tdata := []byte(encoding)\n\tswitch sig.Algorithm {\n\tcase Ed25519:\n\t\tif pkey, ok := publicKey.(ed25519.PublicKey); ok {\n\t\t\treturn ed25519.Verify(pkey, data, sig.Data.([]byte))\n\t\t}\n\t\tlog.Warn(\"Could not assert type ed25519.PublicKey\", \"publicKeyType\", fmt.Sprintf(\"%T\", publicKey))\n\tcase Ed448:\n\t\tlog.Warn(\"Ed448 not yet Supported!\")\n\tcase Ecdsa256:\n\t\tif pkey, ok := publicKey.(*ecdsa.PublicKey); ok {\n\t\t\tif sig, ok := sig.Data.([]*big.Int); ok && len(sig) == 2 {\n\t\t\t\thash := sha256.Sum256(data)\n\t\t\t\treturn ecdsa.Verify(pkey, hash[:], sig[0], sig[1])\n\t\t\t}\n\t\t\tlog.Warn(\"Could not assert type []*big.Int\", \"signatureDataType\", fmt.Sprintf(\"%T\", sig.Data))\n\t\t\treturn false\n\t\t}\n\t\tlog.Warn(\"Could not assert type ecdsa.PublicKey\", \"publicKeyType\", fmt.Sprintf(\"%T\", publicKey))\n\tcase Ecdsa384:\n\t\tif pkey, ok := publicKey.(*ecdsa.PublicKey); ok {\n\t\t\tif sig, ok := sig.Data.([]*big.Int); ok && len(sig) == 2 {\n\t\t\t\thash := sha512.Sum384(data)\n\t\t\t\treturn ecdsa.Verify(pkey, hash[:], sig[0], sig[1])\n\t\t\t}\n\t\t\tlog.Warn(\"Could not assert type []*big.Int\", \"signature\", sig.Data)\n\t\t\treturn false\n\t\t}\n\t\tlog.Warn(\"Could not assert type ecdsa.PublicKey\", \"publicKeyType\", fmt.Sprintf(\"%T\", publicKey))\n\tdefault:\n\t\tlog.Warn(\"Signature algorithm type not supported\", \"type\", sig.Algorithm)\n\t}\n\treturn false\n}", "func Verify(publicKey *ecdsa.PublicKey, sig []byte, message []byte) (valid bool, err error) {\n\tsignature := new(common.SchnorrSignature)\n\terr = json.Unmarshal(sig, signature)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"Failed unmashalling schnorr signature [%s]\", err)\n\t}\n\n\t// 1. compute h(m|| s * G - e * P)\n\t// 1.1 compute s * G\n\tcurve := publicKey.Curve\n\tx1, y1 := curve.ScalarBaseMult(signature.S.Bytes())\n\n\t// 1.2 compute e * P\n\tx2, y2 := curve.ScalarMult(publicKey.X, publicKey.Y, signature.E.Bytes())\n\n\t// 1.3 计算-(e * P),如果 e * P = (x,y),则 -(e * P) = (x, -y mod P)\n\tnegativeOne := big.NewInt(-1)\n\ty2 = new(big.Int).Mod(new(big.Int).Mul(negativeOne, y2), curve.Params().P)\n\n\t// 1.4 compute s * G - e * P\n\tx, y := curve.Add(x1, y1, x2, y2)\n\n\te := hash.HashUsingSha256(append(message, elliptic.Marshal(curve, x, y)...))\n\n\tintE := new(big.Int).SetBytes(e)\n\n\t// 2. check the equation\n\t//\treturn bytes.Equal(e, signature.E.Bytes()), nil\n\tif intE.Cmp(signature.E) != 0 {\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func SignatureVerify(publicKey, sig, hash []byte) bool {\n\n\tbytesDecded, _ := base58.DecodeToBig(publicKey)\n\tpubl := splitBig(bytesDecded, 2)\n\tx, y := publ[0], publ[1]\n\n\tbytesDecded, _ = base58.DecodeToBig(sig)\n\tsigg := splitBig(bytesDecded, 2)\n\tr, s := sigg[0], sigg[1]\n\n\tpub := ecdsa.PublicKey{elliptic.P224(), x, y}\n\n\treturn ecdsa.Verify(&pub, hash, r, s)\n}", "func (brs BRSignature) Verify(msghash *[32]byte, pub [][]Point) bool {\n\tn,m,e0,brss := brs.load(pub)\n\tif brss == nil {\n\t\treturn false\n\t}\n\n\tE := sha512.New()\n\n\tfor t := byte(0); t < n; t++ {\n\t\te := Scalar(e0)\n\t\tfor i := byte(0); i < m; i++ {\n\t\t\tz := Scalar(brss[t][i])\n\t\t\tz[31] &= 0x0f\n\t\t\tw := brss[t][i][31] & 0xf0\n\t\t\tpneg := pub[t][i].Negate()\n\t\t\te = chameleon(pneg.MulAdd(&e, &z), msghash, t, (i+1)%m, w)\n\t\t}\n\t\tE.Write(e[:])\n\t}\n\n\tvar e1hash [64]byte\n\tE.Sum(e1hash[:0])\n\tif e0 != reduce(&e1hash) {\n\t\treturn false\n\t}\n\treturn true\n}", "func VerifySignature(key Key, sig Signature, unverified []byte) error {\n\terr := validateKey(key)\n\tif err != nil {\n\t\treturn err\n\t}\n\tsigBytes, err := hex.DecodeString(sig.Sig)\n\tif err != nil {\n\t\treturn err\n\t}\n\thashMapping := getHashMapping()\n\tswitch key.KeyType {\n\tcase rsaKeyType:\n\t\t// We do not need the pemData here, so we can throw it away via '_'\n\t\t_, parsedKey, err := decodeAndParse([]byte(key.KeyVal.Public))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tparsedKey, ok := parsedKey.(*rsa.PublicKey)\n\t\tif !ok {\n\t\t\treturn ErrKeyKeyTypeMismatch\n\t\t}\n\t\tswitch key.Scheme {\n\t\tcase rsassapsssha256Scheme:\n\t\t\thashed := hashToHex(hashMapping[\"sha256\"](), unverified)\n\t\t\terr = rsa.VerifyPSS(parsedKey.(*rsa.PublicKey), crypto.SHA256, hashed, sigBytes, &rsa.PSSOptions{SaltLength: sha256.Size, Hash: crypto.SHA256})\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"%w: %s\", ErrInvalidSignature, err)\n\t\t\t}\n\t\tdefault:\n\t\t\t// supported key schemes will get checked in validateKey\n\t\t\tpanic(\"unexpected Error in VerifySignature function\")\n\t\t}\n\tcase ecdsaKeyType:\n\t\t// We do not need the pemData here, so we can throw it away via '_'\n\t\t_, parsedKey, err := decodeAndParse([]byte(key.KeyVal.Public))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tparsedKey, ok := parsedKey.(*ecdsa.PublicKey)\n\t\tif !ok {\n\t\t\treturn ErrKeyKeyTypeMismatch\n\t\t}\n\t\tcurveSize := parsedKey.(*ecdsa.PublicKey).Curve.Params().BitSize\n\t\tvar hashed []byte\n\t\tif err := matchEcdsaScheme(curveSize, key.Scheme); err != nil {\n\t\t\treturn ErrCurveSizeSchemeMismatch\n\t\t}\n\t\t// implement https://tools.ietf.org/html/rfc5656#section-6.2.1\n\t\t// We determine the curve size and choose the correct hashing\n\t\t// method based on the curveSize\n\t\tswitch {\n\t\tcase curveSize <= 256:\n\t\t\thashed = hashToHex(hashMapping[\"sha256\"](), unverified)\n\t\tcase 256 < curveSize && curveSize <= 384:\n\t\t\thashed = hashToHex(hashMapping[\"sha384\"](), unverified)\n\t\tcase curveSize > 384:\n\t\t\thashed = hashToHex(hashMapping[\"sha512\"](), unverified)\n\t\tdefault:\n\t\t\tpanic(\"unexpected Error in VerifySignature function\")\n\t\t}\n\t\tif ok := ecdsa.VerifyASN1(parsedKey.(*ecdsa.PublicKey), hashed[:], sigBytes); !ok {\n\t\t\treturn ErrInvalidSignature\n\t\t}\n\tcase ed25519KeyType:\n\t\t// We do not need a scheme switch here, because ed25519\n\t\t// only consist of sha256 and curve25519.\n\t\tpubHex, err := hex.DecodeString(key.KeyVal.Public)\n\t\tif err != nil {\n\t\t\treturn ErrInvalidHexString\n\t\t}\n\t\tif ok := ed25519.Verify(pubHex, unverified, sigBytes); !ok {\n\t\t\treturn fmt.Errorf(\"%w: ed25519\", ErrInvalidSignature)\n\t\t}\n\tdefault:\n\t\t// We should never get here, because we call validateKey in the first\n\t\t// line of the function.\n\t\tpanic(\"unexpected Error in VerifySignature function\")\n\t}\n\treturn nil\n}", "func verifySig(sigStr []byte, publicKeyStr []byte, scriptPubKey []byte, tx *types.Transaction, txInIdx int) bool {\n\ttxHash, _ := tx.TxHash()\n\tsig, err := crypto.SigFromBytes(sigStr)\n\tif err != nil {\n\t\tlogger.Errorf(\"Deserialize signature failed. Err: %v(sig: %x, pubkey: %x, \"+\n\t\t\t\"scriptpubkey: %x, txhash: %s, txInIdx: %d)\", err, sigStr, publicKeyStr,\n\t\t\tscriptPubKey, txHash, txInIdx)\n\t\treturn false\n\t}\n\tpublicKey, err := crypto.PublicKeyFromBytes(publicKeyStr)\n\tif err != nil {\n\t\tlogger.Errorf(\"Deserialize public key failed. Err: %v(sig: %x, pubkey: %x, \"+\n\t\t\t\"scriptpubkey: %x, txhash: %s, txInIdx: %d)\", err, sigStr, publicKeyStr,\n\t\t\tscriptPubKey, txHash, txInIdx)\n\t\treturn false\n\t}\n\n\tsigHash, err := CalcTxHashForSig(scriptPubKey, tx, txInIdx)\n\tif err != nil {\n\t\tlogger.Errorf(\"Calculate signature hash failed. Err: %v(sig: %x, pubkey: %x, \"+\n\t\t\t\"scriptpubkey: %x, txhash: %s, txInIdx: %d)\", err, sigStr, publicKeyStr,\n\t\t\tscriptPubKey, txHash, txInIdx)\n\t\treturn false\n\t}\n\n\tif ok := sig.VerifySignature(publicKey, sigHash); !ok {\n\t\tlogger.Errorf(\"verify signatrure failed.(sig: %x, raw sig: %x, pubkey: %x, \"+\n\t\t\t\"scriptpubkey: %x, sigHash: %x, txhash: %s, txInIdx: %d)\", sig.Serialize(),\n\t\t\tsigStr, publicKeyStr, scriptPubKey, sigHash[:], txHash, txInIdx)\n\t\treturn false\n\t}\n\treturn true\n}", "func Verify(mesg, testsig, publicKey string) bool {\n\tif len(mesg)*4 > len(publicKey)*3 {\n\t\treturn false\n\t}\n\tvar m, decrypted big.Int\n\tsetBytesReverse(&m, []byte(mesg))\n\tn := base64ToInt(publicKey)\n\tintSig := base64ToInt(testsig)\n\tdecrypted.Exp(intSig, rsaPublicE, n)\n\n\treturn decrypted.Cmp(&m) == 0\n}", "func VerifyRSASignature(publicKey, signature, data []byte, digestType string, pkeyopt map[string]string) (bool, error) {\n\n\tmd, err := GetDigestByName(digestType)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tinf := C.BIO_new(C.BIO_s_mem())\n\tif inf == nil {\n\t\treturn false, errors.New(\"failed allocating input buffer\")\n\t}\n\tdefer C.BIO_free(inf)\n\t_, err = asAnyBio(inf).Write(publicKey)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tpubKey := C.d2i_PUBKEY_bio(inf, nil)\n\tif pubKey == nil {\n\t\treturn false, errors.New(\"failed to load public key\")\n\t}\n\tdefer C.EVP_PKEY_free(pubKey)\n\tctx := C.EVP_PKEY_CTX_new(pubKey, nil)\n\tif ctx == nil {\n\t\treturn false, errors.New(\"failed to setup context\")\n\t}\n\tdefer C.EVP_PKEY_CTX_free(ctx)\n\n\tmdctx := C.EVP_MD_CTX_new()\n\tdefer C.EVP_MD_CTX_free(mdctx)\n\n\tnRes := C.EVP_DigestVerifyInit(mdctx, &ctx, md.ptr, nil, pubKey)\n\tif nRes != 1 {\n\t\treturn false, errors.New(\"unable to init digest verify\")\n\t}\n\n\tif pkeyopt != nil && len(pkeyopt) > 0 {\n\t\t// This is a convenience function for calling X_EVP_PKEY_CTX_ctrl_str. The _Ctype_struct_evp_pkey_ctx_st type is not\n\t\t// exposed, but ctx can be captured in a local function like this.\n\t\tsetKeyOpt := func(pkeyopt map[string]string, k string) error {\n\t\t\tv, ok := pkeyopt[k]\n\t\t\tif !ok {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tck := C.CString(k)\n\t\t\tdefer C.free(unsafe.Pointer(ck))\n\t\t\tcv := C.CString(v)\n\t\t\tdefer C.free(unsafe.Pointer(cv))\n\t\t\tif C.X_EVP_PKEY_CTX_ctrl_str(ctx, ck, cv) <= 0 {\n\t\t\t\treturn fmt.Errorf(\"failed to set %s\", k)\n\t\t\t}\n\t\t\treturn nil\n\t\t}\n\n\t\t// Set RSA padding mode and salt length if they exist. Order matters; mode must be set before salt length.\n\t\tif rsaPaddingMode, ok := pkeyopt[\"rsa_padding_mode\"]; ok {\n\t\t\tif err := setKeyOpt(pkeyopt, \"rsa_padding_mode\"); err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t\tswitch rsaPaddingMode {\n\t\t\tcase \"pss\":\n\t\t\t\tif err := setKeyOpt(pkeyopt, \"rsa_pss_saltlen\"); err != nil {\n\t\t\t\t\treturn false, err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t// Fallback to make sure all pkeyopt get processed. Skips any keys found in pkeyoptSkip.\n\t\tfor k := range pkeyopt {\n\t\t\tif contains(pkeyoptSkip, k) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif err := setKeyOpt(pkeyopt, k); err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t}\n\t}\n\n\tnRes = C.EVP_DigestUpdate(mdctx, unsafe.Pointer((*C.uchar)(&data[0])), C.size_t(len(data)))\n\tif nRes != 1 {\n\t\treturn false, errors.New(\"unable to update digest\")\n\t}\n\n\tnRes = C.EVP_DigestVerifyFinal(mdctx, (*C.uchar)(&signature[0]), C.size_t(len(signature)))\n\tif nRes != 1 {\n\t\treturn false, nil\n\t}\n\n\treturn true, nil\n}", "func verifySig(sigStr []byte, publicKeyStr []byte, scriptPubKey []byte, tx *types.Transaction, txInIdx int) bool {\n\tsig, err := crypto.SigFromBytes(sigStr)\n\tif err != nil {\n\t\tlogger.Debugf(\"Deserialize signature failed\")\n\t\treturn false\n\t}\n\tpublicKey, err := crypto.PublicKeyFromBytes(publicKeyStr)\n\tif err != nil {\n\t\tlogger.Debugf(\"Deserialize public key failed\")\n\t\treturn false\n\t}\n\n\tsigHash, err := CalcTxHashForSig(scriptPubKey, tx, txInIdx)\n\tif err != nil {\n\t\tlogger.Debugf(\"Calculate signature hash failed\")\n\t\treturn false\n\t}\n\n\treturn sig.VerifySignature(publicKey, sigHash)\n}", "func RsaPublicKeyVerify(data string, publicKeyHexOrPem string, signatureHex string) error {\n\t// data is required\n\tif len(data) == 0 {\n\t\treturn errors.New(\"Data To Verify is Required\")\n\t}\n\n\t// get public key\n\tvar publicKey *rsa.PublicKey\n\tvar err error\n\n\tif util.Left(publicKeyHexOrPem, 26) == \"-----BEGIN PUBLIC KEY-----\" && util.Right(publicKeyHexOrPem, 24) == \"-----END PUBLIC KEY-----\" {\n\t\t// get public key from pem\n\t\tpublicKey, err = rsaPublicKeyFromPem(publicKeyHexOrPem)\n\t} else {\n\t\t// get public key from hex\n\t\tpublicKey, err = rsaPublicKeyFromHex(publicKeyHexOrPem)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// convert data to byte array\n\tmsg := []byte(data)\n\n\t// define hash\n\th := sha256.New()\n\th.Write(msg)\n\td := h.Sum(nil)\n\n\tsig, _ := util.HexToByte(signatureHex)\n\n\terr1 := rsa.VerifyPKCS1v15(publicKey, crypto.SHA256, d, sig)\n\n\tif err1 != nil {\n\t\treturn err1\n\t}\n\n\t// verified\n\treturn nil\n}", "func VerifyPubkey(pubkey []byte) int {\n\tif len(pubkey) != 33 {\n\t\treturn -2\n\t}\n\n\tif secp.PubkeyIsValid(pubkey) != 1 {\n\t\treturn -1 // tests parse and validity\n\t}\n\n\treturn 1 //valid\n}", "func verify(pub crypto.PublicKey, hasher crypto.Hash, data, sig []byte) error {\n\tif sig == nil {\n\t\treturn errors.New(\"signature is nil\")\n\t}\n\n\th := hasher.New()\n\tif _, err := h.Write(data); err != nil {\n\t\treturn errors.Wrap(err, \"write\")\n\t}\n\tdigest := h.Sum(nil)\n\n\tswitch pub := pub.(type) {\n\tcase *ecdsa.PublicKey:\n\t\tif !ecdsa.VerifyASN1(pub, digest, sig) {\n\t\t\treturn errors.New(\"verification failed\")\n\t\t}\n\tdefault:\n\t\treturn fmt.Errorf(\"unknown public key type: %T\", pub)\n\t}\n\treturn nil\n}", "func (pk PublicKey) Verify(sig, message []byte, hasher Hasher) (bool, error) {\n\treturn pk.publicKey.Verify(sig, message, hasher)\n}", "func (r *rsaPublicKey) CheckSignature(message []byte, sig []byte) error {\r\n\th := sha256.New()\r\n\th.Write(message)\r\n\td := h.Sum(nil)\r\n\treturn rsa.VerifyPKCS1v15(r.PublicKey, crypto.SHA256, d, sig)\r\n}", "func VerifySignWithPublicKey(src, signed []byte, hash crypto.Hash) (e error) {\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tswitch x := r.(type) {\n\t\t\tcase string:\n\t\t\t\te = errors.New(x)\n\t\t\tcase error:\n\t\t\t\te = x\n\t\t\tdefault:\n\t\t\t\te = errors.New(\"Unknown panic\")\n\t\t\t}\n\t\t}\n\t}()\n\th := hash.New()\n\th.Write(src)\n\thashed := h.Sum(nil)\n\terr := rsa.VerifyPKCS1v15(publicKey, hash, hashed, signed)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func verifyHandshakeSignature(sigType uint8, pubkey crypto.PublicKey, hashFunc crypto.Hash, signed, sig []byte) error {\n\tswitch sigType {\n\tcase signatureECDSA:\n\t\tpubKey, ok := pubkey.(*ecdsa.PublicKey)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"expected an ECDSA public key, got %T\", pubkey)\n\t\t}\n\t\tif !ecdsa.VerifyASN1(pubKey, signed, sig) {\n\t\t\treturn errors.New(\"ECDSA verification failure\")\n\t\t}\n\tcase signatureEd25519:\n\t\tpubKey, ok := pubkey.(ed25519.PublicKey)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"expected an Ed25519 public key, got %T\", pubkey)\n\t\t}\n\t\tif !ed25519.Verify(pubKey, signed, sig) {\n\t\t\treturn errors.New(\"Ed25519 verification failure\")\n\t\t}\n\tcase signaturePKCS1v15:\n\t\tpubKey, ok := pubkey.(*rsa.PublicKey)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"expected an RSA public key, got %T\", pubkey)\n\t\t}\n\t\tif err := rsa.VerifyPKCS1v15(pubKey, hashFunc, signed, sig); err != nil {\n\t\t\treturn err\n\t\t}\n\tcase signatureRSAPSS:\n\t\tpubKey, ok := pubkey.(*rsa.PublicKey)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"expected an RSA public key, got %T\", pubkey)\n\t\t}\n\t\tsignOpts := &rsa.PSSOptions{SaltLength: rsa.PSSSaltLengthEqualsHash}\n\t\tif err := rsa.VerifyPSS(pubKey, hashFunc, signed, sig, signOpts); err != nil {\n\t\t\treturn err\n\t\t}\n\tdefault:\n\t\treturn errors.New(\"internal error: unknown signature type\")\n\t}\n\treturn nil\n}", "func Verify(pub *ecdsa.PublicKey, hash []byte, r, s *big.Int) bool {\n\treturn ecdsa.Verify(pub, hash, r, s)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
calculateKeyPair converts a Montgomery private key k to a twisted Edwards public key and private key (A, a) as defined in calculate_key_pair(k): E = kB A.y = E.y A.s = 0 if E.s == 1: a = k (mod q) else: a = k (mod q) return A, a
func (p PrivateKey) calculateKeyPair() ([]byte, *edwards25519.Scalar, error) { var pA edwards25519.Point var sa edwards25519.Scalar k, err := (&edwards25519.Scalar{}).SetBytesWithClamping(p) if err != nil { return nil, nil, err } pub := pA.ScalarBaseMult(k).Bytes() signBit := (pub[31] & 0x80) >> 7 if signBit == 1 { sa.Negate(k) // Set sig bit to 0 pub[31] &= 0x7F } else { sa.Set(k) } return pub, &sa, nil }
[ "func GenerateKeyPair(group *schnorr.Group) (*SecKey, *PubKey) {\n\ts1 := common.GetRandomInt(group.Q)\n\ts2 := common.GetRandomInt(group.Q)\n\th1 := group.Exp(group.G, s1)\n\th2 := group.Exp(group.G, s2)\n\n\treturn NewSecKey(s1, s2), NewPubKey(h1, h2)\n}", "func DeriveKeyPair(input string) (pubKey, privateKey *[32]byte) {\n\tprivateKey = new([32]byte)\n\tpubKey = new([32]byte)\n\tprivateKeySlice := sha512.Sum512([]byte(input))\n\tcopy(privateKey[:], privateKeySlice[0:32])\n\tcurve25519.ScalarBaseMult(pubKey, privateKey)\n\treturn\n}", "func GenerateKeyPair(bits int) (keypair *KeyPair, err error) {\n\tkeypair = new(KeyPair)\n\tkeypair.PublicKey = new(PublicKey)\n\tkeypair.PrivateKey = new(PrivateKey)\n\n\tif bits == 0 {\n\t\terr = errors.New(\"RSA modulus size must not be zero.\")\n\t\treturn\n\t}\n\tif bits%8 != 0 {\n\t\terr = errors.New(\"RSA modulus size must be a multiple of 8.\")\n\t\treturn\n\t}\n\n\tfor limit := 0; limit < 1000; limit++ {\n\t\tvar tempKey *rsa.PrivateKey\n\t\ttempKey, err = rsa.GenerateKey(rand.Reader, bits)\n\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tif len(tempKey.Primes) != 2 {\n\t\t\terr = errors.New(\"RSA package generated a weird set of primes (i.e. not two)\")\n\t\t\treturn\n\t\t}\n\n\t\tp := tempKey.Primes[0]\n\t\tq := tempKey.Primes[1]\n\n\t\tif p.Cmp(q) == 0 {\n\t\t\terr = errors.New(\"RSA keypair factors were equal. This is really unlikely dependent on the bitsize and it appears something horrible has happened.\")\n\t\t\treturn\n\t\t}\n\t\tif gcd := new(big.Int).GCD(nil, nil, p, q); gcd.Cmp(big.NewInt(1)) != 0 {\n\t\t\terr = errors.New(\"RSA primes were not relatively prime!\")\n\t\t\treturn\n\t\t}\n\n\t\tmodulus := new(big.Int).Mul(p, q)\n\n\t\tpublicExp := big.NewInt(3)\n\t\t//publicExp := big.NewInt(65537)\n\n\t\t//totient = (p-1) * (q-1)\n\t\ttotient := new(big.Int)\n\t\ttotient.Sub(p, big.NewInt(1))\n\t\ttotient.Mul(totient, new(big.Int).Sub(q, big.NewInt(1)))\n\n\t\tif gcd := new(big.Int).GCD(nil, nil, publicExp, totient); gcd.Cmp(big.NewInt(1)) != 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\tprivateExp := new(big.Int).ModInverse(publicExp, totient)\n\t\tkeypair.PublicKey.Modulus = modulus\n\t\tkeypair.PrivateKey.Modulus = modulus\n\t\tkeypair.PublicKey.PublicExp = publicExp\n\t\tkeypair.PrivateKey.PrivateExp = privateExp\n\t\treturn\n\t}\n\terr = errors.New(\"Failed to generate a within the limit!\")\n\treturn\n\n}", "func generateKeyPair() (priv, pub *[keySize]byte, err error) {\n\tpub, priv, err = box.GenerateKey(rand.Reader)\n\treturn\n}", "func NewKeyPair(suite suites.Suite, random cipher.Stream) (kyber.Scalar, kyber.Point) {\n\tx := suite.G2().Scalar().Pick(random)\n\tX := suite.G2().Point().Mul(x, nil)\n\treturn x, X\n}", "func generateKeyPair() (publicKey, privateKey *[32]byte, err error) {\n\treturn box.GenerateKey(rand.Reader)\n}", "func GenerateKeyPair() (*rsa.PrivateKey, *rsa.PublicKey, error) {\n\tprivKey, err := rsa.GenerateKey(rand.Reader, 2048)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\treturn privKey, &privKey.PublicKey, nil\n}", "func NewKeyPair() (*keyPair, error) {\n\tprivKey, err := rsa.GenerateKey(rand.Reader, 2048)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tprivKey.Precompute()\n\n\tpubKey := &privKey.PublicKey\n\treturn &keyPair{Private: privKey, Public: pubKey}, nil\n}", "func GetAccountKeyPairFor(name string) (string, string) {\n\n\tar := AccountsRepository()\n\tpk1, ok := ar.publicKey[name]\n\tvar puk, prk string\n\tif ok {\n\t\tpuk = pk1\n\t} else {\n\t\tpuk = \"\"\n\t}\n\tpk2, ok := ar.privateKey[name]\n\tif ok {\n\t\tprk = pk2\n\t} else {\n\t\tprk = \"\"\n\t}\n\treturn puk, prk\n}", "func possibleK(pair messagePair, pub *dsa.PublicKey) *big.Int {\n\tz1 := new(big.Int).SetBytes(pair.fst.sum)\n\tz2 := new(big.Int).SetBytes(pair.snd.sum)\n\n\tz1.Sub(z1, z2)\n\tz2.Sub(pair.fst.s, pair.snd.s)\n\tz2.ModInverse(z2, pub.Q)\n\tk := z1.Mul(z1, z2)\n\n\treturn k.Mod(k, pub.Q)\n}", "func newKeyPair() (ecdsa.PrivateKey, []byte) {\n\tcurve := elliptic.P256()\n\n\tpriKey, err := ecdsa.GenerateKey(curve, rand.Reader)\n\tif err != nil {\n\t\tlog.Panicln(err)\n\t}\n\tpubKey := append(priKey.PublicKey.X.Bytes(), priKey.PublicKey.Y.Bytes()...)\n\n\treturn *priKey, pubKey\n}", "func GeneratePairingKey() (priv []byte, g1Pub []byte, g2Pub []byte, err error) {\n\tprivInt, err := rand.Int(rand.Reader, bn256.Order)\n\tif err != nil {\n\t\treturn nil, nil, nil, err\n\t}\n\n\tg1PubInt := new(bn256.G1).ScalarBaseMult(privInt)\n\tg2PubInt := new(bn256.G2).ScalarBaseMult(privInt)\n\n\tpriv = privInt.Bytes()\n\tg1Pub = g1PubInt.Marshal()\n\tg2Pub = g2PubInt.Marshal()\n\treturn\n}", "func (n *nauth) GenerateKeyPair(passphrase string) ([]byte, []byte, error) {\n\tpriv, err := rsa.GenerateKey(rand.Reader, 2048)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tprivDer := x509.MarshalPKCS1PrivateKey(priv)\n\tprivBlock := pem.Block{\n\t\tType: \"RSA PRIVATE KEY\",\n\t\tHeaders: nil,\n\t\tBytes: privDer,\n\t}\n\tprivPem := pem.EncodeToMemory(&privBlock)\n\n\tpub, err := ssh.NewPublicKey(&priv.PublicKey)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tpubBytes := ssh.MarshalAuthorizedKey(pub)\n\treturn privPem, pubBytes, nil\n}", "func (s Seed) deriveKeyPair(index uint64) (keypair [64]byte) {\n\tbuf := make([]byte, len(s.siadSeed)+8)\n\tn := copy(buf, s.siadSeed[:])\n\tbinary.LittleEndian.PutUint64(buf[n:], index)\n\tseed := blake2b.Sum256(buf)\n\tcopy(keypair[:], ed25519.NewKeyFromSeed(seed[:]))\n\treturn\n}", "func NewPair(p *big.Int, g int64) (private, public *big.Int) {\n\tprivate = PrivateKey(p)\n\tpublic = PublicKey(private, p, g)\n\treturn private, public\n}", "func generateKeypair() ([]byte, []byte, error) {\n\tprivateKey, err := rsa.GenerateKey(rand.Reader, 2048)\n\tif err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"failed to generate SSH private key: %v\", err)\n\t}\n\tprivatePEM := pem.EncodeToMemory(&pem.Block{\n\t\tType: \"RSA PRIVATE KEY\",\n\t\tHeaders: nil,\n\t\tBytes: x509.MarshalPKCS1PrivateKey(privateKey),\n\t})\n\tpublicKey, err := cssh.NewPublicKey(&privateKey.PublicKey)\n\tif err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"failed to generate SSH public key: %v\", err)\n\t}\n\tpublicPEM := cssh.MarshalAuthorizedKey(publicKey)\n\treturn privatePEM, publicPEM, nil\n}", "func generateKeyPair(algo string, ecCurve string) (privateKey interface{}, publicKey interface{}, err error) {\n\n // Make them case-insensitive\n switch strings.ToUpper(algo) {\n // If RSA, generate a pair of RSA keys\n case \"RSA\":\n // rsa.GenerateKey(): https://golang.org/pkg/crypto/rsa/#GenerateKey\n // Return value is of type *rsa.PrivateKey\n privateKey, err = rsa.GenerateKey(rand.Reader, 2048) // by default create a 2048 bit key\n\n // If ECDSA, use a provided curve\n case \"ECDSA\":\n // First check if ecCurve is provided\n if ecCurve == \"\" {\n return nil, nil, errors.New(\"ECDSA needs a curve\")\n }\n // Then generate the key based on the curve\n // Curves: https://golang.org/pkg/crypto/elliptic/#Curve\n // ecdsa.GenerateKey(): https://golang.org/pkg/crypto/ecdsa/#GenerateKey\n // Return value is of type *ecdsa.PrivateKey\n switch strings.ToUpper(ecCurve) {\n case \"P224\":\n privateKey, err = ecdsa.GenerateKey(elliptic.P224(), rand.Reader)\n case \"P256\":\n privateKey, err = ecdsa.GenerateKey(elliptic.P256(), rand.Reader)\n case \"P384\":\n \tprivateKey, err = ecdsa.GenerateKey(elliptic.P384(), rand.Reader)\n case \"P521\":\n \tprivateKey, err = ecdsa.GenerateKey(elliptic.P521(), rand.Reader)\n\n // If the curve is invalid\n default:\n return nil, nil, errors.New(\"Unrecognized curve, valid values are P224, P256, P384 and P521\")\n }\n\n // If neither RSA nor ECDSA return an error\n default:\n return nil, nil, errors.New(\"Unrecognized algorithm, valid options are RSA and ECDSA\")\n }\n\n // If we get here, then input parameters have been valid\n // Check if key generation has been successful by checking err\n if err != nil {\n return nil, nil, err\n }\n\n // Exporting the public key (needed later)\n switch tempPrivKey:= privateKey.(type) {\n case *rsa.PrivateKey:\n publicKey = &tempPrivKey.PublicKey\n case *ecdsa.PrivateKey:\n publicKey = &tempPrivKey.PublicKey\n }\n\n return privateKey, publicKey, err // or just return\n}", "func NewPair(p *big.Int, g int64) (private, public *big.Int) {\n\tprivKey := PrivateKey(p)\n\tpubKey := PublicKey(privKey, p, g)\n\treturn privKey, pubKey\n}", "func GenKeyPair(n, rtH, chanH uint32, c uint16, w uint16) (*PrivateKey, *PublicKey, error) {\n\treturn GenerateKeyPair(InitParam(n, rtH, chanH, 0, c, w), 0)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
convertMont converts from a Montgomery ucoordinate to a twisted Edwards point P, according to convert_mont(u): umasked = u (mod 2|p|) P.y = u_to_y(umasked) P.s = 0 return P
func convertMont(u PublicKey) (*edwards25519.Point, error) { um, err := (&field.Element{}).SetBytes(u) if err != nil { return nil, err } // y = (u - 1)/(u + 1) a := new(field.Element).Subtract(um, one) b := new(field.Element).Add(um, one) y := new(field.Element).Multiply(a, b.Invert(b)).Bytes() // Set sign to 0 y[31] &= 0x7F return (&edwards25519.Point{}).SetBytes(y) }
[ "func (p *Poly) toMont() {\n\tvar f int16 = int16((uint64(1) << 32) % uint64(q))\n\tfor i := 0; i < n; i++ {\n\t\tp[i] = montgomeryReduce(int32(p[i]) * int32(f))\n\t}\n}", "func (z *Element22) FromMont() *Element22 {\n\n\t// the following lines implement z = z * 1\n\t// with a modified CIOS montgomery multiplication\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 2085129623399436079\n\t\tC := madd0(m, 9062599614324828209, z[0])\n\t\tC, z[0] = madd2(m, 952425709649632109, z[1], C)\n\t\tC, z[1] = madd2(m, 13987751354083916656, z[2], C)\n\t\tC, z[2] = madd2(m, 9476693002504986527, z[3], C)\n\t\tC, z[3] = madd2(m, 17899356805776864267, z[4], C)\n\t\tC, z[4] = madd2(m, 2607080593922027197, z[5], C)\n\t\tC, z[5] = madd2(m, 6852504016717314360, z[6], C)\n\t\tC, z[6] = madd2(m, 366248478184989226, z[7], C)\n\t\tC, z[7] = madd2(m, 2672987780203805083, z[8], C)\n\t\tC, z[8] = madd2(m, 14115032483094903896, z[9], C)\n\t\tC, z[9] = madd2(m, 8062699450825609015, z[10], C)\n\t\tC, z[10] = madd2(m, 8413249848292746549, z[11], C)\n\t\tC, z[11] = madd2(m, 11172154229712803058, z[12], C)\n\t\tC, z[12] = madd2(m, 18137346262305431037, z[13], C)\n\t\tC, z[13] = madd2(m, 123227702747754650, z[14], C)\n\t\tC, z[14] = madd2(m, 7409464670784690235, z[15], C)\n\t\tC, z[15] = madd2(m, 243347369443125979, z[16], C)\n\t\tC, z[16] = madd2(m, 200317109320159479, z[17], C)\n\t\tC, z[17] = madd2(m, 17492726232193822651, z[18], C)\n\t\tC, z[18] = madd2(m, 17666595880400198649, z[19], C)\n\t\tC, z[19] = madd2(m, 1619463007483089584, z[20], C)\n\t\tC, z[20] = madd2(m, 7910025299994333900, z[21], C)\n\t\tz[21] = C\n\t}\n\n\t// if z > q --> z -= q\n\tif !(z[21] < 7910025299994333900 || (z[21] == 7910025299994333900 && (z[20] < 1619463007483089584 || (z[20] == 1619463007483089584 && (z[19] < 17666595880400198649 || (z[19] == 17666595880400198649 && (z[18] < 17492726232193822651 || (z[18] == 17492726232193822651 && (z[17] < 200317109320159479 || (z[17] == 200317109320159479 && (z[16] < 243347369443125979 || (z[16] == 243347369443125979 && (z[15] < 7409464670784690235 || (z[15] == 7409464670784690235 && (z[14] < 123227702747754650 || (z[14] == 123227702747754650 && (z[13] < 18137346262305431037 || (z[13] == 18137346262305431037 && (z[12] < 11172154229712803058 || (z[12] == 11172154229712803058 && (z[11] < 8413249848292746549 || (z[11] == 8413249848292746549 && (z[10] < 8062699450825609015 || (z[10] == 8062699450825609015 && (z[9] < 14115032483094903896 || (z[9] == 14115032483094903896 && (z[8] < 2672987780203805083 || (z[8] == 2672987780203805083 && (z[7] < 366248478184989226 || (z[7] == 366248478184989226 && (z[6] < 6852504016717314360 || (z[6] == 6852504016717314360 && (z[5] < 2607080593922027197 || (z[5] == 2607080593922027197 && (z[4] < 17899356805776864267 || (z[4] == 17899356805776864267 && (z[3] < 9476693002504986527 || (z[3] == 9476693002504986527 && (z[2] < 13987751354083916656 || (z[2] == 13987751354083916656 && (z[1] < 952425709649632109 || (z[1] == 952425709649632109 && (z[0] < 9062599614324828209))))))))))))))))))))))))))))))))))))))))))) {\n\t\tvar b uint64\n\t\tz[0], b = bits.Sub64(z[0], 9062599614324828209, 0)\n\t\tz[1], b = bits.Sub64(z[1], 952425709649632109, b)\n\t\tz[2], b = bits.Sub64(z[2], 13987751354083916656, b)\n\t\tz[3], b = bits.Sub64(z[3], 9476693002504986527, b)\n\t\tz[4], b = bits.Sub64(z[4], 17899356805776864267, b)\n\t\tz[5], b = bits.Sub64(z[5], 2607080593922027197, b)\n\t\tz[6], b = bits.Sub64(z[6], 6852504016717314360, b)\n\t\tz[7], b = bits.Sub64(z[7], 366248478184989226, b)\n\t\tz[8], b = bits.Sub64(z[8], 2672987780203805083, b)\n\t\tz[9], b = bits.Sub64(z[9], 14115032483094903896, b)\n\t\tz[10], b = bits.Sub64(z[10], 8062699450825609015, b)\n\t\tz[11], b = bits.Sub64(z[11], 8413249848292746549, b)\n\t\tz[12], b = bits.Sub64(z[12], 11172154229712803058, b)\n\t\tz[13], b = bits.Sub64(z[13], 18137346262305431037, b)\n\t\tz[14], b = bits.Sub64(z[14], 123227702747754650, b)\n\t\tz[15], b = bits.Sub64(z[15], 7409464670784690235, b)\n\t\tz[16], b = bits.Sub64(z[16], 243347369443125979, b)\n\t\tz[17], b = bits.Sub64(z[17], 200317109320159479, b)\n\t\tz[18], b = bits.Sub64(z[18], 17492726232193822651, b)\n\t\tz[19], b = bits.Sub64(z[19], 17666595880400198649, b)\n\t\tz[20], b = bits.Sub64(z[20], 1619463007483089584, b)\n\t\tz[21], _ = bits.Sub64(z[21], 7910025299994333900, b)\n\t}\n\treturn z\n}", "func (z *E12) ToMont() *E12 {\n\tz.C0.ToMont()\n\tz.C1.ToMont()\n\treturn z\n}", "func (z *E12) FromMont() *E12 {\n\tz.C0.FromMont()\n\tz.C1.FromMont()\n\treturn z\n}", "func (z *Element29) FromMont() *Element29 {\n\n\t// the following lines implement z = z * 1\n\t// with a modified CIOS montgomery multiplication\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 564333287126902609\n\t\tC := madd0(m, 10495171074974219855, z[0])\n\t\tC, z[0] = madd2(m, 12186246344324844492, z[1], C)\n\t\tC, z[1] = madd2(m, 12986593952034689168, z[2], C)\n\t\tC, z[2] = madd2(m, 4485867796469714651, z[3], C)\n\t\tC, z[3] = madd2(m, 2856192544440570130, z[4], C)\n\t\tC, z[4] = madd2(m, 16684960377341087475, z[5], C)\n\t\tC, z[5] = madd2(m, 6168443939922053920, z[6], C)\n\t\tC, z[6] = madd2(m, 15211678537843601168, z[7], C)\n\t\tC, z[7] = madd2(m, 4136621548862493093, z[8], C)\n\t\tC, z[8] = madd2(m, 11772484457930480875, z[9], C)\n\t\tC, z[9] = madd2(m, 1624439382184385293, z[10], C)\n\t\tC, z[10] = madd2(m, 17673142019601735854, z[11], C)\n\t\tC, z[11] = madd2(m, 18382841015356938091, z[12], C)\n\t\tC, z[12] = madd2(m, 16403499969676496523, z[13], C)\n\t\tC, z[13] = madd2(m, 602485813262065246, z[14], C)\n\t\tC, z[14] = madd2(m, 11227388612910627318, z[15], C)\n\t\tC, z[15] = madd2(m, 11704114044424477697, z[16], C)\n\t\tC, z[16] = madd2(m, 6473159202362305986, z[17], C)\n\t\tC, z[17] = madd2(m, 17094834390420404065, z[18], C)\n\t\tC, z[18] = madd2(m, 13593751014174640077, z[19], C)\n\t\tC, z[19] = madd2(m, 2418034124407931402, z[20], C)\n\t\tC, z[20] = madd2(m, 1811136756773556623, z[21], C)\n\t\tC, z[21] = madd2(m, 16876362074790388108, z[22], C)\n\t\tC, z[22] = madd2(m, 5703058630810439316, z[23], C)\n\t\tC, z[23] = madd2(m, 8899594887613302780, z[24], C)\n\t\tC, z[24] = madd2(m, 11519780057184870984, z[25], C)\n\t\tC, z[25] = madd2(m, 462160053616308140, z[26], C)\n\t\tC, z[26] = madd2(m, 7949737452137819570, z[27], C)\n\t\tC, z[27] = madd2(m, 7203848216053748556, z[28], C)\n\t\tz[28] = C\n\t}\n\n\t// if z > q --> z -= q\n\tif !(z[28] < 7203848216053748556 || (z[28] == 7203848216053748556 && (z[27] < 7949737452137819570 || (z[27] == 7949737452137819570 && (z[26] < 462160053616308140 || (z[26] == 462160053616308140 && (z[25] < 11519780057184870984 || (z[25] == 11519780057184870984 && (z[24] < 8899594887613302780 || (z[24] == 8899594887613302780 && (z[23] < 5703058630810439316 || (z[23] == 5703058630810439316 && (z[22] < 16876362074790388108 || (z[22] == 16876362074790388108 && (z[21] < 1811136756773556623 || (z[21] == 1811136756773556623 && (z[20] < 2418034124407931402 || (z[20] == 2418034124407931402 && (z[19] < 13593751014174640077 || (z[19] == 13593751014174640077 && (z[18] < 17094834390420404065 || (z[18] == 17094834390420404065 && (z[17] < 6473159202362305986 || (z[17] == 6473159202362305986 && (z[16] < 11704114044424477697 || (z[16] == 11704114044424477697 && (z[15] < 11227388612910627318 || (z[15] == 11227388612910627318 && (z[14] < 602485813262065246 || (z[14] == 602485813262065246 && (z[13] < 16403499969676496523 || (z[13] == 16403499969676496523 && (z[12] < 18382841015356938091 || (z[12] == 18382841015356938091 && (z[11] < 17673142019601735854 || (z[11] == 17673142019601735854 && (z[10] < 1624439382184385293 || (z[10] == 1624439382184385293 && (z[9] < 11772484457930480875 || (z[9] == 11772484457930480875 && (z[8] < 4136621548862493093 || (z[8] == 4136621548862493093 && (z[7] < 15211678537843601168 || (z[7] == 15211678537843601168 && (z[6] < 6168443939922053920 || (z[6] == 6168443939922053920 && (z[5] < 16684960377341087475 || (z[5] == 16684960377341087475 && (z[4] < 2856192544440570130 || (z[4] == 2856192544440570130 && (z[3] < 4485867796469714651 || (z[3] == 4485867796469714651 && (z[2] < 12986593952034689168 || (z[2] == 12986593952034689168 && (z[1] < 12186246344324844492 || (z[1] == 12186246344324844492 && (z[0] < 10495171074974219855))))))))))))))))))))))))))))))))))))))))))))))))))))))))) {\n\t\tvar b uint64\n\t\tz[0], b = bits.Sub64(z[0], 10495171074974219855, 0)\n\t\tz[1], b = bits.Sub64(z[1], 12186246344324844492, b)\n\t\tz[2], b = bits.Sub64(z[2], 12986593952034689168, b)\n\t\tz[3], b = bits.Sub64(z[3], 4485867796469714651, b)\n\t\tz[4], b = bits.Sub64(z[4], 2856192544440570130, b)\n\t\tz[5], b = bits.Sub64(z[5], 16684960377341087475, b)\n\t\tz[6], b = bits.Sub64(z[6], 6168443939922053920, b)\n\t\tz[7], b = bits.Sub64(z[7], 15211678537843601168, b)\n\t\tz[8], b = bits.Sub64(z[8], 4136621548862493093, b)\n\t\tz[9], b = bits.Sub64(z[9], 11772484457930480875, b)\n\t\tz[10], b = bits.Sub64(z[10], 1624439382184385293, b)\n\t\tz[11], b = bits.Sub64(z[11], 17673142019601735854, b)\n\t\tz[12], b = bits.Sub64(z[12], 18382841015356938091, b)\n\t\tz[13], b = bits.Sub64(z[13], 16403499969676496523, b)\n\t\tz[14], b = bits.Sub64(z[14], 602485813262065246, b)\n\t\tz[15], b = bits.Sub64(z[15], 11227388612910627318, b)\n\t\tz[16], b = bits.Sub64(z[16], 11704114044424477697, b)\n\t\tz[17], b = bits.Sub64(z[17], 6473159202362305986, b)\n\t\tz[18], b = bits.Sub64(z[18], 17094834390420404065, b)\n\t\tz[19], b = bits.Sub64(z[19], 13593751014174640077, b)\n\t\tz[20], b = bits.Sub64(z[20], 2418034124407931402, b)\n\t\tz[21], b = bits.Sub64(z[21], 1811136756773556623, b)\n\t\tz[22], b = bits.Sub64(z[22], 16876362074790388108, b)\n\t\tz[23], b = bits.Sub64(z[23], 5703058630810439316, b)\n\t\tz[24], b = bits.Sub64(z[24], 8899594887613302780, b)\n\t\tz[25], b = bits.Sub64(z[25], 11519780057184870984, b)\n\t\tz[26], b = bits.Sub64(z[26], 462160053616308140, b)\n\t\tz[27], b = bits.Sub64(z[27], 7949737452137819570, b)\n\t\tz[28], _ = bits.Sub64(z[28], 7203848216053748556, b)\n\t}\n\treturn z\n}", "func (z *Element17) FromMont() *Element17 {\n\n\t// the following lines implement z = z * 1\n\t// with a modified CIOS montgomery multiplication\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\t{\n\t\t// m = z[0]n'[0] mod W\n\t\tm := z[0] * 8334087986022555761\n\t\tC := madd0(m, 14074317334421561199, z[0])\n\t\tC, z[0] = madd2(m, 3603097598639507118, z[1], C)\n\t\tC, z[1] = madd2(m, 1901064316570948514, z[2], C)\n\t\tC, z[2] = madd2(m, 13592020221286190892, z[3], C)\n\t\tC, z[3] = madd2(m, 5795418031312419601, z[4], C)\n\t\tC, z[4] = madd2(m, 512834486801673566, z[5], C)\n\t\tC, z[5] = madd2(m, 4918065994611781387, z[6], C)\n\t\tC, z[6] = madd2(m, 3333387418770116569, z[7], C)\n\t\tC, z[7] = madd2(m, 11268225615708655620, z[8], C)\n\t\tC, z[8] = madd2(m, 1049792113801022996, z[9], C)\n\t\tC, z[9] = madd2(m, 8783558615805171218, z[10], C)\n\t\tC, z[10] = madd2(m, 12764383670427348163, z[11], C)\n\t\tC, z[11] = madd2(m, 5284929597638693549, z[12], C)\n\t\tC, z[12] = madd2(m, 3026685772878610893, z[13], C)\n\t\tC, z[13] = madd2(m, 1747888123470247313, z[14], C)\n\t\tC, z[14] = madd2(m, 12413919554062974026, z[15], C)\n\t\tC, z[15] = madd2(m, 7176756318483791390, z[16], C)\n\t\tz[16] = C\n\t}\n\n\t// if z > q --> z -= q\n\tif !(z[16] < 7176756318483791390 || (z[16] == 7176756318483791390 && (z[15] < 12413919554062974026 || (z[15] == 12413919554062974026 && (z[14] < 1747888123470247313 || (z[14] == 1747888123470247313 && (z[13] < 3026685772878610893 || (z[13] == 3026685772878610893 && (z[12] < 5284929597638693549 || (z[12] == 5284929597638693549 && (z[11] < 12764383670427348163 || (z[11] == 12764383670427348163 && (z[10] < 8783558615805171218 || (z[10] == 8783558615805171218 && (z[9] < 1049792113801022996 || (z[9] == 1049792113801022996 && (z[8] < 11268225615708655620 || (z[8] == 11268225615708655620 && (z[7] < 3333387418770116569 || (z[7] == 3333387418770116569 && (z[6] < 4918065994611781387 || (z[6] == 4918065994611781387 && (z[5] < 512834486801673566 || (z[5] == 512834486801673566 && (z[4] < 5795418031312419601 || (z[4] == 5795418031312419601 && (z[3] < 13592020221286190892 || (z[3] == 13592020221286190892 && (z[2] < 1901064316570948514 || (z[2] == 1901064316570948514 && (z[1] < 3603097598639507118 || (z[1] == 3603097598639507118 && (z[0] < 14074317334421561199))))))))))))))))))))))))))))))))) {\n\t\tvar b uint64\n\t\tz[0], b = bits.Sub64(z[0], 14074317334421561199, 0)\n\t\tz[1], b = bits.Sub64(z[1], 3603097598639507118, b)\n\t\tz[2], b = bits.Sub64(z[2], 1901064316570948514, b)\n\t\tz[3], b = bits.Sub64(z[3], 13592020221286190892, b)\n\t\tz[4], b = bits.Sub64(z[4], 5795418031312419601, b)\n\t\tz[5], b = bits.Sub64(z[5], 512834486801673566, b)\n\t\tz[6], b = bits.Sub64(z[6], 4918065994611781387, b)\n\t\tz[7], b = bits.Sub64(z[7], 3333387418770116569, b)\n\t\tz[8], b = bits.Sub64(z[8], 11268225615708655620, b)\n\t\tz[9], b = bits.Sub64(z[9], 1049792113801022996, b)\n\t\tz[10], b = bits.Sub64(z[10], 8783558615805171218, b)\n\t\tz[11], b = bits.Sub64(z[11], 12764383670427348163, b)\n\t\tz[12], b = bits.Sub64(z[12], 5284929597638693549, b)\n\t\tz[13], b = bits.Sub64(z[13], 3026685772878610893, b)\n\t\tz[14], b = bits.Sub64(z[14], 1747888123470247313, b)\n\t\tz[15], b = bits.Sub64(z[15], 12413919554062974026, b)\n\t\tz[16], _ = bits.Sub64(z[16], 7176756318483791390, b)\n\t}\n\treturn z\n}", "func (p *Poly) fromMont() {\n\tinv := uint64(8265825)\n\tfor i := uint(0); i < n; i++ {\n\t\tp[i] = int32((uint64(p[i]) * inv) % q)\n\t}\n}", "func (p *Poly) tomont() {\n\tfor i := 0; i < n; i++ {\n\t\tp[i] = montgomeryReduce(int64(p[i]))\n\t}\n}", "func convertTemperature(fromUOM, toUOM string, value float64) float64 {\n\tfromUOM = resolveTemperatureSynonyms(fromUOM)\n\ttoUOM = resolveTemperatureSynonyms(toUOM)\n\tif fromUOM == toUOM {\n\t\treturn value\n\t}\n\t// convert to Kelvin\n\tswitch fromUOM {\n\tcase \"F\":\n\t\tvalue = (value-32)/1.8 + 273.15\n\tcase \"C\":\n\t\tvalue += 273.15\n\tcase \"Rank\":\n\t\tvalue /= 1.8\n\tcase \"Reau\":\n\t\tvalue = value*1.25 + 273.15\n\t}\n\t// convert from Kelvin\n\tswitch toUOM {\n\tcase \"F\":\n\t\tvalue = (value-273.15)*1.8 + 32\n\tcase \"C\":\n\t\tvalue -= 273.15\n\tcase \"Rank\":\n\t\tvalue *= 1.8\n\tcase \"Reau\":\n\t\tvalue = (value - 273.15) * 0.8\n\t}\n\treturn value\n}", "func (f *Fpdf) UnitToPointConvert(u float64) (pt float64) {\n\treturn u * f.k\n}", "func NotationToCoord(algebra string) Coord {\n\tif len(algebra) != 2 {\n\t\tpanic(\"Algebraic notation must be 2 characters precisely; got: '\" + algebra + \"'\")\n\t}\n\talgebra = strings.ToUpper(algebra)\n\n\tvar c Coord\n\tfile := algebra[0]\n\trank := algebra[1]\n\n\t// Remember, these are ASCII code points, not numbers\n\tif file < 65 || file > 72 || rank < 48 || rank > 57 {\n\t\tpanic(\"Bad position (\" + algebra + \")\")\n\t}\n\n\tc.Row = int(rank - 48 - 1)\n\tc.Col = int(file - 65)\n\n\treturn c\n}", "func ASwissCoordToStruct(coord string) (*SwissCoord, error) {\n\n\tcompact := strings.ToUpper(strings.TrimSpace(coord))\n\tvar rights, heights string\n\tvar coordType, oldcoordType SwissCoordType\n\tvar right, height float64\n\tvar err error\n\nL1:\n\tfor i, index := 0, 0; i < 2; i++ {\n\t\tindex = strings.Index(compact, \" \")\n\t\tif index == -1 {\n\t\t\tindex = len(compact)\n\t\t}\n\n\t\tswitch compact[:2] {\n\t\tcase \"X:\":\n\t\t\tcoordType = LV03\n\t\t\theights = compact[2:index]\n\t\tcase \"Y:\":\n\t\t\tcoordType = LV03\n\t\t\trights = compact[2:index]\n\t\tcase \"E:\":\n\t\t\tcoordType = LV95\n\t\t\trights = compact[2:index]\n\t\tcase \"N:\":\n\t\t\tcoordType = LV95\n\t\t\theights = compact[2:index]\n\t\tdefault:\n\t\t\terr = cartconvert.ErrSyntax\n\t\t\tbreak L1\n\t\t}\n\n\t\tif oldcoordType != coordType {\n\t\t\terr = cartconvert.ErrSyntax\n\t\t\tbreak L1\n\t\t}\n\n\t\tif i == 1 {\n\t\t\tbreak L1\n\t\t}\n\t\tcompact = compact[index+len(\" \"):]\n\t\tcompact = strings.TrimLeft(compact, \" \")\n\t\toldcoordType = coordType\n\t}\n\n\tif err == nil {\n\n\t\tright, err = strconv.ParseFloat(rights, 64)\n\t\tif err == nil {\n\n\t\t\theight, err = strconv.ParseFloat(heights, 64)\n\t\t\tif err == nil {\n\t\t\t\treturn &SwissCoord{Easting: right, Northing: height, CoordType: coordType, El: cartconvert.Bessel1841Ellipsoid}, nil\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil, err\n}", "func ToCoordinate(l latlong.LatLonger) Coordinate {\n\trlat, rlon := rad(l.Lat()), rad(l.Lon())\n\n\treturn Coordinate{\n\t\tX: deg(math.Cos(rlat) * math.Cos(rlon)),\n\t\tY: deg(math.Cos(rlat) * math.Sin(rlon)),\n\t\tZ: deg(math.Sin(rlat)),\n\t}\n}", "func (f *Fpdf) PointToUnitConvert(pt float64) (u float64) {\n\treturn pt / f.k\n}", "func newCoord(instruction string, xwards bool) {\n\tsteps, _ := strconv.Atoi(instruction[1:])\n\tif xwards {\n\t\tif string(instruction[0]) == \"R\" {\n\t\t\tfacing.x = facing.y\n\t\t} else {\n\t\t\tfacing.x = -facing.y\n\t\t}\n\t\twalk(steps, facing.x, coord, xwards)\n\t\tcoord.x += facing.x * steps\n\t} else {\n\t\tif string(instruction[0]) == \"R\" {\n\t\t\tfacing.y = -facing.x\n\t\t} else {\n\t\t\tfacing.y = facing.x\n\t\t}\n\t\twalk(steps, facing.y, coord, xwards)\n\t\tcoord.y += facing.y * steps\n\t}\n}", "func ConvertPt(pt []float64) []int64 {\n\tnewpt := make([]int64, 2)\n\tnewpt[0] = int64(pt[0] * math.Pow(10.0, 7.0))\n\tnewpt[1] = int64(pt[1] * math.Pow(10.0, 7.0))\n\treturn newpt\n}", "func stToUV(s float64) float64 {\n\tif s >= 0.5 {\n\t\treturn (1 / 3.) * (4*s*s - 1)\n\t}\n\treturn (1 / 3.) * (1 - 4*(1-s)*(1-s))\n}", "func PToXY(c ...float64) ([]float64, error) {\n\tif len(c) < 2 {\n\t\treturn c, ErrCoordsRequire2Values\n\t}\n\t// log.Println(\"Lon/Lat\", c)\n\t//x, y := PLonToX(c[0]), PLatToY(c[1])\n\n\tcrds := []float64{PLonToX(c[0]), PLatToY(c[1])}\n\tcrds = append(crds, c[2:]...)\n\treturn crds, nil\n}", "func (s *OrganizationsService) ConvertMemberToOutsideCollaborator(ctx context.Context, org string, user string) (*Response, error) {\n\tu := fmt.Sprintf(\"orgs/%v/outside_collaborators/%v\", org, user)\n\treq, err := s.client.NewRequest(\"PUT\", u, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.client.Do(ctx, req, nil)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Read is a shortcut function to read and parse projects
func Read(path string) (*Project, error) { var error error var data []byte data, error = ioutil.ReadFile(path) if error != nil { return nil, error } var project = &Project{} error = json.Unmarshal(data, project) if error != nil { return nil, error } return project, nil }
[ "func ReadProject(projectFile, namespace, context, kubeConfig string, variables map[string]string, variableFiles []string, includeResources []string, excludeResources []string) (*Project, error) {\n\tproject := &Project{\n\t\tcontext: context,\n\t\tkubeConfig: kubeConfig,\n\t}\n\terr := project.resolveCommandlineVariables(variables, variableFiles)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tprojectConfig, err := ReadProjectConfig(projectFile, project.variables)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tproject.config = projectConfig\n\tproject.deleteNamespaceConfig = projectConfig.DeleteNamespace\n\tproject.resolveProjectRoot(projectFile, projectConfig.RootDir)\n\tproject.resolveNamespace(namespace, projectConfig.Namespace)\n\tproject.resolveVariables(projectConfig.Variables)\n\terr = project.resolveResourceGraph(projectConfig.ResourceGroups, includeResources, excludeResources)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn project, nil\n}", "func (a *Client) ReadProject(params *ReadProjectParams) (*ReadProjectOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewReadProjectParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"readProject\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/projects/{uuid}\",\n\t\tProducesMediaTypes: []string{\"application/release-manager.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/release-manager.v1+json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &ReadProjectReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*ReadProjectOK), nil\n\n}", "func dataSourceProjectRead(d *schema.ResourceData, meta interface{}) error {\n config := meta.(*PConfig)\n\n name := d.Get(\"name\").(string)\n platform := d.Get(\"platform\").(string)\n resourcePath := fmt.Sprintf(\"api/v3/%s/projects/\", platform)\n response, err := config.doNormalRequest(platform, resourcePath, \"GET\", nil)\n\n if err != nil {\n return fmt.Errorf(\"Unable to list projects: %v\", err)\n }\n\n var data []map[string]interface{}\n err = json.Unmarshal([]byte(response), &data)\n\n if err != nil {\n return err\n }\n\n for i := range data {\n if data[i][\"name\"] == name {\n return dataSourceProjectAttributes(d, data[i])\n }\n }\n\n return fmt.Errorf(\"Unable to retrieve project %s: %v\", name, err)\n}", "func (c *RollbarAPIClient) ReadProject(projectID int) (*Project, error) {\n\tu := c.BaseURL + pathProjectRead\n\n\tl := log.With().\n\t\tInt(\"projectID\", projectID).\n\t\tLogger()\n\tl.Debug().Msg(\"Reading project from API\")\n\n\tresp, err := c.Resty.R().\n\t\tSetResult(projectResponse{}).\n\t\tSetError(ErrorResult{}).\n\t\tSetPathParams(map[string]string{\n\t\t\t\"projectID\": strconv.Itoa(projectID),\n\t\t}).\n\t\tGet(u)\n\tif err != nil {\n\t\tl.Err(err).Msg(\"Error reading project\")\n\t\treturn nil, err\n\t}\n\terr = errorFromResponse(resp)\n\tif err != nil {\n\t\tl.Err(err).Send()\n\t\treturn nil, err\n\t}\n\tpr := resp.Result().(*projectResponse)\n\t// FIXME: This is a workaround for a known bug in the API\n\t// https://github.com/rollbar/terraform-provider-rollbar/issues/23\n\tif pr.Result.Name == \"\" {\n\t\tl.Warn().Msg(\"Project not found\")\n\t\treturn nil, ErrNotFound\n\t}\n\tl.Debug().Msg(\"Project successfully read\")\n\treturn &pr.Result, nil\n\n}", "func ReadProjectFile(dir fs.Directory) ([]byte, error) {\n\tif !dir.IsProject() {\n\t\treturn nil, fmt.Errorf(\"expected project file at %s\", dir.ProjectPath())\n\t}\n\n\tdat, err := ioutil.ReadFile(dir.ProjectPath())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn dat, nil\n}", "func TestRead(t *testing.T) {\n\tgoodConfig := Read(\"./test/good-project\")\n\tif goodConfig.ContextRoot != \"something/else\" {\n\t\tt.Log(\"good project config is incorrect: \" + goodConfig.ContextRoot)\n\t\tt.Fail()\n\t}\n\tbadConfig := Read(\"./test/bad-project\")\n\tif badConfig.ContextRoot != \"./test/bad-project\" {\n\t\tt.Log(\"bad project config is incorrect: \" + badConfig.ContextRoot)\n\t\tt.Fail()\n\t}\n\tmissingConfig := Read(\"./test/missing-project\")\n\tif missingConfig.ContextRoot != \"./test/missing-project\" {\n\t\tt.Log(\"missing project config is incorrect: \" + missingConfig.ContextRoot)\n\t\tt.Fail()\n\t}\n}", "func readProjectConfig(c context.Context, task proto.Message) error {\n\tprojectID := task.(*internal.ReadProjectConfigTask).ProjectId\n\n\tctx, cancel := context.WithTimeout(c, 150*time.Second)\n\tdefer cancel()\n\n\tjobs, err := globalCatalog.GetProjectJobs(ctx, projectID)\n\tif err != nil {\n\t\tlogging.WithError(err).Errorf(c, \"Failed to query for a list of jobs\")\n\t\treturn err\n\t}\n\n\tif err := globalEngine.UpdateProjectJobs(ctx, projectID, jobs); err != nil {\n\t\tlogging.WithError(err).Errorf(c, \"Failed to update some jobs\")\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func dataSourceIdentityProjectV3Read(ctx context.Context, d *schema.ResourceData, meta interface{}) diag.Diagnostics {\n\tconfig := meta.(*Config)\n\n\tidentityClient, err := config.IdentityV3Client(GetRegion(d, config))\n\tif err != nil {\n\t\treturn diag.Errorf(\"Error creating OpenStack identity client: %s\", err)\n\t}\n\n\tenabled := d.Get(\"enabled\").(bool)\n\tisDomain := d.Get(\"is_domain\").(bool)\n\tlistOpts := projects.ListOpts{\n\t\tDomainID: d.Get(\"domain_id\").(string),\n\t\tEnabled: &enabled,\n\t\tIsDomain: &isDomain,\n\t\tName: d.Get(\"name\").(string),\n\t\tParentID: d.Get(\"parent_id\").(string),\n\t}\n\n\tvar project projects.Project\n\tvar allProjects []projects.Project\n\tallPages, err := projects.List(identityClient, listOpts).AllPages()\n\tif err != nil {\n\t\tuserID := config.UserID\n\t\tlog.Printf(\"[DEBUG] Will try to find project with users.ListProjects as I am unable to query openstack_identity_project_v3: %s. Trying listing userprojects.\", err)\n\t\tif userID == \"\" {\n\t\t\ttokenInfo, tokenErr := getTokenInfo(identityClient)\n\t\t\tif tokenErr != nil {\n\t\t\t\treturn diag.Errorf(\"Error when getting token info: %s\", err)\n\t\t\t}\n\t\t\tuserID = tokenInfo.userID\n\t\t}\n\t\t// Search for all the projects using the users.ListProjects API call and filter them\n\t\tallPages, err = users.ListProjects(identityClient, userID).AllPages()\n\t\tif err != nil {\n\t\t\treturn diag.Errorf(\"Unable to query openstack_identity_project_v3: %s\", err)\n\t\t}\n\t\tallProjects, err = projects.ExtractProjects(allPages)\n\t\tif err != nil {\n\t\t\treturn diag.Errorf(\"Unable to retrieve openstack_identity_project_v3: %s\", err)\n\t\t}\n\t\tallProjects = filterProjects(allProjects, listOpts)\n\t} else {\n\t\tallProjects, err = projects.ExtractProjects(allPages)\n\t\tif err != nil {\n\t\t\treturn diag.Errorf(\"Unable to retrieve openstack_identity_project_v3: %s\", err)\n\t\t}\n\t}\n\n\tif len(allProjects) < 1 {\n\t\treturn diag.Errorf(\"Your openstack_identity_project_v3 query returned no results. \" +\n\t\t\t\"Please change your search criteria and try again\")\n\t}\n\n\tif len(allProjects) > 1 {\n\t\treturn diag.Errorf(\"Your openstack_identity_project_v3 query returned more than one result %#v\", allProjects)\n\t}\n\n\tproject = allProjects[0]\n\n\tdataSourceIdentityProjectV3Attributes(d, &project)\n\n\treturn nil\n}", "func (w WatWorkspace) Read(name string) ([]byte, error) {\n\tpath := filepath.Join(w.root, kWatDirName, name)\n\tcontents, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"ioutil.ReadFile: %v (while reading file '%s')\", err, name)\n\t}\n\treturn contents, nil\n}", "func NewReadProjectOK() *ReadProjectOK {\n\n\treturn &ReadProjectOK{}\n}", "func (g *projectGateway) ReadProjectAction(params project.ReadProjectParams) middleware.Responder {\n\treadRsp, err := g.projectClient.Read(context.TODO(), &proto.ReadRequest{\n\t\tUuid: string(params.UUID),\n\t})\n\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn project.NewReadProjectInternalServerError()\n\t}\n\n\tif uint32(codes.OK) == readRsp.Status {\n\t\tfmt.Println(fmt.Sprintf(\"project.client read: ok. Id = %v\", params.UUID))\n\t} else if uint32(codes.NotFound) == readRsp.Status {\n\t\treturn project.NewReadProjectNotFound()\n\t}\n\n\tpr := &models.Project{\n\t\tUUID: strfmt.UUID(readRsp.Project.Uuid),\n\t\tName: readRsp.Project.Name,\n\t\tDescription: readRsp.Project.Description,\n\t}\n\n\treturn project.NewReadProjectOK().WithPayload(pr)\n}", "func ReadProjectConfig(projPath string) (*ThrapConfig, error) {\n\tfilename := filepath.Join(projPath, consts.WorkDir, consts.ConfigFile)\n\treturn ReadThrapConfig(filename)\n}", "func (s *workspaces) Read(ctx context.Context, organization, workspace string) (*Workspace, error) {\n\treturn s.ReadWithOptions(ctx, organization, workspace, nil)\n}", "func Parse(fileName string) Project {\n\tdir, _ := os.Getwd()\n\tdefer os.Chdir(dir)\n\tprojectDir := filepath.Dir(fileName)\n\tfile, err := os.Open(fileName)\n\tos.Chdir(projectDir)\n\tcheck(err)\n\n\tprojectLines, patternFiles, options := splitFileText(file)\n\n\tparsedTempo := parseTempo(options)\n\n\tprojectOptions := grid.Options{Tempo: parsedTempo}\n\n\tparsedPatternFiles := map[string]patternfile.PatternFile{}\n\n\tfor key, fileName := range patternFiles {\n\t\tparsedPatternFile := patternfile.Parse(fileName, projectOptions)\n\t\tparsedPatternFiles[key] = parsedPatternFile\n\t}\n\n\tparts := make([]part.Part, len(projectLines))\n\n\tendOfLastPart := portmidi.Timestamp(0)\n\tfor i, projectLine := range projectLines {\n\n\t\tnewPart := part.Parse(projectLine, parsedPatternFiles, endOfLastPart)\n\t\tendOfLastPart = findLastPartEventTimestamp(newPart) + 1\n\t\tparts[i] = newPart\n\t}\n\n\treturn Project{\n\t\tProjectLines: projectLines,\n\t\tParts: parts,\n\t\tLength: endOfLastPart - 1,\n\t\tTempo: parsedTempo,\n\t}\n\n}", "func Decode(r io.Reader) (*Project, error) {\n\tdec := newDecoder(r)\n\treturn dec.decode()\n}", "func NewReadProjectNotFound() *ReadProjectNotFound {\n\n\treturn &ReadProjectNotFound{}\n}", "func LoadProjects() error {\n\tprojects = make(map[string]Project)\n\tfile, err := os.Open(projectFile)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdecoder := json.NewDecoder(file)\n\tconfigs := []projectConfig{}\n\tif err = decoder.Decode(&configs); err != nil {\n\t\treturn err\n\t}\n\tfor _, config := range configs {\n\t\tprojects[config.Name] = &project{config, NewQueue()}\n\t}\n\tif len(projects) == 0 {\n\t\treturn errors.New(\"no projects defined\")\n\t}\n\treturn nil\n}", "func readBUILD(ctx context.Context, workspaceRoot, buildFilePath string) (*build.File, error) {\n\tnormalizedG3Path, err := getAbsoluteBUILDPath(workspaceRoot, buildFilePath)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to resolve workspace relative path: %s\", err)\n\t}\n\tdata, err := platform.ReadFile(ctx, buildFilePath)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn &build.File{Path: normalizedG3Path, Type: build.TypeBuild}, nil\n\t\t}\n\t\treturn nil, fmt.Errorf(\"reading %q: %s\", buildFilePath, err)\n\t}\n\treturn build.ParseBuild(normalizedG3Path, data)\n}", "func ReadAll() (p *Page, err error) {\n\tdCmn := config.SourceDir + sep + \"pages\" + sep + \"common\" + sep\n\tdOs := config.SourceDir + sep + \"pages\" + sep + config.OSName() + sep\n\tpaths := []string{dCmn, dOs}\n\tp = &Page{Name: \"Search All\"}\n\tp.Tips = make([]*Tip, 0)\n\tfor _, pt := range paths {\n\t\tfiles, err := ioutil.ReadDir(pt)\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tfor _, f := range files {\n\t\t\tif strings.HasSuffix(f.Name(), \".md\") {\n\t\t\t\tpage, err := Read([]string{f.Name()[:len(f.Name())-3]})\n\t\t\t\tif err != nil {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tp.Tips = append(p.Tips, page.Tips...)\n\t\t\t}\n\t\t}\n\t}\n\treturn p, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Clone wraps html/template.Clone to also clone the name
func (t *Template) Clone() (*Template, error) { var tmpl, err = t.Template.Clone() return &Template{tmpl, t.Name}, err }
[ "func (t *Template) Clone() (*Template, error) {\n\tt.nameSpace.mu.Lock()\n\tdefer t.nameSpace.mu.Unlock()\n\tif t.escaped {\n\t\treturn nil, fmt.Errorf(\"html/template: cannot Clone %q after it has executed\", t.Name())\n\t}\n\ttextClone, err := t.text.Clone()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret := &Template{\n\t\tfalse,\n\t\ttextClone,\n\t\ttextClone.Tree,\n\t\t&nameSpace{\n\t\t\tset: make(map[string]*Template),\n\t\t},\n\t}\n\tfor _, x := range textClone.Templates() {\n\t\tname := x.Name()\n\t\tsrc := t.set[name]\n\t\tif src == nil || src.escaped {\n\t\t\treturn nil, fmt.Errorf(\"html/template: cannot Clone %q after it has executed\", t.Name())\n\t\t}\n\t\tx.Tree = x.Tree.Copy()\n\t\tret.set[name] = &Template{\n\t\t\tfalse,\n\t\t\tx,\n\t\t\tx.Tree,\n\t\t\tret.nameSpace,\n\t\t}\n\t}\n\treturn ret, nil\n}", "func (tc *STemplateController) Clone(clone_name string, recursive bool) (*srv_tmpl.ServiceTemplate, error) {\n\turl := urlTemplateAction(tc.ID)\n\taction := make(map[string]interface{})\n\n\taction[\"action\"] = map[string]interface{}{\n\t\t\"perform\": \"clone\",\n\t\t\"params\": map[string]interface{}{\n\t\t\t\"name\": clone_name,\n\t\t\t\"recursive\": recursive,\n\t\t},\n\t}\n\n\t//Get response\n\tresponse, err := tc.c.ClientFlow.HTTPMethod(\"POST\", url, action)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !response.status {\n\t\treturn nil, errors.New(response.body)\n\t}\n\n\t//Build Service from response\n\tstemplate := &srv_tmpl.ServiceTemplate{}\n\tstemplate_str, err := json.Marshal(response.BodyMap()[\"DOCUMENT\"])\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = json.Unmarshal(stemplate_str, stemplate)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn stemplate, nil\n}", "func (b *Buildtemplate) Clone(source buildv1alpha1.BuildTemplate, clientset *client.ConfigSet) (*buildv1alpha1.BuildTemplate, error) {\n\tsource.SetName(\"\")\n\tsource.SetGenerateName(b.Name + \"-\")\n\tsource.SetNamespace(b.Namespace)\n\tsource.SetOwnerReferences([]metav1.OwnerReference{})\n\tsource.SetResourceVersion(\"\")\n\tsource.Kind = \"BuildTemplate\"\n\tif len(clientset.Registry.Secret) != 0 {\n\t\taddSecretVolume(clientset.Registry.Secret, &source)\n\t\tsetEnvConfig(clientset.Registry.Secret, &source)\n\t}\n\treturn createBuildTemplate(source, clientset)\n}", "func execmTemplateClone(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret, ret1 := args[0].(*template.Template).Clone()\n\tp.Ret(1, ret, ret1)\n}", "func (t *TRoot) Clone() *TRoot {\n\tvar clone, _ = t.template.Clone()\n\treturn &TRoot{clone, t.Path}\n}", "func (nameObject Name) Clone() Object {\n\treturn nameObject\n}", "func (s *HTTPServer) Clone() model.Part { s0 := *s; return &s0 }", "func (i *IContainer) Clone(w http.ResponseWriter, r *http.Request) *IClone {\n\treturn &IClone{\n\t\tIContainer: i,\n\t\tw: w,\n\t\tr: r,\n\t\tmutex: &sync.RWMutex{},\n\t\tthreadData: make(map[string]interface{}),\n\t}\n}", "func Clone(n *html.Node) *html.Node {\n\tnew := &html.Node{\n\t\tType: n.Type,\n\t\tDataAtom: n.DataAtom,\n\t\tData: n.Data,\n\t\tNamespace: n.Namespace,\n\t\tAttr: slices.Clone(n.Attr),\n\t}\n\tfor c := n.FirstChild; c != nil; c = c.NextSibling {\n\t\tc2 := Clone(c)\n\t\tnew.AppendChild(c2)\n\t}\n\treturn new\n}", "func (z *zfsctl) Clone(ctx context.Context, name string, properties map[string]string, source string) *execute {\n\targs := []string{\"clone\", \"-p\"}\n\tif properties != nil {\n\t\tkv := \"-o \"\n\t\tfor k, v := range properties {\n\t\t\tkv += fmt.Sprintf(\"%s=%s \", k, v)\n\t\t}\n\t\targs = append(args, kv)\n\t}\n\targs = append(args, source, name)\n\treturn &execute{ctx: ctx, name: z.cmd, args: args}\n}", "func CloneHandler(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tpaste := vars[\"pasteId\"]\n\n\tp := getPaste(paste)\n\n\tloggy(p.Paste)\n\n\t// Clone page struct\n\tpage := &Page{\n\t\tBody: template.HTML(p.Paste),\n\t\tPasteTitle: \"Copy of \" + p.Title,\n\t\tTitle: \"Copy of \" + p.Title,\n\t\tUserKey: getUserKey(r),\n\t}\n\n\terr := templates.ExecuteTemplate(w, \"index.html\", page)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t}\n}", "func TemplateNodeCopy(t *parse.TemplateNode,) parse.Node", "func (op RollupOp) Clone() RollupOp {\n\tnewName := make([]byte, len(op.newName))\n\tcopy(newName, op.newName)\n\treturn RollupOp{\n\t\tType: op.Type,\n\t\tTags: xbytes.ArrayCopy(op.Tags),\n\t\tAggregationID: op.AggregationID,\n\t\tnewName: newName,\n\t\tnewNameTemplated: op.newNameTemplated,\n\t}\n}", "func (llrb *LLRB) Clone(name string) *LLRB {\n\tif !llrb.lock() {\n\t\treturn nil\n\t}\n\n\tnewllrb := NewLLRB(llrb.name, llrb.setts)\n\tnewllrb.llrbstats = llrb.llrbstats\n\tnewllrb.h_upsertdepth = llrb.h_upsertdepth.Clone()\n\tnewllrb.seqno = llrb.seqno\n\n\tnewllrb.setroot(newllrb.clonetree(llrb.getroot()))\n\n\tllrb.unlock()\n\treturn newllrb\n}", "func Clone(url string) {\n\thg(\"clone %s\", url)\n}", "func (this *Selection) AppendClones(template *html.Node) *Selection {\n\tfor _, parent := range this.Nodes {\n\t\tparent.AppendChild(cloneNode(template))\n\t}\n\treturn this\n}", "func Clone(l *logrus.Logger, u *url.URL, prefix string, dest string) error {\n\tl.Infof(\"Cloning... %s into %s\", u.String(), dest)\n\tf := memfs.New()\n\tc := git.CloneOptions{\n\t\tURL: u.String(),\n\t\tReferenceName: plumbing.ReferenceName(\"refs/heads/master\"),\n\t\tDepth: 1,\n\t}\n\t_, err := git.Clone(memory.NewStorage(), f, &c)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcp := Copier{\n\t\tdest: dest,\n\t\tprefix: prefix,\n\t\tlogger: l,\n\t}\n\treturn cp.WriteResource(f)\n}", "func (r *InstancesService) Clone(project string, instancesclonerequest *InstancesCloneRequest) *InstancesCloneCall {\n\tc := &InstancesCloneCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.project = project\n\tc.instancesclonerequest = instancesclonerequest\n\treturn c\n}", "func CloneTemplates() {\n\t_, err := git.PlainClone(\".templates\", false, &git.CloneOptions{\n\t\tURL: \"http://10.1.38.31/afougerouse/templates.git\",\n\t\tProgress: os.Stdout,\n\t})\n\tif err != nil {\n\t\tfmt.Errorf(\"Impossible de récupérer les templates\")\n\t\tos.Exit(1)\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Root creates a new TRoot for use in spawning templates. The name should match the main layout's name (as defined in the layout template) so execution of templates doesn't require a template.Lookup call, which can be somewhat error prone.
func Root(name, path string) *TRoot { var tmpl = &Template{template.New(name), name} var t = &TRoot{tmpl, path} return t }
[ "func NewRootTemplate() *template.Template {\n\t// Create template\n\tvar tmpl = template.New(\"root\")\n\n\t// Make sure template execution fails if a key is missing\n\ttmpl = tmpl.Option(\"missingkey=error\")\n\n\t// Add sprig functions\n\ttmpl = tmpl.Funcs(sprig.TxtFuncMap())\n\n\t// Add global functions\n\ttmpl = tmpl.Funcs(GetGlobalFuncMap())\n\n\t// Add placeholders for package-specific functions\n\ttmpl = tmpl.Funcs(GetPackageFuncMap(nil))\n\n\treturn tmpl\n}", "func (*Root) Name() (name string) { return \"/\" }", "func (r *Root) Name() string { return \"/\" }", "func InitRootTemplate(templatePath string) {\n\tvar templateFunctions = tmpl.FuncMap{\n\t\t\"IncludeCSS\": webutil.IncludeCSS,\n\t\t\"RawCSS\": webutil.RawCSS,\n\t\t\"IncludeJS\": webutil.IncludeJS,\n\t\t\"RawJS\": webutil.RawJS,\n\t\t\"ImageURL\": webutil.ImageURL,\n\t\t\"HomePath\": webutil.HomePath,\n\t\t\"FullPath\": webutil.FullPath,\n\t\t\"ProdURL\": func() string { return webutil.ProductionURL },\n\t\t\"Comment\": HTMLComment,\n\t\t\"TimeString\": func(t time.Time) string { return t.Format(\"2006-01-02 15:04\") },\n\t\t\"dtstr\": func(t time.Time) string { return t.Format(\"on Jan 2, 2006 at 3:04pm\") },\n\t\t\"actionVerb\": actionVerb,\n\t\t\"nl2br\": func(s string) template.HTML {\n\t\t\tvar escaped = template.HTMLEscaper(s)\n\t\t\tvar replaced = strings.Replace(escaped, \"\\n\", \"<br />\", -1)\n\t\t\treturn template.HTML(replaced)\n\t\t},\n\t\t\"ErrorHTML\": errorHTML,\n\t\t\"ErrorListHTML\": errorListHTML,\n\t\t\"IIIFInfoURL\": webutil.IIIFInfoURL,\n\t\t\"raw\": func(s string) template.HTML { return template.HTML(s) },\n\t\t\"debug\": func() bool { return settings.DEBUG },\n\t\t\"dict\": dict,\n\t\t\"option\": option,\n\t\t\"log\": func(val any) string { logger.Debugf(\"%#v\", val); return \"\" },\n\n\t\t// This hack helps with dynamic heading - Go's templating system seems to\n\t\t// be confused when we have something like \"<{{.Something}}>\" - it decides\n\t\t// the brackets, despite not being in a variable, need to be escaped.\n\t\t\"Open\": func(s string) template.HTML { return template.HTML(\"<\" + s + \">\") },\n\t\t\"Close\": func(s string) template.HTML { return template.HTML(\"</\" + s + \">\") },\n\n\t\t// We have functions for our privileges since they need to be \"global\" and\n\t\t// easily verified at template compile time\n\t\t\"ListTitles\": func() *privilege.Privilege { return privilege.ListTitles },\n\t\t\"ModifyTitles\": func() *privilege.Privilege { return privilege.ModifyTitles },\n\t\t\"ManageMOCs\": func() *privilege.Privilege { return privilege.ManageMOCs },\n\t\t\"ViewMetadataWorkflow\": func() *privilege.Privilege { return privilege.ViewMetadataWorkflow },\n\t\t\"EnterIssueMetadata\": func() *privilege.Privilege { return privilege.EnterIssueMetadata },\n\t\t\"ReviewIssueMetadata\": func() *privilege.Privilege { return privilege.ReviewIssueMetadata },\n\t\t\"ReviewOwnMetadata\": func() *privilege.Privilege { return privilege.ReviewOwnMetadata },\n\t\t\"ReviewUnfixableIssues\": func() *privilege.Privilege { return privilege.ReviewUnfixableIssues },\n\t\t\"ListUsers\": func() *privilege.Privilege { return privilege.ListUsers },\n\t\t\"ModifyUsers\": func() *privilege.Privilege { return privilege.ModifyUsers },\n\t\t\"ViewUploadedIssues\": func() *privilege.Privilege { return privilege.ViewUploadedIssues },\n\t\t\"ModifyUploadedIssues\": func() *privilege.Privilege { return privilege.ModifyUploadedIssues },\n\t\t\"SearchIssues\": func() *privilege.Privilege { return privilege.SearchIssues },\n\t\t\"ViewBatchStatus\": func() *privilege.Privilege { return privilege.ViewBatchStatus },\n\t\t\"ViewQCReadyBatches\": func() *privilege.Privilege { return privilege.ViewQCReadyBatches },\n\t\t\"ApproveQCReadyBatches\": func() *privilege.Privilege { return privilege.ApproveQCReadyBatches },\n\t\t\"RejectQCReadyBatches\": func() *privilege.Privilege { return privilege.RejectQCReadyBatches },\n\t\t\"ViewQCPassedBatches\": func() *privilege.Privilege { return privilege.ViewQCPassedBatches },\n\t\t\"LoadBatches\": func() *privilege.Privilege { return privilege.LoadBatches },\n\t\t\"PurgeBatches\": func() *privilege.Privilege { return privilege.PurgeBatches },\n\t\t\"ArchiveBatches\": func() *privilege.Privilege { return privilege.ArchiveBatches },\n\t\t\"ModifyValidatedLCCNs\": func() *privilege.Privilege { return privilege.ModifyValidatedLCCNs },\n\t\t\"ListAuditLogs\": func() *privilege.Privilege { return privilege.ListAuditLogs },\n\t}\n\n\t// Set up the layout and then our global templates\n\tLayout = tmpl.Root(\"layout\", templatePath)\n\tLayout.Funcs(templateFunctions)\n\tLayout.MustReadPartials(\"layout.go.html\")\n\tInsufficientPrivileges = Layout.MustBuild(\"insufficient-privileges.go.html\")\n\tEmpty = Layout.MustBuild(\"empty.go.html\")\n\tHome = Layout.MustBuild(\"home.go.html\")\n}", "func Root(name string) *Manager {\n\tmanager, err := ResourceManager.Root(name)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn manager\n}", "func NewDefaultRoot() *Root {\n\treturn &Root{\n\t\tCoordinator: *NewDefaultCoordinator(),\n\t\tQuery: *NewDefaultQuery(),\n\t\tHTTP: HTTP{\n\t\t\tPort: 3000,\n\t\t\tIdleTimeout: ltoml.Duration(time.Minute * 2),\n\t\t\tReadTimeout: ltoml.Duration(time.Second * 5),\n\t\t\tWriteTimeout: ltoml.Duration(time.Second * 5),\n\t\t},\n\t\tMonitor: *NewDefaultMonitor(),\n\t\tLogging: *NewDefaultLogging(),\n\t}\n}", "func NewRoot(name string, invalidState int) *DSSRoot {\n\troot := &DSSRoot{Name: name}\n\troot.bottom = newDSSNode(invalidState, &pseudo{\"bottom\"})\n\troot.bottom.pathcnt = 1\n\troot.stacks = make([]*Stack, 0, 10)\n\troot.reservoir = ssl.New()\n\treturn root\n}", "func RootRouter(responseWriter http.ResponseWriter, request *http.Request) {\n\tTemplateInput := getTemplateInputFromRequest(responseWriter, request)\n\treplyWithTemplate(\"indextemplate.html\", TemplateInput, responseWriter, request)\n}", "func (s Manifest) NewRoot() (protocol.ContentId, error) {\n\tss, err := protocol.NewContentId(capnp.Struct(s).Segment())\n\tif err != nil {\n\t\treturn protocol.ContentId{}, err\n\t}\n\terr = capnp.Struct(s).SetPtr(1, capnp.Struct(ss).ToPtr())\n\treturn ss, err\n}", "func (r *Root) Root() (fs.Node, error) {\n\treturn newDir(nil, r.registry), nil\n}", "func NewRoot(tx *bolt.Tx) *Root {\n\treturn &Root{tx}\n}", "func (tree *DNFTree) CreateRoot(phi br.ClauseSet, isFinal bool) int {\n\treturn tree.CreateNodeEntry(phi, 0, isFinal)\n}", "func rootHandler(w http.ResponseWriter, r *http.Request, title string) {\n\tp, err := loadRoot(title)\n\n\tp.Body = template.HTML(blackfriday.MarkdownCommon([]byte(p.Body)))\n\tp.Body = template.HTML(convertWikiMarkup([]byte(p.Body)))\n\n\terr = templates.ExecuteTemplate(w, \"root.html\", p)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t}\n}", "func New(name string) *Template {\n\ttmpl := &Template{\n\t\tfalse,\n\t\ttemplate.New(name),\n\t\tnil,\n\t\t&nameSpace{\n\t\t\tset: make(map[string]*Template),\n\t\t},\n\t}\n\ttmpl.set[name] = tmpl\n\treturn tmpl\n}", "func (ap * SystemVariables ) GetRootname() string {\n\tap.rwlock.RLock()\n\tdefer ap.rwlock.RUnlock()\n\treturn ap.rootname\n}", "func NewRoot() *Root {\n\tvar r Root\n\tr.d = make(types.Dict)\n\treturn &r\n}", "func NewRoot(repo restic.Repository, cfg Config) *Root {\n\tdebug.Log(\"NewRoot(), config %v\", cfg)\n\n\troot := &Root{\n\t\trepo: repo,\n\t\tcfg: cfg,\n\t\tblobCache: bloblru.New(blobCacheSize),\n\t}\n\n\tif !cfg.OwnerIsRoot {\n\t\troot.uid = uint32(os.Getuid())\n\t\troot.gid = uint32(os.Getgid())\n\t}\n\n\t// set defaults, if PathTemplates is not set\n\tif len(cfg.PathTemplates) == 0 {\n\t\tcfg.PathTemplates = []string{\n\t\t\t\"ids/%i\",\n\t\t\t\"snapshots/%T\",\n\t\t\t\"hosts/%h/%T\",\n\t\t\t\"tags/%t/%T\",\n\t\t}\n\t}\n\n\troot.SnapshotsDir = NewSnapshotsDir(root, rootInode, rootInode, NewSnapshotsDirStructure(root, cfg.PathTemplates, cfg.TimeTemplate), \"\")\n\n\treturn root\n}", "func NewRoot() *Root {\n\treturn ExtendRoot(nil)\n}", "func (t *TRoot) Template() *Template {\n\treturn t.Clone().template\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Funcs allows adding template function maps to TRoots; this should be done before creating any templates, or else previously created templates won't get the newest function maps
func (t *TRoot) Funcs(fnList FuncMap) *TRoot { t.template.Funcs(template.FuncMap(fnList)) return t }
[ "func (tmpl Template) Funcs(funcMap template.FuncMap) Template {\n\ttmpl.funcMap = funcMap\n\treturn tmpl\n}", "func (t *Tmpl) Funcs(funcMap template.FuncMap) {\n\tt.Template = t.Template.Funcs(funcMap)\n\tt.funcs = funcMap\n}", "func (docTemplate *DocTemplate) AddFunctions(funcMap template.FuncMap) {\n\tdocTemplate.Template = docTemplate.Template.Funcs(funcMap)\n}", "func (app *App) TemplateFuncs(funcs ...template.FuncMap) *App {\n\tapp.templateFuncs = append(app.templateFuncs, funcs...)\n\treturn app\n}", "func FuncMap() template.FuncMap {\n\tf := sprig.TxtFuncMap()\n\tdelete(f, \"env\")\n\tdelete(f, \"expandenv\")\n\n\t// Add some extra functionality\n\textra := template.FuncMap{\n\t\t\"toToml\": chartutil.ToToml,\n\t\t\"toYaml\": chartutil.ToYaml,\n\t\t\"fromYaml\": chartutil.FromYaml,\n\t\t\"toJson\": chartutil.ToJson,\n\t\t\"fromJson\": chartutil.FromJson,\n\n\t\t// This is a placeholder for the \"include\" function, which is\n\t\t// late-bound to a template. By declaring it here, we preserve the\n\t\t// integrity of the linter.\n\t\t\"include\": func(string, interface{}) string { return \"not implemented\" },\n\t\t\"required\": func(string, interface{}) interface{} { return \"not implemented\" },\n\t\t\"tpl\": func(string, interface{}) interface{} { return \"not implemented\" },\n\t}\n\n\tfor k, v := range extra {\n\t\tf[k] = v\n\t}\n\n\treturn f\n}", "func NewFuncMap() template.FuncMap {\n\treturn map[string]any{\n\t\t\"ctx\": func() any { return nil }, // template context function\n\n\t\t\"DumpVar\": dumpVar,\n\n\t\t// -----------------------------------------------------------------\n\t\t// html/template related functions\n\t\t\"dict\": dict, // it's lowercase because this name has been widely used. Our other functions should have uppercase names.\n\t\t\"Eval\": Eval,\n\t\t\"Safe\": Safe,\n\t\t\"Escape\": html.EscapeString,\n\t\t\"QueryEscape\": url.QueryEscape,\n\t\t\"JSEscape\": template.JSEscapeString,\n\t\t\"Str2html\": Str2html, // TODO: rename it to SanitizeHTML\n\t\t\"URLJoin\": util.URLJoin,\n\t\t\"DotEscape\": DotEscape,\n\n\t\t\"PathEscape\": url.PathEscape,\n\t\t\"PathEscapeSegments\": util.PathEscapeSegments,\n\n\t\t// utils\n\t\t\"StringUtils\": NewStringUtils,\n\t\t\"SliceUtils\": NewSliceUtils,\n\t\t\"JsonUtils\": NewJsonUtils,\n\n\t\t// -----------------------------------------------------------------\n\t\t// svg / avatar / icon\n\t\t\"svg\": svg.RenderHTML,\n\t\t\"EntryIcon\": base.EntryIcon,\n\t\t\"MigrationIcon\": MigrationIcon,\n\t\t\"ActionIcon\": ActionIcon,\n\n\t\t\"SortArrow\": SortArrow,\n\n\t\t// -----------------------------------------------------------------\n\t\t// time / number / format\n\t\t\"FileSize\": base.FileSize,\n\t\t\"CountFmt\": base.FormatNumberSI,\n\t\t\"TimeSince\": timeutil.TimeSince,\n\t\t\"TimeSinceUnix\": timeutil.TimeSinceUnix,\n\t\t\"DateTime\": timeutil.DateTime,\n\t\t\"Sec2Time\": util.SecToTime,\n\t\t\"LoadTimes\": func(startTime time.Time) string {\n\t\t\treturn fmt.Sprint(time.Since(startTime).Nanoseconds()/1e6) + \"ms\"\n\t\t},\n\n\t\t// -----------------------------------------------------------------\n\t\t// setting\n\t\t\"AppName\": func() string {\n\t\t\treturn setting.AppName\n\t\t},\n\t\t\"AppSubUrl\": func() string {\n\t\t\treturn setting.AppSubURL\n\t\t},\n\t\t\"AssetUrlPrefix\": func() string {\n\t\t\treturn setting.StaticURLPrefix + \"/assets\"\n\t\t},\n\t\t\"AppUrl\": func() string {\n\t\t\t// The usage of AppUrl should be avoided as much as possible,\n\t\t\t// because the AppURL(ROOT_URL) may not match user's visiting site and the ROOT_URL in app.ini may be incorrect.\n\t\t\t// And it's difficult for Gitea to guess absolute URL correctly with zero configuration,\n\t\t\t// because Gitea doesn't know whether the scheme is HTTP or HTTPS unless the reverse proxy could tell Gitea.\n\t\t\treturn setting.AppURL\n\t\t},\n\t\t\"AppVer\": func() string {\n\t\t\treturn setting.AppVer\n\t\t},\n\t\t\"AppDomain\": func() string { // documented in mail-templates.md\n\t\t\treturn setting.Domain\n\t\t},\n\t\t\"AssetVersion\": func() string {\n\t\t\treturn setting.AssetVersion\n\t\t},\n\t\t\"DefaultShowFullName\": func() bool {\n\t\t\treturn setting.UI.DefaultShowFullName\n\t\t},\n\t\t\"ShowFooterTemplateLoadTime\": func() bool {\n\t\t\treturn setting.Other.ShowFooterTemplateLoadTime\n\t\t},\n\t\t\"AllowedReactions\": func() []string {\n\t\t\treturn setting.UI.Reactions\n\t\t},\n\t\t\"CustomEmojis\": func() map[string]string {\n\t\t\treturn setting.UI.CustomEmojisMap\n\t\t},\n\t\t\"MetaAuthor\": func() string {\n\t\t\treturn setting.UI.Meta.Author\n\t\t},\n\t\t\"MetaDescription\": func() string {\n\t\t\treturn setting.UI.Meta.Description\n\t\t},\n\t\t\"MetaKeywords\": func() string {\n\t\t\treturn setting.UI.Meta.Keywords\n\t\t},\n\t\t\"EnableTimetracking\": func() bool {\n\t\t\treturn setting.Service.EnableTimetracking\n\t\t},\n\t\t\"DisableGitHooks\": func() bool {\n\t\t\treturn setting.DisableGitHooks\n\t\t},\n\t\t\"DisableWebhooks\": func() bool {\n\t\t\treturn setting.DisableWebhooks\n\t\t},\n\t\t\"DisableImportLocal\": func() bool {\n\t\t\treturn !setting.ImportLocalPaths\n\t\t},\n\t\t\"DefaultTheme\": func() string {\n\t\t\treturn setting.UI.DefaultTheme\n\t\t},\n\t\t\"NotificationSettings\": func() map[string]any {\n\t\t\treturn map[string]any{\n\t\t\t\t\"MinTimeout\": int(setting.UI.Notification.MinTimeout / time.Millisecond),\n\t\t\t\t\"TimeoutStep\": int(setting.UI.Notification.TimeoutStep / time.Millisecond),\n\t\t\t\t\"MaxTimeout\": int(setting.UI.Notification.MaxTimeout / time.Millisecond),\n\t\t\t\t\"EventSourceUpdateTime\": int(setting.UI.Notification.EventSourceUpdateTime / time.Millisecond),\n\t\t\t}\n\t\t},\n\t\t\"MermaidMaxSourceCharacters\": func() int {\n\t\t\treturn setting.MermaidMaxSourceCharacters\n\t\t},\n\n\t\t// -----------------------------------------------------------------\n\t\t// render\n\t\t\"RenderCommitMessage\": RenderCommitMessage,\n\t\t\"RenderCommitMessageLinkSubject\": RenderCommitMessageLinkSubject,\n\n\t\t\"RenderCommitBody\": RenderCommitBody,\n\t\t\"RenderCodeBlock\": RenderCodeBlock,\n\t\t\"RenderIssueTitle\": RenderIssueTitle,\n\t\t\"RenderEmoji\": RenderEmoji,\n\t\t\"RenderEmojiPlain\": emoji.ReplaceAliases,\n\t\t\"ReactionToEmoji\": ReactionToEmoji,\n\t\t\"RenderNote\": RenderNote,\n\n\t\t\"RenderMarkdownToHtml\": RenderMarkdownToHtml,\n\t\t\"RenderLabel\": RenderLabel,\n\t\t\"RenderLabels\": RenderLabels,\n\n\t\t// -----------------------------------------------------------------\n\t\t// misc\n\t\t\"ShortSha\": base.ShortSha,\n\t\t\"ActionContent2Commits\": ActionContent2Commits,\n\t\t\"IsMultilineCommitMessage\": IsMultilineCommitMessage,\n\t\t\"CommentMustAsDiff\": gitdiff.CommentMustAsDiff,\n\t\t\"MirrorRemoteAddress\": mirrorRemoteAddress,\n\n\t\t\"FilenameIsImage\": FilenameIsImage,\n\t\t\"TabSizeClass\": TabSizeClass,\n\t}\n}", "func (v *View) AddFuncMap(funcMap template.FuncMap) {\n v.funcMap = funcMap\n}", "func AddFuncMap(key string, fn interface{}) error {\n\tbeegoTplFuncMap[key] = fn\n\treturn nil\n}", "func (t *template) addTemplateFunctions() map[string]interface{} {\n\treturn map[string]interface{}{\n\t\t\"Random\": func(values ...string) string {\n\t\t\treturn values[rand.Intn(len(values))]\n\t\t},\n\t\t\"Username\": func(user map[string]string) string {\n\t\t\tif _, ok := user[\"name\"]; !ok {\n\t\t\t\treturn \"<no value>\"\n\t\t\t}\n\n\t\t\treturn user[\"name\"]\n\t\t},\n\t\t\"Subtract\": func(a, b int) int {\n\t\t\treturn a - b\n\t\t},\n\t}\n}", "func AddFuncMap(key string, fn interface{}) error {\n\ttplFuncMap[key] = fn\n\treturn nil\n}", "func TmplFunctionsMap() template.FuncMap {\n\tfuncMap := template.FuncMap{\n\t\t\"envOrDef\": envOrDefault,\n\t\t\"env\": env,\n\t\t\"fileMD5\": fileMD5,\n\t\t\"Iterate\": Iterate,\n\t}\n\treturn funcMap\n}", "func WithFuncs(funcs gotemplate.FuncMap) Opt {\n\treturn func(t *gotemplate.Template) (*gotemplate.Template, error) {\n\t\treturn t.Funcs(funcs), nil\n\t}\n}", "func AddTemplateFuncsNamespace(ns func(d *deps.Deps) *TemplateFuncsNamespace) {\n\tTemplateFuncsNamespaceRegistry = append(TemplateFuncsNamespaceRegistry, ns)\n}", "func (e *Engine) alterFuncMap(t *template.Template, referenceTpls map[string]renderable) template.FuncMap {\n\t// Clone the func map because we are adding context-specific functions.\n\tvar funcMap template.FuncMap = map[string]interface{}{}\n\tfor k, v := range e.FuncMap {\n\t\tfuncMap[k] = v\n\t}\n\n\tincludedNames := make(map[string]int)\n\n\t// Add the 'include' function here so we can close over t.\n\tfuncMap[\"include\"] = func(name string, data interface{}) (string, error) {\n\t\tbuf := bytes.NewBuffer(nil)\n\t\tif v, ok := includedNames[name]; ok {\n\t\t\tif v > recursionMaxNums {\n\t\t\t\treturn \"\", errors.Wrapf(fmt.Errorf(\"unable to execute template\"), \"rendering template has a nested reference name: %s\", name)\n\t\t\t}\n\t\t\tincludedNames[name]++\n\t\t} else {\n\t\t\tincludedNames[name] = 1\n\t\t}\n\t\tif err := t.ExecuteTemplate(buf, name, data); err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tincludedNames[name]--\n\t\treturn buf.String(), nil\n\t}\n\n\t// Add the 'required' function here\n\tfuncMap[\"required\"] = func(warn string, val interface{}) (interface{}, error) {\n\t\tif val == nil {\n\t\t\tif e.LintMode {\n\t\t\t\t// Don't fail on missing required values when linting\n\t\t\t\tlog.Printf(\"[INFO] Missing required value: %s\", warn)\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\t// Convert nil to \"\" in case required is piped into other functions\n\t\t\treturn \"\", fmt.Errorf(warn)\n\t\t} else if _, ok := val.(string); ok {\n\t\t\tif val == \"\" {\n\t\t\t\tif e.LintMode {\n\t\t\t\t\t// Don't fail on missing required values when linting\n\t\t\t\t\tlog.Printf(\"[INFO] Missing required value: %s\", warn)\n\t\t\t\t\treturn val, nil\n\t\t\t\t}\n\t\t\t\treturn val, fmt.Errorf(warn)\n\t\t\t}\n\t\t}\n\t\treturn val, nil\n\t}\n\n\t// Add the 'tpl' function here\n\tfuncMap[\"tpl\"] = func(tpl string, vals chartutil.Values) (string, error) {\n\t\tbasePath, err := vals.PathValue(\"Template.BasePath\")\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"Cannot retrieve Template.Basepath from values inside tpl function: %s (%s)\", tpl, err.Error())\n\t\t}\n\n\t\tr := renderable{\n\t\t\ttpl: tpl,\n\t\t\tvals: vals,\n\t\t\tbasePath: basePath.(string),\n\t\t}\n\n\t\ttemplates := map[string]renderable{}\n\t\ttemplateName, err := vals.PathValue(\"Template.Name\")\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"Cannot retrieve Template.Name from values inside tpl function: %s (%s)\", tpl, err.Error())\n\t\t}\n\n\t\ttemplates[templateName.(string)] = r\n\n\t\tresult, err := e.renderWithReferences(templates, referenceTpls)\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"Error during tpl function execution for %q: %s\", tpl, err.Error())\n\t\t}\n\t\treturn result[templateName.(string)], nil\n\t}\n\n\treturn funcMap\n}", "func funcMap() template.FuncMap {\n\tfmap := sprig.TxtFuncMap()\n\n\tconfig := kubeConfig()\n\n\tfmap[\"lookup\"] = NewLookupFunction(config)\n\n\treturn fmap\n}", "func createFuncMap() template.FuncMap {\n\tfuncs := sprig.TxtFuncMap()\n\tfuncs[\"Registry\"] = func(registry string) string {\n\t\treturn registry\n\t}\n\treturn funcs\n}", "func (t *Templates) AddTemplateFunc(functionName string, function interface{}) {\n\tt.funcMap[functionName] = function\n}", "func (f *BaseFormatter) AddFuncMap(funcMap template.FuncMap) {\n\tfor k, v := range funcMap {\n\t\tf.funcMap[k] = v\n\t}\n}", "func MakeTemplateFuncs(fo FuncOptions, superUsers []string) (map[string]interface{}, error) {\n\tr := map[string]interface{}{\n\t\t// IsSuperUser returns true if the given user Id has super user privileges.\n\t\t\"IsSuperUser\": func(userName string) bool {\n\t\t\treturn len(superUsers) == 0 || util.StringSliceContains(superUsers, userName)\n\t\t},\n\t\t// Gravatar returns a Gravatar URL for the given email string.\n\t\t\"Gravatar\": func(email string) string {\n\t\t\th := md5.New()\n\t\t\t_, err := io.WriteString(h, email)\n\t\t\tgrip.Warning(err)\n\t\t\treturn fmt.Sprintf(\"http://www.gravatar.com/avatar/%x?s=50\", h.Sum(nil))\n\t\t},\n\n\t\t// DateFormat returns a time Formatted to the given layout and timezone.\n\t\t// If the timezone is unset, it defaults to \"New_York.\"\n\t\t\"DateFormat\": func(when time.Time, layout string, timezone string) string {\n\t\t\tif len(timezone) == 0 {\n\t\t\t\ttimezone = \"America/New_York\" // I ♥ NY\n\t\t\t}\n\t\t\tloc, err := time.LoadLocation(timezone)\n\t\t\tif err != nil {\n\t\t\t\treturn when.Format(layout)\n\t\t\t}\n\n\t\t\twhenTZ := when.In(loc)\n\t\t\treturn whenTZ.Format(layout)\n\t\t},\n\n\t\t// Static returns a link to a static file.\n\t\t\"Static\": func(filetype, filename string) string {\n\t\t\treturn fmt.Sprintf(\"/static/%s/%s\", filetype, filename)\n\t\t},\n\n\t\t// RemoveANSI strips out ANSI color sequences in cases where it doesn't make sense to include\n\t\t// them, e.g. raw task logs\n\t\t\"RemoveANSI\": func(line string) string {\n\t\t\tre, err := regexp.Compile(\"\\x1B\\\\[([0-9]{1,2}(;[0-9]{1,2})?)?[m|K]\")\n\t\t\tif err != nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn re.ReplaceAllString(line, \"\")\n\t\t},\n\n\t\t// Is Prod returns whether or not Evergreen is running in \"production.\"\n\t\t// Currently this is only used to toggle the use of minified css files.\n\t\t\"IsProd\": func() bool {\n\t\t\treturn fo.IsProd\n\t\t},\n\n\t\t// GetTimezone returns the timezone for a user.\n\t\t// Defaults to \"New_York\".\n\t\t\"GetTimezone\": func(u *user.DBUser) string {\n\t\t\tif u != nil && u.Settings.Timezone != \"\" {\n\t\t\t\treturn u.Settings.Timezone\n\t\t\t}\n\t\t\treturn \"America/New_York\"\n\t\t},\n\n\t\t// MutableVar creates an unset MutableVar.\n\t\t\"MutableVar\": func() interface{} {\n\t\t\treturn &MutableVar{\"\"}\n\t\t},\n\n\t\t// Trunc cuts off a string to be n characters long.\n\t\t\"Trunc\": func(s string, n int) string {\n\t\t\tif n > len(s) {\n\t\t\t\treturn s\n\t\t\t}\n\t\t\treturn s[0:n]\n\t\t},\n\n\t\t// UrlFor generates a URL for the given route.\n\t\t\"UrlFor\": func(name string, pairs ...interface{}) (*url.URL, error) {\n\t\t\tsize := len(pairs)\n\t\t\tstrPairs := make([]string, size)\n\t\t\tfor i := 0; i < size; i++ {\n\t\t\t\tif v, ok := pairs[i].(string); ok {\n\t\t\t\t\tstrPairs[i] = v\n\t\t\t\t} else {\n\t\t\t\t\tstrPairs[i] = fmt.Sprint(pairs[i])\n\t\t\t\t}\n\t\t\t}\n\n\t\t\troute := fo.Router.Get(name)\n\t\t\tif route == nil {\n\t\t\t\treturn nil, errors.Errorf(\"UrlFor: can't find a route named %v\", name)\n\t\t\t}\n\n\t\t\treturn route.URL(strPairs...)\n\t\t},\n\n\t\t// HelpUrl returns the address of the Evergreen help page,\n\t\t// if one is set.\n\t\t\"HelpUrl\": func() string {\n\t\t\tif fo.HelpHome != \"\" {\n\t\t\t\treturn fo.HelpHome\n\t\t\t}\n\t\t\treturn defaultHelpURL\n\t\t},\n\t}\n\n\tstaticsMD5, err := DirectoryChecksum(fo.WebHome)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tr[\"StaticsMD5\"] = func() string {\n\t\treturn staticsMD5\n\t}\n\treturn r, nil\n\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Clone creates a copy of the TRoot for ease of creating sublayouts. Since TRoots cannot be executed externally, we don't have the possibility of returning an error.
func (t *TRoot) Clone() *TRoot { var clone, _ = t.template.Clone() return &TRoot{clone, t.Path} }
[ "func (t *Template) Clone() (*Template, error) {\n\tt.nameSpace.mu.Lock()\n\tdefer t.nameSpace.mu.Unlock()\n\tif t.escaped {\n\t\treturn nil, fmt.Errorf(\"html/template: cannot Clone %q after it has executed\", t.Name())\n\t}\n\ttextClone, err := t.text.Clone()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret := &Template{\n\t\tfalse,\n\t\ttextClone,\n\t\ttextClone.Tree,\n\t\t&nameSpace{\n\t\t\tset: make(map[string]*Template),\n\t\t},\n\t}\n\tfor _, x := range textClone.Templates() {\n\t\tname := x.Name()\n\t\tsrc := t.set[name]\n\t\tif src == nil || src.escaped {\n\t\t\treturn nil, fmt.Errorf(\"html/template: cannot Clone %q after it has executed\", t.Name())\n\t\t}\n\t\tx.Tree = x.Tree.Copy()\n\t\tret.set[name] = &Template{\n\t\t\tfalse,\n\t\t\tx,\n\t\t\tx.Tree,\n\t\t\tret.nameSpace,\n\t\t}\n\t}\n\treturn ret, nil\n}", "func (t *Template) Clone() (*Template, error) {\n\tvar tmpl, err = t.Template.Clone()\n\treturn &Template{tmpl, t.Name}, err\n}", "func (t Tree) Clone() Tree {\n\treturn Tree{CloneNode(t.n)}\n}", "func (w *WebGLRenderTarget) Clone() *WebGLRenderTarget {\n\tw.p.Call(\"clone\")\n\treturn w\n}", "func (b *Buildtemplate) Clone(source buildv1alpha1.BuildTemplate, clientset *client.ConfigSet) (*buildv1alpha1.BuildTemplate, error) {\n\tsource.SetName(\"\")\n\tsource.SetGenerateName(b.Name + \"-\")\n\tsource.SetNamespace(b.Namespace)\n\tsource.SetOwnerReferences([]metav1.OwnerReference{})\n\tsource.SetResourceVersion(\"\")\n\tsource.Kind = \"BuildTemplate\"\n\tif len(clientset.Registry.Secret) != 0 {\n\t\taddSecretVolume(clientset.Registry.Secret, &source)\n\t\tsetEnvConfig(clientset.Registry.Secret, &source)\n\t}\n\treturn createBuildTemplate(source, clientset)\n}", "func (t *Tile) Clone() Tile {\n\tn := *t\n\tb := t.Bounds()\n\tn.Image = image.NewRGBA(b)\n\tdraw.Draw(n.Image, b, t, b.Min, draw.Src)\n\treturn n\n}", "func (v *Vbox) Clone() Widget {\n\tc := *v\n\treturn &c\n}", "func (t *Trie) Clone() (*Trie, error) {\n\treturn &Trie{rootHash: t.rootHash, storage: t.storage, needChangelog: t.needChangelog}, nil\n}", "func (t *TaskBox[T, U, C, CT, TF]) Clone() *TaskBox[T, U, C, CT, TF] {\n\tnewBox := NewTaskBox[T, U, C, CT, TF](t.constArgs, t.contextFunc, t.wg, t.task, t.resultCh, t.taskID)\n\treturn &newBox\n}", "func (r *View) Clone() *View {\n\treturn r.CloneLimit(r.size)\n}", "func (t *FaultDomainTree) Copy() *FaultDomainTree {\n\tif t == nil {\n\t\treturn nil\n\t}\n\n\ttCopy := NewFaultDomainTree().\n\t\tWithNodeDomain(t.Domain).\n\t\tWithID(t.ID)\n\tfor _, c := range t.Children {\n\t\ttCopy.Children = append(tCopy.Children, c.Copy())\n\t}\n\n\treturn tCopy\n}", "func (i *IContainer) Clone(w http.ResponseWriter, r *http.Request) *IClone {\n\treturn &IClone{\n\t\tIContainer: i,\n\t\tw: w,\n\t\tr: r,\n\t\tmutex: &sync.RWMutex{},\n\t\tthreadData: make(map[string]interface{}),\n\t}\n}", "func execmTemplateClone(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret, ret1 := args[0].(*template.Template).Clone()\n\tp.Ret(1, ret, ret1)\n}", "func (t *TRoot) Template() *Template {\n\treturn t.Clone().template\n}", "func (cte *CTE) Clone() *CTE {\n\tif cte == nil {\n\t\treturn nil\n\t}\n\tother := *cte\n\tother.TableName = cte.TableName.Clone()\n\tother.Columns = cloneIdents(cte.Columns)\n\tother.Select = cte.Select.Clone()\n\treturn &other\n}", "func (m *Mocker) Clone(t *testing.T) (clone *Mocker) {\n\tm.Close()\n\n\tclone = New(t)\n\n\tclone.handlers = m.deepCopyHandlers()\n\n\treturn\n}", "func (lt *PhysicalTopN) Clone() (PhysicalPlan, error) {\n\tcloned := new(PhysicalTopN)\n\t*cloned = *lt\n\tbase, err := lt.basePhysicalPlan.cloneWithSelf(cloned)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcloned.basePhysicalPlan = *base\n\tcloned.ByItems = make([]*util.ByItems, 0, len(lt.ByItems))\n\tfor _, it := range lt.ByItems {\n\t\tcloned.ByItems = append(cloned.ByItems, it.Clone())\n\t}\n\tcloned.PartitionBy = make([]property.SortItem, 0, len(lt.PartitionBy))\n\tfor _, it := range lt.PartitionBy {\n\t\tcloned.PartitionBy = append(cloned.PartitionBy, it.Clone())\n\t}\n\treturn cloned, nil\n}", "func (tree *Tree) GetCopy() *Tree {\n\tnewTree := &Tree{\n\t\tmaxEntry: tree.maxEntry,\n\t\tminEntry: tree.minEntry,\n\t\tdistCalc: tree.distCalc,\n\t\tObjectCount: tree.ObjectCount,\n\t\tsplitMecha: tree.splitMecha,\n\t}\n\n\tvar newRoot node\n\troot := tree.root\n\tnewEntryList := copyEntryList(tree.root)\n\tif root.isLeaf() {\n\t\tnewRoot = &leaf{\n\t\t\tradius: root.getRadius(),\n\t\t\tcentroidObject: root.getCentroidObject(),\n\t\t\tentryList: newEntryList,\n\t\t}\n\t} else {\n\t\tnewRoot = &branch{\n\t\t\tradius: root.getRadius(),\n\t\t\tcentroidObject: root.getCentroidObject(),\n\t\t\tentryList: newEntryList,\n\t\t}\n\t}\n\tfor idx := range newEntryList {\n\t\tnewEntryList[idx].setParent(newRoot)\n\t}\n\tnewTree.root = newRoot\n\treturn newTree\n}", "func (t Topology) Copy() Topology {\n\treturn Topology{\n\t\tNodes: t.Nodes.Copy(),\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
ReadPartials parses the given files into the TRoot instance for gathering things like the toplevel layout, navigation elements, etc. The list of files is relative to the TRoot's Path. Returns on the first error encountered, if any.
func (t *TRoot) ReadPartials(files ...string) error { for _, file := range files { var _, err = t.template.ParseFiles(filepath.Join(t.Path, file)) if err != nil { return err } } return nil }
[ "func (t *TRoot) MustReadPartials(files ...string) {\n\tvar err = t.ReadPartials(files...)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func PartialLatexFiles(path string) {\n\t//read all files in directory sections\n\t// this must be settleable and discoverable\n\tvar counter int\n\n\ttype Content struct {\n\t\tfileName string\n\t\tContents string\n\t}\n\n\tvar contentsList map[string]string\n\tcontentsList = make(map[string]string)\n\n\tList := new(Content)\n\n\t//append(s []T, x ...T)\n\tfiles, _ := ioutil.ReadDir(\"./sections\")\n\tfor _, f := range files {\n\t\tif f.Name() == \"main.tex\" {\n\t\t\tfmt.Println(\"We found a main file\")\n\t\t}\n\t\tfmt.Println(f.Name())\n\t\tS1, err := ioutil.ReadFile(\"./sections/\" + f.Name())\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\n\t\tList.fileName = f.Name()\n\t\tList.Contents = string(S1)\n\t\tfmt.Println(string(S1))\n\n\t\tcontentsList[f.Name()] = string(S1)\n\t\tcounter++\n\n\t}\n\tfmt.Println(\"TEST\", contentsList[\"main.tex\"])\n\t//inFile, _ := ioutil.ReadFile(path)\n\t//fmt.Println(\"CONCATENATION:\", Y.contents)\n\t//fmt.Printf(\"Found %v files\", counter)\n}", "func loadPartials() (map[string]string, error) {\n\tg := make(map[string]string)\n\t//load resources from paths\n\tfor key, path := range paths {\n\t\tbody, err := ioutil.ReadFile(path)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tg[key] = string(body)\n\t}\n\treturn g, nil\n}", "func Read(files []string) (documents []Document) {\n\n\tfor _, fp := range files {\n\t\tf, err := ioutil.ReadFile(fp)\n\n\t\tif err != nil {\n\t\t\tfmt.Println(\"There was an error reading the file\", err)\n\t\t\tos.Exit(-1)\n\t\t}\n\n\t\tyamlDocumentsInFile := bytes.SplitN(f, []byte(\"---\\n\"), -1)\n\t\t//fmt.Printf(\"%q\\n\", yamlDocumentsInFile)\n\n\t\tif (len(yamlDocumentsInFile) % 2) != 0 {\n\t\t\tfmt.Println(\"File \", fp, \" has an odd number of documents. File must consist of pairs of preamble and template documents, in order.\")\n\t\t\tos.Exit(-1)\n\t\t}\n\n\t\tfor i := 0; i < len(yamlDocumentsInFile); i += 2 {\n\n\t\t\tdoc := Document{}\n\t\t\terr = yaml.Unmarshal(yamlDocumentsInFile[i], &doc.Preamble)\n\t\t\tdoc.Template = string(yamlDocumentsInFile[i+1])\n\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(\"There was an error unmarshaling yaml\", err)\n\t\t\t\tos.Exit(-1)\n\t\t\t}\n\n\t\t\t//fmt.Printf(\"%+v\\n\", doc)\n\n\t\t\t// Perform type conversions to handle lists of maps or single map\n\t\t\tswitch p := doc.Preamble.ReadParams.(type) {\n\t\t\tcase []interface{}:\n\t\t\t\tfor _, params := range p {\n\n\t\t\t\t\t// We cannot derive a map[string]inteface{} from interface{} directly\n\t\t\t\t\tparamsMap, _ := params.(map[interface{}]interface{})\n\n\t\t\t\t\ttParams := typeCastMap(paramsMap)\n\n\t\t\t\t\tdocument := Document{}\n\t\t\t\t\tdocument.Preamble.Params = tParams\n\t\t\t\t\tdocument.Template = doc.Template\n\n\t\t\t\t\tdocuments = append(documents, document)\n\t\t\t\t}\n\t\t\tcase interface{}:\n\t\t\t\t// We cannot derive a map[string]inteface{} from interface{} directly\n\t\t\t\ttParams := p.(map[interface{}]interface{})\n\n\t\t\t\tdoc.Preamble.Params = typeCastMap(tParams)\n\n\t\t\t\tdocuments = append(documents, doc)\n\t\t\tdefault:\n\t\t\t\tfmt.Printf(\"I don't know how to deal with type %T %+v!\\n\", p, p)\n\t\t\t\tos.Exit(-1)\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\treturn\n}", "func ReadPartialReport(scope beam.Scope, partialReportFile string) beam.PCollection {\n\tallFiles := ioutils.AddStrInPath(partialReportFile, \"*\")\n\tlines := textio.ReadSdf(scope, allFiles)\n\treturn beam.ParDo(scope, &parseEncryptedPartialReportFn{}, lines)\n}", "func (g *Group) AddPartialFiles(dir string) error {\n\treturn collectTemplateFiles(dir, g.AddPartialFile)\n}", "func LoadTemplates(relativePath string, pOpt *ParseOptions) {\n\t// Initializes the template map\n\ttemplates = make(map[string]*template.Template)\n\n\t// Save Path to Base file\n\tpOpt.BasePath = relativePath\n\n\t// Check if every option is set\n\tif pOpt.BaseName == \"\" {\n\t\tpOpt.BaseName = DefaultParseOptions.BaseName\n\t}\n\n\tif pOpt.Delimiter == \"\" {\n\t\tpOpt.Delimiter = DefaultParseOptions.Delimiter\n\t}\n\n\tif pOpt.Ext == \"\" {\n\t\tpOpt.Ext = DefaultParseOptions.Ext\n\t}\n\n\tif pOpt.NonBaseFolder == \"\" {\n\t\tpOpt.NonBaseFolder = DefaultParseOptions.NonBaseFolder\n\t}\n\n\t// Start checking the main dir of the views\n\tcheckDir(relativePath, pOpt, false)\n}", "func (s *server) loadTemplates() error {\n includePath := \"templates/\"\n layoutPath := \"templates/layout/\"\n\n if s.templates == nil {\n s.templates = make(map[string]*template.Template)\n }\n\n layoutFiles, err := filepath.Glob(layoutPath + \"*.tmpl\")\n if err != nil {\n log.Println(\"failed to get included templates\")\n return err\n }\n\n includeFiles, err := filepath.Glob(includePath + \"*.tmpl\")\n if err != nil {\n log.Println(\"failed to get layout templates\")\n return err\n }\n\n mainTemplate := template.New(\"main\")\n mainTemplate, err = mainTemplate.Parse(mainTmpl)\n if err != nil {\n log.Println(\"failed to parse main template\")\n return err\n }\n\n for _, file := range includeFiles {\n fileName := filepath.Base(file)\n files := append(layoutFiles, file)\n s.templates[fileName], err = mainTemplate.Clone()\n if err != nil {\n return err\n }\n s.templates[fileName] = template.Must(\n s.templates[fileName].ParseFiles(files...))\n }\n\n s.bufpool = bpool.NewBufferPool(64)\n return nil\n}", "func (conf *Config) ReadSpecs() error {\n\tconf.specs = make([]*tomlSpec, 0, len(conf.specFiles))\n\tconf.indexes = make(map[string]*indexSpec, len(conf.specFiles))\n\tfor _, path := range conf.specFiles {\n\t\tspec, err := ReadSpec(path)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"couldn't read spec '%s': %v\", path, err)\n\t\t}\n\t\t// here is where we put overrides like setting the prefix\n\t\t// from command-line parameters before doing more validation and\n\t\t// populating inferred fields.\n\t\tif spec.Prefix == \"\" {\n\t\t\tspec.Prefix = conf.Prefix\n\t\t}\n\t\terr = spec.CleanupIndexes(conf)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tconf.specs = append(conf.specs, spec)\n\t}\n\tfor _, spec := range conf.specs {\n\t\terr := spec.CleanupWorkloads(conf)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (p *Parse) ParseFiles() (e error) {\n\tvar wg sync.WaitGroup\n\tfor _, fname := range p.Files {\n\t\twg.Add(1)\n\t\tgo func(fname string) {\n\t\t\tdefer wg.Done()\n\t\t\tfset := token.NewFileSet() // positions are relative to fset\n\n\t\t\t// Parse the file given in arguments\n\t\t\tf, err := parser.ParseFile(fset, fname, nil, parser.ParseComments)\n\t\t\tif err != nil {\n\t\t\t\te = err\n\t\t\t\treturn\n\t\t\t}\n\t\t\tbs, err := ioutil.ReadFile(fname)\n\t\t\tif err != nil {\n\t\t\t\te = err\n\t\t\t\treturn\n\t\t\t}\n\t\t\tstructMap, baseMap := p.parseTypes(f)\n\t\t\t// Parse structs\n\t\t\tstructKeys := make([]string, 0, len(structMap))\n\t\t\tfor k := range structMap {\n\t\t\t\tstructKeys = append(structKeys, k)\n\t\t\t}\n\t\t\tsort.Strings(structKeys)\n\t\t\tp.Lock()\n\t\t\tfor _, structName := range structKeys {\n\t\t\t\tp.mappings[structName] = p.parseStruct(structMap[structName], structName, bs)\n\t\t\t}\n\t\t\tp.Unlock()\n\t\t\tbaseKeys := make([]string, 0, len(baseMap))\n\t\t\tfor k := range baseMap {\n\t\t\t\tbaseKeys = append(baseKeys, k)\n\t\t\t}\n\t\t\tsort.Strings(baseKeys)\n\t\t\tp.Lock()\n\t\t\tfor _, baseName := range baseKeys {\n\t\t\t\tp.baseMappings[baseName] = field{\n\t\t\t\t\ttyp: baseMap[baseName],\n\t\t\t\t\tname: baseName,\n\t\t\t\t}\n\t\t\t}\n\t\t\tp.Unlock()\n\t\t}(fname)\n\t}\n\twg.Wait()\n\treturn nil\n}", "func (contrl *MailController) LoadTemplateFiles(filenames ...string) {\n\tcontrl.HTMLTemplate = template.Must(template.ParseFiles(filenames...))\n}", "func ReadAll() (p *Page, err error) {\n\tdCmn := config.SourceDir + sep + \"pages\" + sep + \"common\" + sep\n\tdOs := config.SourceDir + sep + \"pages\" + sep + config.OSName() + sep\n\tpaths := []string{dCmn, dOs}\n\tp = &Page{Name: \"Search All\"}\n\tp.Tips = make([]*Tip, 0)\n\tfor _, pt := range paths {\n\t\tfiles, err := ioutil.ReadDir(pt)\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tfor _, f := range files {\n\t\t\tif strings.HasSuffix(f.Name(), \".md\") {\n\t\t\t\tpage, err := Read([]string{f.Name()[:len(f.Name())-3]})\n\t\t\t\tif err != nil {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tp.Tips = append(p.Tips, page.Tips...)\n\t\t\t}\n\t\t}\n\t}\n\treturn p, nil\n}", "func (r *Renderer) LoadFiles(files ...string) {\n\tif !r.initialized {\n\t\tpanicErr(fmt.Errorf(\"please call Initialize(), before load templates\"))\n\t}\n\n\tfor _, file := range files {\n\t\text := filepath.Ext(file)\n\t\tif r.IsValidExt(ext) {\n\t\t\tname := filepath.ToSlash(file[0 : len(file)-len(ext)])\n\t\t\tr.loadTemplateFile(name, file)\n\t\t}\n\t}\n}", "func (tg ReadFiles) TestReadProcedures(t *testing.T) {\n\tfilePath := fmt.Sprintf(\"%s/procedures/workstation.md\", util.GetRootPath())\n\tfileInfo, _ := os.Lstat(filePath)\n\tpath.Procedures = func() ([]path.File, error) {\n\t\treturn []path.File{\n\t\t\t{FullPath: filePath, Info: fileInfo},\n\t\t}, nil\n\t}\n\n\tdocuments, err := ReadProcedures()\n\tif err != nil {\n\t\tt.Fatalf(`ReadProcedures() returned an error %v`, err)\n\t}\n\tif len(documents) != 1 {\n\t\tt.Fatal(`Invalid number of documents`)\n\t}\n\tif documents[0].FullPath != filePath {\n\t\tt.Fatalf(`Invalid document path %s`, documents[0].FullPath)\n\t}\n}", "func loadTemplates() {\n\n\tfmt.Println(\"About to load templates\")\n\n\t// get layouts\n\tlayouts, err := filepath.Glob(\"templates/layouts/*.layout\")\n\tpanicOnError(err)\n\n\t// get list of main pages\n\tpages, err := filepath.Glob(\"templates/pages/*.html\")\n\tpanicOnError(err)\n\n\tfor _, page := range pages {\n\t\tfiles := append(layouts, page)\n\t\ttemplateName := filepath.Base(page)\n\n\t\tnewTemplate := template.Must(template.ParseFiles(files...))\n\t\tnewTemplate.Option(\"missingkey=default\")\n\n\t\tappTemplates[templateName] = newTemplate\n\t}\n\n\t// loaded templates\n\tfor file, _ := range appTemplates {\n\t\tfmt.Printf(\"Loaded Template: %s\\n\", file)\n\t\tfmt.Printf(\"loaded: %s\\n\", file)\n\t}\n\n}", "func loadInitialFiles(t *testing.T, data dataSection) int32 {\n\tfilesDocs := make([]interface{}, 0, len(data.Files))\n\tchunksDocs := make([]interface{}, 0, len(data.Chunks))\n\tvar chunkSize int32\n\n\tfor _, v := range data.Files {\n\t\tdocBytes, err := v.MarshalJSON()\n\t\ttesthelpers.RequireNil(t, err, \"error converting raw message to bytes: %s\", err)\n\t\tdoc := bsonx.Doc{}\n\t\terr = bson.UnmarshalExtJSON(docBytes, false, &doc)\n\t\ttesthelpers.RequireNil(t, err, \"error creating file document: %s\", err)\n\n\t\t// convert length from int32 to int64\n\t\tif length, err := doc.LookupErr(\"length\"); err == nil {\n\t\t\tdoc = doc.Delete(\"length\")\n\t\t\tdoc = doc.Append(\"length\", bsonx.Int64(int64(length.Int32())))\n\t\t}\n\t\tif cs, err := doc.LookupErr(\"chunkSize\"); err == nil {\n\t\t\tchunkSize = cs.Int32()\n\t\t}\n\n\t\tfilesDocs = append(filesDocs, doc)\n\t}\n\n\tfor _, v := range data.Chunks {\n\t\tdocBytes, err := v.MarshalJSON()\n\t\ttesthelpers.RequireNil(t, err, \"error converting raw message to bytes: %s\", err)\n\t\tdoc := bsonx.Doc{}\n\t\terr = bson.UnmarshalExtJSON(docBytes, false, &doc)\n\t\ttesthelpers.RequireNil(t, err, \"error creating file document: %s\", err)\n\n\t\t// convert data $hex to binary value\n\t\tif hexStr, err := doc.LookupErr(\"data\", \"$hex\"); err == nil {\n\t\t\thexBytes := convertHexToBytes(t, hexStr.StringValue())\n\t\t\tdoc = doc.Delete(\"data\")\n\t\t\tdoc = append(doc, bsonx.Elem{\"data\", bsonx.Binary(0x00, hexBytes)})\n\t\t}\n\n\t\t// convert n from int64 to int32\n\t\tif n, err := doc.LookupErr(\"n\"); err == nil {\n\t\t\tdoc = doc.Delete(\"n\")\n\t\t\tdoc = append(doc, bsonx.Elem{\"n\", bsonx.Int32(n.Int32())})\n\t\t}\n\n\t\tchunksDocs = append(chunksDocs, doc)\n\t}\n\n\tif len(filesDocs) > 0 {\n\t\t_, err := files.InsertMany(ctx, filesDocs)\n\t\ttesthelpers.RequireNil(t, err, \"error inserting into files: %s\", err)\n\t\t_, err = expectedFiles.InsertMany(ctx, filesDocs)\n\t\ttesthelpers.RequireNil(t, err, \"error inserting into expected files: %s\", err)\n\t}\n\n\tif len(chunksDocs) > 0 {\n\t\t_, err := chunks.InsertMany(ctx, chunksDocs)\n\t\ttesthelpers.RequireNil(t, err, \"error inserting into chunks: %s\", err)\n\t\t_, err = expectedChunks.InsertMany(ctx, chunksDocs)\n\t\ttesthelpers.RequireNil(t, err, \"error inserting into expected chunks: %s\", err)\n\t}\n\n\treturn chunkSize\n}", "func ReadFromFiles(paths ...string) (c Config, err error) {\n\tfor _, v := range paths {\n\t\t_, err = toml.DecodeFile(v, &c)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func (s *Site) readFiles(dir, base string) error {\n\treturn filepath.Walk(dir, func(filename string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\trel := utils.MustRel(base, filename)\n\t\tswitch {\n\t\tcase info.IsDir() && s.Exclude(rel):\n\t\t\treturn filepath.SkipDir\n\t\tcase info.IsDir():\n\t\t\treturn nil\n\t\tcase s.Exclude(rel):\n\t\t\treturn nil\n\t\tcase strings.HasPrefix(rel, \"_\"):\n\t\t\treturn nil\n\t\t}\n\t\tdefaultFrontmatter := s.cfg.GetFrontMatterDefaults(\"\", rel)\n\t\td, err := pages.NewFile(s, filename, filepath.ToSlash(rel), defaultFrontmatter)\n\t\tif err != nil {\n\t\t\treturn utils.WrapPathError(err, filename)\n\t\t}\n\t\ts.AddDocument(d, true)\n\t\tif p, ok := d.(Page); ok {\n\t\t\ts.nonCollectionPages = append(s.nonCollectionPages, p)\n\t\t}\n\t\treturn nil\n\t})\n}", "func (ui *GUI) loadTemplates() error {\n\tvar templates []string\n\tfindTemplate := func(path string, f os.FileInfo, err error) error {\n\t\t// If path doesn't exist, or other error with path, return error so\n\t\t// that Walk will quit and return the error to the caller.\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif !f.IsDir() && strings.HasSuffix(f.Name(), \".html\") {\n\t\t\ttemplates = append(templates, path)\n\t\t}\n\t\treturn nil\n\t}\n\n\terr := filepath.Walk(ui.cfg.GUIDir, findTemplate)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\thttpTemplates := template.New(\"template\").Funcs(template.FuncMap{\n\t\t\"hashString\": util.HashString,\n\t\t\"upper\": strings.ToUpper,\n\t\t\"percentString\": util.PercentString,\n\t})\n\n\t// Since template.Must panics with non-nil error, it is much more\n\t// informative to pass the error to the caller to log it and exit\n\t// gracefully.\n\thttpTemplates, err = httpTemplates.ParseFiles(templates...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tui.templates = template.Must(httpTemplates, nil)\n\treturn nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
MustReadPartials calls ReadPartials and panics on any error
func (t *TRoot) MustReadPartials(files ...string) { var err = t.ReadPartials(files...) if err != nil { panic(err) } }
[ "func (t *TRoot) ReadPartials(files ...string) error {\n\tfor _, file := range files {\n\t\tvar _, err = t.template.ParseFiles(filepath.Join(t.Path, file))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func mustReadFull(r io.Reader, b []byte) (int, error) {\n\tn, err := tryReadFull(r, b)\n\tif err == io.EOF {\n\t\terr = io.ErrUnexpectedEOF\n\t}\n\treturn n, err\n}", "func (w *wapcHost) requireRead(ctx context.Context, mem api.Memory, fieldName string, offset, byteCount uint32) []byte {\n\tbuf, ok := mem.Read(ctx, offset, byteCount)\n\tif !ok {\n\t\tw.t.Fatalf(\"out of memory reading %s\", fieldName)\n\t}\n\treturn buf\n}", "func TestExtractPartialRead(t *testing.T) {\n\trc := mutate.Extract(invalidImage{})\n\tif _, err := io.Copy(io.Discard, io.LimitReader(rc, 1)); err != nil {\n\t\tt.Errorf(\"Could not read one byte from reader\")\n\t}\n\tif err := rc.Close(); err != nil {\n\t\tt.Errorf(\"rc.Close: %v\", err)\n\t}\n}", "func (r *readerWithStats) MustReadFull(data []byte) {\n\tfs.MustReadData(r.r, data)\n\tr.bytesRead += uint64(len(data))\n}", "func (r *ReaderAt) MustReadAt(p []byte, off int64) {\n\tif len(p) == 0 {\n\t\treturn\n\t}\n\tif off < 0 {\n\t\tlogger.Panicf(\"off=%d cannot be negative\", off)\n\t}\n\tend := off + int64(len(p))\n\tif len(r.mmapData) == 0 || (len(p) > 8*1024 && !r.isInPageCache(off, end)) {\n\t\t// Read big blocks directly from file.\n\t\t// This could be faster than reading these blocks from mmap,\n\t\t// since it triggers less page faults.\n\t\tn, err := r.f.ReadAt(p, off)\n\t\tif err != nil {\n\t\t\tlogger.Panicf(\"FATAL: cannot read %d bytes at offset %d of file %q: %s\", len(p), off, r.f.Name(), err)\n\t\t}\n\t\tif n != len(p) {\n\t\t\tlogger.Panicf(\"FATAL: unexpected number of bytes read; got %d; want %d\", n, len(p))\n\t\t}\n\t\tif len(r.mmapData) > 0 {\n\t\t\tr.markInPageCache(off, end)\n\t\t}\n\t} else {\n\t\tif off > int64(len(r.mmapData)-len(p)) {\n\t\t\tlogger.Panicf(\"off=%d is out of allowed range [0...%d] for len(p)=%d\", off, len(r.mmapData)-len(p), len(p))\n\t\t}\n\t\tsrc := r.mmapData[off:]\n\t\tif r.isInPageCache(off, end) {\n\t\t\t// It is safe copying the data with copy(), since it is likely it is in the page cache.\n\t\t\t// This is up to 4x faster than copyMmap() below.\n\t\t\tcopy(p, src)\n\t\t} else {\n\t\t\t// The data may be missing in the page cache, so it is better to copy it via cgo trick\n\t\t\t// in order to avoid P stalls in Go runtime.\n\t\t\t// See https://medium.com/@valyala/mmap-in-go-considered-harmful-d92a25cb161d for details.\n\t\t\tcopyMmap(p, src)\n\t\t\tr.markInPageCache(off, end)\n\t\t}\n\t}\n\treadCalls.Inc()\n\treadBytes.Add(len(p))\n}", "func ReadPartialReport(scope beam.Scope, partialReportFile string) beam.PCollection {\n\tallFiles := ioutils.AddStrInPath(partialReportFile, \"*\")\n\tlines := textio.ReadSdf(scope, allFiles)\n\treturn beam.ParDo(scope, &parseEncryptedPartialReportFn{}, lines)\n}", "func (g Group) MustGetPartial(name string) *Partial {\n\tpartial, found := g.partials[name]\n\tif !found {\n\t\tpanic(\"Could not find partial named \" + name)\n\t}\n\treturn partial\n}", "func MustRead(file string) []byte {\n\tb, err := Read(file)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn b\n}", "func (e PartialContent) IsPartialContent() {}", "func (bsr *blockStreamReader) MustInitFromFilePart(path string) {\n\tbsr.reset()\n\n\t// Files in the part are always read without OS cache pollution,\n\t// since they are usually deleted after the merge.\n\tconst nocache = true\n\n\tmetaindexPath := filepath.Join(path, metaindexFilename)\n\tindexPath := filepath.Join(path, indexFilename)\n\tcolumnsHeaderPath := filepath.Join(path, columnsHeaderFilename)\n\ttimestampsPath := filepath.Join(path, timestampsFilename)\n\tfieldValuesPath := filepath.Join(path, fieldValuesFilename)\n\tfieldBloomFilterPath := filepath.Join(path, fieldBloomFilename)\n\tmessageValuesPath := filepath.Join(path, messageValuesFilename)\n\tmessageBloomFilterPath := filepath.Join(path, messageBloomFilename)\n\n\tbsr.ph.mustReadMetadata(path)\n\n\t// Open data readers\n\tmetaindexReader := filestream.MustOpen(metaindexPath, nocache)\n\tindexReader := filestream.MustOpen(indexPath, nocache)\n\tcolumnsHeaderReader := filestream.MustOpen(columnsHeaderPath, nocache)\n\ttimestampsReader := filestream.MustOpen(timestampsPath, nocache)\n\tfieldValuesReader := filestream.MustOpen(fieldValuesPath, nocache)\n\tfieldBloomFilterReader := filestream.MustOpen(fieldBloomFilterPath, nocache)\n\tmessageValuesReader := filestream.MustOpen(messageValuesPath, nocache)\n\tmessageBloomFilterReader := filestream.MustOpen(messageBloomFilterPath, nocache)\n\n\t// Initialize streamReaders\n\tbsr.streamReaders.init(metaindexReader, indexReader, columnsHeaderReader, timestampsReader,\n\t\tfieldValuesReader, fieldBloomFilterReader, messageValuesReader, messageBloomFilterReader)\n\n\t// Read metaindex data\n\tbsr.indexBlockHeaders = mustReadIndexBlockHeaders(bsr.indexBlockHeaders[:0], &bsr.streamReaders.metaindexReader)\n}", "func (bsr *blockStreamReader) MustInitFromInmemoryPart(mp *inmemoryPart) {\n\tbsr.reset()\n\n\tbsr.ph = mp.ph\n\n\t// Initialize streamReaders\n\tmetaindexReader := mp.metaindex.NewReader()\n\tindexReader := mp.index.NewReader()\n\tcolumnsHeaderReader := mp.columnsHeader.NewReader()\n\ttimestampsReader := mp.timestamps.NewReader()\n\tfieldValuesReader := mp.fieldValues.NewReader()\n\tfieldBloomFilterReader := mp.fieldBloomFilter.NewReader()\n\tmessageValuesReader := mp.messageValues.NewReader()\n\tmessageBloomFilterReader := mp.messageBloomFilter.NewReader()\n\n\tbsr.streamReaders.init(metaindexReader, indexReader, columnsHeaderReader, timestampsReader,\n\t\tfieldValuesReader, fieldBloomFilterReader, messageValuesReader, messageBloomFilterReader)\n\n\t// Read metaindex data\n\tbsr.indexBlockHeaders = mustReadIndexBlockHeaders(bsr.indexBlockHeaders[:0], &bsr.streamReaders.metaindexReader)\n}", "func (f *FakelogicalReader) ReadNotCalled() bool {\n\treturn len(f.ReadCalls) == 0\n}", "func MustReadInt(r io.Reader) int {\n\tvar res int\n\t_, err := fmt.Fscanf(r, \"%d\", &res)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Failed to read int: %v\", err))\n\t}\n\treturn res\n}", "func (_this *StreamingReadBuffer) RequireAndRetry(position int, requiredByteCount int, operation func(positionOffset int)) {\n\toffset := _this.RequireBytes(position, requiredByteCount)\n\toperation(position + offset)\n}", "func (fr *Frame) mustRead() (n int) {\n\tn = int(fr.raw[1] & 127)\n\tswitch n {\n\tcase 127:\n\t\tn = 8\n\tcase 126:\n\t\tn = 2\n\tdefault:\n\t\tn = 0\n\t}\n\treturn\n}", "func requestBodyRemains(rc io.ReadCloser) bool {\n\tif rc == eofReader {\n\t\treturn false\n\t}\n\tswitch v := rc.(type) {\n\tcase *expectContinueReader:\n\t\treturn requestBodyRemains(v.readCloser)\n\tcase *body:\n\t\treturn v.bodyRemains()\n\tdefault:\n\t\tpanic(\"unexpected type \" + fmt.Sprintf(\"%T\", rc))\n\t}\n}", "func requestBodyRemains(rc io.ReadCloser) bool {\n\tif rc == NoBody {\n\t\treturn false\n\t}\n\tswitch v := rc.(type) {\n\tcase *expectContinueReader:\n\t\treturn requestBodyRemains(v.readCloser)\n\tcase *body:\n\t\treturn v.bodyRemains()\n\tdefault:\n\t\tpanic(\"unexpected type \" + fmt.Sprintf(\"%T\", rc))\n\t}\n}", "func MustReadAll(r io.Reader) []byte {\n\tall, err := io.ReadAll(r)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn all\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Build clones the root (for layout, funcs, etc) and parses the given file in the clone. The returned template is the clone, and is safe to alter without worrying about breaking the root.
func (t *TRoot) Build(path string) (*Template, error) { var tNew, err = t.template.Clone() if err != nil { return nil, err } _, err = tNew.ParseFiles(filepath.Join(t.Path, path)) if err != nil { return nil, err } tNew.Name = path return tNew, nil }
[ "func (t *TRoot) Clone() *TRoot {\n\tvar clone, _ = t.template.Clone()\n\treturn &TRoot{clone, t.Path}\n}", "func Parse(src []byte) (*Template, error) {\n\tt := New()\n\terr := t.Parse(src)\n\treturn t, err\n}", "func (t *Template) Clone() (*Template, error) {\n\tt.nameSpace.mu.Lock()\n\tdefer t.nameSpace.mu.Unlock()\n\tif t.escaped {\n\t\treturn nil, fmt.Errorf(\"html/template: cannot Clone %q after it has executed\", t.Name())\n\t}\n\ttextClone, err := t.text.Clone()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret := &Template{\n\t\tfalse,\n\t\ttextClone,\n\t\ttextClone.Tree,\n\t\t&nameSpace{\n\t\t\tset: make(map[string]*Template),\n\t\t},\n\t}\n\tfor _, x := range textClone.Templates() {\n\t\tname := x.Name()\n\t\tsrc := t.set[name]\n\t\tif src == nil || src.escaped {\n\t\t\treturn nil, fmt.Errorf(\"html/template: cannot Clone %q after it has executed\", t.Name())\n\t\t}\n\t\tx.Tree = x.Tree.Copy()\n\t\tret.set[name] = &Template{\n\t\t\tfalse,\n\t\t\tx,\n\t\t\tx.Tree,\n\t\t\tret.nameSpace,\n\t\t}\n\t}\n\treturn ret, nil\n}", "func newTemplateFile(c []byte) (*templatefile, error) {\n\ttf := &templatefile{\n\t\tcontents: c,\n\t}\n\n\tr := bytes.NewReader(tf.contents)\n\tpos := 0\n\tvar line []byte\n\tfor {\n\t\tch, l, err := r.ReadRune()\n\t\tpos += l\n\n\t\t// read until first line or EOF\n\t\tif ch == '\\n' || err == io.EOF {\n\t\t\tline = c[0:pos]\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif len(line) < 10 {\n\t\treturn tf, nil\n\t}\n\n\t// if we have a match, strip first line of content\n\tif m := extendsRegex.FindSubmatch(line); m != nil {\n\t\ttf.layout = string(m[1])\n\t\ttf.contents = c[len(line):]\n\t}\n\n\treturn tf, nil\n}", "func (t Template) Build(norepo bool) error {\n\t// Make dirs\n\tfor _, dir := range t.Dirs {\n\t\tdir = filepath.Join(t.Root, dir)\n\t\terr := os.MkdirAll(dir, 0755)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfmt.Printf(\"> Created dir: %s\\n\", dir)\n\t}\n\t// Make files\n\tfor _, file := range t.Files {\n\t\t_, filename := splitFilename(file)\n\t\tcontent := loadContent(t.Name, filename)\n\t\tcontent = replaceTokens(content, t.Root)\n\t\tfile = filepath.Join(t.Root, file)\n\t\terr := ioutil.WriteFile(file, content, 0644)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfmt.Printf(\"> Created file: %s\\n\", file)\n\t}\n\tif norepo {\n\t\treturn nil\n\t}\n\treturn t.CreateRepo()\n}", "func (t *TRoot) Template() *Template {\n\treturn t.Clone().template\n}", "func ParseFile(filename string) (*Template, error) {\n\tt := New()\n\terr := t.ParseFile(filename)\n\treturn t, err\n}", "func parseTemplate(filename string) *appTemplate {\n\ttmpl := template.Must(template.ParseFiles(\"templates/base.html\"))\n\n\t// Put the named file into a template called \"body\"\n\tpath := filepath.Join(\"templates\", filename)\n\tb, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\tApp.Log.Errorf(\"could not read template: %v\", err)\n\t\tpanic(fmt.Errorf(\"could not read template: %v\", err))\n\t}\n\ttemplate.Must(tmpl.New(\"body\").Parse(string(b)))\n\n\treturn &appTemplate{tmpl.Lookup(\"base.html\")}\n}", "func NewTemplateFromRootTemplate(templateFileName string) (*Template, error) {\n\ttemplateDir := filepath.Dir(templateFileName)\n\tcontent, err := ioutil.ReadFile(templateFileName)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"cannot read template file (%s): %s\", err, templateFileName)\n\t}\n\n\tvar c map[string]interface{}\n\terr = yaml.Unmarshal([]byte(content), &c)\n\tif err != nil {\n\t\tlog.Fatalf(\"Cannot parse template: %v\", err)\n\t}\n\n\t// For each of the imports, grab the import file\n\tvar imports []string\n\tif c[\"imports\"] != nil {\n\t\tfor _, importFile := range c[\"imports\"].([]interface{}) {\n\t\t\tvar fileName = importFile.(map[string]interface{})[\"path\"].(string)\n\t\t\timports = append(imports, templateDir+\"/\"+fileName)\n\t\t}\n\t}\n\treturn NewTemplateFromFileNames(templateFileName, imports[0:])\n}", "func Parse(filePath string, preprocessor Preprocessor, options protocols.ExecuterOptions) (*Template, error) {\n\tif value, err := parsedTemplatesCache.Has(filePath); value != nil {\n\t\treturn value.(*Template), err\n\t}\n\n\ttemplate := &Template{}\n\n\tf, err := os.Open(filePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer f.Close()\n\n\tdata, err := ioutil.ReadAll(f)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdata = template.expandPreprocessors(data)\n\tif preprocessor != nil {\n\t\tdata = preprocessor.Process(data)\n\t}\n\n\terr = yaml.Unmarshal(data, template)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif utils.IsBlank(template.Info.Name) {\n\t\treturn nil, errors.New(\"no template name field provided\")\n\t}\n\tif template.Info.Authors.IsEmpty() {\n\t\treturn nil, errors.New(\"no template author field provided\")\n\t}\n\n\t// Setting up variables regarding template metadata\n\toptions.TemplateID = template.ID\n\toptions.TemplateInfo = template.Info\n\toptions.TemplatePath = filePath\n\n\t// If no requests, and it is also not a workflow, return error.\n\tif len(template.RequestsDNS)+len(template.RequestsHTTP)+len(template.RequestsFile)+len(template.RequestsNetwork)+len(template.RequestsHeadless)+len(template.Workflows) == 0 {\n\t\treturn nil, fmt.Errorf(\"no requests defined for %s\", template.ID)\n\t}\n\n\t// Compile the workflow request\n\tif len(template.Workflows) > 0 {\n\t\tcompiled := &template.Workflow\n\n\t\tcompileWorkflow(filePath, preprocessor, &options, compiled, options.WorkflowLoader)\n\t\ttemplate.CompiledWorkflow = compiled\n\t\ttemplate.CompiledWorkflow.Options = &options\n\t}\n\n\t// Compile the requests found\n\trequests := []protocols.Request{}\n\tif len(template.RequestsDNS) > 0 && !options.Options.OfflineHTTP {\n\t\tfor _, req := range template.RequestsDNS {\n\t\t\trequests = append(requests, req)\n\t\t}\n\t\ttemplate.Executer = executer.NewExecuter(requests, &options)\n\t}\n\tif len(template.RequestsHTTP) > 0 {\n\t\tif options.Options.OfflineHTTP {\n\t\t\toperatorsList := []*operators.Operators{}\n\n\t\tmainLoop:\n\t\t\tfor _, req := range template.RequestsHTTP {\n\t\t\t\tfor _, path := range req.Path {\n\t\t\t\t\tif !(strings.EqualFold(path, \"{{BaseURL}}\") || strings.EqualFold(path, \"{{BaseURL}}/\")) {\n\t\t\t\t\t\tbreak mainLoop\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\toperatorsList = append(operatorsList, &req.Operators)\n\t\t\t}\n\t\t\tif len(operatorsList) > 0 {\n\t\t\t\toptions.Operators = operatorsList\n\t\t\t\ttemplate.Executer = executer.NewExecuter([]protocols.Request{&offlinehttp.Request{}}, &options)\n\t\t\t}\n\t\t} else {\n\t\t\tfor _, req := range template.RequestsHTTP {\n\t\t\t\trequests = append(requests, req)\n\t\t\t}\n\t\t\ttemplate.Executer = executer.NewExecuter(requests, &options)\n\t\t}\n\t}\n\tif len(template.RequestsFile) > 0 && !options.Options.OfflineHTTP {\n\t\tfor _, req := range template.RequestsFile {\n\t\t\trequests = append(requests, req)\n\t\t}\n\t\ttemplate.Executer = executer.NewExecuter(requests, &options)\n\t}\n\tif len(template.RequestsNetwork) > 0 && !options.Options.OfflineHTTP {\n\t\tfor _, req := range template.RequestsNetwork {\n\t\t\trequests = append(requests, req)\n\t\t}\n\t\ttemplate.Executer = executer.NewExecuter(requests, &options)\n\t}\n\tif len(template.RequestsHeadless) > 0 && !options.Options.OfflineHTTP && options.Options.Headless {\n\t\tfor _, req := range template.RequestsHeadless {\n\t\t\trequests = append(requests, req)\n\t\t}\n\t\ttemplate.Executer = executer.NewExecuter(requests, &options)\n\t}\n\tif template.Executer != nil {\n\t\terr := template.Executer.Compile()\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"could not compile request\")\n\t\t}\n\t\ttemplate.TotalRequests += template.Executer.Requests()\n\t}\n\tif template.Executer == nil && template.CompiledWorkflow == nil {\n\t\treturn nil, ErrCreateTemplateExecutor\n\t}\n\ttemplate.Path = filePath\n\n\tparsedTemplatesCache.Store(filePath, template, err)\n\treturn template, nil\n}", "func Root(name, path string) *TRoot {\n\tvar tmpl = &Template{template.New(name), name}\n\tvar t = &TRoot{tmpl, path}\n\n\treturn t\n}", "func (b *Buildtemplate) Clone(source buildv1alpha1.BuildTemplate, clientset *client.ConfigSet) (*buildv1alpha1.BuildTemplate, error) {\n\tsource.SetName(\"\")\n\tsource.SetGenerateName(b.Name + \"-\")\n\tsource.SetNamespace(b.Namespace)\n\tsource.SetOwnerReferences([]metav1.OwnerReference{})\n\tsource.SetResourceVersion(\"\")\n\tsource.Kind = \"BuildTemplate\"\n\tif len(clientset.Registry.Secret) != 0 {\n\t\taddSecretVolume(clientset.Registry.Secret, &source)\n\t\tsetEnvConfig(clientset.Registry.Secret, &source)\n\t}\n\treturn createBuildTemplate(source, clientset)\n}", "func (t *Template) Parse(src string) (*Template, error) {\n\tt.nameSpace.mu.Lock()\n\tt.escaped = false\n\tt.nameSpace.mu.Unlock()\n\tret, err := t.text.Parse(src)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// In general, all the named templates might have changed underfoot.\n\t// Regardless, some new ones may have been defined.\n\t// The template.Template set has been updated; update ours.\n\tt.nameSpace.mu.Lock()\n\tdefer t.nameSpace.mu.Unlock()\n\tfor _, v := range ret.Templates() {\n\t\tname := v.Name()\n\t\ttmpl := t.set[name]\n\t\tif tmpl == nil {\n\t\t\ttmpl = t.new(name)\n\t\t}\n\t\t// Restore our record of this text/template to its unescaped original state.\n\t\ttmpl.escaped = false\n\t\ttmpl.text = v\n\t\ttmpl.Tree = v.Tree\n\t}\n\treturn t, nil\n}", "func New(o Options) (*Template, error) {\n\t// Init vars\n\tt := Template{\n\t\tname: o.Name,\n\t\tfilePath: o.FilePath,\n\t\tcontent: o.Content,\n\t\tdata: o.Data,\n\t}\n\tif t.name == \"\" {\n\t\tt.name = fmt.Sprintf(\"%p\", &t) // use pointer\n\t}\n\n\t// If the file path is not empty then\n\tif t.filePath != \"\" {\n\t\t// Read the file and set the template content\n\t\tb, err := os.ReadFile(t.filePath)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to create template due to %s\", err.Error())\n\t\t}\n\t\tt.content = string(b)\n\t}\n\n\t// If the content is not empty then\n\tif t.content != \"\" {\n\t\tvar err error\n\t\tt.template, err = template.New(t.name).Funcs(template.FuncMap{\n\t\t\t\"env\": tplFuncEnv,\n\t\t\t\"time\": tplFuncTime,\n\t\t\t\"exec\": tplFuncExec,\n\t\t}).Parse(t.content)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to parse template due to %s\", err.Error())\n\t\t}\n\t}\n\n\treturn &t, nil\n}", "func Parse(rawtemplate string) (template *Template, err error) {\n\ttemplate = new(Template)\n\ttemplate.raw = rawtemplate\n\tsplit := strings.Split(rawtemplate, \"{\")\n\ttemplate.parts = make([]templatePart, len(split)*2-1)\n\tfor i, s := range split {\n\t\tif i == 0 {\n\t\t\tif strings.Contains(s, \"}\") {\n\t\t\t\terr = errors.New(\"unexpected }\")\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttemplate.parts[i].raw = s\n\t\t} else {\n\t\t\tsubsplit := strings.Split(s, \"}\")\n\t\t\tif len(subsplit) != 2 {\n\t\t\t\terr = errors.New(\"malformed template\")\n\t\t\t\tbreak\n\t\t\t}\n\t\t\texpression := subsplit[0]\n\t\t\ttemplate.parts[i*2-1], err = parseExpression(expression)\n\t\t\tif err != nil {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttemplate.parts[i*2].raw = subsplit[1]\n\t\t}\n\t}\n\tif err != nil {\n\t\ttemplate = nil\n\t}\n\treturn template, err\n}", "func (page *LandingPage) ParseTemplate() *template.Template {\n\ttmpl, err := template.ParseFiles(\"templates/landing-page.html\")\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"error\": err,\n\t\t}).Panic(\"LandingPage template error\")\n\t}\n\n\tlog.WithFields(log.Fields{\n\t\t\"tree\": tmpl.Tree.Name,\n\t}).Debug(\"Parsed template\")\n\n\treturn tmpl\n}", "func (t *Template) Clone() (*Template, error) {\n\tvar tmpl, err = t.Template.Clone()\n\treturn &Template{tmpl, t.Name}, err\n}", "func populateTemplate() map[string]*template.Template {\n\tresult := make(map[string]*template.Template)\n\tconst basePath = \"template\"\n\n\t// Load the template by PareFiles\n\tlayout := template.Must(template.PareFiles(basePath + \"/_layout.html\"))\n\n\t// Load template that layout gonna use by ParseFiles in the layout pull header and footer template\n\ttemplate.Must(layout.ParseFiles(basePath+ \"/_header.html\", basePath+ \"/_footer.html\"))\n\n\t// Load the actual template\n\t// All the content templates will be defined inside content directory\n\t// Open with the `os` commend from os package\n\tdir, err := os.Open(basePath + \"/content\")\n\n\t// Error check\n\tif err != nil {\n\t\tpanic(\"Failed to open template blocks directory: \" + err.Error())\n\t}\n\n\t// Read all of content by `dir` command\n\tfis, err := dir.Readdir(-1)\n\n\t// Error check\n\tif err != nil {\n\t\tpanic(\"Failed to read contents of content directory: \" + err.Error())\n\t}\n\n\t// Loop all files\n\tfor _, fi := range fis {\n\n\t\t// Open the file where is pointing to\n\t\tf, err := os.Open(basePath + \"/content/\" + fi.Name())\n\n\t\t// Error check\n\t\tif err != nil {\n\t\t\tpanic(\"Failed to open templates'\" + fi.Name() + \"'\")\n\t\t}\n\n\t\t// Read the content\n\t\tcontent, err := ioutil.ReadAll(f)\n\t\tif err != nil {\n\t\t\tpanic(\"Failed to open content from file '\" + fi.Name() + \"'\")\n\t\t}\n\n\t\t// Close file\n\t\tf.Close()\n\n\t\t// Create the actual template itself\n\t\t// Clone method on the layout template\n\t\t// He takes the layout template + all children and clone that into `tmpl` object\n\n\t\ttmpl := template.Must(layout.Clone())\n\n\t\t// `tmpl` object he's ready\n\t\t// Parse the content been read to that file\n\t\t_, err = tmpl.Parse(string(content))\n\n\t\t// Error check as Parse a template might fails\n\t\tif err != nil {\n\t\t\tpanic(\"Failed to parse content of '\" + fi.Name() + \"' as template\")\n\t\t}\n\n\t\t// Add that template to result map\n\t\tresult[fi.Name()] = tmpl\n\t}\n\treturn result\n}", "func (g *Generator) ParseFile(path string) (*template.Template, error) {\n\treturn g.generateTemplate(path, nil)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
MustBuild calls Build and panics on any error
func (t *TRoot) MustBuild(path string) *Template { var tmpl, err = t.Build(path) if err != nil { panic(err) } return tmpl }
[ "func (runner Runner) RequiresBuild() bool {\n\treturn true\n}", "func (creator Creator) RequiresBuild() bool {\n\treturn false\n}", "func (builder *Builder) MustBuild() string {\n\tpolicy, err := builder.Build()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn policy\n}", "func (b *OpenAPIBuilder) MustBuild(options ...Option) OpenAPI {\n\tv, err := b.Build()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn v\n}", "func (installer Installer) RequiresBuild() bool {\n\treturn false\n}", "func Build() error {\n\tif strings.Contains(runtime.Version(), \"1.8\") {\n\t\t// Go 1.8 doesn't play along with go test ./... and /vendor.\n\t\t// We could fix that, but that would take time.\n\t\tfmt.Printf(\"Skip Build on %s\\n\", runtime.Version())\n\t\treturn nil\n\t}\n\n\t// TODO: Add lint after fixing errors\n\tmg.Deps(Fmt, Vet, TestRace)\n\treturn nil\n}", "func MustBuilds(action string) gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tcurrent := Current(c)\n\n\t\tif current.Admin {\n\t\t\tc.Next()\n\t\t\treturn\n\t\t}\n\n\t\tswitch {\n\t\tcase action == \"display\":\n\t\t\tif allowBuildDisplay(c) {\n\t\t\t\tc.Next()\n\t\t\t\treturn\n\t\t\t}\n\t\tcase action == \"change\":\n\t\t\tif allowBuildChange(c) {\n\t\t\t\tc.Next()\n\t\t\t\treturn\n\t\t\t}\n\t\tcase action == \"delete\":\n\t\t\tif allowBuildDelete(c) {\n\t\t\t\tc.Next()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tAbortUnauthorized(c)\n\t}\n}", "func (builder testBuilder) Build(config *s2iapi.Config) (*s2iapi.Result, error) {\n\treturn nil, builder.buildError\n}", "func (b *RequestBodyBuilder) MustBuild(options ...Option) RequestBody {\n\tv, err := b.Build()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn v\n}", "func (s *Schema) MustBuild() *graphql.Schema {\n\tbuilt, err := s.Build()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn built\n}", "func (e *Expect) Build(ctx *context.Context) (assert.Assertion, error) {\n\texpectBody, err := ctx.ExecuteTemplate(e.Body)\n\tif err != nil {\n\t\treturn nil, errors.Errorf(\"invalid expect response: %s\", err)\n\t}\n\tassertion := protocol.CreateAssertion(expectBody)\n\n\treturn assert.AssertionFunc(func(v interface{}) error {\n\t\tmessage, callErr, err := extract(v)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := e.assertCode(callErr); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := assertion.Assert(message); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}), nil\n}", "func (b *Builder) Build(ctx context.Context, app *AppContext) error {\n\tif err := buildComponents(ctx, app); err != nil {\n\t\treturn fmt.Errorf(\"error building components: %v\", err)\n\t}\n\treturn nil\n}", "func (g Golang) Build(gopath string, meta Metadata, skipTargets string, local bool) (err error) {\n\tlogrus.Debugf(\"Checking to see that gox is installed.\")\n\n\t// Install gox if it's not already there\n\tif _, err := os.Stat(filepath.Join(gopath, \"bin/gox\")); os.IsNotExist(err) {\n\t\terr = GoxInstall(gopath)\n\t\tif err != nil {\n\t\t\terr = errors.Wrap(err, \"Failed to install gox\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvar wd string\n\n\tif local {\n\t\twd, err = os.Getwd()\n\t\tif err != nil {\n\t\t\terr = errors.Wrapf(err, \"failed getting CWD\")\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\twd = fmt.Sprintf(\"%s/src/%s\", gopath, meta.Package)\n\n\t\tlogrus.Debugf(\"Changing working directory to: %s\", wd)\n\n\t\terr = os.Chdir(wd)\n\n\t\tif err != nil {\n\t\t\terr = errors.Wrapf(err, \"changing working dir to %q\", wd)\n\t\t\treturn err\n\t\t}\n\t}\n\n\tgox := fmt.Sprintf(\"%s/bin/gox\", gopath)\n\n\tlogrus.Debugf(\"Gox is: %s\", gox)\n\n\tvar metadatapath string\n\tif local {\n\t\tmetadatapath = fmt.Sprintf(\"%s/%s\", wd, METADATA_FILENAME)\n\n\t} else {\n\t\tmetadatapath = fmt.Sprintf(\"%s/src/%s/%s\", gopath, meta.Package, METADATA_FILENAME)\n\t}\n\n\tmd, err := ReadMetadata(metadatapath)\n\tif err != nil {\n\t\terr = errors.Wrap(err, \"Failed to read metadata file from checked out code\")\n\t\treturn err\n\t}\n\n\tskipTargetsMap := make(map[string]int)\n\n\tif skipTargets != \"\" {\n\t\ttargetsList := strings.Split(skipTargets, \",\")\n\n\t\tfor _, t := range targetsList {\n\t\t\tskipTargetsMap[t] = 1\n\t\t}\n\t}\n\n\tfor _, target := range md.BuildInfo.Targets {\n\t\t// skip this target if we're told to do so\n\t\t_, skip := skipTargetsMap[target.Name]\n\t\tif skip {\n\t\t\tcontinue\n\t\t}\n\n\t\tlogrus.Debugf(\"Building target: %q in dir %s\", target.Name, wd)\n\n\t\t// This gets weird because go's exec shell doesn't like the arg format that gox expects\n\t\t// Building it thusly keeps the various quoting levels straight\n\n\t\trunenv := os.Environ()\n\n\t\tif !local {\n\t\t\tgopathenv := fmt.Sprintf(\"GOPATH=%s\", gopath)\n\t\t\trunenv = append(runenv, gopathenv)\n\t\t}\n\n\t\t// allow user to turn off go modules\n\t\tif !target.Legacy {\n\t\t\trunenv = append(runenv, \"GO111MODULE=on\")\n\t\t}\n\n\t\tcgo := \"\"\n\t\t// build with cgo if we're told to do so.\n\t\tif target.Cgo {\n\t\t\tcgo = \" -cgo\"\n\t\t}\n\n\t\tfor k, v := range target.Flags {\n\t\t\trunenv = append(runenv, fmt.Sprintf(\"%s=%s\", k, v))\n\t\t\tlogrus.Debugf(\"Build Flag: %s=%s\", k, v)\n\t\t}\n\n\t\tldflags := \"\"\n\t\tif target.Ldflags != \"\" {\n\t\t\tldflags = fmt.Sprintf(\" -ldflags %q \", target.Ldflags)\n\t\t\tlogrus.Debugf(\"LD Flag: %s\", ldflags)\n\t\t}\n\n\t\t// Interesting idea, but breaks multiple binary builds such as dbt. To properly implement, we'd have to find and handle each binary instead of relying on the './...'.\n\t\t//outputTemplate := fmt.Sprintf(\"%s_{{.OS}}_{{.Arch}}\", meta.Name)\n\t\t//args := gox + cgo + ldflags + ` -osarch=\"` + target.Name + `\"` + ` -output=\"` + outputTemplate + `\"` + \" ./...\"\n\n\t\targs := gox + cgo + ldflags + ` -osarch=\"` + target.Name + `\"` + \" ./...\"\n\n\t\tlogrus.Debugf(\"Running gox with: %s in dir %s\", args, wd)\n\n\t\t// Calling it through sh makes everything happy\n\t\tcmd := exec.Command(\"sh\", \"-c\", args)\n\n\t\tcmd.Env = runenv\n\n\t\tcmd.Stdout = os.Stdout\n\t\tcmd.Stderr = os.Stderr\n\t\tcmd.Stdin = os.Stdin\n\n\t\terr = cmd.Run()\n\t\tif err != nil {\n\t\t\terr = errors.Wrapf(err, \"failed building target %s\", target.Name)\n\t\t\treturn err\n\t\t}\n\n\t\tlogrus.Debugf(\"Gox build of target %s complete and successful.\", target.Name)\n\t}\n\n\terr = BuildExtras(md, wd)\n\tif err != nil {\n\t\terr = errors.Wrapf(err, \"Failed to build extras\")\n\t\treturn err\n\t}\n\n\treturn err\n}", "func (b *Builder) ShouldBuild() bool {\n\tif osFlag != \"\" && !osFilter[b.OS] {\n\t\treturn false\n\t}\n\tif archFlag != \"\" && !archFilter[b.Arch] {\n\t\treturn false\n\t}\n\tif cmdFlag != \"\" && !cmdFilter[b.Cmd] {\n\t\treturn false\n\t}\n\treturn true\n}", "func shouldBuild(ctxt *build.Context, content []byte, allTags map[string]bool) (shouldBuild, binaryOnly bool, err error) {\n\t// Identify leading run of // comments and blank lines,\n\t// which must be followed by a blank line.\n\t// Also identify any //go:build comments.\n\tcontent, goBuild, sawBinaryOnly, err := parseFileHeader(content)\n\tif err != nil {\n\t\treturn false, false, err\n\t}\n\n\t// If //go:build line is present, it controls.\n\t// Otherwise fall back to +build processing.\n\tswitch {\n\tcase goBuild != nil:\n\t\tx, err := constraint.Parse(string(goBuild))\n\t\tif err != nil {\n\t\t\treturn false, false, fmt.Errorf(\"parsing //go:build line: %v\", err)\n\t\t}\n\t\tshouldBuild = eval(ctxt, x, allTags)\n\n\tdefault:\n\t\tshouldBuild = true\n\t\tp := content\n\t\tfor len(p) > 0 {\n\t\t\tline := p\n\t\t\tif i := bytes.IndexByte(line, '\\n'); i >= 0 {\n\t\t\t\tline, p = line[:i], p[i+1:]\n\t\t\t} else {\n\t\t\t\tp = p[len(p):]\n\t\t\t}\n\t\t\tline = bytes.TrimSpace(line)\n\t\t\tif !bytes.HasPrefix(line, bSlashSlash) || !bytes.Contains(line, bPlusBuild) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ttext := string(line)\n\t\t\tif !constraint.IsPlusBuild(text) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif x, err := constraint.Parse(text); err == nil {\n\t\t\t\tif !eval(ctxt, x, allTags) {\n\t\t\t\t\tshouldBuild = false\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn shouldBuild, sawBinaryOnly, nil\n}", "func (e *Expect) Build(ctx *context.Context) (assert.Assertion, error) {\n\texpectBody, err := ctx.ExecuteTemplate(e.Body)\n\tif err != nil {\n\t\treturn nil, errors.WrapPathf(err, \"body\", \"invalid expect response\")\n\t}\n\tassertion := assert.Build(expectBody)\n\n\treturn assert.AssertionFunc(func(v interface{}) error {\n\t\tres, ok := v.(response)\n\t\tif !ok {\n\t\t\treturn errors.Errorf(\"expected response but got %T\", v)\n\t\t}\n\t\tif err := e.assertCode(res.status); err != nil {\n\t\t\treturn errors.WithPath(err, \"code\")\n\t\t}\n\t\tif err := e.assertHeader(res.Header); err != nil {\n\t\t\treturn errors.WithPath(err, \"header\")\n\t\t}\n\t\tif err := assertion.Assert(res.Body); err != nil {\n\t\t\treturn errors.WithPath(err, \"body\")\n\t\t}\n\t\treturn nil\n\t}), nil\n}", "func Must(err error) {\n\tif err != nil {\n\t\tDie(err)\n\t}\n}", "func Must(err error) bool {\n\tif err != nil {\n\t\tif panicOnErrorMode {\n\t\t\tpanic(err)\n\t\t} else {\n\t\t\tklog.Errorf(\"%s\", err)\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func executeBuild() {\n\tfmt.Println(\"Building ...\")\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Template returns an empty template associated with this TRoot so we can use it for errors without fake "empty" templates
func (t *TRoot) Template() *Template { return t.Clone().template }
[ "func NewTemplateEmpty() (template *Template) {\n\ttemplate = &Template{\n\t\tbase: \"\",\n\t}\n\treturn template\n}", "func NewRootTemplate() *template.Template {\n\t// Create template\n\tvar tmpl = template.New(\"root\")\n\n\t// Make sure template execution fails if a key is missing\n\ttmpl = tmpl.Option(\"missingkey=error\")\n\n\t// Add sprig functions\n\ttmpl = tmpl.Funcs(sprig.TxtFuncMap())\n\n\t// Add global functions\n\ttmpl = tmpl.Funcs(GetGlobalFuncMap())\n\n\t// Add placeholders for package-specific functions\n\ttmpl = tmpl.Funcs(GetPackageFuncMap(nil))\n\n\treturn tmpl\n}", "func (c *MockedHTTPContext) Template() texttemplate.TemplateEngine {\n\tif c.MockedTemplate != nil {\n\t\treturn c.MockedTemplate()\n\t}\n\treturn nil\n}", "func T(args ...interface{}) Template {\n\treturn Template(e(Error{}, args...))\n}", "func newTemplate() *template.Template {\n\treturn template.New(\"jobTemplate\").Delims(\"[[\", \"]]\").Option(\"missingkey=error\")\n}", "func (f *ImpliedFact) Template() *Template {\n\ttplptr := C.EnvFactDeftemplate(f.env.env, f.factptr)\n\treturn createTemplate(f.env, tplptr)\n}", "func InitRootTemplate(templatePath string) {\n\tvar templateFunctions = tmpl.FuncMap{\n\t\t\"IncludeCSS\": webutil.IncludeCSS,\n\t\t\"RawCSS\": webutil.RawCSS,\n\t\t\"IncludeJS\": webutil.IncludeJS,\n\t\t\"RawJS\": webutil.RawJS,\n\t\t\"ImageURL\": webutil.ImageURL,\n\t\t\"HomePath\": webutil.HomePath,\n\t\t\"FullPath\": webutil.FullPath,\n\t\t\"ProdURL\": func() string { return webutil.ProductionURL },\n\t\t\"Comment\": HTMLComment,\n\t\t\"TimeString\": func(t time.Time) string { return t.Format(\"2006-01-02 15:04\") },\n\t\t\"dtstr\": func(t time.Time) string { return t.Format(\"on Jan 2, 2006 at 3:04pm\") },\n\t\t\"actionVerb\": actionVerb,\n\t\t\"nl2br\": func(s string) template.HTML {\n\t\t\tvar escaped = template.HTMLEscaper(s)\n\t\t\tvar replaced = strings.Replace(escaped, \"\\n\", \"<br />\", -1)\n\t\t\treturn template.HTML(replaced)\n\t\t},\n\t\t\"ErrorHTML\": errorHTML,\n\t\t\"ErrorListHTML\": errorListHTML,\n\t\t\"IIIFInfoURL\": webutil.IIIFInfoURL,\n\t\t\"raw\": func(s string) template.HTML { return template.HTML(s) },\n\t\t\"debug\": func() bool { return settings.DEBUG },\n\t\t\"dict\": dict,\n\t\t\"option\": option,\n\t\t\"log\": func(val any) string { logger.Debugf(\"%#v\", val); return \"\" },\n\n\t\t// This hack helps with dynamic heading - Go's templating system seems to\n\t\t// be confused when we have something like \"<{{.Something}}>\" - it decides\n\t\t// the brackets, despite not being in a variable, need to be escaped.\n\t\t\"Open\": func(s string) template.HTML { return template.HTML(\"<\" + s + \">\") },\n\t\t\"Close\": func(s string) template.HTML { return template.HTML(\"</\" + s + \">\") },\n\n\t\t// We have functions for our privileges since they need to be \"global\" and\n\t\t// easily verified at template compile time\n\t\t\"ListTitles\": func() *privilege.Privilege { return privilege.ListTitles },\n\t\t\"ModifyTitles\": func() *privilege.Privilege { return privilege.ModifyTitles },\n\t\t\"ManageMOCs\": func() *privilege.Privilege { return privilege.ManageMOCs },\n\t\t\"ViewMetadataWorkflow\": func() *privilege.Privilege { return privilege.ViewMetadataWorkflow },\n\t\t\"EnterIssueMetadata\": func() *privilege.Privilege { return privilege.EnterIssueMetadata },\n\t\t\"ReviewIssueMetadata\": func() *privilege.Privilege { return privilege.ReviewIssueMetadata },\n\t\t\"ReviewOwnMetadata\": func() *privilege.Privilege { return privilege.ReviewOwnMetadata },\n\t\t\"ReviewUnfixableIssues\": func() *privilege.Privilege { return privilege.ReviewUnfixableIssues },\n\t\t\"ListUsers\": func() *privilege.Privilege { return privilege.ListUsers },\n\t\t\"ModifyUsers\": func() *privilege.Privilege { return privilege.ModifyUsers },\n\t\t\"ViewUploadedIssues\": func() *privilege.Privilege { return privilege.ViewUploadedIssues },\n\t\t\"ModifyUploadedIssues\": func() *privilege.Privilege { return privilege.ModifyUploadedIssues },\n\t\t\"SearchIssues\": func() *privilege.Privilege { return privilege.SearchIssues },\n\t\t\"ViewBatchStatus\": func() *privilege.Privilege { return privilege.ViewBatchStatus },\n\t\t\"ViewQCReadyBatches\": func() *privilege.Privilege { return privilege.ViewQCReadyBatches },\n\t\t\"ApproveQCReadyBatches\": func() *privilege.Privilege { return privilege.ApproveQCReadyBatches },\n\t\t\"RejectQCReadyBatches\": func() *privilege.Privilege { return privilege.RejectQCReadyBatches },\n\t\t\"ViewQCPassedBatches\": func() *privilege.Privilege { return privilege.ViewQCPassedBatches },\n\t\t\"LoadBatches\": func() *privilege.Privilege { return privilege.LoadBatches },\n\t\t\"PurgeBatches\": func() *privilege.Privilege { return privilege.PurgeBatches },\n\t\t\"ArchiveBatches\": func() *privilege.Privilege { return privilege.ArchiveBatches },\n\t\t\"ModifyValidatedLCCNs\": func() *privilege.Privilege { return privilege.ModifyValidatedLCCNs },\n\t\t\"ListAuditLogs\": func() *privilege.Privilege { return privilege.ListAuditLogs },\n\t}\n\n\t// Set up the layout and then our global templates\n\tLayout = tmpl.Root(\"layout\", templatePath)\n\tLayout.Funcs(templateFunctions)\n\tLayout.MustReadPartials(\"layout.go.html\")\n\tInsufficientPrivileges = Layout.MustBuild(\"insufficient-privileges.go.html\")\n\tEmpty = Layout.MustBuild(\"empty.go.html\")\n\tHome = Layout.MustBuild(\"home.go.html\")\n}", "func Template(str string, data interface{}) (string, error) {\n\ttmpl, err := template.New(\"test\").Funcs(fmap).Parse(str)\n\tif err == nil {\n\t\tvar doc bytes.Buffer\n\t\terr = tmpl.Execute(&doc, data)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn strings.Replace(doc.String(), \"<no value>\", \"\", -1), nil\n\t}\n\treturn \"\", err\n}", "func (trs *Transformations) Template() (trans *Transformation) {\n\ttrans = nil\n\n\tif trs != nil {\n\t\ttrans = &trs.Tmpl\n\t}\n\n\treturn trans\n}", "func (repeated Repeated) Template() (Template, error) {\n\tvar template Template\n\n\t// check if all element types are the same\n\t// TODO: remove once \"oneOf\" support is added\n\tfor position := range repeated {\n\t\tif position == 0 {\n\t\t\ttemplate = repeated[position]\n\n\t\t\tcontinue\n\t\t}\n\n\t\tif err := template.Compare(repeated[position]); err != nil {\n\t\t\treturn Template{}, fmt.Errorf(\"all the elements inside the array must have the same type: %w\", err)\n\t\t}\n\t}\n\n\t// get rid of default value if scalar type\n\tif template.Scalar != nil {\n\t\ttemplate = template.Clone()\n\t\ttemplate.Scalar.Default = nil\n\t}\n\n\treturn template, nil\n}", "func Root(name, path string) *TRoot {\n\tvar tmpl = &Template{template.New(name), name}\n\tvar t = &TRoot{tmpl, path}\n\n\treturn t\n}", "func NewDummyTemplate() TemplateEngine {\n\treturn DummyTemplate{}\n}", "func MustInstantiatePlainTemplate(templateStr string, customFuncMap template.FuncMap) *template.Template {\n\ttpl, err := instantiateTemplate(templateStr, plainFuncs, customFuncMap)\n\tutils.Must(err)\n\treturn tpl\n}", "func (hc *HipChat) Template() *template.Template {\n\treturn hc.template\n}", "func (tg *TemplateGenerator) Template(name string) (tpl *template.Template) {\n\tif len(tg.funcmap) > 0 {\n\t\ttpl = template.New(name).Funcs(tg.funcmap)\n\t} else {\n\t\ttpl = template.New(name)\n\t}\n\treturn tpl\n\n}", "func schemaTemplate() *template.Template {\n\tif localTemplate == nil {\n\t\tlocalTemplate = template.Must(template.New(\"Schema Template\").Parse(constTemp))\n\t}\n\treturn localTemplate\n}", "func New(cfg Config) *XTemplate {\n\n\txt := new(XTemplate)\n\txt.cache = make(map[string]*template.Template)\n\txt.rootFolder = cfg.RootFolder\n\tif xt.rootFolder == \"\" {\n\t\txt.rootFolder = \"./templates\"\n\t}\n\txt.partialsFolder = cfg.PartialsFolder\n\tif xt.partialsFolder == \"\" {\n\t\txt.partialsFolder = filepath.Join(xt.rootFolder, \"_partials\")\n\t}\n\txt.componentsFolder = cfg.ComponentsFolder\n\tif xt.componentsFolder == \"\" {\n\t\txt.componentsFolder = filepath.Join(xt.rootFolder, \"_components\")\n\t}\n\n\txt.shared = template.New(\"\")\n\txt.ext = cfg.Ext\n\tif xt.ext == \"\" {\n\t\txt.ext = \"html\"\n\t}\n\n\tfuncs := template.FuncMap{\n\t\t\"args\": args,\n\t\t\"kwargs\": kwargs,\n\t\t\"title\": capitalize,\n\t\t\"lower\": lower,\n\t\t\"upper\": upper,\n\t\t\"json\": marshalJSON,\n\t\t\"tag\": tags,\n\t\t\"nocache\": NoCache,\n\t\t\"ifEmpty\": IfEmpty,\n\t\t\"formatDate\": formatDate,\n\t\t\"formatCDate\": formatCDate,\n\t\t\"isEmpty\": IsEmpty,\n\t}\n\n\txt.funcs = funcs\n\tif len(cfg.Funcs) > 0 {\n\t\tfor k, v := range cfg.Funcs {\n\t\t\txt.funcs[k] = v\n\t\t}\n\t}\n\n\txt.shared.Funcs(xt.funcs)\n\treturn xt\n}", "func DefaultTmpl() *template.Template {\n\ttmpl, err := template.New(\"sidecar\").Parse(sidecarContainer)\n\tif err != nil {\n\t\topenlogging.Error(\"get default template failed: \" + err.Error())\n\t}\n\treturn tmpl\n}", "func (o ServiceSpecOutput) Template() RevisionTemplatePtrOutput {\n\treturn o.ApplyT(func(v ServiceSpec) *RevisionTemplate { return v.Template }).(RevisionTemplatePtrOutput)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
NewEnvironment creates new type environment with current analysis pass.
func NewEnvironment(pass *analysis.Pass) *Environment { return &Environment{ //TODO make it private and reject object named __val ExplicitRefinementMap: map[types.Object]types.Type{}, ImplicitRefinementMap: map[types.Object]types.Type{}, funArgRefinementMap: map[string]types.Type{}, Scope: nil, Pos: token.NoPos, Pass: pass, } }
[ "func NewEnvironment(name string, parent *Environment) *Environment {\n\treturn &Environment{\n\t\tname: name,\n\t\tparent: parent,\n\t\tdict: make(map[string]*Symbol),\n\t}\n}", "func NewEnvironment(ctx *pulumi.Context,\n\tname string, args *EnvironmentArgs, opts ...pulumi.ResourceOption) (*Environment, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.EnvironmentId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'EnvironmentId'\")\n\t}\n\tif args.InfrastructureSpec == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InfrastructureSpec'\")\n\t}\n\tif args.LakeId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'LakeId'\")\n\t}\n\treplaceOnChanges := pulumi.ReplaceOnChanges([]string{\n\t\t\"environmentId\",\n\t\t\"lakeId\",\n\t\t\"location\",\n\t\t\"project\",\n\t})\n\topts = append(opts, replaceOnChanges)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Environment\n\terr := ctx.RegisterResource(\"google-native:dataplex/v1:Environment\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewEnvironment() *Environment {\n\ts := make(map[string]Object)\n\treturn &Environment{store: s}\n}", "func NewEnvironment() *Environment {\n\treturn GlobalEnvironment().WithoutFunctions().WithoutMethods()\n}", "func newEnvironment(uid int64) Environment {\n\tenv := Environment{\n\t\tcr: newCursor(db),\n\t\tuid: uid,\n\t\tcontext: types.NewContext(),\n\t\tcache: newCache(),\n\t}\n\treturn env\n}", "func NewEnvironment(ctx context.Context) *Environment {\n\tctx, cancel := context.WithCancel(ctx)\n\te := &Environment{\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\tgenerator: NewIDGenerator(),\n\t\tstopCh: make(chan struct{}),\n\t}\n\treturn e\n}", "func NewEnvironment() *LateXEnviro {\n\te := new(LateXEnviro)\n\treturn e\n}", "func New() (e Env) { return }", "func NewEnvironment() (*Environment, error) {\n\tptr := spEnv()\n\tif ptr == nil {\n\t\treturn nil, errors.New(\"sp_env failed\")\n\t}\n\treturn &Environment{varStore: newVarStore(ptr, 4)}, nil\n}", "func (e *Env) NewEnv() *Env {\n\treturn &Env{\n\t\tenv: make(map[string]interface{}),\n\t\tparent: e,\n\t\tbuiltin: e.builtin,\n\t\tglobal: e.global,\n\t\tfuncArg: make(map[string]interface{}),\n\t\t//importFunc: e.importFunc,\n\t\tfileInfo: e.fileInfo,\n\t}\n}", "func NewEnvironment(testFileName string, testSource []byte) *Environment {\n\tenv := &Environment{\n\t\tmockedFunctions: []string{},\n\t\tschemaDir: filepath.Dir(testFileName),\n\t\ttestFileName: testFileName,\n\t\ttestSource: testSource,\n\t}\n\treturn env\n}", "func NewEnvironment(env ...string) *Environment {\n\te := Environment{\n\t\tHidden: &Environment{},\n\t}\n\tfor _, keyvalue := range env {\n\t\tpair := strings.SplitN(keyvalue, \"=\", 2)\n\t\te.Add(pair[0], pair[1])\n\t}\n\n\treturn &e\n}", "func NewEnv(context *libcoap.Context) *Env {\n return &Env{\n context,\n nil,\n make(chan Event, 32),\n nil,\n }\n}", "func newEnv(envPath, provider, name, tfStatePath string, logger hclog.Logger) (*environment, error) {\n\t// Make sure terraform is on the PATH\n\ttf, err := exec.LookPath(\"terraform\")\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to lookup terraform binary: %v\", err)\n\t}\n\n\tlogger = logger.Named(\"provision\").With(\"provider\", provider, \"name\", name)\n\n\t// set the path to the terraform module\n\ttfPath := path.Join(envPath, provider, name)\n\tlogger.Debug(\"using tf path\", \"path\", tfPath)\n\tif _, err := os.Stat(tfPath); os.IsNotExist(err) {\n\t\treturn nil, fmt.Errorf(\"failed to lookup terraform configuration dir %s: %v\", tfPath, err)\n\t}\n\n\t// set the path to state file\n\ttfState := path.Join(tfStatePath, fmt.Sprintf(\"e2e.%s.%s.tfstate\", provider, name))\n\n\tenv := &environment{\n\t\tprovider: provider,\n\t\tname: name,\n\t\ttf: tf,\n\t\ttfPath: tfPath,\n\t\ttfState: tfState,\n\t\tlogger: logger,\n\t}\n\treturn env, nil\n}", "func NewEnvironment(jsonData string) (*Environment, error) {\n\t// initialize env with input data\n\tenv := new(Environment)\n\terr := serialize.CopyFromJSON(jsonData, env)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn env, nil\n}", "func NewEnvironment(ctx context.Context, existing *env.DoltEnv) (Environment, error) {\n\tmfs, err := getMigrateFS(existing.FS)\n\tif err != nil {\n\t\treturn Environment{}, err\n\t}\n\n\tif err = initMigrationDB(ctx, existing, existing.FS, mfs); err != nil {\n\t\treturn Environment{}, err\n\t}\n\n\tmdb, err := doltdb.LoadDoltDB(ctx, targetFormat, doltdb.LocalDirDoltDB, mfs)\n\tif err != nil {\n\t\treturn Environment{}, err\n\t}\n\n\tconfig, err := env.LoadDoltCliConfig(env.GetCurrentUserHomeDir, mfs)\n\tif err != nil {\n\t\treturn Environment{}, err\n\t}\n\n\tmigration := &env.DoltEnv{\n\t\tVersion: existing.Version,\n\t\tConfig: config,\n\t\tRepoState: existing.RepoState,\n\t\tDoltDB: mdb,\n\t\tFS: mfs,\n\t\t//urlStr: urlStr,\n\t\t//hdp: hdp,\n\t}\n\n\treturn Environment{\n\t\tMigration: migration,\n\t\tExisting: existing,\n\t}, nil\n}", "func (f *EnvTestFixture) NewEnv() *hermit.Env {\n\tenvDir, err := ioutil.TempDir(\"\", \"\")\n\trequire.NoError(f.t, err)\n\tlog, _ := ui.NewForTesting()\n\terr = hermit.Init(log, envDir, \"\", f.State.Root(), hermit.Config{})\n\trequire.NoError(f.t, err)\n\tenv, err := hermit.OpenEnv(envDir, f.State, envars.Envars{}, f.Server.Client())\n\trequire.NoError(f.t, err)\n\treturn env\n}", "func NewEnvironment(ctx *pulumi.Context,\n\tname string, args *EnvironmentArgs, opts ...pulumi.ResourceOption) (*Environment, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.EnvironmentId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'EnvironmentId'\")\n\t}\n\treplaceOnChanges := pulumi.ReplaceOnChanges([]string{\n\t\t\"environmentId\",\n\t\t\"location\",\n\t\t\"project\",\n\t})\n\topts = append(opts, replaceOnChanges)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Environment\n\terr := ctx.RegisterResource(\"google-native:notebooks/v1:Environment\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func newTestEnv(ctx context.Context, t *testing.T, pipelineInfo *pps.PipelineInfo, realEnv *realenv.RealEnv) *testEnv {\n\tlogger := logs.New(pctx.Child(ctx, t.Name()))\n\tworkerDir := filepath.Join(realEnv.Directory, \"worker\")\n\tdriver, err := driver.NewDriver(\n\t\tctx,\n\t\trealEnv.ServiceEnv,\n\t\trealEnv.PachClient,\n\t\tpipelineInfo,\n\t\tworkerDir,\n\t)\n\trequire.NoError(t, err)\n\n\tctx, cancel := pctx.WithCancel(realEnv.PachClient.Ctx())\n\tt.Cleanup(cancel)\n\tdriver = driver.WithContext(ctx)\n\n\treturn &testEnv{\n\t\tRealEnv: realEnv,\n\t\tlogger: logger,\n\t\tdriver: &testDriver{driver},\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
ToQuery converts the Request to URL Query.
func (r AnnounceRequest) ToQuery() (vs url.Values) { vs = make(url.Values, 9) vs.Set("info_hash", r.InfoHash.BytesString()) vs.Set("peer_id", r.PeerID.BytesString()) vs.Set("uploaded", strconv.FormatInt(r.Uploaded, 10)) vs.Set("downloaded", strconv.FormatInt(r.Downloaded, 10)) vs.Set("left", strconv.FormatInt(r.Left, 10)) if r.IP != "" { vs.Set("ip", r.IP) } if r.Event > 0 { vs.Set("event", strconv.FormatInt(int64(r.Event), 10)) } if r.Port > 0 { vs.Set("port", strconv.FormatUint(uint64(r.Port), 10)) } if r.NumWant != 0 { vs.Set("numwant", strconv.FormatUint(uint64(r.NumWant), 10)) } if r.Key != 0 { vs.Set("key", strconv.FormatInt(int64(r.Key), 10)) } // BEP 23 if r.Compact { vs.Set("compact", "1") } else { vs.Set("compact", "0") } return }
[ "func (filter TestRunFilter) ToQuery() (q url.Values) {\n\tu := url.URL{}\n\tq = u.Query()\n\tif !filter.SHAs.EmptyOrLatest() {\n\t\tfor _, sha := range filter.SHAs {\n\t\t\tq.Add(\"sha\", sha)\n\t\t}\n\t}\n\tif filter.Labels != nil && filter.Labels.Cardinality() > 0 {\n\t\tfor label := range filter.Labels.Iter() {\n\t\t\tq.Add(\"label\", label.(string))\n\t\t}\n\t}\n\tif len(filter.Products) > 0 {\n\t\tfor _, p := range filter.Products {\n\t\t\tq.Add(\"product\", p.String())\n\t\t}\n\t}\n\tif filter.Aligned != nil {\n\t\tq.Set(\"aligned\", strconv.FormatBool(*filter.Aligned))\n\t}\n\tif filter.MaxCount != nil {\n\t\tq.Set(\"max-count\", fmt.Sprintf(\"%v\", *filter.MaxCount))\n\t}\n\tif filter.Offset != nil {\n\t\tq.Set(\"offset\", fmt.Sprintf(\"%v\", *filter.Offset))\n\t}\n\tif filter.From != nil {\n\t\tq.Set(\"from\", filter.From.Format(time.RFC3339))\n\t}\n\tif filter.To != nil {\n\t\tq.Set(\"to\", filter.From.Format(time.RFC3339))\n\t}\n\tif filter.View != nil {\n\t\tq.Set(\"view\", *filter.View)\n\t}\n\treturn q\n}", "func (tir TreeInstanceRequest) ToQueryURI() string {\n\tvar (\n\t\tparams = &url.Values{}\n\t\tt = reflect.TypeOf(tir)\n\t\tv = reflect.ValueOf(tir)\n\t\tfv string\n\t)\n\tfor i := 0; i < t.NumField(); i++ {\n\t\tfv = v.Field(i).Interface().(string)\n\t\tif fv != \"\" {\n\t\t\tparams.Set(t.Field(i).Tag.Get(_query), fv)\n\t\t}\n\t}\n\treturn params.Encode()\n}", "func (f *Filter) ToQuery() string {\n\treturn fmt.Sprintf(\"last_knowledge_of_server=%d\", f.LastKnowledgeOfServer)\n}", "func (p SearchParameters) ToQuery() string {\n\treturn paramsToQuery(p)\n}", "func (o AvailablePhoneNumbersOptions) ToQueryString() (url.Values, error) {\n\treturn query.Values(o)\n}", "func (req *Request) Query() string {\n\treturn req.q\n}", "func (r *Request) Query(q map[string]string) *Request {\n\tr.query = q\n\treturn r\n}", "func ToHistoryQueryWorkflowRequest(t *history.QueryWorkflowRequest) *types.HistoryQueryWorkflowRequest {\n\tif t == nil {\n\t\treturn nil\n\t}\n\treturn &types.HistoryQueryWorkflowRequest{\n\t\tDomainUUID: t.GetDomainUUID(),\n\t\tRequest: ToQueryWorkflowRequest(t.Request),\n\t}\n}", "func (opts ShowOpts) ToShowQuery() (string, error) {\n\tq, err := gophercloud.BuildQueryString(opts)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn q.String(), nil\n}", "func (p *Params) EncodeToQuery() string {\n\treturn \"\"\n}", "func (r *Request) GetQuery(p string) string {\n\treturn r.natReqObj.QueryStringParameters[p]\n}", "func Query(q Mappable) *QueryRequest {\n\treturn &QueryRequest{q}\n}", "func QueryToJSON(queryString string) ([]byte, error) {\r\n\r\n\tparsedQuery, err := url.ParseQuery(queryString)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\tjsonString, err := json.Marshal(parsedQuery)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\treturn jsonString, nil\r\n}", "func (opts DeleteOpts) ToDeleteQuery() (string, error) {\n\tq, err := golangsdk.BuildQueryString(opts)\n\treturn q.String(), err\n}", "func (c *Context) RequestQuery() *orm.RequestQuery {\n\trq := &orm.RequestQuery{\n\t\tOffset: 0,\n\t\tLimit: -1,\n\t\tConditions: make([]map[string]string, 0),\n\t}\n\n\treturn rq.ReadFromContext(c.QueryParams())\n}", "func (c *Comment) QueryRequest() *RequestQuery {\n\treturn NewCommentClient(c.config).QueryRequest(c)\n}", "func (vo VideoOptions) ToQuery() (q string, err error) {\n\tif vo.PerPage == 0 {\n\t\tvo.PerPage = minPerPage\n\t}\n\tif vo.Page < 1 {\n\t\tvo.Page = 1\n\t}\n\tv, err := query.Values(vo)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn v.Encode(), nil\n}", "func (opts GetOpts) ToObjectGetQuery() (string, error) {\n\tq, err := gophercloud.BuildQueryString(opts)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn q.String(), nil\n}", "func (r *Request) Query() url.Values {\n\t// prepare values\n\tvalues := url.Values{}\n\n\t// add include\n\tif len(r.Include) > 0 {\n\t\tvalues.Set(\"include\", strings.Join(r.Include, \",\"))\n\t}\n\n\t// add pagination\n\tif r.PageNumber != 0 {\n\t\tvalues.Set(\"page[number]\", strconv.FormatInt(r.PageNumber, 10))\n\t}\n\tif r.PageSize != 0 {\n\t\tvalues.Set(\"page[size]\", strconv.FormatInt(r.PageSize, 10))\n\t}\n\tif r.PageOffset != 0 {\n\t\tvalues.Set(\"page[offset]\", strconv.FormatInt(r.PageOffset, 10))\n\t}\n\tif r.PageLimit != 0 {\n\t\tvalues.Set(\"page[limit]\", strconv.FormatInt(r.PageLimit, 10))\n\t}\n\tif r.PageBefore != \"\" {\n\t\tvalues.Set(\"page[before]\", r.PageBefore)\n\t}\n\tif r.PageAfter != \"\" {\n\t\tvalues.Set(\"page[after]\", r.PageAfter)\n\t}\n\tif r.Pagination != \"\" {\n\t\tvalues.Set(\"pagination\", r.Pagination)\n\t}\n\n\t// add sorting\n\tif len(r.Sorting) > 0 {\n\t\tvalues.Set(\"sort\", strings.Join(r.Sorting, \",\"))\n\t}\n\n\t// add fields\n\tfor name, fields := range r.Fields {\n\t\tvalues.Set(\"fields[\"+name+\"]\", strings.Join(fields, \",\"))\n\t}\n\n\t// add filters\n\tfor name, filter := range r.Filters {\n\t\tfor _, value := range filter {\n\t\t\tvalues.Add(\"filter[\"+name+\"]\", value)\n\t\t}\n\t}\n\n\t// add search\n\tif r.Search != \"\" {\n\t\tvalues.Set(\"search\", r.Search)\n\t}\n\n\treturn values\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
DecodeFrom reads the []byte data from r and decodes them to sr by bencode. r may be the body of the request from the http client.
func (sr *ScrapeResponse) DecodeFrom(r io.Reader) (err error) { return bencode.NewDecoder(r).Decode(sr) }
[ "func NewDecoder(enc *base64.Encoding, r io.Reader) io.Reader", "func Decode(r io.Reader) (map[string]interface{}, error) {\n\tbuf := bufio.NewReader(r)\n\n\tif firstByte, err := buf.ReadByte(); err != nil {\n\t\treturn nil, err\n\t} else if firstByte != 'd' {\n\t\treturn nil, errors.New(\"bencode data must begin with a dictionary\")\n\t}\n\n\tmp, err := decodeDict(buf)\n\tif err != nil {\n\t\treturn nil, errors.New(\"Decode failed\")\n\t}\n\treturn mp, nil\n}", "func decodeRequest(_ context.Context, r *http.Request) (request interface{}, err error) {\n\tdefer r.Body.Close()\n\treturn nil, nil\n}", "func Decode(r io.Reader, v interface{}, order binary.ByteOrder) error {\n\tb, err := Read(r, order)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = json.Unmarshal(b, v)\n\tif err != nil {\n\t\treturn errors.New(err.Error() + string(b))\n\t}\n\treturn nil\n}", "func Decode(r BytesReader) (interface{}, error) {\n\treturn decodeValue(r)\n}", "func Decode(r io.Reader, v Decoder) error {\n\treader := NewReader(r)\n\terr := v.DecodeMsgpack(reader)\n\treleaseReader(reader)\n\treturn err\n}", "func (subr *SRRecordResponse) Decode(b []byte) (err error) {\n\tbuffer := bytes.NewReader(b)\n\tcrn := make([]byte, 2)\n\tif _, err = buffer.Read(crn); err != nil {\n\t\treturn fmt.Errorf(\"EGTS_SR_RECORD_RESPONSE; Error reading CRN\")\n\t}\n\tsubr.ConfirmedRecordNumber = binary.LittleEndian.Uint16(crn)\n\tif subr.RecordStatus, err = buffer.ReadByte(); err != nil {\n\t\tlog.Println(err, b)\n\t\treturn fmt.Errorf(\"EGTS_SR_RECORD_RESPONSE; Error reading RST\")\n\t}\n\n\treturn nil\n}", "func (f *linkFrame) Decode(r io.Reader) (err error) {\n var m map[string]interface{}\n m, err = BencodeReadMap(r)\n if err == nil {\n v, ok := m[\"m\"]\n if ok {\n switch v.(type) {\n case int64:\n f.method = v.(int64)\n break\n default:\n err = errors.New(\"bad method type\")\n return\n }\n } else {\n // no method?\n err = errors.New(\"no method specified\")\n return\n }\n v, ok = m[\"r\"]\n if ok {\n switch v.(type) {\n case int64:\n f.response = v.(int64)\n break\n default:\n err = errors.New(\"bad response type\")\n return\n }\n }\n v, ok = m[\"p\"]\n if ok {\n switch v.(type) {\n case map[string]interface{}:\n f.param = v.(map[string]interface{})\n break\n default:\n err = errors.New(\"Bad parameters type\")\n return\n }\n } else {\n // no parameters?\n err = errors.New(\"no parameters\")\n return\n }\n }\n return\n}", "func decodeRequest(r io.Reader) *plugin.CodeGeneratorRequest {\n\tvar req plugin.CodeGeneratorRequest\n\tinput, err := ioutil.ReadAll(r)\n\tif err != nil {\n\t\tlog.Fatal(\"unable to read stdin: \" + err.Error())\n\t}\n\tif err := proto.Unmarshal(input, &req); err != nil {\n\t\tlog.Fatal(\"unable to marshal stdin as protobuf: \" + err.Error())\n\t}\n\treturn &req\n}", "func NewDecoder(r io.Reader) io.Reader {}", "func (d *Decoder) Decode() (b B, e error) {\n\tif d.r != nil {\n\t\tbody, e := ioutil.ReadAll(d.r)\n\t\tif e != nil {\n\t\t\treturn b, e\n\t\t}\n\t\td.buf = body\n\t}\n\td.bufLen = len(d.buf)\n\n\td.decode(&b, 1)\n\n\treturn\n}", "func DecodeFromReader(r io.Reader) (WrappedReply, error) {\n\tvar wr WrappedReply\n\terr := json.NewDecoder(r).Decode(&wr)\n\treturn wr, err\n}", "func (m *ModifyBearerRequest) DecodeFromBytes(b []byte) error {\n\tlog.Println(\"ModifyBearerRequest.DecodeFromBytes is deprecated. use ModifyBearerRequest.UnmarshalBinary instead\")\n\treturn m.UnmarshalBinary(b)\n}", "func Decode(r io.Reader, result interface{}) error {\n\t// codec can make many little reads from the reader, so wrap it in a buffered reader\n\t// to keep perf lively\n\treturn codec.NewDecoder(bufio.NewReader(r), DecoderHandle()).Decode(result)\n}", "func Decode(r io.Reader) (image.Image, error)", "func Decode(r io.ReadSeeker) (Decoded, string, error) {\n\trr := asReader(r)\n\tf := sniff(rr)\n\tif f.decode == nil {\n\t\treturn nil, \"\", ErrFormat\n\t}\n\tm, err := f.decode(rr)\n\treturn m, f.name, err\n}", "func (p *RequestPacket) ReadFrom(r io.Reader, b []byte, maxPacketLength uint32) error {\n\tb, err := readPacket(r, b, maxPacketLength)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn p.UnmarshalFrom(NewBuffer(b))\n}", "func (broadcast *Broadcast) decodeFromDecrypted(r io.Reader) error {\n\tbroadcast.bm = &Bitmessage{}\n\terr := broadcast.bm.decodeBroadcast(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar sigLength uint64\n\tif sigLength, err = bmutil.ReadVarInt(r); err != nil {\n\t\treturn err\n\t}\n\tif sigLength > obj.SignatureMaxLength {\n\t\tstr := fmt.Sprintf(\"signature length exceeds max length - \"+\n\t\t\t\"indicates %d, but max length is %d\",\n\t\t\tsigLength, obj.SignatureMaxLength)\n\t\treturn wire.NewMessageError(\"DecodeFromDecrypted\", str)\n\t}\n\tbroadcast.sig = make([]byte, sigLength)\n\t_, err = io.ReadFull(r, broadcast.sig)\n\treturn err\n}", "func (m *Message) decodeBody(r io.Reader, h textproto.MIMEHeader) error {\n\tcth := h.Get(\"Content-Type\")\n\tif cth == \"\" {\n\t\tcth = \"text/plain\"\n\t}\n\tct, ctp, err := mime.ParseMediaType(cth)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"invalid content-type: %q\", cth)\n\t}\n\n\t// Find name.\n\tfilename := ctp[\"name\"]\n\tif filename == \"\" {\n\t\tcdh := h.Get(\"Content-Disposition\")\n\t\tif cdh != \"\" {\n\t\t\t_, cdp, err := mime.ParseMediaType(cdh)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"invalid content-disposition: %q\", cdh)\n\t\t\t}\n\t\t\tfilename = cdp[\"filename\"]\n\n\t\t}\n\t}\n\n\t// If it has filename, add as attachment.\n\tif filename != \"\" {\n\t\tname, err := decodeHeader(filename)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"decode filename: %v\", err)\n\t\t}\n\t\tdata, err := ioutil.ReadAll(decodeTransfer(r, h.Get(\"Content-Transfer-Encoding\")))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"read attachment: %v\", err)\n\t\t}\n\n\t\tm.Parts = append(m.Parts, Part{Name: name, Data: data})\n\t\treturn nil\n\t}\n\n\tif ct == \"text/plain\" || ct == \"text/html\" {\n\t\tbuf := new(bytes.Buffer)\n\t\tfor {\n\t\t\tdata, err := ioutil.ReadAll(decodeTransfer(r, h.Get(\"Content-Transfer-Encoding\")))\n\t\t\tbuf.Write(data)\n\t\t\tif err != nil {\n\t\t\t\tif _, ok := err.(base64.CorruptInputError); ok {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"read body: %v\", err)\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\n\t\tbody, err := decodeCharset(buf.String(), ctp[\"charset\"])\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"charsetDecode: %v\", err)\n\t\t}\n\n\t\tif ct == \"text/html\" {\n\t\t\tm.HTML += body\n\t\t\treturn nil\n\t\t}\n\n\t\tm.Body += body\n\t\treturn nil\n\t}\n\n\tif strings.HasPrefix(ct, \"multipart/\") {\n\t\tr := multipart.NewReader(r, ctp[\"boundary\"])\n\t\tfor {\n\t\t\tp, err := r.NextPart()\n\t\t\tif err != nil {\n\t\t\t\tif err == io.EOF {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"next part: %q\", err)\n\t\t\t}\n\n\t\t\tif err := m.decodeBody(p, p.Header); err != nil {\n\t\t\t\tp.Close() // p.Close is also called automatically by r.NextPart.\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\n\t// TODO: decide what to do with this.\n\t//return fmt.Errorf(\"content-type without filename: %q\", ct)\n\treturn nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
EncodeTo encodes the response to []byte by bencode and write the result into w. w may be http.ResponseWriter.
func (sr ScrapeResponse) EncodeTo(w io.Writer) (err error) { return bencode.NewEncoder(w).Encode(sr) }
[ "func (r *SurveyResponse) EncodeToByte() []byte {\n\tb, _ := json.Marshal(r)\n\treturn b\n}", "func (resp *Response) WriteTo(w http.ResponseWriter) {\n\tif resp.buf != nil {\n\t\tresp.Body = resp.buf.Bytes()\n\t\tresp.buf = nil\n\t}\n\n\tif w != nil {\n\t\t// Write the headers\n\t\tfor k, vs := range resp.Header {\n\t\t\t// Reset existing values\n\t\t\tw.Header().Del(k)\n\t\t\tif len(vs) == 1 {\n\t\t\t\tw.Header().Set(k, resp.Header.Get(k))\n\t\t\t}\n\t\t\tif len(vs) > 1 {\n\t\t\t\tfor _, v := range vs {\n\t\t\t\t\tw.Header().Add(k, v)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif resp.redirect != \"\" {\n\t\t\thttp.Redirect(w, resp.req, resp.redirect, resp.StatusCode)\n\t\t\treturn\n\t\t}\n\n\t\tw.WriteHeader(resp.StatusCode)\n\t\tw.Write(resp.Body)\n\t}\n}", "func EncodeConcatbytesResponse(encoder func(context.Context, http.ResponseWriter) goahttp.Encoder) func(context.Context, http.ResponseWriter, any) error {\n\treturn func(ctx context.Context, w http.ResponseWriter, v any) error {\n\t\tres, _ := v.([]byte)\n\t\tctx = context.WithValue(ctx, goahttp.ContentTypeKey, \"text/html\")\n\t\tenc := encoder(ctx, w)\n\t\tbody := res\n\t\tw.WriteHeader(http.StatusOK)\n\t\treturn enc.Encode(body)\n\t}\n}", "func (p *Poll) EncodeToByte() []byte {\n\tb, _ := json.Marshal(p)\n\treturn b\n}", "func ResponseEncoder(ctx context.Context, w http.ResponseWriter) goahttp.Encoder {\n\tw.Header().Set(\"Content-Type\", \"application/cbor\")\n\treturn cbor.NewEncoder(w)\n}", "func (r *ReminderMetadata) EncodeToByte() []byte {\n\tb, _ := json.Marshal(r)\n\treturn b\n}", "func (p *LokiPromResponse) encode(ctx context.Context) (*http.Response, error) {\n\tsp := opentracing.SpanFromContext(ctx)\n\tvar (\n\t\tb []byte\n\t\terr error\n\t)\n\tif p.Response.Data.ResultType == loghttp.ResultTypeVector {\n\t\tb, err = p.marshalVector()\n\t} else {\n\t\tb, err = p.marshalMatrix()\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif sp != nil {\n\t\tsp.LogFields(otlog.Int(\"bytes\", len(b)))\n\t}\n\n\tresp := http.Response{\n\t\tHeader: http.Header{\n\t\t\t\"Content-Type\": []string{\"application/json\"},\n\t\t},\n\t\tBody: io.NopCloser(bytes.NewBuffer(b)),\n\t\tStatusCode: http.StatusOK,\n\t}\n\treturn &resp, nil\n}", "func Bytes(ctx context.Context, code int, contentType string, b []byte) error {\n\tw := ResponseWriter(ctx)\n\tw.WriteHeader(code)\n\tw.Header().Set(\"Content-Type\", contentType)\n\t_, err := w.Write(b)\n\treturn err\n}", "func encodeResponse(ctx context.Context, w http.ResponseWriter, response interface{}) (err error) {\n\tresp := response.(*common.XmidtResponse)\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Header().Set(common.HeaderWPATID, ctx.Value(common.ContextKeyRequestTID).(string))\n\tcommon.ForwardHeadersByPrefix(\"\", resp.ForwardedHeaders, w.Header())\n\n\tw.WriteHeader(resp.Code)\n\t_, err = w.Write(resp.Body)\n\treturn\n}", "func encodeResponse(resp *plugin.CodeGeneratorResponse, w io.Writer) {\n\toutBytes, err := proto.Marshal(resp)\n\tif err != nil {\n\t\tlog.Fatal(\"unable to marshal response to protobuf: \" + err.Error())\n\t}\n\n\tif _, err := w.Write(outBytes); err != nil {\n\t\tlog.Fatal(\"unable to write protobuf to stdout: \" + err.Error())\n\t}\n}", "func encodeGetUserResponse(ctx context.Context, w http1.ResponseWriter, response interface{}) (err error) {\n\tif f, ok := response.(endpoint.Failure); ok && f.Failed() != nil {\n\t\tErrorEncoder(ctx, f.Failed(), w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\treturn\n}", "func EncodeNewResponse(encoder func(context.Context, http.ResponseWriter) goahttp.Encoder) func(context.Context, http.ResponseWriter, interface{}) error {\n\treturn func(ctx context.Context, w http.ResponseWriter, v interface{}) error {\n\t\tres := v.(string)\n\t\tenc := encoder(ctx, w)\n\t\tbody := res\n\t\tw.WriteHeader(http.StatusOK)\n\t\treturn enc.Encode(body)\n\t}\n}", "func (r *Response) Write(w io.Writer) error", "func encodeGetResponse(ctx context.Context, w http1.ResponseWriter, response interface{}) (err error) {\n\tif f, ok := response.(endpoint.Failure); ok && f.Failed() != nil {\n\t\tErrorEncoder(ctx, f.Failed(), w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\treturn\n}", "func (w *WriterInterceptor) Write(b []byte) (int, error) {\n\tlength := w.response.Header.Get(\"Content-Length\")\n\tif length == \"\" || length == \"0\" {\n\t\tw.buf = b\n\t\treturn w.DoWrite()\n\t}\n\n\tw.response.ContentLength += int64(len(b))\n\tw.buf = append(w.buf, b...)\n\n\t// If not EOF\n\tif cl, _ := strconv.Atoi(length); w.response.ContentLength != int64(cl) {\n\t\treturn len(b), nil\n\t}\n\n\tw.response.Body = ioutil.NopCloser(bytes.NewReader(w.buf))\n\tresm := NewResponseModifier(w.response.Request, w.response)\n\tw.modifier(resm)\n\treturn w.DoWrite()\n}", "func EncodeResponse(ctx context.Context, w http.ResponseWriter, response interface{}) error {\n return json.NewEncoder(w).Encode(response)\n}", "func encodeResponse(ctx context.Context, w http.ResponseWriter, response interface{}) (err error) {\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\treturn\n}", "func encodeByteSlice(w io.Writer, bz []byte) (err error) {\n\terr = encodeVarint(w, int64(len(bz)))\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = w.Write(bz)\n\treturn\n}", "func EncodePickResponse(encoder func(context.Context, http.ResponseWriter) goahttp.Encoder) func(context.Context, http.ResponseWriter, any) error {\n\treturn func(ctx context.Context, w http.ResponseWriter, v any) error {\n\t\tres := v.(sommelierviews.StoredBottleCollection)\n\t\tenc := encoder(ctx, w)\n\t\tbody := NewStoredBottleResponseCollection(res.Projected)\n\t\tw.WriteHeader(http.StatusOK)\n\t\treturn enc.Encode(body)\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
NewClient returns a new HTTPClient. scrapeURL may be empty, which will replace the "announce" in announceURL with "scrape" to generate the scrapeURL.
func NewClient(announceURL, scrapeURL string) *Client { if scrapeURL == "" { scrapeURL = strings.Replace(announceURL, "announce", "scrape", -1) } id := metainfo.NewRandomHash() return &Client{AnnounceURL: announceURL, ScrapeURL: scrapeURL, ID: id} }
[ "func NewScraper(httpClient utils.HTTPClient) Scraper {\n\treturn &soupScraper{\n\t\thttpClient,\n\t}\n}", "func newClient(hecURL url.URL, hecToken, index, hostname string, skipTLSVerify bool) *splunk.Client {\n\thttpTransport := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{\n\t\t\tInsecureSkipVerify: skipTLSVerify,\n\t\t},\n\t}\n\thttpClient := &http.Client{\n\t\tTimeout: httpTimeout,\n\t\tTransport: httpTransport,\n\t}\n\n\tif hecURL.Path == \"\" || hecURL.Path == \"/\" {\n\t\thecURL.Path = eventURLPath\n\t}\n\n\treturn &splunk.Client{\n\t\tHTTPClient: httpClient,\n\t\tURL: hecURL.String(),\n\t\tHostname: hostname,\n\t\tToken: hecToken,\n\t\tIndex: index,\n\t}\n}", "func New() (crawl *Crawl) {\n\tc := &http.Client{\n\t\tTransport: http.DefaultTransport,\n\t}\n\tc.Jar, _ = cookiejar.New(nil)\n\n\tcrawl = &Crawl{\n\t\tClient: c,\n\t\tmutex: new(sync.RWMutex),\n\t\thandlers: make(map[interface{}]Handler),\n\t\tcloseCh: make(chan bool, 1),\n\t\tdoneCh: make(chan bool, 1),\n\t}\n\tcrawl.SetOptions(DefaultOptions)\n\treturn\n}", "func New() *Scraper {\n\treturn &Scraper{\n\t\tclient: &http.Client{Timeout: 10 * time.Second},\n\t}\n}", "func New(startUrl string) (*scraper, error) {\n\treturn NewWithConfig(startUrl, NewConfig())\n}", "func New(url string) *Client {\n\treturn NewWithHTTP(url, http.DefaultClient)\n}", "func NewClient(addr string, c *http.Client) (*Client, error) {\n\tif c == nil {\n\t\tc = &http.Client{Timeout: 2 * time.Second}\n\t}\n\n\tu, err := url.Parse(addr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Client{\n\t\tc: c,\n\t\tu: u,\n\t}, nil\n}", "func NewClient(apiKey string, options ...OptionFunc) *Client {\n\tbaseURL, _ := url.Parse(defaultBaseURL)\n\tc := &Client{\n\t\tclient: &http.Client{\n\t\t\tTimeout: time.Second * 10,\n\t\t},\n\t\tbaseURL: baseURL,\n\t\tapiKey: apiKey,\n\t\tuserAgent: \"github.com/barthr/newsapi\",\n\t}\n\n\tfor _, opt := range options {\n\t\topt(c)\n\t}\n\treturn c\n}", "func New(url string, httpClient *http.Client, customHeaders http.Header) *Client {\n\tif httpClient == nil {\n\t\thttpClient = &http.Client{\n\t\t\tTimeout: defaultHTTPTimeout,\n\t\t}\n\t}\n\n\treturn &Client{\n\t\turl: url,\n\t\thttpClient: httpClient,\n\t\tcustomHeaders: customHeaders,\n\t}\n}", "func NewClient() (*Client, error) {\n\tvar seedTickers = []string{\"AAPL\", \"GOOG\", \"MSFT\"}\n\tjar, err := cookiejar.New(nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\thttpClient := &http.Client{Jar: jar}\n\tc := Client{httpClient: httpClient}\n\n\ti := rand.Intn(len(seedTickers))\n\tticker := seedTickers[i]\n\tcrumb, err := getCrumb(c.httpClient, ticker)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc.crumb = crumb\n\n\treturn &c, nil\n\n}", "func NewClient(varsURL string) *Client {\n\treturn &Client{varsURL}\n}", "func NewScrape(cfg *domain.Config) *Scrape {\n\treturn &Scrape{\n\t\tcfg: cfg,\n\t}\n}", "func New(c *http.Client, url, user, pass string) Client {\n\treturn &client{\n\t\tclient: c,\n\t\turl: url,\n\t\tuser: user,\n\t\tpass: pass,\n\t}\n}", "func NewClient(baseURL string) *Client {\n\tu, err := url.Parse(baseURL)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"service\": \"HTTP-Client\",\n\t\t}).WithError(err).Error(\"failed to create HTTP client\")\n\t}\n\n\tport, err := strconv.Atoi(u.Port())\n\tif err != nil {\n\t\tport = 80\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{\n\t\t\tInsecureSkipVerify: true,\n\t\t},\n\t}\n\n\treturn &Client{\n\t\tBaseURL: baseURL,\n\t\tPort: port,\n\t\tScheme: u.Scheme,\n\t\tContentType: \"application/json\",\n\t\tClient: &http.Client{\n\t\t\tTimeout: 5 * time.Second,\n\t\t\tTransport: tr,\n\t\t},\n\t}\n}", "func NewScrapeServiceClient(httpClient connect_go.HTTPClient, baseURL string, opts ...connect_go.ClientOption) ScrapeServiceClient {\n\tbaseURL = strings.TrimRight(baseURL, \"/\")\n\treturn &scrapeServiceClient{\n\t\ttargets: connect_go.NewClient[v1alpha1.TargetsRequest, v1alpha1.TargetsResponse](\n\t\t\thttpClient,\n\t\t\tbaseURL+\"/parca.scrape.v1alpha1.ScrapeService/Targets\",\n\t\t\topts...,\n\t\t),\n\t}\n}", "func NewFromURL(base *url.URL) *Client {\n\n\tif baseStr := base.String(); len(baseStr) > 0 {\n\n\t\tLogger.Debug(\"Creating Marathon Client from url.URL = %s\", base.String())\n\t\tbaseURL, err := url.Parse(baseStr)\n\t\tif err != nil {\n\t\t\tLogger.Debug(\"Invalid baseURL\")\n\t\t\treturn nil\n\t\t}\n\n\t\t_client := &Client{}\n\t\treturn _client.New(baseURL)\n\t}\n\treturn nil\n}", "func NewClient() *Client {\n\tclientId := os.Getenv(clientIdEnvVar)\n\tif clientId == \"\" {\n\t\tlog.Fatalf(\"Please provide client id by setting %s\", clientIdEnvVar)\n\t}\n\treturn &Client{\n\t\thttp: &http.Client{},\n\t\tbaseAddr: \"https://api.seatGeek.com/2/\",\n\t\tclientId: clientId}\n}", "func NewClient(owner string, url string) *Client {\n\n\tclient := &Client{\n\t\turl: url,\n\t\towner: owner,\n\t}\n\n\treturn client\n}", "func NewClient(address string, httpService httpservice.HTTPService) (*Client, error) {\n\tvar httpClient *http.Client\n\taddressURL, err := url.Parse(address)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to parse marketplace address\")\n\t}\n\tif addressURL.Hostname() == \"localhost\" || addressURL.Hostname() == \"127.0.0.1\" {\n\t\thttpClient = httpService.MakeClient(true)\n\t} else {\n\t\thttpClient = httpService.MakeClient(false)\n\t}\n\n\treturn &Client{\n\t\taddress: address,\n\t\thttpClient: httpClient,\n\t}, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Announce sends a Announce request to the tracker.
func (t *Client) Announce(c context.Context, req AnnounceRequest) ( resp AnnounceResponse, err error) { if req.PeerID.IsZero() { if t.ID.IsZero() { req.PeerID = metainfo.NewRandomHash() } else { req.PeerID = t.ID } } err = t.send(c, t.AnnounceURL, req.ToQuery(), &resp) return }
[ "func (b *Builder) Announce() (err error) {\n\targs := &rpc.AnnounceArgs{\n\t\tGOOS: b.b.GOOS(),\n\t\tGOARCH: b.b.GOARCH(),\n\t\tType: \"Builder\",\n\t\tURL: b.base,\n\t}\n\treply := new(rpc.AnnounceReply)\n\tif err = b.tcl.Call(\"Tracker.Announce\", args, reply); err != nil {\n\t\treturn\n\t}\n\tb.key = reply.Key\n\treturn\n}", "func (r *Runner) Announce() (err error) {\n\targs := &rpc.AnnounceArgs{\n\t\tGOOS: runtime.GOOS,\n\t\tGOARCH: runtime.GOARCH,\n\t\tType: \"Runner\",\n\t\tURL: r.base,\n\t}\n\treply := new(rpc.AnnounceReply)\n\tif err = r.tcl.Call(\"Tracker.Announce\", args, reply); err != nil {\n\t\treturn\n\t}\n\tr.key = reply.Key\n\treturn\n}", "func SendAnnounce() {\n\tvar pkt MeshPkt\n\tvar annc AnnouncePkt\n\n\tpkt.PktPayload = TypeAnnounce\n\tpkt.SenderID = MySID\n\n\tannc.NumServices = 3\n\tannc.Services = [32]int8{ServiceLookup, ServicePubChat, ServicePrivMsg}\n\n\tvar buffer bytes.Buffer\n\n\tbinary.Write(&buffer, binary.BigEndian, &pkt)\n\tbinary.Write(&buffer, binary.BigEndian, &annc)\n\n\tsendPkt(\"255.255.255.255\", 8032, buffer.Bytes())\n}", "func (tracker *Tracker) HandleAnnounce(ctx *gin.Context) {\n\tstats.RegisterEvent(stats.EV_ANNOUNCE)\n\tr := db.Pool.Get()\n\tdefer r.Close()\n\tif r.Err() != nil {\n\t\tstats.RegisterEvent(stats.EV_ANNOUNCE_FAIL)\n\t\tctx.Error(r.Err()).SetMeta(errMeta(\n\t\t\tMSG_GENERIC_ERROR,\n\t\t\t\"Internal error, HALP\",\n\t\t\tlog.Fields{\"fn\": \"HandleAnnounce\"},\n\t\t\tlog.ErrorLevel,\n\t\t))\n\t\treturn\n\t}\n\n\tlog.Debugln(ctx.Request.RequestURI)\n\tann, err := NewAnnounce(ctx)\n\tif err != nil {\n\t\tstats.RegisterEvent(stats.EV_ANNOUNCE_FAIL)\n\t\tctx.Error(err).SetMeta(errMeta(\n\t\t\tMSG_QUERY_PARSE_FAIL,\n\t\t\t\"Failed to parse announce\",\n\t\t\tlog.Fields{\n\t\t\t\t\"fn\": \"HandleAnnounce\",\n\t\t\t\t\"remote_ip\": ctx.Request.RemoteAddr,\n\t\t\t\t\"uri\": ctx.Request.RequestURI,\n\t\t\t},\n\t\t\tlog.ErrorLevel,\n\t\t))\n\t\treturn\n\t}\n\n\tinfo_hash_hex := fmt.Sprintf(\"%x\", ann.InfoHash)\n\tlog.WithFields(log.Fields{\n\t\t\"ih\": info_hash_hex,\n\t\t\"ip\": ann.IPv4,\n\t\t\"port\": ann.Port,\n\t\t\"up\": util.Bytes(ann.Uploaded),\n\t\t\"dn\": util.Bytes(ann.Downloaded),\n\t\t\"left\": util.Bytes(ann.Left),\n\t\t\"event\": ann.Event,\n\t}).Debug(\"Announce event\")\n\n\tpasskey := ctx.Param(\"passkey\")\n\n\tuser_id := tracker.findUserID(passkey)\n\tif user_id == 0 {\n\t\tstats.RegisterEvent(stats.EV_INVALID_PASSKEY)\n\t\tctx.Error(errors.New(\"Invalid passkey\")).SetMeta(errMeta(\n\t\t\tMSG_INVALID_AUTH,\n\t\t\t\"Invalid passkey supplied\",\n\t\t\tlog.Fields{\"fn\": \"HandleAnnounce\", \"passkey\": passkey},\n\t\t\tlog.ErrorLevel,\n\t\t))\n\t\treturn\n\t}\n\n\tuser := tracker.FindUserByID(user_id)\n\tif !user.CanLeech && ann.Left > 0 {\n\t\tctx.Error(errors.New(\"Leech disabled for user\")).SetMeta(errMeta(\n\t\t\tMSG_GENERIC_ERROR,\n\t\t\t\"Leeching not allowed for user\",\n\t\t\tlog.Fields{\"fn\": \"HandleAnnounce\", \"passkey\": passkey},\n\t\t\tlog.ErrorLevel,\n\t\t))\n\t\treturn\n\t}\n\tif !user.Enabled {\n\t\tctx.Error(errors.New(\"Disabled user\")).SetMeta(errMeta(\n\t\t\tMSG_GENERIC_ERROR,\n\t\t\t\"User disabled\",\n\t\t\tlog.Fields{\"fn\": \"HandleAnnounce\", \"passkey\": passkey},\n\t\t\tlog.ErrorLevel,\n\t\t))\n\t\treturn\n\t}\n\n\tif !tracker.IsValidClient(ann.PeerID) {\n\t\tstats.RegisterEvent(stats.EV_INVALID_CLIENT)\n\t\tctx.Error(errors.New(\"Banned client\")).SetMeta(errMeta(\n\t\t\tMSG_GENERIC_ERROR,\n\t\t\t\"Banned client, check wiki for whitelisted clients\",\n\t\t\tlog.Fields{\n\t\t\t\t\"fn\": \"HandleAnnounce\",\n\t\t\t\t\"user_id\": user.UserID,\n\t\t\t\t\"user_name\": user.Username,\n\t\t\t\t\"peer_id\": ann.PeerID[0:8],\n\t\t\t},\n\t\t\tlog.ErrorLevel,\n\t\t))\n\t\treturn\n\t}\n\n\ttorrent := tracker.FindTorrentByInfoHash(info_hash_hex)\n\tif torrent == nil {\n\t\tstats.RegisterEvent(stats.EV_INVALID_INFOHASH)\n\t\tctx.Error(errors.New(\"Invalid info hash\")).SetMeta(errMeta(\n\t\t\tMSG_INFO_HASH_NOT_FOUND,\n\t\t\t\"Torrent not found, try TPB\",\n\t\t\tlog.Fields{\n\t\t\t\t\"fn\": \"HandleAnnounce\",\n\t\t\t\t\"user_id\": user.UserID,\n\t\t\t\t\"user_name\": user.Username,\n\t\t\t\t\"info_hash\": info_hash_hex,\n\t\t\t},\n\t\t\tlog.WarnLevel,\n\t\t))\n\t\treturn\n\t} else if !torrent.Enabled {\n\t\tstats.RegisterEvent(stats.EV_INVALID_INFOHASH)\n\t\tctx.Error(errors.New(\"Torrent not enabled\")).SetMeta(errMeta(\n\t\t\tMSG_INFO_HASH_NOT_FOUND,\n\t\t\ttorrent.DelReason(),\n\t\t\tlog.Fields{\n\t\t\t\t\"fn\": \"HandleAnnounce\",\n\t\t\t\t\"user_id\": user.UserID,\n\t\t\t\t\"user_name\": user.Username,\n\t\t\t\t\"info_hash\": info_hash_hex,\n\t\t\t},\n\t\t\tlog.WarnLevel,\n\t\t))\n\t\treturn\n\t}\n\n\tpeer := torrent.findPeer(ann.PeerID)\n\tif peer == nil {\n\t\tlog.Debug(\"No existing peer found\")\n\t\tpeer = NewPeer(ann.PeerID, ann.IPv4.String(), ann.Port, torrent, user)\n\t\t// torrent.AddPeer(r, peer)\n\t}\n\n\tpeer_diff := PeerDiff{User: user, Torrent: torrent}\n\t// user update MUST happen after peer update since we rely on the old dl/ul values\n\tpeer.Update(ann, &peer_diff, torrent.Seeders)\n\ttorrent.Update(ann)\n\tuser.Update(ann, &peer_diff, torrent.MultiUp, torrent.MultiDn)\n\n\tif ann.Event == STOPPED {\n\t\tlog.Debug(\"Removing peer due to stop announce\")\n\t\ttorrent.DelPeer(r, peer)\n\t} else {\n\t\tif !torrent.HasPeer(peer) {\n\t\t\ttorrent.AddPeer(r, peer)\n\t\t}\n\t}\n\n\tif ann.Event == STOPPED {\n\t\t// Remove from torrents active peer set\n\t\tr.Send(\"SREM\", torrent.TorrentPeersKey, ann.PeerID)\n\n\t\tr.Send(\"SREM\", user.KeyActive, torrent.TorrentID)\n\n\t\t// Mark the peer as inactive\n\t\tr.Send(\"HSET\", peer.KeyPeer, \"active\", 0)\n\n\t\tr.Send(\"DEL\", peer.KeyTimer)\n\n\t\tif peer.IsHNR() {\n\t\t\tuser.AddHNR(r, torrent.TorrentID)\n\t\t}\n\t} else if ann.Event == COMPLETED {\n\n\t\t// Remove the torrent from the users incomplete set\n\t\tr.Send(\"SREM\", user.KeyIncomplete, torrent.TorrentID)\n\n\t\t// Remove the torrent from the users incomplete set\n\t\tr.Send(\"SADD\", user.KeyComplete, torrent.TorrentID)\n\n\t\t// Remove from the users hnr list if it exists\n\t\tr.Send(\"SREM\", user.KeyHNR, torrent.TorrentID)\n\n\t} else if ann.Event == STARTED {\n\t\t// Make sure we account for a user completing a torrent outside of\n\t\t// our view, or resuming from previously completions\n\t\tif peer.IsSeeder() {\n\t\t\tr.Send(\"SREM\", user.KeyHNR, torrent.TorrentID)\n\t\t\tr.Send(\"SREM\", user.KeyIncomplete, torrent.TorrentID)\n\t\t\tr.Send(\"SADD\", user.KeyComplete, torrent.TorrentID)\n\t\t} else {\n\t\t\tr.Send(\"SREM\", user.KeyComplete, torrent.TorrentID)\n\t\t\tr.Send(\"SADD\", user.KeyIncomplete, torrent.TorrentID)\n\t\t}\n\t}\n\n\tif ann.Event != STOPPED {\n\n\t\t// Add peer to torrent active peers\n\t\tr.Send(\"SADD\", torrent.TorrentPeersKey, ann.PeerID)\n\n\t\t// Add to users active torrent set\n\t\tr.Send(\"SADD\", user.KeyActive, torrent.TorrentID)\n\n\t\t// Refresh the peers expiration timer\n\t\t// If this expires, the peer reaper takes over and removes the\n\t\t// peer from torrents in the case of a non-clean client shutdown\n\t\tr.Send(\"SETEX\", peer.KeyTimer, conf.Config.ReapInterval, 1)\n\t}\n\tr.Flush()\n\n\tSyncEntityC <- torrent\n\tSyncEntityC <- user\n\n\tdict := bencode.Dict{\n\t\t\"complete\": torrent.Seeders,\n\t\t\"incomplete\": torrent.Leechers,\n\t\t\"interval\": conf.Config.AnnInterval,\n\t\t\"min interval\": conf.Config.AnnIntervalMin,\n\t}\n\n\tpeers := torrent.GetPeers(ann.NumWant, peer.GetCoord())\n\tif peers != nil {\n\t\tdict[\"peers\"] = MakeCompactPeers(peers, ann.PeerID)\n\t} else {\n\t\tdict[\"peers\"] = []byte{}\n\t}\n\tvar out_bytes bytes.Buffer\n\tencoder := bencode.NewEncoder(&out_bytes)\n\n\ter_msg_encoded := encoder.Encode(dict)\n\tif er_msg_encoded != nil {\n\t\tstats.RegisterEvent(stats.EV_ANNOUNCE_FAIL)\n\t\tctx.Error(er_msg_encoded).SetMeta(errMeta(\n\t\t\tMSG_GENERIC_ERROR,\n\t\t\t\"Internal error\",\n\t\t\tlog.Fields{\n\t\t\t\t\"fn\": \"HandleAnnounce\",\n\t\t\t\t\"user_id\": user.UserID,\n\t\t\t\t\"user_name\": user.Username,\n\t\t\t\t\"info_hash\": info_hash_hex,\n\t\t\t},\n\t\t\tlog.DebugLevel,\n\t\t))\n\t\treturn\n\t}\n\n\tctx.String(MSG_OK, out_bytes.String())\n}", "func Announce(url string) (*AnnounceResponse, error) {\n\tres, err := http.Get(url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbuf := new(bytes.Buffer)\n\tbuf.ReadFrom(res.Body)\n\tannRes := new(AnnounceResponse)\n\terr = bencode.Unmarshal(buf, annRes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn annRes, nil\n}", "func (u UDPTracker) Announce(query url.Values, file data.FileRecord) []byte {\n\t// Create UDP announce response\n\tannounce := udp.AnnounceResponse{\n\t\tAction: 1,\n\t\tTransID: u.TransID,\n\t\tInterval: uint32(common.Static.Config.Interval),\n\t\tLeechers: uint32(file.Leechers()),\n\t\tSeeders: uint32(file.Seeders()),\n\t}\n\n\t// Convert to UDP byte buffer\n\tannounceBuf, err := announce.MarshalBinary()\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn u.Error(\"Could not create UDP announce response\")\n\t}\n\n\t// Numwant\n\tnumwant, err := strconv.Atoi(query.Get(\"numwant\"))\n\tif err != nil {\n\t\tnumwant = 50\n\t}\n\n\t// Add compact peer list\n\tres := bytes.NewBuffer(announceBuf)\n\terr = binary.Write(res, binary.BigEndian, file.PeerList(query.Get(\"ip\"), numwant))\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn u.Error(\"Could not create UDP announce response\")\n\t}\n\n\treturn res.Bytes()\n}", "func (t *HttpTracker) Announce(req *Request) (resp *Response, err error) {\n\t//if req == nil {\n\t//\treturn\n\t//}\n\t// http client\n\tvar client http.Client\n\n\tclient.Transport = &http.Transport{\n\t\tDial: func(_, _ string) (c net.Conn, e error) {\n\t\t\tvar a net.Addr\n\t\t\tt.resolving.Lock()\n\t\t\tif t.shouldResolve() {\n\t\t\t\tvar h, p string\n\t\t\t\t// XXX: hack\n\t\t\t\tif strings.Index(t.u.Host, \":\") == -1 {\n\t\t\t\t\tt.u.Host += \":80\"\n\t\t\t\t}\n\t\t\t\th, p, e = net.SplitHostPort(t.u.Host)\n\t\t\t\tif e == nil {\n\t\t\t\t\ta, e = req.GetNetwork().Lookup(h, p)\n\t\t\t\t\tif e == nil {\n\t\t\t\t\t\tt.addr = a\n\t\t\t\t\t\tt.lastResolved = time.Now()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\ta = t.addr\n\t\t\t}\n\t\t\tt.resolving.Unlock()\n\t\t\tif e == nil {\n\t\t\t\tc, e = req.GetNetwork().Dial(a.Network(), a.String())\n\t\t\t}\n\t\t\treturn\n\t\t},\n\t}\n\n\tresp = new(Response)\n\tinterval := 30\n\t// build query\n\tvar u *url.URL\n\tu, err = url.Parse(t.u.String())\n\tif err == nil {\n\t\tv := u.Query()\n\t\tn := req.GetNetwork()\n\t\ta := n.Addr()\n\t\thost, _, _ := net.SplitHostPort(a.String())\n\t\tif a.Network() == \"i2p\" {\n\t\t\thost += \".i2p\"\n\t\t\treq.Compact = true\n\t\t}\n\t\tv.Add(\"ip\", host)\n\t\tv.Add(\"info_hash\", string(req.Infohash.Bytes()))\n\t\tv.Add(\"peer_id\", string(req.PeerID.Bytes()))\n\t\tv.Add(\"port\", fmt.Sprintf(\"%d\", req.Port))\n\t\tv.Add(\"numwant\", fmt.Sprintf(\"%d\", req.NumWant))\n\t\tv.Add(\"left\", fmt.Sprintf(\"%d\", req.Left))\n\t\tif req.Event != Nop {\n\t\t\tv.Add(\"event\", req.Event.String())\n\t\t}\n\t\tv.Add(\"downloaded\", fmt.Sprintf(\"%d\", req.Downloaded))\n\t\tv.Add(\"uploaded\", fmt.Sprintf(\"%d\", req.Uploaded))\n\n\t\t// compact response\n\t\tif req.Compact || u.Path != \"/a\" {\n\t\t\treq.Compact = true\n\t\t\tv.Add(\"compact\", \"1\")\n\t\t}\n\t\tu.RawQuery = v.Encode()\n\t\tvar r *http.Response\n\t\tlog.Debugf(\"%s announcing\", t.Name())\n\t\tr, err = client.Get(u.String())\n\t\tif err == nil {\n\t\t\tdefer r.Body.Close()\n\t\t\tdec := bencode.NewDecoder(r.Body)\n\t\t\tif req.Compact {\n\t\t\t\tcresp := new(compactHttpAnnounceResponse)\n\t\t\t\terr = dec.Decode(cresp)\n\t\t\t\tif err == nil {\n\t\t\t\t\tinterval = cresp.Interval\n\t\t\t\t\tvar cpeers string\n\n\t\t\t\t\t_, ok := cresp.Peers.(string)\n\t\t\t\t\tif ok {\n\t\t\t\t\t\tcpeers = cresp.Peers.(string)\n\t\t\t\t\t\tl := len(cpeers) / 32\n\t\t\t\t\t\tfor l > 0 {\n\t\t\t\t\t\t\tvar p common.Peer\n\t\t\t\t\t\t\t// TODO: bounds check\n\t\t\t\t\t\t\tcopy(p.Compact[:], cpeers[(l-1)*32:l*32])\n\t\t\t\t\t\t\tresp.Peers = append(resp.Peers, p)\n\t\t\t\t\t\t\tl--\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfullpeers, ok := cresp.Peers.([]interface{})\n\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\tfor idx := range fullpeers {\n\t\t\t\t\t\t\t\t// XXX: this is horribad :DDDDDDDDD\n\t\t\t\t\t\t\t\tvar peer map[string]interface{}\n\t\t\t\t\t\t\t\tpeer, ok = fullpeers[idx].(map[string]interface{})\n\t\t\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\t\t\tvar p common.Peer\n\t\t\t\t\t\t\t\t\tp.IP = fmt.Sprintf(\"%s\", peer[\"ip\"])\n\t\t\t\t\t\t\t\t\tport, ok := peer[\"port\"].(int64)\n\t\t\t\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\t\t\t\tp.Port = int(port)\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\tresp.Peers = append(resp.Peers, p)\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\n\t\t\t\t\tif len(cresp.Error) > 0 {\n\t\t\t\t\t\terr = errors.New(cresp.Error)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// decode non compact response\n\t\t\t\terr = dec.Decode(resp)\n\t\t\t\tinterval = resp.Interval\n\t\t\t\tif len(resp.Error) > 0 {\n\t\t\t\t\terr = errors.New(resp.Error)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tif err == nil {\n\t\tlog.Infof(\"%s got %d peers for %s\", t.Name(), len(resp.Peers), req.Infohash.Hex())\n\t} else {\n\t\tlog.Warnf(\"%s got error while announcing: %s\", t.Name(), err)\n\t}\n\tif interval == 0 {\n\t\tinterval = 60\n\t}\n\tresp.NextAnnounce = time.Now().Add(time.Second * time.Duration(interval))\n\treturn\n}", "func (m *DHTModule) Announce(key string) {\n\n\tif m.IsAttached() {\n\t\tif err := m.Client.DHT().Announce(key); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\treturn\n\t}\n\n\tm.dht.Announce(announcer.ObjTypeAny, \"\", []byte(key), nil)\n}", "func (s *serverWrapper) handleAnnounce() http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tdecoder := json.NewDecoder(r.Body)\n\t\tvar v pb.AnnounceDetails\n\t\terr := decoder.Decode(&v)\n\t\tif err != nil {\n\t\t\tlog.Printf(invalidJSONErrorWithPrint, err)\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\t// We use their logged in GlobalID, since the client shouldn't\n\t\t// need to care about that detail.\n\t\tuid, err := s.getSessionGlobalID(r)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Access denied in handleAnnounce: %v\", err)\n\t\t\tw.WriteHeader(http.StatusForbidden)\n\t\t\treturn\n\t\t}\n\n\t\tv.AnnouncerId = uid\n\t\tts := ptypes.TimestampNow()\n\t\tv.AnnounceTime = ts\n\n\t\tctx, cancel := context.WithTimeout(context.Background(), defaultTimeoutDuration)\n\t\tdefer cancel()\n\t\tresp, err := s.announce.SendAnnounceActivity(ctx, &v)\n\t\tif resp.ResultType == pb.ResultType_ERROR {\n\t\t\tlog.Printf(\"Could not send announce: %#v\", resp.Error)\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\treturn\n\t\t} else if resp.ResultType == pb.ResultType_ERROR_400 {\n\t\t\tlog.Printf(\"Could not send announce: %#v\", resp.Error)\n\t\t\tw.WriteHeader(http.StatusBadRequest)\n\t\t\treturn\n\t\t} else if err != nil {\n\t\t\tlog.Printf(\"Could not send announce: %#v\", err)\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t}\n}", "func Announce() {\n\tUpdateFileIndex()\n\n\tvar buffer bytes.Buffer\n\n\tbuffer.WriteString(\"forban;name;\")\n\tbuffer.WriteString(MyName)\n\tbuffer.WriteString(\";uuid;\")\n\tbuffer.WriteString(MyUuid)\n\tbuffer.WriteString(\";hmac;\")\n\tbuffer.WriteString(GetIndexHmac())\n\n\tif DisableIPv4 == false {\n\t\tdst, err := net.ResolveUDPAddr(\"udp\", \"255.255.255.255:12555\")\n\t\tif err != nil {\n\t\t\tlog.Fatal(\"NET \", err)\n\t\t}\n\n\t\tlog.Debug(\"NET Sending IPv4 announcement\")\n\t\tif _, err := ServerConn.WriteTo(buffer.Bytes(), dst); err != nil {\n\t\t\tlog.Fatal(\"NET \", err)\n\t\t}\n\t}\n\n\tif DisableIPv6 == false {\n\t\t//ifaces, _ := net.Interfaces()\n\t\tfor _, iface := range Interfaces {\n\t\t\tdst6, err := net.ResolveUDPAddr(\"udp\", \"[ff02::1%\"+iface+\"]:12555\")\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(\"NET \", err)\n\t\t\t}\n\n\t\t\tlog.Debug(\"NET Sending IPv6 announcement\")\n\t\t\tif _, err := ServerConn.WriteTo(buffer.Bytes(), dst6); err != nil {\n\t\t\t\tlog.Fatal(\"NET \", err)\n\t\t\t}\n\t\t}\n\t}\n}", "func Announce(event Event) {\n\tglobalEvents.Announce(event)\n}", "func (h *Host) managedAnnounce(addr modules.NetAddress) error {\n\t// The wallet needs to be unlocked to add fees to the transaction, and the\n\t// host needs to have an active unlock hash that renters can make payment\n\t// to.\n\tif !h.wallet.Unlocked() {\n\t\treturn errAnnWalletLocked\n\t}\n\n\th.mu.Lock()\n\tpubKey := h.publicKey\n\tsecKey := h.secretKey\n\terr := h.checkUnlockHash()\n\th.mu.Unlock()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Create the announcement that's going to be added to the arbitrary data\n\t// field of the transaction.\n\tsignedAnnouncement, err := modules.CreateAnnouncement(addr, pubKey, secKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Create a transaction, with a fee, that contains the full announcement.\n\ttxnBuilder := h.wallet.StartTransaction()\n\t_, fee := h.tpool.FeeEstimation()\n\tfee = fee.Mul64(500) // Estimated txn size (in bytes) of a host announcement.\n\terr = txnBuilder.FundSiacoins(fee)\n\tif err != nil {\n\t\ttxnBuilder.Drop()\n\t\treturn err\n\t}\n\t_ = txnBuilder.AddMinerFee(fee)\n\t_ = txnBuilder.AddArbitraryData(signedAnnouncement)\n\ttxnSet, err := txnBuilder.Sign(true)\n\tif err != nil {\n\t\ttxnBuilder.Drop()\n\t\treturn err\n\t}\n\n\t// Add the transactions to the transaction pool.\n\terr = h.tpool.AcceptTransactionSet(txnSet)\n\tif err != nil {\n\t\ttxnBuilder.Drop()\n\t\treturn err\n\t}\n\n\th.mu.Lock()\n\th.announced = true\n\th.mu.Unlock()\n\th.log.Printf(\"INFO: Successfully announced as %v\", addr)\n\treturn nil\n}", "func (teamID TeamID) SendAnnounce(message string) error {\n\t// for each agent on the team\n\t// determine which messaging protocols are enabled for gid\n\t// pick optimal\n\n\t// ok, err := SendMessage(gid, message)\n\treturn nil\n}", "func (n *Interface) StartAnnounce() {\n\tn.ad.StartAnnounceDaemon()\n}", "func announcePeriodically(t tracker, transfer *Transfer, cancel <-chan struct{}, startEvent trackerEvent, eventC <-chan trackerEvent, responseC chan<- *announceResponse) {\n\tvar nextAnnounce time.Duration\n\tvar retry = *defaultRetryBackoff\n\n\tannounce := func(e trackerEvent) {\n\t\tr, err := t.Announce(transfer, e, cancel)\n\t\tif err != nil {\n\t\t\tr = &announceResponse{Error: err}\n\t\t\tnextAnnounce = retry.NextBackOff()\n\t\t} else {\n\t\t\tretry.Reset()\n\t\t\tnextAnnounce = r.Interval\n\t\t}\n\t\tselect {\n\t\tcase responseC <- r:\n\t\tcase <-cancel:\n\t\t\treturn\n\t\t}\n\t}\n\n\tannounce(startEvent)\n\tfor {\n\t\tselect {\n\t\tcase <-time.After(nextAnnounce):\n\t\t\tannounce(eventNone)\n\t\tcase e := <-eventC:\n\t\t\tannounce(e)\n\t\tcase <-cancel:\n\t\t\treturn\n\t\t}\n\t}\n}", "func (c *chatRoom) announce(msg string) {\n\tc.messages <- \"* \" + msg + \" *\"\n}", "func (g *GateKeeper) announce() error {\n\tm, err := json.Marshal(g.Meta)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog.Debug().Msg(\"Starting to announce API to etcd\")\n\t_, err = (*g.etcd).Set(context.Background(), \"/meta/gatekeeper\", string(m), nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog.Info().Msg(\"Gatekeeper registered in etcd.\")\n\treturn nil\n}", "func (duel *Duel) Announce() {\n\tfmt.Printf(\"%s\\n\\nversus\\n\\n%s\\n\\n\", duel.fighter1, duel.fighter2)\n}", "func TestAnnounceRequest(t *testing.T) {\n\tlog.Println(\"TestAnnounceRequest()\")\n\n\t// Generate mock AnnounceRequest\n\tannounce := AnnounceRequest{\n\t\tConnID: 1,\n\t\tAction: 1,\n\t\tTransID: 1,\n\t\tInfoHash: []byte(\"abcdef0123456789abcd\"),\n\t\tPeerID: []byte(\"abcdef0123456789abcd\"),\n\t\tDownloaded: 0,\n\t\tLeft: 0,\n\t\tUploaded: 0,\n\t\tEvent: 0,\n\t\tIP: 1234,\n\t\tKey: 1234,\n\t\tNumwant: 50,\n\t\tPort: 8080,\n\t}\n\n\t// Marshal to binary representation\n\tout, err := announce.MarshalBinary()\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to marshal AnnounceRequest to binary: %s\", err.Error())\n\t}\n\n\t// Unmarshal announce from binary representation\n\tannounce2 := new(AnnounceRequest)\n\tif err := announce2.UnmarshalBinary(out); err != nil {\n\t\tt.Fatalf(\"Failed to unmarshal AnnounceRequest from binary: %s\", err.Error())\n\t}\n\n\t// Verify announces are identical\n\tif announce.ConnID != announce2.ConnID || !bytes.Equal(announce.InfoHash, announce2.InfoHash) {\n\t\tt.Fatalf(\"AnnounceRequest results do not match\")\n\t}\n\n\t// Convert announce to url.Values map\n\tquery := announce.ToValues()\n\n\t// Verify conversion occurred properly\n\tif query.Get(\"downloaded\") != \"0\" || query.Get(\"numwant\") != \"50\" {\n\t\tt.Fatalf(\"AnnounceRequest values map results are not correct\")\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Scrape sends a Scrape request to the tracker.
func (t *Client) Scrape(c context.Context, infohashes []metainfo.Hash) ( resp ScrapeResponse, err error) { hs := make([]string, len(infohashes)) for i, h := range infohashes { hs[i] = h.BytesString() } err = t.send(c, t.ScrapeURL, url.Values{"info_hash": hs}, &resp) return }
[ "func (s *Scraper) Scrape() {\n\ts.trigger <- false\n}", "func (c *Client) Scrape() {\n\tfor range time.Tick(c.interval) {\n\n\t\tallstats := c.getStatistics()\n\t\t//Set the metrics\n\t\tc.setMetrics(allstats.status, allstats.stats, allstats.logStats)\n\n\t\tlog.Printf(\"New tick of statistics: %s\", allstats.stats.ToString())\n\t}\n}", "func (s *scraper) Scrape() {\n\tgo s.process()\n\ts.scrapeStartUrl()\n\ts.wg.Wait() // wait until all links are scraped.\n\tclose(s.done)\n}", "func ScrapeAndSend(url string) ([]byte, error) {\n\tlog.Println(\"url received is : \", url)\n\n\tproductInfo := doScrape(url)\n\n\tdata, err := json.Marshal(productInfo)\n\tif err != nil {\n\t\tlog.Println(\"error while marsheling the scrape results. Error : \", err.Error())\n\t}\n\n\tgo func([]byte) {\n\t\terr = sendToPersist(data)\n\t\tif err != nil {\n\t\t\tlog.Println(\"error while sending the scrape results. Error : \", err.Error())\n\t\t}\n\t}(data)\n\n\treturn data, nil\n}", "func (s *Scraper) Scrape() {\n\tdefer elapsedTime(\"Scrape()\", time.Now())\n\tlog.Println(\"Beginning scrape...\")\n\n\trespBytes, err := s.getFitBarkData(hourly)\n\n\tif err != nil {\n\t\tlog.Fatalln(\"ERROR: \", err)\n\t}\n\n\thourlyResponse := &fitBarkResponseHourly{}\n\trespErr := json.Unmarshal(respBytes, hourlyResponse)\n\n\tif respErr != nil {\n\t\tlog.Fatalln(\"ERROR: Failed unmarshalling FitBark Hourly response: \", err)\n\t}\n\n\tlog.Println(fmt.Sprintf(\"Hourly data retrieved successfully. Records: %d\", len(hourlyResponse.ActivitySeries.Records)))\n\n\thourlyResponse.formatDates()\n\n\thourlyResponse.save(s.config.db)\n\n\trespBytes, err = s.getFitBarkData(daily)\n\n\tif err != nil {\n\t\tlog.Fatalln(\"ERROR: \", err)\n\t}\n\n\tdailyResponse := &fitBarkResponseDaily{}\n\trespErr = json.Unmarshal(respBytes, dailyResponse)\n\n\tif respErr != nil {\n\t\tlog.Fatalln(\"ERROR: Failed unmarshalling FitBark Daily response: \", err)\n\t}\n\n\tlog.Println(fmt.Sprintf(\"Daily data retrieved successfully. Records: %d\", len(dailyResponse.ActivitySeries.Records)))\n\n\tdailyResponse.save(s.config.db)\n\n}", "func (s *Scraper) Scrape(message string) {\n buffer := strings.NewReader(message)\n\n attempts := 0\n\n for {\n resp, err := http.Post(s.uri, \"application/json\", buffer)\n\n if err != nil {\n WriteStderr([]byte(fmt.Sprintf(\"waiting %s ... %d\\n\", s.uri, attempts)))\n time.Sleep(500 * time.Millisecond)\n attempts += 1\n\n if attempts > maxAttempts {\n panic(err)\n }\n\n continue\n } else {\n resp.Body.Close()\n break\n }\n }\n}", "func (n *NewsPaper) Scrape() {\n\tvar err error\n\tvar resp *http.Response\n\tvar fd *os.File\n\tvar out string\n\tclog := logrus.WithField(\"newspaper\", n.Name)\n\n\tc := &http.Client{\n\t\tTimeout: time.Second * 10,\n\t}\n\tif resp, err = c.Get(n.URL); err != nil {\n\t\tclog.WithError(err).Error(\"Couldn't scrape\")\n\t\treturn\n\t}\n\tdefer resp.Body.Close()\n\n\tif fd, err = n.CreateDumpFile(); err != nil {\n\t\tclog.WithError(err).Error(\"Couldn't create dump file\")\n\t\treturn\n\t}\n\tout, err = html2text.FromReader(resp.Body)\n\tif _, err = fd.WriteString(out); err != nil {\n\t\tclog.WithError(err).Error(\"Couldn't copy ouput\")\n\t}\n}", "func (scraper *Scraper) Scrape(sess *session.Session) map[string][]*ScrapeResult {\n\tstatus := \"success\"\n\tstart := time.Now()\n\n\tresponse, err := scraper.Fn(sess)\n\tif err != nil {\n\t\tlog.Error(err)\n\t\tstatus = \"error\"\n\t}\n\n\tduration := time.Since(start)\n\tmetrics.ScrapeDurationHistogram.WithLabelValues(scraper.ID, status).Observe(duration.Seconds())\n\n\treturn response\n}", "func (c Crawler) Crawl() {\n\tc.processing()\n\tdefer c.done()\n\n\tq := c.fetcher.Start()\n\n\tlog.Printf(\"type=startCrawling jobUUID=%s url=%s\\n\", c.jobUUID(), c.msg.URL)\n\tq.SendStringGet(c.msg.URL)\n\n\tq.Close()\n\tlog.Printf(\"type=endCrawling jobUUID=%s url=%s\\n\", c.jobUUID(), c.msg.URL)\n}", "func (scraper *Scraper) Scrape() (err error) {\n\tdefer func() {\n\t\tif err == nil {\n\t\t\terr = scraper.Save()\n\t\t}\n\t}()\n\n\tscraper.GeocodeInfo, err = geocoding.Geocode(scraper.Address)\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif scraper.GeocodeInfo.Status != \"OK\" {\n\t\terr = errors.New(\"geocode failed\")\n\n\t\treturn\n\t}\n\n\tlat := scraper.GeocodeInfo.Results[0].Geometry.Location.Latitude\n\tlng := scraper.GeocodeInfo.Results[0].Geometry.Location.Longitude\n\n\tscraper.NBNInfo, err = nbn.RolloutInfo(lat, lng)\n\n\tif err != nil {\n\t\tlog.Printf(\"could not get NBN rollout info for %v\", scraper.Address)\n\t}\n\n\tif scraper.NBNInfo.ServingArea.ServiceStatus == \"\" {\n\t\tscraper.NBNInfo.ServingArea.ServiceStatus = \"unavailable\"\n\t}\n\n\tscraper.RealEstateComAuInfo, err = realestatecomau.GetInfo(scraper.Address)\n\n\tif err != nil {\n\t\tlog.Printf(\"could not get real estate info for %v: %v\", scraper.Address, err)\n\t} else {\n\t\terr = scraper.RealEstateComAuInfo.GetImages()\n\n\t\tif err != nil {\n\t\t\tlog.Printf(\"could not get real estate images for %v\", scraper.Address)\n\t\t}\n\n\t\terr = scraper.RealEstateComAuInfo.GetInspections()\n\n\t\tif err != nil {\n\t\t\tlog.Printf(\"could not get real estate inspections for %v\", scraper.Address)\n\t\t}\n\t}\n\n\tscraper.ADSLInfo, err = adsl.Lookup(scraper.Address)\n\n\tif err != nil {\n\t\tlog.Printf(\"could not get ADSL info for %v\", scraper.Address)\n\t}\n\n\tscraper.MapsEmbed = fmt.Sprintf(MapsEmbedURL, GoogleAPIKey, url.QueryEscape(scraper.Address))\n\n\tscraper.NearbyAmenitiesInfo, err = GetNearbyAmenitiesInfo(lat, lng)\n\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\n\t// no critical errors at this point.\n\terr = nil\n\n\treturn\n}", "func (t *trackerTorrent) scrape() (response bmap) {\n\tresponse = make(bmap)\n\tcompleteCount, incompleteCount := t.countPeers()\n\tresponse[\"complete\"] = completeCount\n\tresponse[\"incomplete\"] = incompleteCount\n\tresponse[\"downloaded\"] = t.downloaded\n\tif t.name != \"\" {\n\t\tresponse[\"name\"] = t.name\n\t}\n\treturn\n}", "func scrape(u string, p string, selector string, ch chan string, chFinished chan bool) {\n\n\tfmt.Printf(\"Scraping url with POST args %s \\n\", p)\n\n\t// I think there's probably a better way to get the query in proper format -- I am using this for POST requests\n\tuv, err := url.ParseQuery(p)\n\tcheck(err, \"ERROR: Could not do url.ParseQuery for \\\"\"+p+\"\\\"\")\n\n\t// POST to the desired url\n\tresp, err := http.PostForm(\n\t\tu,\n\t\tuv)\n\tem := \"ERROR: Failed to scrape \\\"\" + u + \"\\\"\"\n\tcheck(err, em)\n\n\tdefer func() {\n\t\t// Notify that we're done after this function\n\t\tchFinished <- true\n\t}()\n\n\t// get the response ready for goquery\n\tdoc, err := goquery.NewDocumentFromResponse(resp)\n\tcheck(err, \"ERROR: There was an issue making the doc. \\n\")\n\n\t// get the html in the selector from the doc\n\ttext, err := doc.Find(selector).Html()\n\tcheck(err, \"ERROR: There was an issue selection \" + selector + \" from the doc. \\n\")\n\n\t// send it to the channel\n\tch <- text\n}", "func (rcvr *iisReceiver) scrape(ctx context.Context) (pmetric.Metrics, error) {\n\tvar errs error\n\tnow := pcommon.NewTimestampFromTime(time.Now())\n\n\tfor _, wr := range rcvr.watcherRecorders {\n\t\tcounterValues, err := wr.watcher.ScrapeData()\n\t\tif err != nil {\n\t\t\trcvr.params.Logger.Warn(\"some performance counters could not be scraped; \", zap.Error(err))\n\t\t\tcontinue\n\t\t}\n\t\tvalue := 0.0\n\t\tfor _, counterValue := range counterValues {\n\t\t\tvalue += counterValue.Value\n\t\t}\n\t\twr.recorder(rcvr.metricBuilder, now, value)\n\t}\n\n\treturn rcvr.metricBuilder.Emit(), errs\n}", "func (ns *NewsPapers) Scrape() {\n\tlogrus.Info(\"Started scraping\")\n\tns.CreateDirectories()\n\tvar wg sync.WaitGroup\n\tfor _, n := range ns.NewsPapers {\n\t\twg.Add(1)\n\t\tgo func(n *NewsPaper) {\n\t\t\tdefer wg.Done()\n\t\t\tn.Scrape()\n\t\t}(n)\n\t}\n\twg.Wait()\n\tlogrus.Info(\"Done scraping\")\n}", "func (c *Crawler) CrawlAndScrape() {\n\n\tvar wg sync.WaitGroup\n\n\t//Crawl BaseURL to find the urls\n\tc.crawlURL(c.BaseURL)\n\n\tfor {\n\t\te := c.CrawlQueue.Front()\n\n\t\tif e == nil {\n\t\t\tbreak\n\t\t}\n\n\t\tc.crawlURL(e.Value.(string))\n\n\t\tc.CrawlQueue.Remove(e)\n\n\t\tfmt.Printf(\"URLS Queue: %d \\n\", c.CrawlQueue.Len())\n\t\tfmt.Printf(\"URLS Crawled: %d \\n\", c.CrawledCount)\n\n\t\t//Save progress every 20 crawls\n\t\tif c.CrawledCount%20 == 0 {\n\t\t\tgo c.saveProgress(&wg)\n\t\t\twg.Add(1)\n\t\t}\n\t}\n\n\t//Save progress one last time\n\tgo c.saveProgress(&wg)\n\twg.Add(1)\n\n\twg.Wait()\n\n}", "func (day timeURL) Scrape() (jobs []r.Scraper, results []r.Broadcaster, err error) {\n\tbcs, err := day.parseBroadcastsFromURL()\n\tif nil == err {\n\t\tfor _, bc := range bcs {\n\t\t\tresults = append(results, bc)\n\t\t}\n\t}\n\treturn\n}", "func (c CustomScraper) Scrape(_ *http.Request, closer io.ReadCloser) ([]*url.URL, error) {\n\tdefer closer.Close()\n\n\tvar links []*url.URL\n\n\tz := html.NewTokenizer(closer)\n\n\tfor {\n\t\ttt := z.Next()\n\n\t\tif tt == html.ErrorToken {\n\t\t\treturn links, nil\n\t\t}\n\n\t\tif tt == html.TextToken {\n\t\t\ttoken := z.Token()\n\n\t\t\tfmt.Println(strings.TrimSpace(token.Data))\n\t\t}\n\t}\n}", "func (s *snowflakeMetricsScraper) scrape(ctx context.Context) (pmetric.Metrics, error) {\n\terrs := &scrapererror.ScrapeErrors{}\n\n\tnow := pcommon.NewTimestampFromTime(time.Now())\n\n\t// each client call has its own scrape function\n\n\ts.scrapeBillingMetrics(ctx, now, *errs)\n\ts.scrapeWarehouseBillingMetrics(ctx, now, *errs)\n\ts.scrapeLoginMetrics(ctx, now, *errs)\n\ts.scrapeHighLevelQueryMetrics(ctx, now, *errs)\n\ts.scrapeDBMetrics(ctx, now, *errs)\n\ts.scrapeSessionMetrics(ctx, now, *errs)\n\ts.scrapeSnowpipeMetrics(ctx, now, *errs)\n\ts.scrapeStorageMetrics(ctx, now, *errs)\n\n\trb := s.mb.NewResourceBuilder()\n\trb.SetSnowflakeAccountName(s.conf.Account)\n\treturn s.mb.Emit(metadata.WithResource(rb.Emit())), errs.Combine()\n}", "func (c *crawling) crawl(s site) {\n\turls := c.crawlSite(s) // the core crawl process\n\tc.Feed(urls, s.URL, s.Depth-1) // new urls enter crawling - circular feedback\n\ttime.Sleep(c.Delay) // have a gentle nap\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
build state function that will check if a job responsible for building function image succeeded or failed; if a job is not running start one
func buildStateFnCheckImageJob(expectedJob batchv1.Job) stateFn { return func(ctx context.Context, r *reconciler, s *systemState) (stateFn, error) { labels := s.internalFunctionLabels() err := r.client.ListByLabel(ctx, s.instance.GetNamespace(), labels, &s.jobs) if err != nil { return nil, errors.Wrap(err, "while listing jobs") } jobLen := len(s.jobs.Items) if jobLen == 0 { return buildStateFnInlineCreateJob(expectedJob), nil } jobFailed := s.jobFailed(backoffLimitExceeded) conditionStatus := getConditionStatus( s.instance.Status.Conditions, serverlessv1alpha2.ConditionBuildReady, ) if jobFailed && conditionStatus == corev1.ConditionFalse { return stateFnInlineDeleteJobs, nil } if jobFailed { r.result = ctrl.Result{ RequeueAfter: time.Minute * 5, Requeue: true, } condition := serverlessv1alpha2.Condition{ Type: serverlessv1alpha2.ConditionBuildReady, Status: corev1.ConditionFalse, LastTransitionTime: metav1.Now(), Reason: serverlessv1alpha2.ConditionReasonJobFailed, Message: fmt.Sprintf("Job %s failed, it will be re-run", s.jobs.Items[0].Name), } return buildStatusUpdateStateFnWithCondition(condition), nil } s.image = s.buildImageAddress(r.cfg.docker.PullAddress) jobChanged := s.fnJobChanged(expectedJob) if !jobChanged { return stateFnCheckDeployments, nil } if jobLen > 1 || !equalJobs(s.jobs.Items[0], expectedJob) { return stateFnInlineDeleteJobs, nil } expectedLabels := expectedJob.GetLabels() if !mapsEqual(s.jobs.Items[0].GetLabels(), expectedLabels) { return buildStateFnInlineUpdateJobLabels(expectedLabels), nil } return stateFnUpdateJobStatus, nil } }
[ "func (c *FunctionController) startImageBuildJob(funcObj *kubelessApi.Function, or []metav1.OwnerReference) (string, bool, error) {\n\timagePullSecret, err := c.clientset.CoreV1().Secrets(funcObj.ObjectMeta.Namespace).Get(\"kubeless-registry-credentials\", metav1.GetOptions{})\n\tif err != nil {\n\t\treturn \"\", false, fmt.Errorf(\"Unable to locate registry credentials to build function image: %v\", err)\n\t}\n\treg, err := registry.New(*imagePullSecret)\n\tif err != nil {\n\t\treturn \"\", false, fmt.Errorf(\"Unable to retrieve registry information: %v\", err)\n\t}\n\t// Use function content and deps as tag (digested)\n\ttag := fmt.Sprintf(\"%x\", sha256.Sum256([]byte(fmt.Sprintf(\"%v%v\", funcObj.Spec.Function, funcObj.Spec.Deps))))\n\timageName := fmt.Sprintf(\"%s/%s\", reg.Creds.Username, funcObj.ObjectMeta.Name)\n\t// Check if image already exists\n\texists, err := reg.ImageExists(imageName, tag)\n\tif err != nil {\n\t\treturn \"\", false, fmt.Errorf(\"Unable to check is target image exists: %v\", err)\n\t}\n\tregURL, err := url.Parse(reg.Endpoint)\n\tif err != nil {\n\t\treturn \"\", false, fmt.Errorf(\"Unable to parse registry URL: %v\", err)\n\t}\n\timage := fmt.Sprintf(\"%s/%s:%s\", regURL.Host, imageName, tag)\n\tif !exists {\n\t\ttlsVerify := true\n\t\tif c.config.Data[\"function-registry-tls-verify\"] == \"false\" {\n\t\t\ttlsVerify = false\n\t\t}\n\t\terr = utils.EnsureFuncImage(c.clientset, funcObj, c.langRuntime, or, imageName, tag, c.config.Data[\"builder-image\"], regURL.Host, imagePullSecret.Name, c.config.Data[\"provision-image\"], tlsVerify, c.imagePullSecrets)\n\t\tif err != nil {\n\t\t\treturn \"\", false, fmt.Errorf(\"Unable to create image build job: %v\", err)\n\t\t}\n\t} else {\n\t\t// Image already exists\n\t\treturn image, false, nil\n\t}\n\treturn image, true, nil\n}", "func (st *buildStatus) start() {\n\tsetStatus(st.BuilderRev, st)\n\tgo func() {\n\t\terr := st.build()\n\t\tif err == errSkipBuildDueToDeps {\n\t\t\tst.setDone(true)\n\t\t} else {\n\t\t\tif err != nil {\n\t\t\t\tfmt.Fprintf(st, \"\\n\\nError: %v\\n\", err)\n\t\t\t\tlog.Println(st.BuilderRev, \"failed:\", err)\n\t\t\t}\n\t\t\tst.setDone(err == nil)\n\t\t\tpool.CoordinatorProcess().PutBuildRecord(st.buildRecord())\n\t\t}\n\t\tmarkDone(st.BuilderRev)\n\t}()\n}", "func checkForJobSuccess(org, repo string, targetBuildNum int, client *circleci.Client) (err error) {\n\tcheckAttempts := 0\n\tcheckLimit := 60\n\tcheckInterval := 5 * time.Second\n\tlogger.Infof(\"Polling CircleCI for status of build: %d\", targetBuildNum)\n\tfor {\n\t\tvar build *circleci.Build\n\t\tif build, err = client.GetBuild(org, repo, targetBuildNum); err != nil {\n\t\t\treturn\n\t\t}\n\t\tif build.Status == \"success\" {\n\t\t\tlogger.Infof(\"Detected success of CircleCI build: %d\", targetBuildNum)\n\t\t\tbreak\n\t\t} else if build.Status == \"failed\" {\n\t\t\treturn fmt.Errorf(\"CircleCI job: %d has failed\", targetBuildNum)\n\t\t}\n\t\tcheckAttempts++\n\t\tif checkAttempts == checkLimit {\n\t\t\treturn fmt.Errorf(\"Unable to verify CircleCI job was a success: https://circleci.com/gh/%s/%s/%d\",\n\t\t\t\torg, repo, targetBuildNum)\n\t\t}\n\t\ttime.Sleep(checkInterval)\n\t}\n\treturn\n}", "func ci_workflow() {\n\n}", "func executeJenkinsJobAndWaitForCompletion(eventMapConfig *EventMappingConfig, actionConfig *ActionConfig, jenkinsServerConfig *JenkinsServerConfig) (bool, *KeptnResultArtifact, error) {\n\n\t// before we execute the job we save current time in the eventMap\n\teventMapConfig.startedAt = time.Now()\n\teventMapConfig.finishedAt = time.Now()\n\n\tjob, buildNumber, err := executeJenkinsJob(actionConfig, jenkinsServerConfig)\n\tif err != nil {\n\t\treturn false, nil, nil\n\t}\n\n\t// lets see if we have to wait for the completion. If not we just return true!\n\tif (len(eventMapConfig.OnSuccess) == 0) && (len(eventMapConfig.OnFailure) == 0) {\n\t\treturn true, nil, nil\n\t}\n\n\t// TODO - make sure we are polling the currently started job that might still be in queue -> https://github.com/bndr/gojenkins/issues/161\n\tlog.Printf(\"Waiting until %d is finished\", buildNumber)\n\n\t// now - lets check correct timeout values\n\ttimeout := eventMapConfig.Timeout\n\tif timeout <= 0 {\n\t\ttimeout = DEFAULT_TIMEOUT\n\t}\n\tif timeout > MAX_TIMEOUT {\n\t\ttimeout = MAX_TIMEOUT\n\t}\n\n\t// query the job state until its done or until we run into our max wait\n\tvar lastBuild *gojenkins.Build = nil\n\ttimeleft := timeout\n\tfor timeleft > 0 {\n\t\tstart := time.Now()\n\n\t\t// first we sleep before we try to fetch the job state\n\t\ttime.Sleep(DEFAULT_WAIT_RETRY * time.Second)\n\n\t\t// first we need to poll the job to get the latest data\n\t\tjob.Poll()\n\n\t\t// then we query last build object\n\t\tlastBuild, err = job.GetLastBuild()\n\t\tif err != nil {\n\t\t\treturn false, nil, fmt.Errorf(\"Couldnt retrieve last build from job %s. Error: %s\", actionConfig.JenkinsJob, err.Error())\n\t\t}\n\n\t\t// now lets check the status\n\t\tif !lastBuild.IsRunning() {\n\t\t\tbuildResult := lastBuild.GetResult()\n\t\t\tlog.Printf(\"Build %d finished with status: %s\", lastBuild.GetBuildNumber(), buildResult)\n\t\t\teventMapConfig.finishedAt = time.Now()\n\n\t\t\t// lets check if there is a keptn.result.yaml in the build artifacts\n\t\t\tvar keptnResult *KeptnResultArtifact\n\t\t\tkeptnResult, err = getJenkinsBuildArtifacts(eventMapConfig, lastBuild)\n\t\t\tif keptnResult != nil {\n\t\t\t\t// we have additional results from the Jenkins Pipeline\n\t\t\t\tlog.Printf(\"Retrieved jenkins.conf.yaml!\", keptnResult)\n\t\t\t}\n\n\t\t\treturn buildResult == \"SUCCESS\", keptnResult, nil\n\t\t}\n\n\t\tlog.Printf(\"Build %d still running. Checking again in %ds\", lastBuild.GetBuildNumber(), DEFAULT_WAIT_RETRY)\n\n\t\t// adjust our timeout with the time this iteration took\n\t\tt := time.Now()\n\t\telapsed := t.Sub(start)\n\t\ttimeleft = timeleft - int(elapsed.Seconds())\n\t}\n\n\tlogMessage := fmt.Sprintf(\"Job %s did not finish within %d seconds\", actionConfig.JenkinsJob, timeout)\n\tlog.Printf(logMessage)\n\teventMapConfig.finishedAt = time.Now()\n\treturn false, nil, errors.New(logMessage)\n}", "func (st *buildStatus) start() {\n\tsetStatus(st.builderRev, st)\n\tgo func() {\n\t\terr := st.build()\n\t\tif err != nil {\n\t\t\tfmt.Fprintf(st, \"\\n\\nError: %v\\n\", err)\n\t\t\tlog.Println(st.builderRev, \"failed:\", err)\n\t\t}\n\t\tst.setDone(err == nil)\n\t\tst.buildRecord().put()\n\t\tmarkDone(st.builderRev)\n\t}()\n}", "func (pkgw *packageWatcher) build(buildCache *cache.Cache, srcpkg *crd.Package) {\n\n\t// Ignore non-pending state packages.\n\tif srcpkg.Status.BuildStatus != fission.BuildStatusPending {\n\t\treturn\n\t}\n\n\t// Ignore duplicate build requests\n\tkey := fmt.Sprintf(\"%v-%v\", srcpkg.Metadata.Name, srcpkg.Metadata.ResourceVersion)\n\terr, _ := buildCache.Set(key, srcpkg)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer buildCache.Delete(key)\n\n\tlog.Printf(\"Start build for package %v with resource version %v\", srcpkg.Metadata.Name, srcpkg.Metadata.ResourceVersion)\n\n\tpkg, err := updatePackage(pkgw.fissionClient, srcpkg, fission.BuildStatusRunning, \"\", nil)\n\tif err != nil {\n\t\te := fmt.Sprintf(\"Error setting package pending state: %v\", err)\n\t\tlog.Println(e)\n\t\tupdatePackage(pkgw.fissionClient, srcpkg, fission.BuildStatusFailed, e, nil)\n\t\treturn\n\t}\n\n\tenv, err := pkgw.fissionClient.Environments(pkg.Spec.Environment.Namespace).Get(pkg.Spec.Environment.Name)\n\tif errors.IsNotFound(err) {\n\t\tupdatePackage(pkgw.fissionClient, pkg,\n\t\t\tfission.BuildStatusFailed, \"Environment not existed\", nil)\n\t\treturn\n\t}\n\n\t// Do health check for environment builder pod\n\tfor i := 0; i < 15; i++ {\n\t\t// Informer store is not able to use label to find the pod,\n\t\t// iterate all available environment builders.\n\t\titems := pkgw.podStore.List()\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Error retrieving pod information for env %v: %v\", err, env.Metadata.Name)\n\t\t\treturn\n\t\t}\n\n\t\tif len(items) == 0 {\n\t\t\tlog.Printf(\"Environment \\\"%v\\\" builder pod is not existed yet, retry again later.\", pkg.Spec.Environment.Name)\n\t\t\ttime.Sleep(time.Duration(i*1) * time.Second)\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, item := range items {\n\t\t\tpod := item.(*apiv1.Pod)\n\n\t\t\t// In order to support backward compatibility, for all builder images created in default env,\n\t\t\t// the pods will be created in fission-builder namespace\n\t\t\tbuilderNs := pkgw.builderNamespace\n\t\t\tif env.Metadata.Namespace != metav1.NamespaceDefault {\n\t\t\t\tbuilderNs = env.Metadata.Namespace\n\t\t\t}\n\n\t\t\t// Filter non-matching pods\n\t\t\tif pod.ObjectMeta.Labels[LABEL_ENV_NAME] != env.Metadata.Name ||\n\t\t\t\tpod.ObjectMeta.Labels[LABEL_ENV_NAMESPACE] != builderNs ||\n\t\t\t\tpod.ObjectMeta.Labels[LABEL_ENV_RESOURCEVERSION] != env.Metadata.ResourceVersion {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Pod may become \"Running\" state but still failed at health check, so use\n\t\t\t// pod.Status.ContainerStatuses instead of pod.Status.Phase to check pod readiness states.\n\t\t\tpodIsReady := true\n\n\t\t\tfor _, cStatus := range pod.Status.ContainerStatuses {\n\t\t\t\tpodIsReady = podIsReady && cStatus.Ready\n\t\t\t}\n\n\t\t\tif !podIsReady {\n\t\t\t\tlog.Printf(\"Environment \\\"%v\\\" builder pod is not ready, retry again later.\", pkg.Spec.Environment.Name)\n\t\t\t\ttime.Sleep(time.Duration(i*1) * time.Second)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\t// Add the package getter rolebinding to builder sa\n\t\t\t// we continue here if role binding was not setup succeesffully. this is because without this, the fetcher wont be able to fetch the source pkg into the container and\n\t\t\t// the build will fail eventually\n\t\t\terr := fission.SetupRoleBinding(pkgw.k8sClient, fission.PackageGetterRB, pkg.Metadata.Namespace, fission.PackageGetterCR, fission.ClusterRole, fission.FissionBuilderSA, builderNs)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Error : %v in setting up the role binding %s for pkg : %s.%s\", err, fission.PackageGetterRB, pkg.Metadata.Name, pkg.Metadata.Namespace)\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tlog.Printf(\"Setup rolebinding for sa : %s.%s for pkg : %s.%s\", fission.FissionBuilderSA, builderNs, pkg.Metadata.Name, pkg.Metadata.Namespace)\n\t\t\t}\n\n\t\t\tctx := context.Background()\n\t\t\tuploadResp, buildLogs, err := buildPackage(ctx, pkgw.fissionClient, builderNs, pkgw.storageSvcUrl, pkg)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Error building package %v: %v\", pkg.Metadata.Name, err)\n\t\t\t\tupdatePackage(pkgw.fissionClient, pkg, fission.BuildStatusFailed, buildLogs, nil)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tlog.Printf(\"Start updating info of package: %v\", pkg.Metadata.Name)\n\n\t\t\tfnList, err := pkgw.fissionClient.\n\t\t\t\tFunctions(metav1.NamespaceAll).List(metav1.ListOptions{})\n\t\t\tif err != nil {\n\t\t\t\te := fmt.Sprintf(\"Error getting function list: %v\", err)\n\t\t\t\tlog.Println(e)\n\t\t\t\tbuildLogs += fmt.Sprintf(\"%v\\n\", e)\n\t\t\t\tupdatePackage(pkgw.fissionClient, pkg, fission.BuildStatusFailed, buildLogs, nil)\n\t\t\t}\n\n\t\t\t// A package may be used by multiple functions. Update\n\t\t\t// functions with old package resource version\n\t\t\tfor _, fn := range fnList.Items {\n\t\t\t\tif fn.Spec.Package.PackageRef.Name == pkg.Metadata.Name &&\n\t\t\t\t\tfn.Spec.Package.PackageRef.Namespace == pkg.Metadata.Namespace &&\n\t\t\t\t\tfn.Spec.Package.PackageRef.ResourceVersion != pkg.Metadata.ResourceVersion {\n\t\t\t\t\tfn.Spec.Package.PackageRef.ResourceVersion = pkg.Metadata.ResourceVersion\n\t\t\t\t\t// update CRD\n\t\t\t\t\t_, err = pkgw.fissionClient.Functions(fn.Metadata.Namespace).Update(&fn)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\te := fmt.Sprintf(\"Error updating function package resource version: %v\", err)\n\t\t\t\t\t\tlog.Println(e)\n\t\t\t\t\t\tbuildLogs += fmt.Sprintf(\"%v\\n\", e)\n\t\t\t\t\t\tupdatePackage(pkgw.fissionClient, pkg, fission.BuildStatusFailed, buildLogs, nil)\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t_, err = updatePackage(pkgw.fissionClient, pkg,\n\t\t\t\tfission.BuildStatusSucceeded, buildLogs, uploadResp)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Error update package info: %v\", err)\n\t\t\t\tupdatePackage(pkgw.fissionClient, pkg, fission.BuildStatusFailed, buildLogs, nil)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tlog.Printf(\"Completed build request for package: %v\", pkg.Metadata.Name)\n\t\t\treturn\n\t\t}\n\t}\n\t// build timeout\n\tupdatePackage(pkgw.fissionClient, pkg,\n\t\tfission.BuildStatusFailed, \"Build timeout due to environment builder not ready\", nil)\n\n\tlog.Printf(\"Max retries exceeded in building the source pkg : %s.%s, timeout due to environment builder not ready\",\n\t\tpkg.Metadata.Name, pkg.Metadata.Namespace)\n\n\treturn\n}", "func checkJobHealth(ctx context.Context, job *batchv1.Job) (bool, error) {\n\tif job.Status.Succeeded > 0 {\n\t\treturn true, nil\n\t}\n\n\t// Since we have filled in \"activeDeadlineSeconds\",\n\t// the Job will 'Active == 0' if and only if it exceeds the deadline or if the update image could not be pulled.\n\t// Failed jobs will be recreated in the next run.\n\tif job.Status.Active == 0 {\n\t\tklog.V(2).Infof(\"No active pods for job %s in namespace %s\", job.Name, job.Namespace)\n\t\tfailed, reason, message := hasJobFailed(job)\n\t\t// If there is more than one failed job pod then get the cause for failure\n\t\tif job.Status.Failed > 0 {\n\t\t\tfailureReason := \"DeadlineExceeded\"\n\t\t\tfailureMessage := \"Job was active longer than specified deadline\"\n\t\t\tif failed {\n\t\t\t\tfailureReason, failureMessage = reason, message\n\t\t\t}\n\t\t\treturn true, fmt.Errorf(\"deadline exceeded, reason: %q, message: %q\", failureReason, failureMessage)\n\t\t}\n\n\t\t// When the update image cannot be pulled then the pod is not marked as Failed, but the status condition is set\n\t\t// after the job deadline is exceeded.\n\t\tif failed {\n\t\t\tif reason == \"DeadlineExceeded\" {\n\t\t\t\treturn true, fmt.Errorf(\"deadline exceeded, reason: %q, message: %q\", reason, message)\n\t\t\t} else {\n\t\t\t\tklog.V(2).Infof(\"Ignoring job %s in namespace %s with condition Failed=True because %s: %s\", job.Name, job.Namespace, reason, message)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false, nil\n}", "func shouldWeDoABuild(builder interface {\n\tIsBuildRunning(*mcfgv1.MachineConfigPool) (bool, error)\n}, oldPool, curPool *mcfgv1.MachineConfigPool) (bool, error) {\n\t// If we don't have a layered pool, we should not build.\n\tpoolStateSuggestsBuild := canPoolBuild(curPool) &&\n\t\t// If we have a config change or we're missing an image pullspec label, we\n\t\t// should do a build.\n\t\t(isPoolConfigChange(oldPool, curPool) || !hasImagePullspecAnnotation(curPool)) &&\n\t\t// If we're missing a build pod reference, it likely means we don't need to\n\t\t// do a build.\n\t\t!machineConfigPoolHasBuildRef(curPool)\n\n\tif !poolStateSuggestsBuild {\n\t\treturn false, nil\n\t}\n\n\t// If a build is found running, we should not do a build.\n\tisRunning, err := builder.IsBuildRunning(curPool)\n\n\treturn !isRunning, err\n}", "func validateBuildRunToSucceed(testBuild *utils.TestBuild, testBuildRun *buildv1alpha1.BuildRun) {\n\ttrueCondition := corev1.ConditionTrue\n\tfalseCondition := corev1.ConditionFalse\n\n\t// Ensure the BuildRun has been created\n\terr := testBuild.CreateBR(testBuildRun)\n\tExpect(err).ToNot(HaveOccurred(), \"Failed to create BuildRun\")\n\n\t// Ensure a BuildRun eventually moves to a succeeded TRUE status\n\tnextStatusLog := time.Now().Add(60 * time.Second)\n\tEventually(func() corev1.ConditionStatus {\n\t\ttestBuildRun, err = testBuild.LookupBuildRun(types.NamespacedName{Name: testBuildRun.Name, Namespace: testBuild.Namespace})\n\t\tExpect(err).ToNot(HaveOccurred(), \"Error retrieving a buildRun\")\n\n\t\tif testBuildRun.Status.GetCondition(buildv1alpha1.Succeeded) == nil {\n\t\t\treturn corev1.ConditionUnknown\n\t\t}\n\n\t\tExpect(testBuildRun.Status.GetCondition(buildv1alpha1.Succeeded).Status).ToNot(Equal(falseCondition), \"BuildRun status doesn't move to Succeeded\")\n\n\t\tnow := time.Now()\n\t\tif now.After(nextStatusLog) {\n\t\t\tLogf(\"Still waiting for build run '%s' to succeed.\", testBuildRun.Name)\n\t\t\tnextStatusLog = time.Now().Add(60 * time.Second)\n\t\t}\n\n\t\treturn testBuildRun.Status.GetCondition(buildv1alpha1.Succeeded).Status\n\n\t}, time.Duration(1100*getTimeoutMultiplier())*time.Second, 5*time.Second).Should(Equal(trueCondition), \"BuildRun did not succeed\")\n\n\t// Verify that the BuildSpec is still available in the status\n\tExpect(testBuildRun.Status.BuildSpec).ToNot(BeNil(), \"BuildSpec is not available in the status\")\n\n\tLogf(\"Test build '%s' is completed after %v !\", testBuildRun.GetName(), testBuildRun.Status.CompletionTime.Time.Sub(testBuildRun.Status.StartTime.Time))\n}", "func (s *githubHook) buildStatus(eventType, commit string, payload []byte, proj *brigade.Project, status *github.RepoStatus) {\n\t// If we need an SSH key, set it here\n\tif proj.Repo.SSHKey != \"\" {\n\t\tkey, err := ioutil.TempFile(\"\", \"\")\n\t\tif err != nil {\n\t\t\tlog.Printf(\"error creating ssh key cache: %s\", err)\n\t\t\treturn\n\t\t}\n\t\tkeyfile := key.Name()\n\t\tdefer os.Remove(keyfile)\n\t\tif _, err := key.WriteString(proj.Repo.SSHKey); err != nil {\n\t\t\tlog.Printf(\"error writing ssh key cache: %s\", err)\n\t\t\treturn\n\t\t}\n\t\tos.Setenv(\"BRIGADE_REPO_KEY\", keyfile)\n\t\tdefer os.Unsetenv(\"BRIGADE_REPO_KEY\") // purely defensive... not really necessary\n\t}\n\n\tmsg := \"Building\"\n\tsvc := StatusContext\n\tstatus.State = &StatePending\n\tstatus.Description = &msg\n\tstatus.Context = &svc\n\tif err := s.createStatus(commit, proj, status); err != nil {\n\t\t// For this one, we just log an error and continue.\n\t\tlog.Printf(\"Error setting status to %s: %s\", *status.State, err)\n\t}\n\tif err := s.build(eventType, commit, payload, proj); err != nil {\n\t\tlog.Printf(\"Build failed: %s\", err)\n\t\tmsg = truncAt(err.Error(), 140)\n\t\tstatus.State = &StateFailure\n\t\tstatus.Description = &msg\n\t} else {\n\t\tmsg = \"Brigade build passed\"\n\t\tstatus.State = &StateSuccess\n\t\tstatus.Description = &msg\n\t}\n\tif err := s.createStatus(commit, proj, status); err != nil {\n\t\t// For this one, we just log an error and continue.\n\t\tlog.Printf(\"After build, error setting status to %s: %s\", *status.State, err)\n\t}\n}", "func (builder *Builder) Check() {\n\tbuilder.logger.Infof(\"Checking build queue\")\n\tjob, err := builder.jobDAO.Claim(builder.db)\n\n\tif err != nil {\n\t\tpanic(\"failed to query for job\")\n\t}\n\n\trepo, err := builder.repoDAO.Get(builder.db, job.Name)\n\n\tif err != nil {\n\t\tbuilder.jobDAO.Release(builder.db, job)\n\t\tpanic(fmt.Sprintf(\"failed to get repository for given job %s\", job.Name))\n\t}\n\n\terr = builder.Spawn(job, repo)\n\n\tif err != nil {\n\t\tbuilder.jobDAO.Release(builder.db, job)\n\t\tpanic(fmt.Sprintf(`failed to spawn job \"%s\" for repository \"%s\"`, job.Name, repo.Name))\n\t}\n}", "func (st *buildStatus) isTry() bool { return st.trySet != nil }", "func (r *ClusterInstallationReconciler) checkUpdateJob(\n\tmattermost *mattermostv1alpha1.ClusterInstallation,\n\tdesired *appsv1.Deployment,\n\treqLogger logr.Logger,\n) (*batchv1.Job, error) {\n\treqLogger.Info(fmt.Sprintf(\"Running Mattermost update image job check for image %s\", mattermost.GetMattermostAppContainerFromDeployment(desired).Image))\n\tjob, err := r.Resources.FetchMattermostUpdateJob(mattermost.Namespace)\n\tif err != nil {\n\t\t// Unable to fetch job\n\t\tif k8sErrors.IsNotFound(err) {\n\t\t\t// Job is not running, let's launch\n\t\t\treqLogger.Info(\"Launching update image job\")\n\t\t\tif err = r.Resources.LaunchMattermostUpdateJob(mattermost, mattermost.Namespace, desired, reqLogger, nil); err != nil {\n\t\t\t\treturn nil, errors.Wrap(err, \"Launching update image job failed\")\n\t\t\t}\n\t\t\treturn nil, errors.New(\"Began update image job\")\n\t\t}\n\n\t\treturn nil, errors.Wrap(err, \"failed to determine if an update image job is already running\")\n\t}\n\n\t// Job is either running or completed\n\n\t// If desired deployment image does not match the one used by update job, restart it.\n\tisSameImage, err := r.isMainContainerImageSame(\n\t\tmattermost,\n\t\tdesired.Spec.Template.Spec.Containers,\n\t\tjob.Spec.Template.Spec.Containers,\n\t)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to compare image of update job and desired deployment\")\n\t}\n\tif !isSameImage {\n\t\treqLogger.Info(\"Mattermost image changed, restarting update job\")\n\t\terr := r.Resources.RestartMattermostUpdateJob(mattermost, job, desired, reqLogger, nil)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to restart update job\")\n\t\t}\n\n\t\treturn nil, errors.New(\"Restarted update image job\")\n\t}\n\n\tif job.Status.CompletionTime == nil {\n\t\treturn nil, errors.New(\"update image job still running\")\n\t}\n\n\t// Job is completed, can check completion status\n\n\tif job.Status.Failed > 0 {\n\t\treturn job, errors.New(\"update image job failed\")\n\t}\n\n\treqLogger.Info(\"Update image job ran successfully\")\n\n\treturn job, nil\n}", "func (s RepoState) IsTryJob() bool {\n\treturn s.Patch.Full()\n}", "func (ctrl *Controller) markBuildSucceeded(pool *mcfgv1.MachineConfigPool) error {\n\tklog.Infof(\"Build succeeded for MachineConfigPool %s, config %s\", pool.Name, pool.Spec.Configuration.Name)\n\n\t// Get the final image pullspec.\n\timagePullspec, err := ctrl.imageBuilder.FinalPullspec(pool)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not get final image pullspec for pool %s: %w\", pool.Name, err)\n\t}\n\n\tif imagePullspec == \"\" {\n\t\treturn fmt.Errorf(\"image pullspec empty for pool %s\", pool.Name)\n\t}\n\n\t// Perform the post-build cleanup.\n\tif err := ctrl.postBuildCleanup(pool, false); err != nil {\n\t\treturn fmt.Errorf(\"could not do post-build cleanup: %w\", err)\n\t}\n\n\t// Set the annotation or field to point to the newly-built container image.\n\tklog.V(4).Infof(\"Setting new image pullspec for %s to %s\", pool.Name, imagePullspec)\n\tif pool.Annotations == nil {\n\t\tpool.Annotations = map[string]string{}\n\t}\n\tpool.Annotations[ctrlcommon.ExperimentalNewestLayeredImageEquivalentConfigAnnotationKey] = imagePullspec\n\n\t// Remove the build object reference from the MachineConfigPool since we're\n\t// not using it anymore.\n\tdeleteBuildRefFromMachineConfigPool(pool)\n\n\t// Adjust the MachineConfigPool status to indicate success.\n\tsetMCPBuildConditions(pool, []mcfgv1.MachineConfigPoolCondition{\n\t\t{\n\t\t\tType: mcfgv1.MachineConfigPoolBuildFailed,\n\t\t\tStatus: corev1.ConditionFalse,\n\t\t},\n\t\t{\n\t\t\tType: mcfgv1.MachineConfigPoolBuildSuccess,\n\t\t\tReason: \"BuildSucceeded\",\n\t\t\tStatus: corev1.ConditionTrue,\n\t\t},\n\t\t{\n\t\t\tType: mcfgv1.MachineConfigPoolBuilding,\n\t\t\tStatus: corev1.ConditionFalse,\n\t\t},\n\t\t{\n\t\t\tType: mcfgv1.MachineConfigPoolDegraded,\n\t\t\tStatus: corev1.ConditionFalse,\n\t\t},\n\t})\n\n\t// Perform the MachineConfigPool update.\n\treturn ctrl.updatePoolAndSyncStatus(pool, ctrl.syncAvailableStatus)\n}", "func (l *LayeredPoolState) IsBuildSuccess() bool {\n\treturn mcfgv1.IsMachineConfigPoolConditionTrue(l.pool.Status.Conditions, mcfgv1.MachineConfigPoolBuildSuccess)\n}", "func (cc *ConfigsController) needsBuild(ctx context.Context, st store.RStore) (*ctrltiltfile.BuildEntry, bool) {\n\tstate := st.RLockState()\n\tdefer st.RUnlockState()\n\n\t// Don't start the next build until the previous action has been recorded,\n\t// so that we don't accidentally repeat the same build.\n\tif cc.loadStartedCount != state.StartedTiltfileLoadCount {\n\t\treturn nil, false\n\t}\n\n\t// Don't start the next build if the last completion hasn't been recorded yet.\n\tfor _, ms := range state.TiltfileStates {\n\t\tisRunning := !ms.CurrentBuild.StartTime.IsZero()\n\t\tif isRunning {\n\t\t\treturn nil, false\n\t\t}\n\t}\n\n\tfor _, name := range state.TiltfileDefinitionOrder {\n\t\ttf, ok := state.Tiltfiles[name.String()]\n\t\tif !ok {\n\t\t\tcontinue\n\t\t}\n\n\t\ttfState, ok := state.TiltfileStates[name]\n\t\tif !ok {\n\t\t\tcontinue\n\t\t}\n\n\t\tvar reason model.BuildReason\n\t\tlastStartTime := tfState.LastBuild().StartTime\n\t\tif !tfState.StartedFirstBuild() {\n\t\t\treason = reason.With(model.BuildReasonFlagInit)\n\t\t}\n\n\t\thasPendingChanges, _ := tfState.HasPendingChanges()\n\t\tif hasPendingChanges {\n\t\t\treason = reason.With(model.BuildReasonFlagChangedFiles)\n\t\t}\n\n\t\tif state.UserConfigState.ArgsChangeTime.After(lastStartTime) {\n\t\t\treason = reason.With(model.BuildReasonFlagTiltfileArgs)\n\t\t}\n\n\t\tif state.ManifestInTriggerQueue(name) {\n\t\t\treason = reason.With(tfState.TriggerReason)\n\t\t}\n\n\t\tif reason == model.BuildReasonNone {\n\t\t\tcontinue\n\t\t}\n\n\t\tfilesChanged := []string{}\n\t\tfor _, st := range tfState.BuildStatuses {\n\t\t\tfor k := range st.PendingFileChanges {\n\t\t\t\tfilesChanged = append(filesChanged, k)\n\t\t\t}\n\t\t}\n\t\tfilesChanged = sliceutils.DedupedAndSorted(filesChanged)\n\n\t\treturn &ctrltiltfile.BuildEntry{\n\t\t\tName: name,\n\t\t\tFilesChanged: filesChanged,\n\t\t\tBuildReason: reason,\n\t\t\tUserConfigState: state.UserConfigState,\n\t\t\tTiltfilePath: tf.Spec.Path,\n\t\t\tCheckpointAtExecStart: state.LogStore.Checkpoint(),\n\t\t}, true\n\t}\n\n\treturn nil, false\n}", "func (w *worker) callStatus() modules.WorkerStatus {\n\tdownloadQueue := w.staticJobLowPrioReadQueue\n\tdownloadQueue.mu.Lock()\n\tdownloadOnCoolDown := downloadQueue.onCooldown()\n\tdownloadTerminated := downloadQueue.killed\n\tdownloadQueueSize := downloadQueue.jobs.Len()\n\tdownloadCoolDownTime := downloadQueue.cooldownUntil.Sub(time.Now())\n\n\tvar downloadCoolDownErr string\n\tif downloadQueue.recentErr != nil {\n\t\tdownloadCoolDownErr = downloadQueue.recentErr.Error()\n\t}\n\tdownloadQueue.mu.Unlock()\n\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\n\tuploadOnCoolDown, uploadCoolDownTime := w.onUploadCooldown()\n\tvar uploadCoolDownErr string\n\tif w.uploadRecentFailureErr != nil {\n\t\tuploadCoolDownErr = w.uploadRecentFailureErr.Error()\n\t}\n\n\tmaintenanceOnCooldown, maintenanceCoolDownTime, maintenanceCoolDownErr := w.staticMaintenanceState.managedMaintenanceCooldownStatus()\n\tvar mcdErr string\n\tif maintenanceCoolDownErr != nil {\n\t\tmcdErr = maintenanceCoolDownErr.Error()\n\t}\n\n\t// Update the worker cache before returning a status.\n\tw.staticTryUpdateCache()\n\tcache := w.staticCache()\n\treturn modules.WorkerStatus{\n\t\t// Contract Information\n\t\tContractID: cache.staticContractID,\n\t\tContractUtility: cache.staticContractUtility,\n\t\tHostPubKey: w.staticHostPubKey,\n\n\t\t// Download information\n\t\tDownloadCoolDownError: downloadCoolDownErr,\n\t\tDownloadCoolDownTime: downloadCoolDownTime,\n\t\tDownloadOnCoolDown: downloadOnCoolDown,\n\t\tDownloadQueueSize: downloadQueueSize,\n\t\tDownloadTerminated: downloadTerminated,\n\n\t\t// Upload information\n\t\tUploadCoolDownError: uploadCoolDownErr,\n\t\tUploadCoolDownTime: uploadCoolDownTime,\n\t\tUploadOnCoolDown: uploadOnCoolDown,\n\t\tUploadQueueSize: w.unprocessedChunks.Len(),\n\t\tUploadTerminated: w.uploadTerminated,\n\n\t\t// Job Queues\n\t\tDownloadSnapshotJobQueueSize: int(w.staticJobDownloadSnapshotQueue.callStatus().size),\n\t\tUploadSnapshotJobQueueSize: int(w.staticJobUploadSnapshotQueue.callStatus().size),\n\n\t\t// Maintenance Cooldown Information\n\t\tMaintenanceOnCooldown: maintenanceOnCooldown,\n\t\tMaintenanceCoolDownError: mcdErr,\n\t\tMaintenanceCoolDownTime: maintenanceCoolDownTime,\n\n\t\t// Account Information\n\t\tAccountBalanceTarget: w.staticBalanceTarget,\n\t\tAccountStatus: w.staticAccount.managedStatus(),\n\n\t\t// Price Table Information\n\t\tPriceTableStatus: w.staticPriceTableStatus(),\n\n\t\t// Read Job Information\n\t\tReadJobsStatus: w.callReadJobStatus(),\n\n\t\t// HasSector Job Information\n\t\tHasSectorJobsStatus: w.callHasSectorJobStatus(),\n\n\t\t// ReadRegistry Job Information\n\t\tReadRegistryJobsStatus: w.callReadRegistryJobsStatus(),\n\n\t\t// UpdateRegistry Job Information\n\t\tUpdateRegistryJobsStatus: w.callUpdateRegistryJobsStatus(),\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
Save changes the configuration of the hostonly network.
func (n *hostOnlyNetwork) Save(vbox VBoxManager) error { if err := n.SaveIPv4(vbox); err != nil { return err } if n.DHCP { vbox.vbm("hostonlyif", "ipconfig", n.Name, "--dhcp") // not implemented as of VirtualBox 4.3 } return nil }
[ "func (d *HostNicDriver) saveConfig() error {\n\tconfigFile := fmt.Sprintf(\"%s/%s\", configDir, \"config.json\")\n\tdata, err := json.Marshal(d.networks)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(configFile, data, os.FileMode(0644))\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Debug(\"Save config [%+v] to [%s]\", d.networks, configFile)\n\treturn nil\n}", "func (system *System) saveEthernetConfig() error {\n\tsystem.ethMux.Lock()\n\n\tdata, err := json.MarshalIndent(system.EthernetConfig(), \"\", \" \")\n\tif err != nil {\n\t\tlog.Print(\"Error while serializing JSON to store network configuration\")\n\t\treturn err\n\t}\n\n\tfile := path.Join(viper.GetString(\"directories.config\"), \"ethernet.json\")\n\terr = ioutil.WriteFile(file, data, 0644)\n\tif err != nil {\n\t\tlog.Print(\"Failed to write the ethernet configuration to the config file: \" + err.Error())\n\t\treturn err\n\t}\n\n\tsystem.ethMux.Unlock()\n\treturn nil\n}", "func (nm *networkManager) saveIPConfig(hostIf *net.Interface, extIf *externalInterface) error {\n\t// Save the default routes on the interface.\n\troutes, err := netlink.GetIpRoute(&netlink.Route{Dst: &net.IPNet{}, LinkIndex: hostIf.Index})\n\tif err != nil {\n\t\tlog.Printf(\"[net] Failed to query routes: %v.\", err)\n\t\treturn err\n\t}\n\n\tfor _, r := range routes {\n\t\tif r.Dst == nil {\n\t\t\tif r.Family == unix.AF_INET {\n\t\t\t\textIf.IPv4Gateway = r.Gw\n\t\t\t} else if r.Family == unix.AF_INET6 {\n\t\t\t\textIf.IPv6Gateway = r.Gw\n\t\t\t}\n\t\t}\n\n\t\textIf.Routes = append(extIf.Routes, (*route)(r))\n\t}\n\n\t// Save global unicast IP addresses on the interface.\n\taddrs, err := hostIf.Addrs()\n\tfor _, addr := range addrs {\n\t\tipAddr, ipNet, err := net.ParseCIDR(addr.String())\n\t\tipNet.IP = ipAddr\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tif !ipAddr.IsGlobalUnicast() {\n\t\t\tcontinue\n\t\t}\n\n\t\textIf.IPAddresses = append(extIf.IPAddresses, ipNet)\n\n\t\tlog.Printf(\"[net] Deleting IP address %v from interface %v.\", ipNet, hostIf.Name)\n\n\t\terr = netlink.DeleteIpAddress(hostIf.Name, ipAddr, ipNet)\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t}\n\n\tlog.Printf(\"[net] Saved interface IP configuration %+v.\", extIf)\n\n\treturn err\n}", "func (c *Config) Save() error {\n\treturn c.backend.Save(c)\n}", "func SaveGateway(svc *providers.Service, host *api.Host, networkID string) error {\n\tm, err := NewGateway(svc, networkID)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Update network\n\tn := NewNetwork(svc)\n\tok, err := n.ReadByID(networkID)\n\tif !ok || err != nil {\n\t\treturn fmt.Errorf(\"metadata about the '%s' doesn't exist anymore\", networkID)\n\t}\n\tnet := n.Get()\n\tnet.GatewayID = host.ID\n\terr = n.Write()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn m.Carry(host).Write()\n}", "func (client *Client) complementHost(host *model.Host, server *servers.Server) error {\n\tnetworks, addresses, ipv4, ipv6, err := client.collectAddresses(host)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Updates intrinsic data of host if needed\n\tif host.ID == \"\" {\n\t\thost.ID = server.ID\n\t}\n\tif host.Name == \"\" {\n\t\thost.Name = server.Name\n\t}\n\n\thost.LastState = toHostState(server.Status)\n\n\t// Updates Host Property propsv1.HostDescription\n\thostDescriptionV1 := propsv1.NewHostDescription()\n\terr = host.Properties.Get(HostProperty.DescriptionV1, hostDescriptionV1)\n\tif err != nil {\n\t\treturn err\n\t}\n\thostDescriptionV1.Created = server.Created\n\thostDescriptionV1.Updated = server.Updated\n\terr = host.Properties.Set(HostProperty.DescriptionV1, hostDescriptionV1)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Updates Host Property propsv1.HostSizing\n\thostSizingV1 := propsv1.NewHostSizing()\n\terr = host.Properties.Get(HostProperty.SizingV1, hostSizingV1)\n\tif err != nil {\n\t\treturn err\n\t}\n\thostSizingV1.AllocatedSize = client.toHostSize(server.Flavor)\n\terr = host.Properties.Set(HostProperty.SizingV1, hostSizingV1)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Updates Host Property HostNetwork\n\thostNetworkV1 := propsv1.NewHostNetwork()\n\terr = host.Properties.Get(HostProperty.NetworkV1, hostNetworkV1)\n\tif err != nil {\n\t\treturn nil\n\t}\n\tif hostNetworkV1.PublicIPv4 == \"\" {\n\t\thostNetworkV1.PublicIPv4 = ipv4\n\t}\n\tif hostNetworkV1.PublicIPv6 == \"\" {\n\t\thostNetworkV1.PublicIPv6 = ipv6\n\t}\n\n\tif len(hostNetworkV1.NetworksByID) > 0 {\n\t\tipv4Addresses := map[string]string{}\n\t\tipv6Addresses := map[string]string{}\n\t\tfor netid, netname := range hostNetworkV1.NetworksByID {\n\t\t\tif ip, ok := addresses[IPVersion.IPv4][netid]; ok {\n\t\t\t\tipv4Addresses[netid] = ip\n\t\t\t} else if ip, ok := addresses[IPVersion.IPv4][netname]; ok {\n\t\t\t\tipv4Addresses[netid] = ip\n\t\t\t} else {\n\t\t\t\tipv4Addresses[netid] = \"\"\n\t\t\t}\n\n\t\t\tif ip, ok := addresses[IPVersion.IPv6][netid]; ok {\n\t\t\t\tipv6Addresses[netid] = ip\n\t\t\t} else if ip, ok := addresses[IPVersion.IPv6][netname]; ok {\n\t\t\t\tipv6Addresses[netid] = ip\n\t\t\t} else {\n\t\t\t\tipv6Addresses[netid] = \"\"\n\t\t\t}\n\t\t}\n\t\thostNetworkV1.IPv4Addresses = ipv4Addresses\n\t\thostNetworkV1.IPv6Addresses = ipv6Addresses\n\t} else {\n\t\tnetworksByID := map[string]string{}\n\t\tipv4Addresses := map[string]string{}\n\t\tipv6Addresses := map[string]string{}\n\t\tfor _, netid := range networks {\n\t\t\tnetworksByID[netid] = \"\"\n\n\t\t\tif ip, ok := addresses[IPVersion.IPv4][netid]; ok {\n\t\t\t\tipv4Addresses[netid] = ip\n\t\t\t} else {\n\t\t\t\tipv4Addresses[netid] = \"\"\n\t\t\t}\n\n\t\t\tif ip, ok := addresses[IPVersion.IPv6][netid]; ok {\n\t\t\t\tipv6Addresses[netid] = ip\n\t\t\t} else {\n\t\t\t\tipv6Addresses[netid] = \"\"\n\t\t\t}\n\t\t}\n\t\thostNetworkV1.NetworksByID = networksByID\n\t\t// IPvxAddresses are here indexed by names... At least we have them...\n\t\thostNetworkV1.IPv4Addresses = ipv4Addresses\n\t\thostNetworkV1.IPv6Addresses = ipv6Addresses\n\t}\n\n\t// Updates network name and relationships if needed\n\tfor netid, netname := range hostNetworkV1.NetworksByID {\n\t\tif netname == \"\" {\n\t\t\tnet, err := client.GetNetwork(netid)\n\t\t\tif err != nil {\n\t\t\t\tlog.Errorf(\"failed to get network '%s'\", netid)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\thostNetworkV1.NetworksByID[netid] = net.Name\n\t\t\thostNetworkV1.NetworksByName[net.Name] = netid\n\t\t}\n\t}\n\n\treturn host.Properties.Set(HostProperty.NetworkV1, hostNetworkV1)\n}", "func (ctx *BoltDbContext) SaveHost(host Host) error {\n\treturn ctx.db.Update(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(hostsBucketName))\n\t\terr := b.Put([]byte(host.Address), host.GobEncode())\n\t\treturn err\n\t})\n}", "func (hc *Hailconfig) Save() error {\n\terr := hc.f.Reset()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to reset\")\n\t}\n\treturn toml.NewEncoder(hc.f).Encode(&hc.config)\n}", "func (tc TunnelConfig) Save(path string) error {\n\treturn MarshalToFile(path, &tc)\n}", "func (hdb *HostDB) save() error {\n\treturn hdb.persist.save(hdb.persistData())\n}", "func (n *Node) saveConfig() error {\n\tconfig, err := readJSON()\n\tif err != nil {\n\t\treturn err\n\t}\n\tn.mutex.Lock()\n\tpeers := n.server.peers\n\tn.mutex.Unlock()\n\tn.server.savePeers(config, peers)\n\treturn writeJSON(config)\n}", "func SaveNetwork(svc *providers.Service, net *api.Network) error {\n\treturn NewNetwork(svc).Carry(net).Write()\n}", "func (n *NetworkBuilder) Save(writer io.Writer) error {\n\terr := json.NewEncoder(writer).Encode(n.Network)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn err\n}", "func (c *Config) save() {\n\tconst file = \"access.json\"\n\n\tc.logger.Printf(\"Save file %s\\n\", file)\n\n\tcfg := conf{\n\t\tIP: c.GetString(\"ip\"),\n\t\tPort: c.GetString(\"port\"),\n\t\tToken: c.GetString(\"token\"),\n\t\tWait: c.GetBool(\"wait\"),\n\t}\n\n\tb, err := json.Marshal(cfg)\n\tif err != nil {\n\t\tc.logger.Error(err)\n\t}\n\n\tif err = ioutil.WriteFile(file, b, 0644); err != nil {\n\t\tc.logger.Error(err)\n\t}\n}", "func (n *hostOnlyNetwork) SaveIPv4(vbox VBoxManager) error {\n\tif n.IPv4.IP != nil && n.IPv4.Mask != nil {\n\t\tif err := vbox.vbm(\"hostonlyif\", \"ipconfig\", n.Name, \"--ip\", n.IPv4.IP.String(), \"--netmask\", net.IP(n.IPv4.Mask).String()); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func SetHost(value string) {\n\tconfig.host = value\n}", "func (s *store) RegisterHost(host Host) error {\n\tconfig := host.Config()\n\tdir, err := host.Dir()\n\tif err != nil {\n\t\treturn err\n\t}\n\tbytes, err := json.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjsonPath := filepath.Join(dir, HostConfigJSONFile)\n\treturn ioutil.WriteFile(jsonPath, bytes, 0644)\n}", "func saveRemoteConfig(conf *types.NetConf,\n\targs *skel.CmdArgs,\n\tipResult *current.Result,\n\tconnData connectionData) error {\n\n\tvar err error\n\n\t// Populate the configData with input data, which will be written to container.\n\tconfigData, err := populateUserspaceConfigData(conf, args, ipResult)\n\tif err != nil {\n\t\tlogging.Errorf(\"ERROR: saveRemoteConfig: Failure to retrieve pod - %v\", err)\n\t\treturn err\n\t}\n\n\t// Wrtie configData to the annotations, which will be read by container.\n\tconnData.pod, err = annotations.WritePodAnnotation(connData.kubeClient, connData.pod, configData)\n\tif err != nil {\n\t\tlogging.Errorf(\"ERROR: saveRemoteConfig: Failure to write annotations - %v\", err)\n\t\treturn err\n\t}\n\n\treturn err\n}", "func WriteHostConfig(host string, content []byte) error {\n\tp := path.Join(configDir, host)\n\n\tencrypted, err := encrypt(content)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(p, encrypted, 0644)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
SaveIPv4 changes the ipv4 configuration of the hostonly network.
func (n *hostOnlyNetwork) SaveIPv4(vbox VBoxManager) error { if n.IPv4.IP != nil && n.IPv4.Mask != nil { if err := vbox.vbm("hostonlyif", "ipconfig", n.Name, "--ip", n.IPv4.IP.String(), "--netmask", net.IP(n.IPv4.Mask).String()); err != nil { return err } } return nil }
[ "func (tt *ServiceTraceroute) SetIPv4() {\n\ttt.Configuration.IPVersion = V4\n}", "func (f *FSEIDFields) SetIPv4Flag() {\n\tf.Flags |= 0x02\n}", "func (in *ActionVpsCreateInput) SetIpv4(value int64) *ActionVpsCreateInput {\n\tin.Ipv4 = value\n\n\tif in._selectedParameters == nil {\n\t\tin._selectedParameters = make(map[string]interface{})\n\t}\n\n\tin._selectedParameters[\"Ipv4\"] = nil\n\treturn in\n}", "func (n *hostOnlyNetwork) Save(vbox VBoxManager) error {\n\tif err := n.SaveIPv4(vbox); err != nil {\n\t\treturn err\n\t}\n\n\tif n.DHCP {\n\t\tvbox.vbm(\"hostonlyif\", \"ipconfig\", n.Name, \"--dhcp\") // not implemented as of VirtualBox 4.3\n\t}\n\n\treturn nil\n}", "func copyIPv4ToProtoIPv4(l3 *layers.IPv4, proto *trace.ProtoIPv4) {\n\tproto.Version = l3.Version\n\tproto.IHL = l3.IHL\n\tproto.TOS = l3.TOS\n\tproto.Length = l3.Length\n\tproto.Id = l3.Id\n\tproto.Flags = uint8(l3.Flags)\n\tproto.FragOffset = l3.FragOffset\n\tproto.TTL = l3.TTL\n\tproto.Protocol = l3.Protocol.String()\n\tproto.Checksum = l3.Checksum\n\tproto.SrcIP = l3.SrcIP.String()\n\tproto.DstIP = l3.DstIP.String()\n\t// TODO: IPv4 options if IHL > 5\n}", "func (m *manager) IPv4Set(family nft.Family, table string, name string, ips ...string) error {\n\t//nft add set ip nat host { type ipv4_addr\\; }\n\t//nft add element ip nat host { 172^C9.0.1, 172.18.0.1 }\n\n\t_, err := m.api.System(\"nft\", \"add\", \"set\", string(family), table, name, \"{\", \"type\", \"ipv4_addr;\", \"}\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(ips) == 0 {\n\t\treturn nil\n\t}\n\n\ts := strings.Join(ips, \", \")\n\t_, err = m.api.System(\"nft\", \"add\", \"element\", string(family), table, name, \"{\", s, \"}\")\n\n\treturn err\n}", "func (i Internet) Ipv4() string {\n\tips := make([]string, 0, 4)\n\n\tips = append(ips, strconv.Itoa(i.Faker.IntBetween(1, 255)))\n\tfor j := 0; j < 3; j++ {\n\t\tips = append(ips, strconv.Itoa(i.Faker.IntBetween(0, 255)))\n\t}\n\n\treturn strings.Join(ips, \".\")\n}", "func IPv4() (string, error) {\n\tconn, err := net.Dial(\"udp\", \"8.8.8.8:80\")\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Failed to determine your IP\")\n\t}\n\tlocalAddr := conn.LocalAddr().(*net.UDPAddr)\n\tmyIP := localAddr.IP.String()\n\tconn.Close()\n\treturn myIP, nil\n}", "func hostToIP4(hostname string) (string, bool) {\n\tip := net.ParseIP(hostname)\n\tif ip == nil {\n\t\tt, err := net.ResolveIPAddr(\"ip4\", hostname)\n\t\tif err != nil {\n\t\t\tlogging.Error.Printf(\"cannot translate hostname %q into an ip4 address\", hostname)\n\t\t\treturn hostname, false\n\t\t}\n\t\tip = t.IP\n\t}\n\treturn ip.String(), true\n}", "func (soc *Socket) SetIpv4only(value bool) error {\n\tval := 0\n\tif value {\n\t\tval = 1\n\t}\n\treturn soc.setInt(C.ZMQ_IPV4ONLY, val)\n}", "func IPv4(name string) (string, error) {\n\ti, err := net.InterfaceByName(name)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\taddrs, err := i.Addrs()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor _, a := range addrs {\n\t\tif ipn, ok := a.(*net.IPNet); ok {\n\t\t\tif ipn.IP.To4() != nil {\n\t\t\t\treturn ipn.IP.String(), nil\n\t\t\t}\n\t\t}\n\t}\n\n\treturn \"\", fmt.Errorf(\"no IPv4 found for interface: %q\", name)\n}", "func V4Address(ip uint32) Address {\n\treturn Address{\n\t\tnetip.AddrFrom4([4]byte{\n\t\t\tuint8(ip),\n\t\t\tuint8(ip >> 8),\n\t\t\tuint8(ip >> 16),\n\t\t\tuint8(ip >> 24),\n\t\t}),\n\t}\n}", "func WanIPv4() (string, error) {\n\treturn wanIP(\"https://v4.ident.me/\")\n}", "func Configure4(iface netlink.Link, packet *dhcp4.Packet) error {\n\tp := NewPacket4(packet)\n\n\tl := p.Lease()\n\tif l == nil {\n\t\treturn fmt.Errorf(\"no lease returned\")\n\t}\n\n\t// Add the address to the iface.\n\tdst := &netlink.Addr{\n\t\tIPNet: l,\n\t}\n\tif err := netlink.AddrReplace(iface, dst); err != nil {\n\t\tif os.IsExist(err) {\n\t\t\treturn fmt.Errorf(\"add/replace %s to %v: %v\", dst, iface, err)\n\t\t}\n\t}\n\n\tif gw := p.Gateway(); gw != nil {\n\t\tr := &netlink.Route{\n\t\t\tLinkIndex: iface.Attrs().Index,\n\t\t\tGw: gw,\n\t\t}\n\n\t\tif err := netlink.RouteReplace(r); err != nil {\n\t\t\treturn fmt.Errorf(\"%s: add %s: %v\", iface.Attrs().Name, r, err)\n\t\t}\n\t}\n\n\tif ips := p.DNS(); ips != nil {\n\t\tif err := WriteDNSSettings(ips); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func NewIPv4(value string) (IPv4, error) {\n\tvar IP = IPv4{value: value}\n\n\tif !IP.validate() {\n\t\treturn IPv4{}, ErrInvalidIPv4\n\t}\n\n\treturn IP, nil\n}", "func IpV4Address() string {\n\tblocks := []string{}\n\tfor i := 0; i < 4; i++ {\n\t\tnumber := seedAndReturnRandom(255)\n\t\tblocks = append(blocks, strconv.Itoa(number))\n\t}\n\n\treturn strings.Join(blocks, \".\")\n}", "func (h *DeviceHeader) SetIP4Address(ip net.IP) {\n\tip4 := ip.To4()\n\tif ip4 == nil {\n\t\tpanic(\"address is not an IPv4 address\")\n\t}\n\tcopy(h.IPAddress[:], ip4[:4])\n}", "func IsIPv4(v *Validate, topStruct reflect.Value, currentStructOrField reflect.Value, field reflect.Value, fieldType reflect.Type, fieldKind reflect.Kind, param string) bool {\n\n\tip := net.ParseIP(field.String())\n\n\treturn ip != nil && ip.To4() != nil\n}", "func ipv4only(addr IPAddr) bool {\n\treturn supportsIPv4 && addr.IP.To4() != nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
createHostonlyAdapter creates a new hostonly network.
func createHostonlyAdapter(vbox VBoxManager) (*hostOnlyNetwork, error) { out, err := vbox.vbmOut("hostonlyif", "create") if err != nil { return nil, err } res := reHostOnlyAdapterCreated.FindStringSubmatch(string(out)) if res == nil { return nil, errors.New("Failed to create host-only adapter") } return &hostOnlyNetwork{Name: res[1]}, nil }
[ "func (client *Client) createHost(request api.HostRequest, isGateway bool) (*api.Host, error) {\n\tif isGateway && !request.PublicIP {\n\t\treturn nil, fmt.Errorf(\"can't create a gateway without public IP\")\n\t}\n\n\t// Validating name of the host\n\tif ok, err := validatehostName(request); !ok {\n\t\treturn nil, fmt.Errorf(\"name '%s' is invalid for a FlexibleEngine Host: %s\", request.Name, openstack.ProviderErrorToString(err))\n\t}\n\n\t// Check name availability\n\tm, err := metadata.LoadHost(providers.FromClient(client), request.Name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif m != nil {\n\t\treturn nil, providers.ResourceAlreadyExistsError(\"Host\", request.Name)\n\t}\n\n\t// Network gateway\n\tvar gw *api.Host\n\t// If the host is not public it has to be created on a network owning a Gateway\n\tif !request.PublicIP {\n\t\tm, err := metadata.LoadGateway(providers.FromClient(client), request.NetworkIDs[0])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif m != nil {\n\t\t\tgw = m.Get()\n\t\t}\n\t}\n\t// If a gateway is created, we need the CIDR for the userdata\n\tvar cidr string\n\tif isGateway {\n\t\tm, err := metadata.LoadNetwork(providers.FromClient(client), request.NetworkIDs[0])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif m == nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to load metadata of network '%s'\", request.NetworkIDs[0])\n\t\t}\n\t\tnetwork := m.Get()\n\t\tcidr = network.CIDR\n\t}\n\n\tvar nets []servers.Network\n\t// Add private networks\n\tfor _, n := range request.NetworkIDs {\n\t\tnets = append(nets, servers.Network{\n\t\t\tUUID: n,\n\t\t})\n\t}\n\n\t// Prepare key pair\n\tkp := request.KeyPair\n\t// If no key pair is supplied create one\n\tif kp == nil {\n\t\tid, _ := uuid.NewV4()\n\t\tname := fmt.Sprintf(\"%s_%s\", request.Name, id)\n\t\tkp, err = client.CreateKeyPair(name)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error creating key pair for host '%s': %s\", request.Name, openstack.ProviderErrorToString(err))\n\t\t}\n\t}\n\n\tuserData, err := userdata.Prepare(client, request, isGateway, kp, gw, cidr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Determine system disk size based on vcpus count\n\ttemplate, err := client.GetTemplate(request.TemplateID)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to get image: %s\", openstack.ProviderErrorToString(err))\n\t}\n\n\tvar diskSize int\n\tif template.HostSize.DiskSize > 0 {\n\t\tdiskSize = template.HostSize.DiskSize\n\t} else if template.HostSize.Cores < 16 {\n\t\tdiskSize = 100\n\t} else if template.HostSize.Cores < 32 {\n\t\tdiskSize = 200\n\t} else {\n\t\tdiskSize = 400\n\t}\n\n\t// Defines boot disk\n\tbootdiskOpts := blockDevice{\n\t\tSourceType: exbfv.SourceImage,\n\t\tDestinationType: exbfv.DestinationVolume,\n\t\tBootIndex: \"0\",\n\t\tDeleteOnTermination: true,\n\t\tUUID: request.ImageID,\n\t\tVolumeType: \"SSD\",\n\t\tVolumeSize: diskSize,\n\t}\n\t// Defines server\n\tsrvOpts := serverCreateOpts{\n\t\tName: request.Name,\n\t\tSecurityGroups: []string{client.SecurityGroup.Name},\n\t\tNetworks: nets,\n\t\tFlavorRef: request.TemplateID,\n\t\tUserData: userData,\n\t}\n\t// Defines host \"Extension bootfromvolume\" options\n\tbdOpts := bootdiskCreateOptsExt{\n\t\tCreateOptsBuilder: srvOpts,\n\t\tBlockDevice: []blockDevice{bootdiskOpts},\n\t}\n\tb, err := bdOpts.ToServerCreateMap()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to build query to create host '%s': %s\", request.Name, openstack.ProviderErrorToString(err))\n\t}\n\tr := servers.CreateResult{}\n\tvar httpResp *http.Response\n\thttpResp, r.Err = client.osclt.Compute.Post(client.osclt.Compute.ServiceURL(\"servers\"), b, &r.Body, &gc.RequestOpts{\n\t\tOkCodes: []int{200, 202},\n\t})\n\tserver, err := r.Extract()\n\tif err != nil {\n\t\tif server != nil {\n\t\t\tservers.Delete(client.osclt.Compute, server.ID)\n\t\t}\n\t\treturn nil, fmt.Errorf(\"query to create host '%s' failed: %s (HTTP return code: %d)\", request.Name, openstack.ProviderErrorToString(err), httpResp.StatusCode)\n\t}\n\n\t// Wait that host is ready, not just that the build is started\n\thost, err := client.WaitHostReady(server.ID, time.Minute*5)\n\tif err != nil {\n\t\tclient.DeleteHost(server.ID)\n\t\treturn nil, fmt.Errorf(\"timeout waiting host '%s' ready: %s\", request.Name, openstack.ProviderErrorToString(err))\n\t}\n\n\t// Fixes the size of bootdisk, FlexibleEngine is used to not give one...\n\thost.Size.DiskSize = diskSize\n\thost.PrivateKey = kp.PrivateKey\n\t//Add gateway ID to host definition\n\tvar gwID string\n\tif gw != nil {\n\t\tgwID = gw.ID\n\t}\n\thost.GatewayID = gwID\n\n\tif request.PublicIP {\n\t\tfip, err := client.attachFloatingIP(host)\n\t\tif err != nil {\n\t\t\tclient.DeleteHost(host.ID)\n\t\t\treturn nil, fmt.Errorf(\"error attaching public IP for host '%s': %s\", request.Name, openstack.ProviderErrorToString(err))\n\t\t}\n\t\tif isGateway {\n\t\t\terr = client.enableHostRouterMode(host)\n\t\t\tif err != nil {\n\t\t\t\tclient.DeleteHost(host.ID)\n\t\t\t\tclient.DeleteFloatingIP(fip.ID)\n\t\t\t\treturn nil, fmt.Errorf(\"error enabling gateway mode of host '%s': %s\", request.Name, openstack.ProviderErrorToString(err))\n\t\t\t}\n\t\t}\n\t}\n\n\treturn host, nil\n}", "func createHostWithIp(nodeId int, ip string, port int) (core.Host, error) {\n\t// Producing private key using nodeId\n\tr := mrand.New(mrand.NewSource(int64(nodeId)))\n\n\tprvKey, _ := ecdsa.GenerateKey(btcec.S256(), r)\n\tsk := (*crypto.Secp256k1PrivateKey)(prvKey)\n\n\t// Starting a peer with default configs\n\topts := []libp2p.Option{\n\t\tlibp2p.ListenAddrStrings(fmt.Sprintf(\"/ip4/0.0.0.0/tcp/%s\", strconv.Itoa(port))),\n\t\tlibp2p.Identity(sk),\n\t\tlibp2p.DefaultTransports,\n\t\tlibp2p.DefaultMuxers,\n\t\tlibp2p.DefaultSecurity,\n\t}\n\n\th, err := libp2p.New(context.Background(), opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn h, nil\n}", "func (s *Set) NewHost(user, address string, port uint16, groups []string) Host {\n}", "func listHostOnlyAdapters(vbox VBoxManager) (map[string]*hostOnlyNetwork, error) {\n\tout, err := vbox.vbmOut(\"list\", \"hostonlyifs\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbyName := map[string]*hostOnlyNetwork{}\n\tbyIP := map[string]*hostOnlyNetwork{}\n\tn := &hostOnlyNetwork{}\n\n\terr = parseKeyValues(out, reColonLine, func(key, val string) error {\n\t\tswitch key {\n\t\tcase \"Name\":\n\t\t\tn.Name = val\n\t\tcase \"GUID\":\n\t\t\tn.GUID = val\n\t\tcase \"DHCP\":\n\t\t\tn.DHCP = (val != \"Disabled\")\n\t\tcase \"IPAddress\":\n\t\t\tn.IPv4.IP = net.ParseIP(val)\n\t\tcase \"NetworkMask\":\n\t\t\tn.IPv4.Mask = parseIPv4Mask(val)\n\t\tcase \"HardwareAddress\":\n\t\t\tmac, err := net.ParseMAC(val)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tn.HwAddr = mac\n\t\tcase \"MediumType\":\n\t\t\tn.Medium = val\n\t\tcase \"Status\":\n\t\t\tn.Status = val\n\t\tcase \"VBoxNetworkName\":\n\t\t\tn.NetworkName = val\n\n\t\t\tif _, present := byName[n.NetworkName]; present {\n\t\t\t\treturn fmt.Errorf(\"VirtualBox is configured with multiple host-only adapters with the same name %q. Please remove one.\", n.NetworkName)\n\t\t\t}\n\t\t\tbyName[n.NetworkName] = n\n\n\t\t\tif len(n.IPv4.IP) != 0 {\n\t\t\t\tif _, present := byIP[n.IPv4.IP.String()]; present {\n\t\t\t\t\treturn fmt.Errorf(\"VirtualBox is configured with multiple host-only adapters with the same IP %q. Please remove one.\", n.IPv4.IP)\n\t\t\t\t}\n\t\t\t\tbyIP[n.IPv4.IP.String()] = n\n\t\t\t}\n\n\t\t\tn = &hostOnlyNetwork{}\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn byName, nil\n}", "func createHost(port int) (core.Host, error) {\n\t// Producing private key\n\tprvKey, _ := ecdsa.GenerateKey(btcec.S256(), rand.Reader)\n\tsk := (*crypto.Secp256k1PrivateKey)(prvKey)\n\n\t// Starting a peer with default configs\n\topts := []libp2p.Option{\n\t\tlibp2p.ListenAddrStrings(fmt.Sprintf(\"/ip4/0.0.0.0/tcp/%d\", port)),\n\t\tlibp2p.Identity(sk),\n\t\tlibp2p.DefaultTransports,\n\t\tlibp2p.DefaultMuxers,\n\t\tlibp2p.DefaultSecurity,\n\t}\n\n\th, err := libp2p.New(context.Background(), opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn h, nil\n}", "func NewHost(host string) Host {\n\treturn Host(host)\n}", "func NewHost(input string) (host Host, err error) {\n\tkeys := strings.Split(input, \" \")\n\tif len(keys) != 3 {\n\t\treturn host, fmt.Errorf(\"invalid host: '%s'\", input)\n\t}\n\n\thost.getNameIP(keys[0])\n\thost.KeyType = keys[1]\n\thost.PubKey = keys[2]\n\n\treturn host, nil\n}", "func (client *Client) CreateHost(request model.HostRequest) (*model.Host, error) {\n\t//msgFail := \"Failed to create Host resource: %s\"\n\tmsgSuccess := fmt.Sprintf(\"Host resource '%s' created successfully\", request.ResourceName)\n\n\tif request.DefaultGateway == nil && !request.PublicIP {\n\t\treturn nil, model.ResourceInvalidRequestError(\"host creation\", \"can't create a host without network and without public access (would be unreachable)\")\n\t}\n\n\t// Validating name of the host\n\tif ok, err := validatehostName(request); !ok {\n\t\treturn nil, fmt.Errorf(\"name '%s' is invalid for a FlexibleEngine Host: %s\", request.ResourceName, openstack.ProviderErrorToString(err))\n\t}\n\n\t// The Default Network is the first of the provided list, by convention\n\tdefaultNetwork := request.Networks[0]\n\tdefaultNetworkID := defaultNetwork.ID\n\tdefaultGateway := request.DefaultGateway\n\tisGateway := defaultGateway == nil && defaultNetwork.Name != model.SingleHostNetworkName\n\tdefaultGatewayID := \"\"\n\tdefaultGatewayPrivateIP := \"\"\n\tif defaultGateway != nil {\n\t\thostNetworkV1 := propsv1.NewHostNetwork()\n\t\terr := defaultGateway.Properties.Get(HostProperty.NetworkV1, hostNetworkV1)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tdefaultGatewayPrivateIP = hostNetworkV1.IPv4Addresses[defaultNetworkID]\n\t\tdefaultGatewayID = defaultGateway.ID\n\t}\n\n\tvar nets []servers.Network\n\t// Add private networks\n\tfor _, n := range request.Networks {\n\t\tnets = append(nets, servers.Network{\n\t\t\tUUID: n.ID,\n\t\t})\n\t}\n\n\t// If no key pair is supplied create one\n\tif request.KeyPair == nil {\n\t\tid, err := uuid.NewV4()\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error creating UID : %v\", err)\n\t\t}\n\n\t\tname := fmt.Sprintf(\"%s_%s\", request.ResourceName, id)\n\t\trequest.KeyPair, err = client.CreateKeyPair(name)\n\t\tif err != nil {\n\t\t\tmsg := fmt.Sprintf(\"failed to create host key pair: %+v\", err)\n\t\t\tlog.Debugf(utils.TitleFirst(msg))\n\t\t}\n\t}\n\n\t// --- prepares data structures for Provider usage ---\n\n\t// Constructs userdata content\n\tuserData, err := userdata.Prepare(client, request, request.KeyPair, defaultNetwork.CIDR)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"failed to prepare user data content: %+v\", err)\n\t\tlog.Debugf(utils.TitleFirst(msg))\n\t\treturn nil, fmt.Errorf(msg)\n\t}\n\n\t// Determine system disk size based on vcpus count\n\ttemplate, err := client.GetTemplate(request.TemplateID)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to get image: %s\", openstack.ProviderErrorToString(err))\n\t}\n\n\t// Determines appropriate disk size\n\tvar diskSize int\n\tif template.HostSize.DiskSize > 0 {\n\t\tdiskSize = template.HostSize.DiskSize\n\t} else if template.HostSize.Cores < 16 {\n\t\tdiskSize = 100\n\t} else if template.HostSize.Cores < 32 {\n\t\tdiskSize = 200\n\t} else {\n\t\tdiskSize = 400\n\t}\n\n\t// Select useable availability zone, the first one in the list\n\tazList, err := client.ListAvailabilityZones(false)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar az string\n\tfor az = range azList {\n\t\tbreak\n\t}\n\tlog.Debugf(\"Selected Availability Zone: '%s'\", az)\n\n\t// Defines boot disk\n\tbootdiskOpts := blockDevice{\n\t\tSourceType: exbfv.SourceImage,\n\t\tDestinationType: exbfv.DestinationVolume,\n\t\tBootIndex: \"0\",\n\t\tDeleteOnTermination: true,\n\t\tUUID: request.ImageID,\n\t\tVolumeType: \"SSD\",\n\t\tVolumeSize: diskSize,\n\t}\n\t// Defines server\n\tsrvOpts := serverCreateOpts{\n\t\tName: request.ResourceName,\n\t\tSecurityGroups: []string{client.SecurityGroup.Name},\n\t\tNetworks: nets,\n\t\tFlavorRef: request.TemplateID,\n\t\tUserData: userData,\n\t\tAvailabilityZone: az,\n\t}\n\t// Defines host \"Extension bootfromvolume\" options\n\tbdOpts := bootdiskCreateOptsExt{\n\t\tCreateOptsBuilder: srvOpts,\n\t\tBlockDevice: []blockDevice{bootdiskOpts},\n\t}\n\tb, err := bdOpts.ToServerCreateMap()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to build query to create host '%s': %s\", request.ResourceName, openstack.ProviderErrorToString(err))\n\t}\n\n\t// --- Initializes model.Host ---\n\n\thost := model.NewHost()\n\thost.PrivateKey = request.KeyPair.PrivateKey // Add PrivateKey to host definition\n\n\thostNetworkV1 := propsv1.NewHostNetwork()\n\thostNetworkV1.IsGateway = isGateway\n\thostNetworkV1.DefaultNetworkID = defaultNetworkID\n\thostNetworkV1.DefaultGatewayID = defaultGatewayID\n\thostNetworkV1.DefaultGatewayPrivateIP = defaultGatewayPrivateIP\n\n\t// Updates Host property NetworkV1\n\terr = host.Properties.Set(HostProperty.NetworkV1, hostNetworkV1)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Adds Host property SizingV1\n\terr = host.Properties.Set(HostProperty.SizingV1, &propsv1.HostSizing{\n\t\t// Note: from there, no idea what was the RequestedSize; caller will have to complement this information\n\t\tTemplate: request.TemplateID,\n\t\tAllocatedSize: template.HostSize,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// --- query provider for host creation ---\n\n\t// Retry creation until success, for 10 minutes\n\tvar (\n\t\thttpResp *http.Response\n\t\tr servers.CreateResult\n\t)\n\n\tretryErr := retry.WhileUnsuccessfulDelay5Seconds(\n\t\tfunc() error {\n\t\t\thttpResp, r.Err = client.osclt.Compute.Post(client.osclt.Compute.ServiceURL(\"servers\"), b, &r.Body, &gc.RequestOpts{\n\t\t\t\tOkCodes: []int{200, 202},\n\t\t\t})\n\t\t\tserver, err := r.Extract()\n\t\t\tif err != nil {\n\t\t\t\tif server != nil {\n\t\t\t\t\tservers.Delete(client.osclt.Compute, server.ID)\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"query to create host '%s' failed: %s (HTTP return code: %d)\", request.ResourceName, openstack.ProviderErrorToString(err), httpResp.StatusCode)\n\t\t\t\t// msg := fmt.Sprintf(msgFail, openstack.ProviderErrorToString(err))\n\t\t\t\t// // TODO Gotcha !!\n\t\t\t\t// log.Debugf(msg)\n\t\t\t\t// return fmt.Errorf(msg)\n\t\t\t}\n\t\t\thost.ID = server.ID\n\n\t\t\tdefer func() {\n\t\t\t\tif err != nil {\n\t\t\t\t\tderr := servers.Delete(client.osclt.Compute, server.ID).ExtractErr()\n\t\t\t\t\tif derr != nil {\n\t\t\t\t\t\tlog.Errorf(\"Failed to delete host '%s': %v\", server.Name, derr)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}()\n\n\t\t\t// Wait that host is ready, not just that the build is started\n\t\t\thost, err = client.WaitHostReady(host, time.Minute*5)\n\t\t\tif err != nil {\n\t\t\t\tswitch err.(type) {\n\t\t\t\tcase model.ErrResourceNotAvailable:\n\t\t\t\t\treturn fmt.Errorf(\"host '%s' is in ERROR state\", request.ResourceName)\n\t\t\t\tdefault:\n\t\t\t\t\treturn fmt.Errorf(\"timeout waiting host '%s' ready: %s\", request.ResourceName, openstack.ProviderErrorToString(err))\n\t\t\t\t\t// msg := fmt.Sprintf(msgFail, openstack.ProviderErrorToString(err))\n\t\t\t\t\t// // TODO Gotcha !!\n\t\t\t\t\t// log.Debugf(msg)\n\t\t\t\t\t// return fmt.Errorf(msg)\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t\t10*time.Minute,\n\t)\n\tif retryErr != nil {\n\t\treturn nil, err\n\t}\n\tif host == nil {\n\t\treturn nil, errors.New(\"unexpected problem creating host\")\n\t}\n\n\t// Starting from here, delete host if exiting with error\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tderr := client.DeleteHost(host.ID)\n\t\t\tif derr != nil {\n\t\t\t\tlog.Warnf(\"Failed to delete host '%s': %v\", host.Name, derr)\n\t\t\t}\n\t\t}\n\t}()\n\n\tif request.PublicIP {\n\t\tfip, err := client.attachFloatingIP(host)\n\t\tif err != nil {\n\t\t\tspew.Dump(err)\n\t\t\treturn nil, fmt.Errorf(\"error attaching public IP for host '%s': %s\", request.ResourceName, openstack.ProviderErrorToString(err))\n\t\t}\n\n\t\t// Starting from here, delete Floating IP if exiting with error\n\t\tdefer func() {\n\t\t\tif err != nil {\n\t\t\t\tderr := client.DeleteFloatingIP(fip.ID)\n\t\t\t\tif derr != nil {\n\t\t\t\t\tlog.Errorf(\"Error deleting Floating IP: %v\", derr)\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\n\t\terr = host.Properties.Get(HostProperty.NetworkV1, hostNetworkV1)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif IPVersion.IPv4.Is(fip.PublicIPAddress) {\n\t\t\thostNetworkV1.PublicIPv4 = fip.PublicIPAddress\n\t\t} else if IPVersion.IPv6.Is(fip.PublicIPAddress) {\n\t\t\thostNetworkV1.PublicIPv6 = fip.PublicIPAddress\n\t\t}\n\n\t\t// Updates Host property NetworkV1 in host instance\n\t\terr = host.Properties.Set(HostProperty.NetworkV1, hostNetworkV1)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif defaultGateway == nil && defaultNetwork.Name != model.SingleHostNetworkName {\n\t\t\terr = client.enableHostRouterMode(host)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"error enabling gateway mode of host '%s': %s\", request.ResourceName, openstack.ProviderErrorToString(err))\n\t\t\t}\n\t\t}\n\t}\n\n\tlog.Infoln(msgSuccess)\n\treturn host, nil\n}", "func addHostOnlyDHCPServer(ifname string, d dhcpServer, vbox VBoxManager) error {\n\tname := dhcpPrefix + ifname\n\n\tdhcps, err := listDHCPServers(vbox)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// On some platforms (OSX), creating a host-only adapter adds a default dhcpserver,\n\t// while on others (Windows?) it does not.\n\tcommand := \"add\"\n\tif dhcp, ok := dhcps[name]; ok {\n\t\tcommand = \"modify\"\n\t\tif (dhcp.IPv4.IP.Equal(d.IPv4.IP)) && (dhcp.IPv4.Mask.String() == d.IPv4.Mask.String()) && (dhcp.LowerIP.Equal(d.LowerIP)) && (dhcp.UpperIP.Equal(d.UpperIP)) && dhcp.Enabled {\n\t\t\t// dhcp is up to date\n\t\t\treturn nil\n\t\t}\n\t}\n\n\targs := []string{\"dhcpserver\", command,\n\t\t\"--netname\", name,\n\t\t\"--ip\", d.IPv4.IP.String(),\n\t\t\"--netmask\", net.IP(d.IPv4.Mask).String(),\n\t\t\"--lowerip\", d.LowerIP.String(),\n\t\t\"--upperip\", d.UpperIP.String(),\n\t}\n\tif d.Enabled {\n\t\targs = append(args, \"--enable\")\n\t} else {\n\t\targs = append(args, \"--disable\")\n\t}\n\n\treturn vbox.vbm(args...)\n}", "func NewBindHostForbidden() *BindHostForbidden {\n\treturn &BindHostForbidden{}\n}", "func NewHost() *Host {\n\tthis := Host{}\n\treturn &this\n}", "func NewPerHost(defaultDialer, bypass Dialer) *PerHost {\n\treturn &PerHost{\n\t\tdef: defaultDialer,\n\t\tbypass: bypass,\n\t}\n}", "func getHostOnlyNetworkInterface(mc *driver.MachineConfig) (string, error) {\n\t// Check if the interface/dhcp exists.\n\tnets, err := HostonlyNets()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tdhcps, err := DHCPs()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor _, n := range nets {\n\t\tif dhcp, ok := dhcps[n.NetworkName]; ok {\n\t\t\tif dhcp.IPv4.IP.Equal(mc.DHCPIP) &&\n\t\t\t\tdhcp.IPv4.Mask.String() == mc.NetMask.String() &&\n\t\t\t\tdhcp.LowerIP.Equal(mc.LowerIP) &&\n\t\t\t\tdhcp.UpperIP.Equal(mc.UpperIP) &&\n\t\t\t\tdhcp.Enabled == mc.DHCPEnabled {\n\t\t\t\treturn n.Name, nil\n\t\t\t}\n\t\t}\n\t}\n\n\t// No existing host-only interface found. Create a new one.\n\thostonlyNet, err := CreateHostonlyNet()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\thostonlyNet.IPv4.IP = mc.HostIP\n\thostonlyNet.IPv4.Mask = mc.NetMask\n\tif err := hostonlyNet.Config(); err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create and add a DHCP server to the host-only network\n\tdhcp := driver.DHCP{}\n\tdhcp.IPv4.IP = mc.DHCPIP\n\tdhcp.IPv4.Mask = mc.NetMask\n\tdhcp.LowerIP = mc.LowerIP\n\tdhcp.UpperIP = mc.UpperIP\n\tdhcp.Enabled = true\n\tif err := AddHostonlyDHCP(hostonlyNet.Name, dhcp); err != nil {\n\t\treturn \"\", err\n\t}\n\treturn hostonlyNet.Name, nil\n}", "func NewTransportConfigHostonly() *TransportConfig {\n\treturn &TransportConfig{\n\t\tComponentNumber: 1,\n\t}\n}", "func NewAdapter(g *gce.Cloud) NetworkEndpointGroupCloud {\n\treturn &cloudProviderAdapter{\n\t\tc: g,\n\t\tnetworkURL: g.NetworkURL(),\n\t\tsubnetworkURL: g.SubnetworkURL(),\n\t}\n}", "func NewHost(address string) (*Host, error) {\n\taddr, err := NewAddress(address)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"Failed to create Host\")\n\t}\n\treturn &Host{Address: addr}, nil\n}", "func createHnsNetwork(backend string, networkAdapter string) (string, error) {\n\tvar network hcsshim.HNSNetwork\n\tif backend == \"vxlan\" {\n\t\t// Ignoring the return because both true and false without an error represent that the firewall rule was created or already exists\n\t\tif _, err := wapi.FirewallRuleAdd(\"OverlayTraffic4789UDP\", \"Overlay network traffic UDP\", \"\", \"4789\", wapi.NET_FW_IP_PROTOCOL_UDP, wapi.NET_FW_PROFILE2_ALL); err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"error creating firewall rules: %v\", err)\n\t\t}\n\t\tlogrus.Infof(\"Creating VXLAN network using the vxlanAdapter: %s\", networkAdapter)\n\t\tnetwork = hcsshim.HNSNetwork{\n\t\t\tType: \"Overlay\",\n\t\t\tName: CalicoHnsNetworkName,\n\t\t\tNetworkAdapterName: networkAdapter,\n\t\t\tSubnets: []hcsshim.Subnet{\n\t\t\t\t{\n\t\t\t\t\tAddressPrefix: \"192.168.255.0/30\",\n\t\t\t\t\tGatewayAddress: \"192.168.255.1\",\n\t\t\t\t\tPolicies: []json.RawMessage{\n\t\t\t\t\t\t[]byte(\"{ \\\"Type\\\": \\\"VSID\\\", \\\"VSID\\\": 9999 }\"),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t} else {\n\t\tnetwork = hcsshim.HNSNetwork{\n\t\t\tType: \"L2Bridge\",\n\t\t\tName: CalicoHnsNetworkName,\n\t\t\tNetworkAdapterName: networkAdapter,\n\t\t\tSubnets: []hcsshim.Subnet{\n\t\t\t\t{\n\t\t\t\t\tAddressPrefix: \"192.168.255.0/30\",\n\t\t\t\t\tGatewayAddress: \"192.168.255.1\",\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t}\n\n\tif _, err := network.Create(); err != nil {\n\t\treturn \"\", fmt.Errorf(\"error creating the %s network: %v\", CalicoHnsNetworkName, err)\n\t}\n\n\t// Check if network exists. If it does not after 5 minutes, fail\n\tfor start := time.Now(); time.Since(start) < 5*time.Minute; {\n\t\tnetwork, err := hcsshim.GetHNSNetworkByName(CalicoHnsNetworkName)\n\t\tif err == nil {\n\t\t\treturn network.ManagementIP, nil\n\t\t}\n\t}\n\n\treturn \"\", fmt.Errorf(\"failed to create %s network\", CalicoHnsNetworkName)\n}", "func (client *Client) CreateHost(request model.HostRequest) (*model.Host, error) {\n\treturn client.osclt.CreateHost(request)\n}", "func (client *Client) CreateHost(request model.HostRequest) (*model.Host, error) {\n\treturn client.feclt.CreateHost(request)\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
listHostOnlyAdapters gets all hostonly adapters in a map keyed by NetworkName.
func listHostOnlyAdapters(vbox VBoxManager) (map[string]*hostOnlyNetwork, error) { out, err := vbox.vbmOut("list", "hostonlyifs") if err != nil { return nil, err } byName := map[string]*hostOnlyNetwork{} byIP := map[string]*hostOnlyNetwork{} n := &hostOnlyNetwork{} err = parseKeyValues(out, reColonLine, func(key, val string) error { switch key { case "Name": n.Name = val case "GUID": n.GUID = val case "DHCP": n.DHCP = (val != "Disabled") case "IPAddress": n.IPv4.IP = net.ParseIP(val) case "NetworkMask": n.IPv4.Mask = parseIPv4Mask(val) case "HardwareAddress": mac, err := net.ParseMAC(val) if err != nil { return err } n.HwAddr = mac case "MediumType": n.Medium = val case "Status": n.Status = val case "VBoxNetworkName": n.NetworkName = val if _, present := byName[n.NetworkName]; present { return fmt.Errorf("VirtualBox is configured with multiple host-only adapters with the same name %q. Please remove one.", n.NetworkName) } byName[n.NetworkName] = n if len(n.IPv4.IP) != 0 { if _, present := byIP[n.IPv4.IP.String()]; present { return fmt.Errorf("VirtualBox is configured with multiple host-only adapters with the same IP %q. Please remove one.", n.IPv4.IP) } byIP[n.IPv4.IP.String()] = n } n = &hostOnlyNetwork{} } return nil }) if err != nil { return nil, err } return byName, nil }
[ "func (this *ZkCluster) OnlyNamedBrokerList() []string {\n\tr := make([]string, 0)\n\tfor _, broker := range this.Brokers() {\n\t\taddr, hasDns := broker.NamedAddr()\n\t\tif hasDns {\n\t\t\tr = append(r, addr)\n\t\t}\n\t}\n\n\tif len(r) == 0 {\n\t\t// however, no brokers has dns records(e,g. in test/dev environment)\n\t\tfor _, broker := range this.Brokers() {\n\t\t\taddr, _ := broker.NamedAddr()\n\t\t\tr = append(r, addr)\n\t\t}\n\t}\n\n\treturn r\n}", "func listDHCPServers(vbox VBoxManager) (map[string]*dhcpServer, error) {\n\tout, err := vbox.vbmOut(\"list\", \"dhcpservers\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tm := map[string]*dhcpServer{}\n\tdhcp := &dhcpServer{}\n\n\terr = parseKeyValues(out, reColonLine, func(key, val string) error {\n\t\tswitch key {\n\t\tcase \"NetworkName\":\n\t\t\tdhcp = &dhcpServer{}\n\t\t\tm[val] = dhcp\n\t\t\tdhcp.NetworkName = val\n\t\tcase \"IP\":\n\t\t\tdhcp.IPv4.IP = net.ParseIP(val)\n\t\tcase \"upperIPAddress\":\n\t\t\tdhcp.UpperIP = net.ParseIP(val)\n\t\tcase \"lowerIPAddress\":\n\t\t\tdhcp.LowerIP = net.ParseIP(val)\n\t\tcase \"NetworkMask\":\n\t\t\tdhcp.IPv4.Mask = parseIPv4Mask(val)\n\t\tcase \"Enabled\":\n\t\t\tdhcp.Enabled = (val == \"Yes\")\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn m, nil\n}", "func AddressesForHost(host string) []string {\n\tss := collection.NewStringSet()\n\tif host == \"\" { // All address on machine\n\t\tif iFaces, err := net.Interfaces(); err == nil {\n\t\t\tfor _, iFace := range iFaces {\n\t\t\t\tconst interesting = net.FlagUp | net.FlagBroadcast\n\t\t\t\tif iFace.Flags&interesting == interesting {\n\t\t\t\t\tvar addrs []net.Addr\n\t\t\t\t\tif addrs, err = iFace.Addrs(); err == nil {\n\t\t\t\t\t\tfor _, addr := range addrs {\n\t\t\t\t\t\t\tvar ip net.IP\n\t\t\t\t\t\t\tswitch v := addr.(type) {\n\t\t\t\t\t\t\tcase *net.IPNet:\n\t\t\t\t\t\t\t\tip = v.IP\n\t\t\t\t\t\t\tcase *net.IPAddr:\n\t\t\t\t\t\t\t\tip = v.IP\n\t\t\t\t\t\t\tdefault:\n\t\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif ip.IsGlobalUnicast() {\n\t\t\t\t\t\t\t\tss.Add(ip.String())\n\t\t\t\t\t\t\t\tvar names []string\n\t\t\t\t\t\t\t\tif names, err = net.LookupAddr(ip.String()); err == nil {\n\t\t\t\t\t\t\t\t\tfor _, name := range names {\n\t\t\t\t\t\t\t\t\t\tif strings.HasSuffix(name, \".\") {\n\t\t\t\t\t\t\t\t\t\t\tname = name[:len(name)-1]\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t\tss.Add(name)\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} else {\n\t\tss.Add(host)\n\t\tif net.ParseIP(host) == nil {\n\t\t\tif ips, err := net.LookupIP(host); err == nil && len(ips) > 0 {\n\t\t\t\tfor _, ip := range ips {\n\t\t\t\t\tss.Add(ip.String())\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tfor _, one := range []string{\"::\", \"::1\", \"127.0.0.1\"} {\n\t\tif ss.Contains(one) {\n\t\t\tdelete(ss, one)\n\t\t\tss.Add(\"localhost\")\n\t\t}\n\t}\n\taddrs := ss.Values()\n\tsort.Slice(addrs, func(i, j int) bool {\n\t\tisName1 := net.ParseIP(addrs[i]) == nil\n\t\tisName2 := net.ParseIP(addrs[j]) == nil\n\t\tif isName1 == isName2 {\n\t\t\treturn txt.NaturalLess(addrs[i], addrs[j], true)\n\t\t}\n\t\treturn isName1\n\t})\n\treturn addrs\n}", "func createHostonlyAdapter(vbox VBoxManager) (*hostOnlyNetwork, error) {\n\tout, err := vbox.vbmOut(\"hostonlyif\", \"create\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres := reHostOnlyAdapterCreated.FindStringSubmatch(string(out))\n\tif res == nil {\n\t\treturn nil, errors.New(\"Failed to create host-only adapter\")\n\t}\n\n\treturn &hostOnlyNetwork{Name: res[1]}, nil\n}", "func GetHostIPNetworks(skipInterfaces []string) ([]*net.IPNet, []net.IP, error) {\n\thostInterfaces, err := net.Interfaces()\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tskipInterfaceMap := make(map[string]bool)\n\tfor _, ifaceName := range skipInterfaces {\n\t\tskipInterfaceMap[ifaceName] = true\n\t}\n\n\terrList := []error{}\n\tvar hostIPNets []*net.IPNet\n\tvar hostIPs []net.IP\n\tfor _, iface := range hostInterfaces {\n\t\tif skipInterfaceMap[iface.Name] {\n\t\t\tcontinue\n\t\t}\n\n\t\tifAddrs, err := iface.Addrs()\n\t\tif err != nil {\n\t\t\terrList = append(errList, err)\n\t\t\tcontinue\n\t\t}\n\t\tfor _, addr := range ifAddrs {\n\t\t\tip, ipNet, err := net.ParseCIDR(addr.String())\n\t\t\tif err != nil {\n\t\t\t\terrList = append(errList, err)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Skip loopback and non IPv4 addrs\n\t\t\tif !ip.IsLoopback() && ip.To4() != nil {\n\t\t\t\thostIPNets = append(hostIPNets, ipNet)\n\t\t\t\thostIPs = append(hostIPs, ip)\n\t\t\t}\n\t\t}\n\t}\n\treturn hostIPNets, hostIPs, kerrors.NewAggregate(errList)\n}", "func (c *Connection) ListExposed(ctx context.Context) ([]Port, error) {\n\tdirs, err := c.client.List(ctx, []string{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresults := make([]Port, 0)\n\n\tfor _, name := range dirs {\n\t\tport, err := parse(name)\n\t\tif err != nil {\n\t\t\t// there are some special files like \".\" and \"README\" to ignore\n\t\t\tcontinue\n\t\t}\n\t\tresults = append(results, *port)\n\t}\n\n\treturn results, nil\n}", "func (p DCNone) List() (map[string]string, error) {\n\tentries := make(map[string]string)\n\tconfig, err := getParsedConfig()\n\tif err != nil {\n\t\treturn entries, err\n\t}\n\tauthsInterface, ok := config[\"auths\"]\n\tif ok {\n\t\tauths, ok := authsInterface.(map[string]interface{})\n\t\tif !ok {\n\t\t\treturn entries, fmt.Errorf(\"Unexpected data: %v: not a hash\\n\", authsInterface)\n\t\t}\n\t\tfor url := range auths {\n\t\t\tusername, _, err := getRecordForServerURL(&config, url)\n\t\t\tif username != \"\" && err == nil {\n\t\t\t\tentries[url] = username\n\t\t\t}\n\t\t}\n\t}\n\treturn entries, nil\n}", "func (c *Eth) ShowList() ([]string, error) {\n c.con.LogQuery(\"(show) list of ethernet interfaces\")\n path := c.xpath(nil)\n return c.con.EntryListUsing(c.con.Show, path[:len(path) - 1])\n}", "func NonTailscaleInterfaces() (map[winipcfg.LUID]*winipcfg.IPAdapterAddresses, error) {\n\tifs, err := winipcfg.GetAdaptersAddresses(windows.AF_UNSPEC, winipcfg.GAAFlagIncludeAllInterfaces)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tret := map[winipcfg.LUID]*winipcfg.IPAdapterAddresses{}\n\tfor _, iface := range ifs {\n\t\tif iface.Description() == tsconst.WintunInterfaceDesc {\n\t\t\tcontinue\n\t\t}\n\t\tret[iface.LUID] = iface\n\t}\n\n\treturn ret, nil\n}", "func (client *Client) ShowHostMaps(host string) ([]Volume, *ResponseStatus, error) {\n\tif len(host) > 0 {\n\t\thost = fmt.Sprintf(\"\\\"%s\\\"\", host)\n\t}\n\tres, status, err := client.FormattedRequest(\"/show/host-maps/%s\", host)\n\tif err != nil {\n\t\treturn nil, status, err\n\t}\n\n\tmappings := make([]Volume, 0)\n\tfor _, rootObj := range res.Objects {\n\t\tif rootObj.Name != \"host-view\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, object := range rootObj.Objects {\n\t\t\tif object.Name == \"volume-view\" {\n\t\t\t\tvol := Volume{}\n\t\t\t\tvol.fillFromObject(&object)\n\t\t\t\tmappings = append(mappings, vol)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn mappings, status, err\n}", "func (d *RestrictedDialer) AllowedHosts() []string {\n\tranges := []string{}\n\tfor _, ipRange := range d.allowedHosts {\n\t\tranges = append(ranges, ipRange.String())\n\t}\n\treturn ranges\n}", "func (h *ConfigHandler) GetHostList(ctx *fasthttp.RequestCtx) {\n\tuser, ok := common.GlobalSession.GetUser(ctx.ID())\n\tif !ok {\n\t\th.WriteJSON(ctx, nil, common.NewNotLoginError())\n\t\treturn\n\t}\n\n\tconf, err := h.Service.GetVPNConfig(context.Background(), &user)\n\tif err != nil {\n\t\th.WriteJSON(ctx, nil, err)\n\t\treturn\n\t}\n\tdata := vpnConfigResponseEncode(conf)\n\th.WriteJSON(ctx, map[string]interface{}{\n\t\t\"list\": data.Hosts,\n\t}, nil)\n\treturn\n}", "func GetAdapters() ([]HCIConfigResult, error) {\n\n\tout, err := cmd.Exec(\"hciconfig\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif len(out) == 0 {\n\t\treturn nil, errors.New(\"hciconfig provided no response\")\n\t}\n\n\tlist := []HCIConfigResult{}\n\tparts := strings.Split(out, \"\\nhci\")\n\n\tfor i, el := range parts {\n\t\tif i > 0 {\n\t\t\tel = \"hci\" + el\n\t\t}\n\t\tcfg := parseControllerInfo(el)\n\t\tlist = append(list, cfg)\n\t}\n\n\t// log.Debugf(\"%++v\", list)\n\n\treturn list, nil\n}", "func (m *MicroService) blackListHost(host string, blacklist bool) {\r\n\tfor idx, inst := range m.Instances {\r\n\t\tif inst.Host == host {\r\n\t\t\tm.blackList(idx, blacklist)\r\n\t\t}\r\n\t}\r\n}", "func (ipb *IPBlocker) HostsToBlock() []RemoteHost {\n\tvar hosts []RemoteHost\n\n\t// key `LocalIP:RemoteIP` formatted string\n\tfor key, portMap := range ipb.IPPortTime {\n\t\tif len(portMap) >= 3 {\n\n\t\t\tvar ports []uint16\n\t\t\tfor port := range portMap {\n\t\t\t\tports = append(ports, port)\n\n\t\t\t\tdelete(ipb.IPPortTime[key], port)\n\t\t\t}\n\n\t\t\tips := strings.Split(key, \":\")\n\t\t\thosts = append(hosts, RemoteHost{\n\t\t\t\tRemoteIP: net.ParseIP(ips[1]),\n\t\t\t\tLocalIP: net.ParseIP(ips[0]),\n\t\t\t\tPorts: ports,\n\t\t\t})\n\t\t}\n\t}\n\n\treturn hosts\n}", "func getHostOnlyNetworkInterface(mc *driver.MachineConfig) (string, error) {\n\t// Check if the interface/dhcp exists.\n\tnets, err := HostonlyNets()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tdhcps, err := DHCPs()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor _, n := range nets {\n\t\tif dhcp, ok := dhcps[n.NetworkName]; ok {\n\t\t\tif dhcp.IPv4.IP.Equal(mc.DHCPIP) &&\n\t\t\t\tdhcp.IPv4.Mask.String() == mc.NetMask.String() &&\n\t\t\t\tdhcp.LowerIP.Equal(mc.LowerIP) &&\n\t\t\t\tdhcp.UpperIP.Equal(mc.UpperIP) &&\n\t\t\t\tdhcp.Enabled == mc.DHCPEnabled {\n\t\t\t\treturn n.Name, nil\n\t\t\t}\n\t\t}\n\t}\n\n\t// No existing host-only interface found. Create a new one.\n\thostonlyNet, err := CreateHostonlyNet()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\thostonlyNet.IPv4.IP = mc.HostIP\n\thostonlyNet.IPv4.Mask = mc.NetMask\n\tif err := hostonlyNet.Config(); err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create and add a DHCP server to the host-only network\n\tdhcp := driver.DHCP{}\n\tdhcp.IPv4.IP = mc.DHCPIP\n\tdhcp.IPv4.Mask = mc.NetMask\n\tdhcp.LowerIP = mc.LowerIP\n\tdhcp.UpperIP = mc.UpperIP\n\tdhcp.Enabled = true\n\tif err := AddHostonlyDHCP(hostonlyNet.Name, dhcp); err != nil {\n\t\treturn \"\", err\n\t}\n\treturn hostonlyNet.Name, nil\n}", "func WhiteListHostFilter(hosts ...string) HostFilter {\n\thostInfos, err := addrsToHosts(hosts, 9042)\n\tif err != nil {\n\t\t// dont want to panic here, but rather not break the API\n\t\tpanic(fmt.Errorf(\"unable to lookup host info from address: %v\", err))\n\t}\n\n\tm := make(map[string]bool, len(hostInfos))\n\tfor _, host := range hostInfos {\n\t\tm[host.ConnectAddress().String()] = true\n\t}\n\n\treturn HostFilterFunc(func(host *HostInfo) bool {\n\t\treturn m[host.ConnectAddress().String()]\n\t})\n}", "func (b *BlubberBlockDirectory) ListHosts(\n\tvoid blubberstore.Empty, hosts *blubberstore.BlockHolderList) error {\n\tvar host string\n\tb.blockMapMtx.RLock()\n\tdefer b.blockMapMtx.RUnlock()\n\n\tfor host, _ = range b.blockHostMap {\n\t\thosts.HostPort = append(hosts.HostPort, host)\n\t}\n\n\treturn nil\n}", "func (h Hostingv4) ListDisks(diskfilter hosting.DiskFilter) ([]hosting.Disk, error) {\n\tfilterv4, err := toDiskFilterv4(diskfilter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfilter, _ := structToMap(filterv4)\n\n\tresponse := []diskv4{}\n\tparams := []interface{}{}\n\tif len(filter) > 0 {\n\t\tparams = append(params, filter)\n\t}\n\t// disk.list and disk.info return the same information\n\terr = h.Send(\"hosting.disk.list\", params, &response)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar disks []hosting.Disk\n\tfor _, disk := range response {\n\t\tdisks = append(disks, fromDiskv4(disk))\n\t}\n\treturn disks, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
removeOrphanDHCPServers removed the DHCP servers linked to no hostonly adapter
func removeOrphanDHCPServers(vbox VBoxManager) error { dhcps, err := listDHCPServers(vbox) if err != nil { return err } if len(dhcps) == 0 { return nil } nets, err := listHostOnlyAdapters(vbox) if err != nil { return err } for name := range dhcps { if strings.HasPrefix(name, dhcpPrefix) { if _, present := nets[name]; !present { if err := vbox.vbm("dhcpserver", "remove", "--netname", name); err != nil { log.Warnf("Unable to remove orphan dhcp server %q: %s", name, err) } } } } return nil }
[ "func DHCPsDelete() error {\n\tdhcps, err := DHCPsGet()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, d := range dhcps {\n\t\tif isUnmanaged(UnmanagedID(d.Ifname), LINKTYPE) {\n\t\t\tlogger.Log.Info(fmt.Sprintf(\"Skipping Unmanaged Link %v DHCP configuration\", d.Ifname))\n\t\t\tcontinue\n\t\t}\n\t\terr = DHCPDelete(d.Ifname)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func listDHCPServers(vbox VBoxManager) (map[string]*dhcpServer, error) {\n\tout, err := vbox.vbmOut(\"list\", \"dhcpservers\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tm := map[string]*dhcpServer{}\n\tdhcp := &dhcpServer{}\n\n\terr = parseKeyValues(out, reColonLine, func(key, val string) error {\n\t\tswitch key {\n\t\tcase \"NetworkName\":\n\t\t\tdhcp = &dhcpServer{}\n\t\t\tm[val] = dhcp\n\t\t\tdhcp.NetworkName = val\n\t\tcase \"IP\":\n\t\t\tdhcp.IPv4.IP = net.ParseIP(val)\n\t\tcase \"upperIPAddress\":\n\t\t\tdhcp.UpperIP = net.ParseIP(val)\n\t\tcase \"lowerIPAddress\":\n\t\t\tdhcp.LowerIP = net.ParseIP(val)\n\t\tcase \"NetworkMask\":\n\t\t\tdhcp.IPv4.Mask = parseIPv4Mask(val)\n\t\tcase \"Enabled\":\n\t\t\tdhcp.Enabled = (val == \"Yes\")\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn m, nil\n}", "func addHostOnlyDHCPServer(ifname string, d dhcpServer, vbox VBoxManager) error {\n\tname := dhcpPrefix + ifname\n\n\tdhcps, err := listDHCPServers(vbox)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// On some platforms (OSX), creating a host-only adapter adds a default dhcpserver,\n\t// while on others (Windows?) it does not.\n\tcommand := \"add\"\n\tif dhcp, ok := dhcps[name]; ok {\n\t\tcommand = \"modify\"\n\t\tif (dhcp.IPv4.IP.Equal(d.IPv4.IP)) && (dhcp.IPv4.Mask.String() == d.IPv4.Mask.String()) && (dhcp.LowerIP.Equal(d.LowerIP)) && (dhcp.UpperIP.Equal(d.UpperIP)) && dhcp.Enabled {\n\t\t\t// dhcp is up to date\n\t\t\treturn nil\n\t\t}\n\t}\n\n\targs := []string{\"dhcpserver\", command,\n\t\t\"--netname\", name,\n\t\t\"--ip\", d.IPv4.IP.String(),\n\t\t\"--netmask\", net.IP(d.IPv4.Mask).String(),\n\t\t\"--lowerip\", d.LowerIP.String(),\n\t\t\"--upperip\", d.UpperIP.String(),\n\t}\n\tif d.Enabled {\n\t\targs = append(args, \"--enable\")\n\t} else {\n\t\targs = append(args, \"--disable\")\n\t}\n\n\treturn vbox.vbm(args...)\n}", "func (s *Server) pruneDeadServers(autopilotConfig *structs.AutopilotConfig) error {\n\t// Find any failed servers\n\tvar failed []string\n\tstaleRaftServers := make(map[string]raft.Server)\n\tif autopilotConfig.CleanupDeadServers {\n\t\tfuture := s.raft.GetConfiguration()\n\t\tif err := future.Error(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfor _, server := range future.Configuration().Servers {\n\t\t\tstaleRaftServers[string(server.Address)] = server\n\t\t}\n\n\t\tfor _, member := range s.serfLAN.Members() {\n\t\t\tvalid, parts := agent.IsConsulServer(member)\n\n\t\t\tif valid {\n\t\t\t\t// Remove this server from the stale list; it has a serf entry\n\t\t\t\tif _, ok := staleRaftServers[parts.Addr.String()]; ok {\n\t\t\t\t\tdelete(staleRaftServers, parts.Addr.String())\n\t\t\t\t}\n\n\t\t\t\tif member.Status == serf.StatusFailed {\n\t\t\t\t\tfailed = append(failed, member.Name)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tremovalCount := len(failed) + len(staleRaftServers)\n\n\t// Nothing to remove, return early\n\tif removalCount == 0 {\n\t\treturn nil\n\t}\n\n\tpeers, err := s.numPeers()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Only do removals if a minority of servers will be affected\n\tif removalCount < peers/2 {\n\t\tfor _, server := range failed {\n\t\t\ts.logger.Printf(\"[INFO] autopilot: Attempting removal of failed server: %v\", server)\n\t\t\tgo s.serfLAN.RemoveFailedNode(server)\n\t\t}\n\n\t\tminRaftProtocol, err := ServerMinRaftProtocol(s.serfLAN.Members())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor _, raftServer := range staleRaftServers {\n\t\t\tvar future raft.Future\n\t\t\tif minRaftProtocol >= 2 {\n\t\t\t\ts.logger.Printf(\"[INFO] autopilot: Attempting removal of stale raft server : %v\", raftServer.ID)\n\t\t\t\tfuture = s.raft.RemoveServer(raftServer.ID, 0, 0)\n\t\t\t} else {\n\t\t\t\ts.logger.Printf(\"[INFO] autopilot: Attempting removal of stale raft server : %v\", raftServer.ID)\n\t\t\t\tfuture = s.raft.RemovePeer(raftServer.Address)\n\t\t\t}\n\t\t\tif err := future.Error(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t} else {\n\t\ts.logger.Printf(\"[DEBUG] autopilot: Failed to remove dead servers: too many dead servers: %d/%d\", removalCount, peers)\n\t}\n\n\treturn nil\n}", "func removeServerInConfig(server string) {\n\tfor k, v := range selfConf.Servers {\n\t\tif v == server {\n\t\t\tselfConf.Servers = selfConf.Servers[:k+copy(selfConf.Servers[k:], selfConf.Servers[k+1:])]\n\t\t}\n\t}\n}", "func (p *Pool) RemoveHostAndPlugins(host string) {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tfor plHostName, pls := range p.hosts {\n\t\tif plHostName == host {\n\t\t\t// stop all plugins\n\t\t\tfor _, pl := range pls.plugins {\n\t\t\t\tpl.stop()\n\t\t\t}\n\t\t\tdelete(p.hosts, host)\n\t\t}\n\t}\n}", "func deleteInexistentServers(ctx context.Context, db *sql.DB, domain *Host) error {\r\n\r\n\tconst selectSQL = \"SELECT * FROM \" + databaseName + \".servers WHERE domain_name = $1;\"\r\n\tserverRows, _ := db.QueryContext(ctx, selectSQL, domain.Name)\r\n\tdefer serverRows.Close()\r\n\r\n\tfor serverRows.Next() {\r\n\t\tvar server Server\r\n\t\terr := serverRows.Scan(&server.Ip, &server.SslGrade, &server.Country, &server.Company, &domain.Name)\r\n\t\tif err != nil {\r\n\t\t\tlog.Fatal(err)\r\n\t\t\treturn err\r\n\t\t}\r\n\r\n\t\tif !containsServer(domain.Servers, server.Ip) {\r\n\t\t\terr := deleteServer(ctx, db, &server)\r\n\t\t\t//A server is deleted, thus the servers changed attribute has to be true\r\n\t\t\tdomain.ServersChanged = true\r\n\t\t\tif err != nil {\r\n\t\t\t\tlog.Fatal(err)\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t}\r\n\treturn nil\r\n\r\n}", "func removeHost(n *libvirt.VirNetwork, ip, mac, name string) error {\n\txmlDesc := getHostXMLDesc(ip, mac, name)\n\tlog.Printf(\"Removing host with XML:\\n%s\", xmlDesc)\n\treturn n.UpdateXMLDesc(xmlDesc, libvirt.VIR_NETWORK_UPDATE_COMMAND_DELETE, libvirt.VIR_NETWORK_SECTION_IP_DHCP_HOST)\n}", "func (ir *IpvsReconciler) CleanupOrphanedIpvs() error {\n\t// TODO: consider a less-greedy lock for clean up\n\t// this lock is for potential race-conditions with the vip interface clean up\n\tir.m.Lock()\n\tdefer ir.m.Unlock()\n\n\terr := ir.cleanupOrphanedIpvsServices()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, svc := range ir.sm.ListServicesWithAddress() {\n\t\terr := ir.cleanupOrphanedIpvsDestinations(svc)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\terr = ir.cleanupOrphanedIPAddresses()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func cleanRemotes(conf *Config) error {\n\tvar lastError error\n\tfor _, backend := range conf.Backends {\n\t\tremote := backend.Remote\n\t\tlog.Printf(\"%q - Cleaning\", remote)\n\t\terr := cleanFs(context.Background(), remote, backend.CleanUp)\n\t\tif err != nil {\n\t\t\tlastError = err\n\t\t\tlog.Printf(\"Failed to purge %q: %v\", remote, err)\n\t\t}\n\t}\n\treturn lastError\n}", "func (g *operationGroup) removeClusterStateServers(hostnames []string) error {\n\tg.Lock()\n\tdefer g.Unlock()\n\n\tsite, err := g.operator.backend().GetSite(g.siteKey.SiteDomain)\n\tif err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\n\tvar servers []storage.Server\n\tfor _, server := range site.ClusterState.Servers {\n\t\tif !utils.StringInSlice(hostnames, server.Hostname) {\n\t\t\tservers = append(servers, server)\n\t\t}\n\t}\n\n\tsite.ClusterState.Servers = servers\n\tif _, err = g.operator.backend().UpdateSite(*site); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\n\treturn nil\n}", "func RemoveDNS(host string) {\n\texecuteCommand(\"sudo\", []string{\n\t\t\"ghost\",\n\t\t\"delete\",\n\t\thost,\n\t})\n}", "func (adminOrg *AdminOrg) removeAllOrgVDCs() error {\n\tfor _, vdcs := range adminOrg.AdminOrg.Vdcs.Vdcs {\n\n\t\tadminVdcUrl := adminOrg.client.VCDHREF\n\t\tsplitVdcId := strings.Split(vdcs.HREF, \"/api/vdc/\")\n\t\tif len(splitVdcId) == 1 {\n\t\t\tadminVdcUrl.Path += \"/admin/vdc/\" + strings.Split(vdcs.HREF, \"/api/admin/vdc/\")[1] + \"/action/disable\"\n\t\t} else {\n\t\t\tadminVdcUrl.Path += \"/admin/vdc/\" + splitVdcId[1] + \"/action/disable\"\n\t\t}\n\n\t\treq := adminOrg.client.NewRequest(map[string]string{}, http.MethodPost, adminVdcUrl, nil)\n\t\t_, err := checkResp(adminOrg.client.Http.Do(req))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error disabling vdc: %s\", err)\n\t\t}\n\t\t// Get admin vdc HREF for normal deletion\n\t\tadminVdcUrl.Path = strings.Split(adminVdcUrl.Path, \"/action/disable\")[0]\n\t\treq = adminOrg.client.NewRequest(map[string]string{\n\t\t\t\"recursive\": \"true\",\n\t\t\t\"force\": \"true\",\n\t\t}, http.MethodDelete, adminVdcUrl, nil)\n\t\tresp, err := checkResp(adminOrg.client.Http.Do(req))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error deleting vdc: %s\", err)\n\t\t}\n\t\ttask := NewTask(adminOrg.client)\n\t\tif err = decodeBody(types.BodyTypeXML, resp, task.Task); err != nil {\n\t\t\treturn fmt.Errorf(\"error decoding task response: %s\", err)\n\t\t}\n\t\tif task.Task.Status == \"error\" {\n\t\t\treturn fmt.Errorf(\"vdc not properly destroyed\")\n\t\t}\n\t\terr = task.WaitTaskCompletion()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"couldn't finish removing vdc %s\", err)\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func deleteNodeVIPs(svcIPs []string, protocol v1.Protocol, sourcePort int32) error {\n\tklog.V(5).Infof(\"Searching to remove Gateway VIPs - %s, %d\", protocol, sourcePort)\n\tgatewayRouters, _, err := gateway.GetOvnGateways()\n\tif err != nil {\n\t\tklog.Errorf(\"Error while searching for gateways: %v\", err)\n\t\treturn err\n\t}\n\n\tfor _, gatewayRouter := range gatewayRouters {\n\t\tvar loadBalancers []string\n\t\tgatewayLB, err := gateway.GetGatewayLoadBalancer(gatewayRouter, protocol)\n\t\tif err != nil {\n\t\t\tklog.Errorf(\"Gateway router %s does not have load balancer (%v)\", gatewayRouter, err)\n\t\t\tcontinue\n\t\t}\n\t\tips := svcIPs\n\t\tif len(ips) == 0 {\n\t\t\tips, err = gateway.GetGatewayPhysicalIPs(gatewayRouter)\n\t\t\tif err != nil {\n\t\t\t\tklog.Errorf(\"Gateway router %s does not have physical ip (%v)\", gatewayRouter, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tloadBalancers = append(loadBalancers, gatewayLB)\n\t\tif config.Gateway.Mode == config.GatewayModeShared {\n\t\t\tworkerNode := util.GetWorkerFromGatewayRouter(gatewayRouter)\n\t\t\tworkerLB, err := loadbalancer.GetWorkerLoadBalancer(workerNode, protocol)\n\t\t\tif err != nil {\n\t\t\t\tklog.Errorf(\"Worker switch %s does not have load balancer (%v)\", workerNode, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tloadBalancers = append(loadBalancers, workerLB)\n\t\t}\n\t\tfor _, loadBalancer := range loadBalancers {\n\t\t\tfor _, ip := range ips {\n\t\t\t\t// With the physical_ip:sourcePort as the VIP, delete an entry in 'load_balancer'.\n\t\t\t\tvip := util.JoinHostPortInt32(ip, sourcePort)\n\t\t\t\tklog.V(5).Infof(\"Removing gateway VIP: %s from load balancer: %s\", vip, loadBalancer)\n\t\t\t\tif err := loadbalancer.DeleteLoadBalancerVIP(loadBalancer, vip); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func DHCPsConfigure(dhcp []Dhcp) error {\n\tfor _, d := range dhcp {\n\t\tif isUnmanaged(UnmanagedID(d.Ifname), LINKTYPE) {\n\t\t\tlogger.Log.Info(fmt.Sprintf(\"Skipping Unmanaged Link %v DHCP configuration\", d.Ifname))\n\t\t\tcontinue\n\t\t}\n\t\terr := DHCPDelete(d.Ifname)\n\t\tif err != nil {\n\t\t\tif _, ok := err.(*NotFoundError); ok != true {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\tif err := DHCPCreate(d); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (h *ValidationHandler) removeServers(swagger *openapi3.Swagger) (*openapi3.Swagger, error) {\n\t// collect API pathPrefix path prefixes\n\tprefixes := make(map[string]struct{}, 0) // a \"set\"\n\tfor _, s := range swagger.Servers {\n\t\tu, err := url.Parse(s.URL)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tprefixes[u.Path] = struct{}{}\n\t}\n\tif len(prefixes) != 1 {\n\t\treturn nil, fmt.Errorf(\"requires a single API pathPrefix path prefix: %v\", prefixes)\n\t}\n\tvar prefix string\n\tfor k := range prefixes {\n\t\tprefix = k\n\t}\n\n\t// update the paths to start with the API pathPrefix path prefixes\n\tpaths := make(openapi3.Paths, 0)\n\tfor key, path := range swagger.Paths {\n\t\tpaths[prefix+key] = path\n\t}\n\tswagger.Paths = paths\n\n\t// now remove the servers\n\tswagger.Servers = nil\n\n\treturn swagger, nil\n}", "func demolish(api libmachine.API, cc config.ClusterConfig, n config.Node, h *host.Host) {\n\tmachineName := config.MachineName(cc, n)\n\tklog.Infof(\"DEMOLISHING %s ...\", machineName)\n\n\t// This will probably fail\n\tif err := stop(h); err != nil {\n\t\tklog.Infof(\"stophost failed (probably ok): %v\", err)\n\t}\n\n\t// For 95% of cases, this should be enough\n\tif err := DeleteHost(api, machineName); err != nil {\n\t\tklog.Warningf(\"deletehost failed: %v\", err)\n\t}\n\n\tif err := deleteHost(api, h, machineName); err != nil {\n\t\tklog.Warningf(\"delete failed (probably ok) %v\", err)\n\t}\n}", "func (self *basicPodManager) DeleteOrphanedMirrorPods() {\n\tpodByFullName, mirrorPodByFullName := self.getFullNameMaps()\n\n\tfor podFullName := range mirrorPodByFullName {\n\t\tif _, ok := podByFullName[podFullName]; !ok {\n\t\t\tself.mirrorClient.DeleteMirrorPod(podFullName)\n\t\t}\n\t}\n}", "func (ipb *IPBlocker) CleanUp() {\n\tlog.Printf(\"Cleaning up iptable entries made by connection watcher\")\n\tfor _, ip := range ipb.BlockedHosts {\n\t\terr := ipb.IP4Table.Delete(Filter, Chain, \"-s\", ip.String(), \"-j\", \"DROP\")\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Failed to remove iptable block for %s: %v\", ip.String(), err)\n\t\t}\n\t}\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
addHostOnlyDHCPServer adds a DHCP server to a hostonly network.
func addHostOnlyDHCPServer(ifname string, d dhcpServer, vbox VBoxManager) error { name := dhcpPrefix + ifname dhcps, err := listDHCPServers(vbox) if err != nil { return err } // On some platforms (OSX), creating a host-only adapter adds a default dhcpserver, // while on others (Windows?) it does not. command := "add" if dhcp, ok := dhcps[name]; ok { command = "modify" if (dhcp.IPv4.IP.Equal(d.IPv4.IP)) && (dhcp.IPv4.Mask.String() == d.IPv4.Mask.String()) && (dhcp.LowerIP.Equal(d.LowerIP)) && (dhcp.UpperIP.Equal(d.UpperIP)) && dhcp.Enabled { // dhcp is up to date return nil } } args := []string{"dhcpserver", command, "--netname", name, "--ip", d.IPv4.IP.String(), "--netmask", net.IP(d.IPv4.Mask).String(), "--lowerip", d.LowerIP.String(), "--upperip", d.UpperIP.String(), } if d.Enabled { args = append(args, "--enable") } else { args = append(args, "--disable") } return vbox.vbm(args...) }
[ "func WithDHCPServer(subnet *net.IPNet) Option {\n\treturn func(d *dnsmasq) {\n\t\td.subnet = subnet\n\t}\n}", "func addHost(n *libvirt.VirNetwork, ip, mac, name string) error {\n\txmlDesc := getHostXMLDesc(ip, mac, name)\n\tlog.Printf(\"Adding host with XML:\\n%s\", xmlDesc)\n\treturn n.UpdateXMLDesc(xmlDesc, libvirt.VIR_NETWORK_UPDATE_COMMAND_ADD_LAST, libvirt.VIR_NETWORK_SECTION_IP_DHCP_HOST)\n}", "func (p *PerHost) AddHost(host string) {\n\tif strings.HasSuffix(host, \".\") {\n\t\thost = host[:len(host)-1]\n\t}\n\tp.bypassHosts = append(p.bypassHosts, host)\n}", "func NewDHCPServer(iface, serverAddr, subnet, gateway, dns string,\n\tduration time.Duration) *DHCPServer {\n\tserver := &DHCPServer{}\n\tlog.Println(\"Configuring DHCP Server...\")\n\tlog.Printf(\" using interface %v\\n\", iface)\n\tserver.iface = iface\n\tlog.Printf(\" using server address %v\\n\", serverAddr)\n\tserver.serverAddr = net.ParseIP(serverAddr).To4()\n\tlog.Printf(\" serving subnet %v\\n\", subnet)\n\t_, ipNet, err := net.ParseCIDR(subnet)\n\tif err != nil {\n\t\tlog.Fatalf(\"subnet %v is not valid...exiting\\n\", subnet)\n\t\tpanic(err)\n\t}\n\tserver.netmask = ipNet\n\tserver.stopAddr = getStopAddr(ipNet)\n\tserver.startAddr = getStartAddr(ipNet)\n\tserver.leaseRange = dhcp4.IPRange(server.startAddr, server.stopAddr)\n\tlog.Printf(\" start/stop addresses are %v/%v total range is %v\\n\",\n\t\tserver.startAddr, server.stopAddr, server.leaseRange)\n\tlog.Printf(\" using gateway address %v\\n\", gateway)\n\tserver.gatewayAddr = net.ParseIP(gateway)\n\tlog.Printf(\" using dns address %v\\n\", dns)\n\tserver.dnsAddr = net.ParseIP(dns)\n\tserver.leases = make(map[string]*dhcplease)\n\tserver.ReserveStaticLeases(server.serverAddr, server.gatewayAddr, server.dnsAddr)\n\tserver.PrintReservedLeases()\n\tserver.leaseDuration = duration\n\tserver.options = GetDefaultDHCPOptions(net.IP(server.netmask.Mask),\n\t\tserver.gatewayAddr.To4(), server.dnsAddr.To4())\n\n\treturn server\n}", "func (lv *Libvirt) AddTransientDHCPHost(newHost *libvirtxml.NetworkDHCPHost, app *App) error {\n\tlv.dhcpLeases.mutex.Lock()\n\tdefer lv.dhcpLeases.mutex.Unlock()\n\n\tlv.dhcpLeases.leases[newHost] = true\n\treturn lv.rebuildDHCPStaticLeases(app)\n}", "func getHostOnlyNetworkInterface(mc *driver.MachineConfig) (string, error) {\n\t// Check if the interface/dhcp exists.\n\tnets, err := HostonlyNets()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tdhcps, err := DHCPs()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor _, n := range nets {\n\t\tif dhcp, ok := dhcps[n.NetworkName]; ok {\n\t\t\tif dhcp.IPv4.IP.Equal(mc.DHCPIP) &&\n\t\t\t\tdhcp.IPv4.Mask.String() == mc.NetMask.String() &&\n\t\t\t\tdhcp.LowerIP.Equal(mc.LowerIP) &&\n\t\t\t\tdhcp.UpperIP.Equal(mc.UpperIP) &&\n\t\t\t\tdhcp.Enabled == mc.DHCPEnabled {\n\t\t\t\treturn n.Name, nil\n\t\t\t}\n\t\t}\n\t}\n\n\t// No existing host-only interface found. Create a new one.\n\thostonlyNet, err := CreateHostonlyNet()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\thostonlyNet.IPv4.IP = mc.HostIP\n\thostonlyNet.IPv4.Mask = mc.NetMask\n\tif err := hostonlyNet.Config(); err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create and add a DHCP server to the host-only network\n\tdhcp := driver.DHCP{}\n\tdhcp.IPv4.IP = mc.DHCPIP\n\tdhcp.IPv4.Mask = mc.NetMask\n\tdhcp.LowerIP = mc.LowerIP\n\tdhcp.UpperIP = mc.UpperIP\n\tdhcp.Enabled = true\n\tif err := AddHostonlyDHCP(hostonlyNet.Name, dhcp); err != nil {\n\t\treturn \"\", err\n\t}\n\treturn hostonlyNet.Name, nil\n}", "func AddDNS(host string, ipaddress string) {\n\texecuteCommand(\"sudo\", []string{\n\t\t\"ghost\",\n\t\t\"add\",\n\t\thost,\n\t\tipaddress,\n\t})\n}", "func (v *Virter) addDHCPEntry(mac string, id uint) (net.IP, error) {\n\tnetwork, err := v.libvirt.NetworkLookupByName(v.networkName)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not get network: %w\", err)\n\t}\n\n\tipNet, err := v.getIPNet(network)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnetworkBaseIP := ipNet.IP.Mask(ipNet.Mask)\n\tip := addToIP(networkBaseIP, id)\n\n\tif !ipNet.Contains(ip) {\n\t\treturn nil, fmt.Errorf(\"computed IP %v is not in network\", ip)\n\t}\n\n\tlog.Printf(\"Add DHCP entry from %v to %v\", mac, ip)\n\terr = v.libvirt.NetworkUpdate(\n\t\tnetwork,\n\t\t// the following 2 arguments are swapped; see\n\t\t// https://github.com/digitalocean/go-libvirt/issues/87\n\t\tuint32(libvirt.NetworkSectionIPDhcpHost),\n\t\tuint32(libvirt.NetworkUpdateCommandAddLast),\n\t\t-1,\n\t\tfmt.Sprintf(\"<host mac='%s' ip='%v'/>\", mac, ip),\n\t\tlibvirt.NetworkUpdateAffectLive|libvirt.NetworkUpdateAffectConfig)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not add DHCP entry: %w\", err)\n\t}\n\n\treturn ip, nil\n}", "func WithDHCPNameServers(dns []string) Option {\n\treturn func(d *dnsmasq) {\n\t\td.dns = dns\n\t}\n}", "func (servers *Servers) AddServer(macAddressStr string) error {\n\tvar macAddress net.HardwareAddr\n\tmacAddress, err := net.ParseMAC(macAddressStr)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = servers.GetServer(macAddress.String())\n\tserver := Server{\n\t\tMacAddress: macAddress, IPAddress: \"change me\", Installed: false, Kernel: \"linux\", SecondMacAddress: \"find me\"}\n\tif err == nil {\n\t\tlog.Warnln(\"The server already exist in the list. Overwrite it.\")\n\t} else {\n\t\tswitch err.(type) {\n\t\tcase *EmptyServerListError:\n\t\t\tlog.Infoln(err)\n\t\t\tbreak\n\t\tcase *UnreconizeServerError:\n\t\t\tbreak\n\t\tcase *NilServerListError:\n\n\t\t\t// A map should not be nil\n\t\t\t// Refence : https://blog.golang.org/go-maps-in-action\n\t\t\treturn err\n\n\t\tdefault:\n\t\t\treturn err\n\t\t}\n\t}\n\n\t(*servers)[macAddress.String()] = &server\n\n\treturn nil\n}", "func (o *DhcpSharednetworkAddInput) HasServerHostaddr() bool {\n\tif o != nil && o.ServerHostaddr != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (h *HostHandler) AddHost(host string) *http.ServeMux {\n\tmux := http.NewServeMux()\n\th.eligibleHosts[host] = mux\n\treturn mux\n}", "func (s *Server) AddHost(input wrpc.AddHostInput, output *wrpc.AddHostOutput) error {\n\treturn s.rpcMiddleware(&input.CommonInput, func() error {\n\t\tep := wrpc.NewEndpoint(string(input.Endpoint), werify.DefaultPort)\n\n\t\ti, _ := s.getHostByEndpoint(ep, true)\n\t\tif i > -1 {\n\t\t\treturn errors.New(\"endpoint already exists in host list\")\n\t\t}\n\n\t\ts.hostMu.Lock()\n\t\tdefer s.hostMu.Unlock()\n\n\t\th := &t.Host{\n\t\t\tEndpoint: ep,\n\t\t\tAdded: time.Now(),\n\t\t\tIsAlive: false,\n\t\t}\n\n\t\terr := s.setIdentifier(h)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\ts.hosts = append(s.hosts, h)\n\n\t\terr = s.healthcheck(h)\n\t\tif err != nil {\n\t\t\t//log.Printf(\"Initial healthcheck failed for %v: %s\", h, err.Error())\n\t\t} else if !h.IsAlive {\n\t\t\tlog.Printf(\"Initial healthcheck not OK for %v\", h)\n\t\t}\n\n\t\toutput.Ok = true\n\t\treturn nil\n\t})\n}", "func AddHCNHostEndpoint(ctx context.Context, i *Endpoint, netns, networkID string, decorates ...HCNEndpointDecorator) error {\n\tif netns == \"\" || networkID == \"\" ||\n\t\t!i.isValid() {\n\t\treturn errors.Errorf(\"invalid HostComputeEndpoint configuration\")\n\t}\n\n\tvar attach = func(ep *hcn.HostComputeEndpoint, isNewlyCreatedEndpoint bool) error {\n\t\t// attach gateway endpoint to host\n\t\tvar condErr error\n\t\tvar err = wait.PollImmediateUntil(100*time.Millisecond, func() (bool, error) {\n\t\t\tcondErr = ep.NamespaceAttach(netns)\n\t\t\tif condErr == nil ||\n\t\t\t\thcn.CheckErrorWithCode(condErr, 0x803B0014) { // if already attached\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t\treturn false, nil\n\t\t}, ctx.Done())\n\t\tif err != nil {\n\t\t\tif condErr == nil {\n\t\t\t\tcondErr = err\n\t\t\t}\n\t\t\treturn errors.Wrapf(condErr, \"failed to attach gateway HostComputeEndpoint %s\", ep.Name)\n\t\t}\n\t\treturn nil\n\t}\n\n\treturn addHCNEndpoint(ctx, i, netns, networkID, attach, decorates...)\n}", "func (s *Server) ServeDHCP(p dhcp4.Packet, msgType dhcp4.MessageType, options dhcp4.Options) dhcp4.Packet {\n\ts.printLeases()\n\n\tswitch msgType {\n\tcase dhcp4.Discover: // Broadcast Packet From Client - Can I have an IP?\n\t\treturn s.handleDiscover(p, options)\n\n\tcase dhcp4.Request: // Broadcast From Client - I'll take that IP (Also start for renewals)\n\t\t// start/renew a lease -- update lease time\n\t\t// some clients (OSX) just go right ahead and do Request first from previously known IP, if they get NAK, they restart full cycle with Discover then Request\n\t\treturn s.handleDHCP4Request(p, options)\n\n\tcase dhcp4.Decline: // Broadcast From Client - Sorry I can't use that IP\n\t\treturn s.handleDecline(p, options)\n\n\tcase dhcp4.Release: // From Client, I don't need that IP anymore\n\t\treturn s.handleRelease(p, options)\n\n\tcase dhcp4.Inform: // From Client, I have this IP and there's nothing you can do about it\n\t\treturn s.handleInform(p, options)\n\n\t// from server -- ignore those but enumerate just in case\n\tcase dhcp4.Offer: // Broadcast From Server - Here's an IP\n\t\tlog.Printf(\"DHCP: received message from %s: Offer\", p.CHAddr())\n\n\tcase dhcp4.ACK: // From Server, Yes you can have that IP\n\t\tlog.Printf(\"DHCP: received message from %s: ACK\", p.CHAddr())\n\n\tcase dhcp4.NAK: // From Server, No you cannot have that IP\n\t\tlog.Printf(\"DHCP: received message from %s: NAK\", p.CHAddr())\n\n\tdefault:\n\t\tlog.Printf(\"DHCP: unknown packet %v from %s\", msgType, p.CHAddr())\n\t\treturn nil\n\t}\n\treturn nil\n}", "func createHostonlyAdapter(vbox VBoxManager) (*hostOnlyNetwork, error) {\n\tout, err := vbox.vbmOut(\"hostonlyif\", \"create\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres := reHostOnlyAdapterCreated.FindStringSubmatch(string(out))\n\tif res == nil {\n\t\treturn nil, errors.New(\"Failed to create host-only adapter\")\n\t}\n\n\treturn &hostOnlyNetwork{Name: res[1]}, nil\n}", "func (client *XenClient) HostAddToLicenseServer(self string, key string, value string) (err error) {\n\t_, err = client.APICall(\"host.add_to_license_server\", self, key, value)\n\tif err != nil {\n\t\treturn\n\t}\n\t// no return result\n\treturn\n}", "func (p *proteusAPI) AddDHCPServiceDeploymentOption(entityId int64, name string, value string, properties string) (int64, error) {\n\tα := struct {\n\t\tM OperationAddDHCPServiceDeploymentOption `xml:\"tns:addDHCPServiceDeploymentOption\"`\n\t}{\n\t\tOperationAddDHCPServiceDeploymentOption{\n\t\t\t&entityId,\n\t\t\t&name,\n\t\t\t&value,\n\t\t\t&properties,\n\t\t},\n\t}\n\n\tγ := struct {\n\t\tM OperationAddDHCPServiceDeploymentOptionResponse `xml:\"addDHCPServiceDeploymentOptionResponse\"`\n\t}{}\n\tif err := p.cli.RoundTripWithAction(\"AddDHCPServiceDeploymentOption\", α, &γ); err != nil {\n\t\treturn 0, err\n\t}\n\treturn *γ.M.Return, nil\n}", "func NewServer(ifname string, addr *net.UDPAddr, handler Handler, opt ...ServerOpt) (*Server, error) {\n\ts := &Server{\n\t\thandler: handler,\n\t\tlogger: EmptyLogger{},\n\t}\n\n\tfor _, o := range opt {\n\t\to(s)\n\t}\n\tif s.conn != nil {\n\t\treturn s, nil\n\t}\n\n\tif addr == nil {\n\t\taddr = &net.UDPAddr{\n\t\t\tIP: net.IPv6unspecified,\n\t\t\tPort: dhcpv6.DefaultServerPort,\n\t\t}\n\t}\n\n\tvar (\n\t\terr error\n\t\tiface *net.Interface\n\t)\n\tif ifname == \"\" {\n\t\tiface = nil\n\t} else {\n\t\tiface, err = net.InterfaceByName(ifname)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\t// no connection provided by the user, create a new one\n\ts.conn, err = NewIPv6UDPConn(ifname, addr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tp := ipv6.NewPacketConn(s.conn)\n\tif addr.IP.IsMulticast() {\n\t\tif err := p.JoinGroup(iface, addr); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else if (addr.IP == nil || addr.IP.IsUnspecified()) && addr.Port == dhcpv6.DefaultServerPort {\n\t\t// For wildcard addresses on the correct port, listen on both multicast\n\t\t// addresses defined in the RFC as a \"default\" behaviour\n\t\tfor _, g := range []net.IP{dhcpv6.AllDHCPRelayAgentsAndServers, dhcpv6.AllDHCPServers} {\n\t\t\tgroup := net.UDPAddr{\n\t\t\t\tIP: g,\n\t\t\t\tPort: dhcpv6.DefaultServerPort,\n\t\t\t}\n\t\t\tif err := p.JoinGroup(iface, &group); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t}\n\t}\n\n\treturn s, nil\n}" ]
{ "objective": { "paired": [], "self": [], "triplet": [ [ "query", "document", "negatives" ] ] } }