query
stringlengths
7
3.85k
document
stringlengths
11
430k
metadata
dict
negatives
sequencelengths
0
101
negative_scores
sequencelengths
0
101
document_score
stringlengths
3
10
document_rank
stringclasses
102 values
Title uses strings.Title to return operand with the first unicode codepoint of each word converted to uppercase.
func Title(operand string) string { return strings.Title(operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ToTitle(r rune) rune", "func Title(in string) string {\n\n\trunes := []rune(in)\n\tlength := len(runes)\n\n\tvar out []rune\n\tfor i := 0; i < length; i++ {\n\t\tif i > 0 && unicode.IsUpper(runes[i]) && ((i+1 < length && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tout = append(out, ' ')\n\t\t}\n\t\tout = append(out, runes[i])\n\t}\n\n\treturn string(out)\n}", "func Title(name string) string {\n\t//name = strings.Replace(name, \"_\", \" \", -1)\n\t//name = strings.Replace(name, \".\", \" \", -1)\n\tname = strings.TrimSpace(name)\n\treturn strings.ToUpper(name)\n}", "func ToTitle(rune int) int {\n\tif rune < 0x80 {\t// quick ASCII check\n\t\tif 'a' <= rune && rune <= 'z' {\t// title case is upper case for ASCII\n\t\t\trune -= 'a' - 'A'\n\t\t}\n\t\treturn rune;\n\t}\n\treturn To(TitleCase, rune);\n}", "func Ucwords(str string) string {\n\treturn strings.Title(str)\n}", "func Title(name string) string {\n\tname = strings.Replace(name, \"_\", \" \", -1)\n\tname = strings.Replace(name, \".\", \" \", -1)\n\tname = strings.TrimSpace(name)\n\treturn strings.ToUpper(name)\n}", "func capitalise(text string) string {\n\tresult := strings.ToLower(text)\n\tresult = strings.Title(result)\n\treturn result\n}", "func SpecialCaseToTitle(special unicode.SpecialCase, r rune) rune", "func (t title) fixCase() string {\n\treturn strings.Title(string(t)) // convert title to a string since type Title is based on a string.\n}", "func Titleize(input string) (titleized string) {\n\tisToUpper := false\n\tfor k, v := range input {\n\t\tif k == 0 {\n\t\t\ttitleized = strings.ToUpper(string(input[0]))\n\t\t} else {\n\t\t\tif isToUpper || unicode.IsUpper(v) {\n\t\t\t\ttitleized += \" \" + strings.ToUpper(string(v))\n\t\t\t\tisToUpper = false\n\t\t\t} else {\n\t\t\t\tif (v == '_') || (v == ' ') {\n\t\t\t\t\tisToUpper = true\n\t\t\t\t} else {\n\t\t\t\t\ttitleized += string(v)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn\n\n}", "func ToTitle(str string) string {\n\treturn strings.Title(str)\n}", "func Titleize(text string) string {\n\tpascalized := Pascalize(text)\n\n\treturn titleRegex.ReplaceAllStringFunc(pascalized, func(s string) string {\n\t\treturn string(s[0]) + \" \" + string(s[1])\n\t})\n}", "func (r *Role) Title() string {\n\t// Uppercase all words, and also ensure \"MARC\" is fully capitalized\n\tvar c = cases.Title(language.AmericanEnglish)\n\treturn c.String(strings.Replace(r.Name, \"marc\", \"MARC\", -1))\n}", "func ToUpper(operand string) string { return strings.ToUpper(operand) }", "func ToTitle(s string) string {\n\treturn strings.ToTitle(s)\n}", "func ToTitleSpecial(c unicode.SpecialCase) MapFunc {\n\treturn func(s string) string { return strings.ToTitleSpecial(c, s) }\n}", "func titleInitial(str string) string {\n\tfor i, v := range str {\n\t\treturn string(unicode.ToTitle(v)) + str[i+1:]\n\t}\n\treturn \"\"\n}", "func TitleCase(n ComponentName) ComponentName {\n\ts := string(n)\n\treturn ComponentName(strings.ToUpper(s[0:1]) + s[1:])\n}", "func Title(s string) string {\n\treturn strings.Title(s)\n}", "func capitalize(s string) string {\n\tif s == \"\" {\n\t\treturn s\n\t}\n\tr, n := utf8.DecodeRuneInString(s)\n\treturn string(unicode.ToTitle(r)) + s[n:]\n}", "func (o AccessLevelCustomExprOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AccessLevelCustomExpr) *string { return v.Title }).(pulumi.StringPtrOutput)\n}", "func EnumValueTitle(v string) string {\n\tv = strings.Replace(v, \"_\", \" \", -1)\n\treturn strings.Title(v)\n}", "func capitalize(label string) string {\n\tfirstLetter := strings.SplitN(label, \"\", 2)\n\tif len(firstLetter) < 1 {\n\t\treturn label\n\t}\n\treturn fmt.Sprintf(\"%v%v\", strings.ToUpper(firstLetter[0]),\n\t\tstrings.TrimPrefix(label, firstLetter[0]))\n}", "func filterTitle(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.Title(stick.CoerceString(val))\n}", "func (o ExprOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Expr) *string { return v.Title }).(pulumi.StringPtrOutput)\n}", "func (o ExprOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Expr) *string { return v.Title }).(pulumi.StringPtrOutput)\n}", "func (o ExprOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Expr) *string { return v.Title }).(pulumi.StringPtrOutput)\n}", "func (o ExprOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Expr) *string { return v.Title }).(pulumi.StringPtrOutput)\n}", "func (o ExprOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Expr) *string { return v.Title }).(pulumi.StringPtrOutput)\n}", "func (o ExprOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Expr) *string { return v.Title }).(pulumi.StringPtrOutput)\n}", "func (o LookupServiceAccountResultOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupServiceAccountResult) string { return v.Title }).(pulumi.StringOutput)\n}", "func capitalize(s string) string {\n\treturn strings.ToUpper(s[:1]) + s[1:]\n}", "func (o LookupServicePerimeterResultOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupServicePerimeterResult) string { return v.Title }).(pulumi.StringOutput)\n}", "func (s *Str) ToTitle() *Str {\n\ts.val = strings.ToTitle(s.val)\n\treturn s\n}", "func titleCase(s string) string {\n\tprev := '_'\n\ts = fixForInitialismCase(s)\n\ttitleCased := strings.Map(\n\t\tfunc(r rune) rune {\n\t\t\tif r == '_' {\n\t\t\t\tprev = r\n\t\t\t\treturn -1\n\t\t\t}\n\t\t\tif prev == '_' {\n\t\t\t\tprev = r\n\t\t\t\treturn unicode.ToUpper(r)\n\t\t\t}\n\t\t\tprev = r\n\t\t\treturn r\n\t\t}, s)\n\t// special cases when a struct name ends in 'Args'/'Result', leads to go name having '_' appended\n\t// https://github.com/apache/thrift/blob/master/compiler/cpp/src/thrift/generate/t_go_generator.cc#L495\n\tif (len(titleCased) >= 4 && titleCased[len(titleCased)-4:] == \"Args\") ||\n\t\t(len(titleCased) >= 6 && s[len(titleCased)-6:] == \"Result\") {\n\t\ttitleCased = titleCased + \"_\"\n\t}\n\treturn titleCased\n}", "func IsTitle(r rune) bool", "func (o AccessLevelsAccessLevelCustomExprOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AccessLevelsAccessLevelCustomExpr) *string { return v.Title }).(pulumi.StringPtrOutput)\n}", "func Abbreviate(s string) string {\n\tvar acro string\n\ts = strings.Replace(s , \"-\" , \" \",-1)\n\tarray := strings.Split(s,\" \")\n\tfor i:= 0 ; i < len(array) ; i++ {\n\t\tacro += strings.ToUpper(string(array[i][0]))\n\t\t}\n\treturn acro\n}", "func Abbreviate(in string) string {\n\tpos := 0\n\tout := \"\"\n\tfor i, v := range in {\n\t\tif v == ' ' || v == '-' || i == len(in)-1 {\n\t\t\tout += strings.ToUpper(in[pos : pos+1])\n\t\t\tpos = i + 1\n\t\t}\n\t}\n\treturn out\n}", "func (o ExprResponseOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ExprResponse) string { return v.Title }).(pulumi.StringOutput)\n}", "func (o ExprResponseOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ExprResponse) string { return v.Title }).(pulumi.StringOutput)\n}", "func (o ExprResponseOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ExprResponse) string { return v.Title }).(pulumi.StringOutput)\n}", "func (o ExprResponseOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ExprResponse) string { return v.Title }).(pulumi.StringOutput)\n}", "func (o ExprResponseOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ExprResponse) string { return v.Title }).(pulumi.StringOutput)\n}", "func (o ExprResponseOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ExprResponse) string { return v.Title }).(pulumi.StringOutput)\n}", "func filterCapitalize(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\ts := stick.CoerceString(val)\n\treturn strings.ToUpper(s[:1]) + s[1:]\n}", "func (tokens Tokens) Capital() string {\n\treturn strings.Join(tokens.Map(strings.Title), \"\")\n}", "func (o ServicePerimetersServicePerimeterOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ServicePerimetersServicePerimeter) string { return v.Title }).(pulumi.StringOutput)\n}", "func (o AccessLevelsAccessLevelOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AccessLevelsAccessLevel) string { return v.Title }).(pulumi.StringOutput)\n}", "func templateFunctionTitle(value string) string {\n\treturn strings.Title(value)\n}", "func upper(in string) string {\n\tout := []rune(in)\n\tout[0] = unicode.ToUpper(out[0])\n\treturn string(out)\n}", "func UpperCamelCase(s string) string {\n\tparts := []string{}\n\tfor _, part := range SplitSymbol(s) {\n\t\tif part == \"\" {\n\t\t\tparts = append(parts, \"_\")\n\t\t\tcontinue\n\t\t}\n\t\tif part == \"s\" && len(parts) > 0 {\n\t\t\tparts[len(parts)-1] += part\n\t\t} else {\n\t\t\tif commonInitialisms[strings.ToUpper(part)] {\n\t\t\t\tpart = strings.ToUpper(part)\n\t\t\t} else {\n\t\t\t\tpart = title(part)\n\t\t\t}\n\t\t\tparts = append(parts, part)\n\t\t}\n\t}\n\treturn strings.Join(parts, \"\")\n}", "func printTitle(datas []*model.InputMetric) {\n\tvar title string\n\t// print title.\n\tfor _, data := range datas {\n\t\ttitle += fmt.Sprintf(\"%-*s\", 33, data.Type)\n\t}\n\tfmt.Printf(\"%c[7;40;36m%s%c[0m\\n\", 0x1B, strings.ToUpper(title), 0x1B)\n}", "func Abbreviate(s string) (retVal string) {\n\ts = strings.ReplaceAll(s, \"-\", \" \")\n\ts = strings.ReplaceAll(s, \"_\", \" \")\n\n\tfor _, v := range strings.Fields(s) {\n\t\tretVal += string(v[0])\n\t}\n\treturn strings.ToUpper(retVal)\n}", "func Title() (string, error) {\n\treturn \"\", nil\n}", "func (t *StringDataType) Title() *StringDataType {\n\treturn t.Formatter(func(s string) string {\n\t\treturn strings.Title(s)\n\t})\n}", "func HeadToUpper(str string) string {\n\tln := len(str)\n\tif ln <= 0 {\n\t\treturn str\n\t}\n\treturn strings.ToUpper(str[0:1]) + str[1:]\n}", "func Abbreviate(s string) string {\n\ts = strings.Replace(s, \"-\", \" \", -1)\n\tparts := strings.Split(s, \" \")\n\tvar abbrv string\n\n\tfor _, part := range parts {\n\t\tabbrv += strings.ToUpper(string(part[0]))\n\t}\n\n\treturn abbrv\n}", "func Abbreviate(str string) string {\n\tstr = strings.Replace(str, \"-\", \" \", -1)\n\twords := strings.Fields(str)\n\tresult := \"\"\n\tfor _, word := range words {\n\t\tresult += string([]byte{word[0]})\n\t}\n\n\treturn strings.ToUpper(result)\n}", "func (l Logger) Title(text ...string) {\n\ttext = append([]string{\"[\"}, text...)\n\ttext = append(text, \"]\")\n\tfmt.Fprintln(l.out, title(strings.Join(text, \" \")))\n}", "func capitalizedWord(word string) string {\n\tvar processedWord = removeAllSymbols(word)\n\treturn \"k\" + strings.ToUpper(processedWord)\n}", "func Abbreviate(s string) string {\n\ts = strings.TrimSpace(s)\n\tvar outPut string\n\tfor i, w := range s {\n\n\t\tif i == 0 {\n\t\t\tfmt.Println(string(w))\n\t\t\toutPut += string(w)\n\t\t} else if !unicode.IsLetter(rune(s[i-1])) && rune(s[i-1]) != 39 {\n\t\t\tif unicode.IsLetter(w) {\n\t\t\t\toutPut += strings.ToUpper(string(w))\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\treturn outPut\n}", "func Title(v string) predicate.Ethnicity {\n\treturn predicate.Ethnicity(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldTitle), v))\n\t})\n}", "func VZEROUPPER() { ctx.VZEROUPPER() }", "func (o IamMemberConditionOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IamMemberCondition) string { return v.Title }).(pulumi.StringOutput)\n}", "func (o AccessLevelCustomExprPtrOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AccessLevelCustomExpr) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Title\n\t}).(pulumi.StringPtrOutput)\n}", "func Abbreviate(s string) string {\n\tvar a string\n\tfor _, w := range regexp.MustCompile(\"\\\\s|\\\\-\").Split(s, 5) {\n\t\ta = a + strings.ToUpper(string([]rune(w)[0]))\n\t}\n\treturn a\n}", "func Abbreviate(word string) (abbr string) {\n\tregex := regexp.MustCompile(\"(\\\\b[A-Za-z])|([A-Z]*)([a-z]+)([A-Z])\")\n\tlowercase := regexp.MustCompile(\"^[^A-Z]{2,}\")\n\tarr := regex.FindAllString(word, -1)\n\tfor i := range arr {\n\t\tabbr = abbr + lowercase.ReplaceAllString(arr[i], \"\")\n\t}\n\treturn strings.ToUpper(abbr)\n}", "func Abbreviate(s string) string {\n\tprior := ' '\n\tacronym := make([]rune, 0)\n\tfor _, c := range s {\n\t\tif prior == ' ' || prior == '-' {\n\t\t\tacronym = append(acronym, unicode.ToUpper(c))\n\t\t}\n\t\tprior = c\n\t}\n\treturn string(acronym)\n}", "func Abbreviate(s string) string {\n\tacronym := \"\"\n\n\tsplitedSpaces := regexp.MustCompile(`[^-A-Za-z\\s]`).ReplaceAllString(ReplaceNonAlpha(s), \"\")\n\n\tfor _, word := range strings.Split(splitedSpaces, \" \") {\n\t\tif len(strings.TrimSpace(word)) > 0 {\n\t\t\tacronym += string(word[0])\n\t\t}\n\t}\n\n\treturn strings.ToUpper(acronym)\n}", "func Abbreviate(s string) string {\n\twordParse := regexp.MustCompile(`[[:alpha:]']+`)\n\twords := wordParse.FindAllString(s, -1)\n\tabbr := make([]byte, len(words))\n\tfor i, w := range words {\n\t\tabbr[i] += strings.ToUpper(w)[0]\n\t}\n\treturn string(abbr)\n}", "func (o BucketIAMMemberConditionOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketIAMMemberCondition) string { return v.Title }).(pulumi.StringOutput)\n}", "func IsTitle(rune int) bool {\n\tif rune < 0x80 {\t// quick ASCII check\n\t\treturn false\n\t}\n\treturn Is(Title, rune);\n}", "func convertToTitle(n int) string {\n\tvar buf bytes.Buffer\n\tfor n > 0 {\n\t\tbuf.WriteByte(byte((n-1)%26) + 'A')\n\t\tn = (n - 1) / 26\n\t}\n\trevTitle := buf.Bytes()\n\ti, j := 0, len(revTitle)-1\n\tfor i < j {\n\t\trevTitle[i], revTitle[j] = revTitle[j], revTitle[i]\n\t\ti++\n\t\tj--\n\t}\n\treturn string(revTitle)\n}", "func Acronym(text string) string {\n\n\ttext = strings.TrimSuffix(text, \"\\n\")\n\n\twordArr := strings.Fields(text)\n\n\tvar acronym string\n\n\tfor _, word := range wordArr {\n\t\tacronym += strings.ToUpper(string(word[0]))\n\t}\n\n\treturn acronym\n}", "func upperFirst(str string) string {\n\tfor i, v := range str {\n\t\treturn string(unicode.ToUpper(v)) + str[i+1:]\n\t}\n\treturn \"\"\n}", "func Abbreviate(s string) string {\n\tphrase:= strings.Replace(s, \"_\", \" \", -1)\n\tphrase = strings.Replace(phrase, \"-\", \" \", -1)\n\twords := strings.Fields(phrase)\n\n\tabbreviation:=\"\"\n\tfor _, element := range words{\n\t\tabbreviation += strings.ToUpper(string(element[0]))\n\t}\n\treturn abbreviation\n}", "func upperFirst(s string) string {\n\treturn strings.ToUpper(string([]rune(s)[0])) + string([]rune(s)[1:])\n}", "func Abbreviate(s string) string {\n\tvar b strings.Builder\n\tfor _, word := range regexp.MustCompile(\"[\\\\s-]+\").Split(s, -1) {\n\t\tletter, err := GetFistLetter(word)\n\t\tif err == nil {\n\t\t\tb.WriteRune(unicode.ToUpper(letter))\n\t\t}\n\t}\n\treturn b.String()\n}", "func Ucfirst(word string) string {\n\tif word == \"\" {\n\t\treturn \"\"\n\t}\n\n\tif len(word) == 1 {\n\t\treturn strings.ToUpper(word)\n\t}\n\n\treturn strings.ToUpper(string(word[0])) + word[1:]\n}", "func (o IamBindingConditionOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IamBindingCondition) string { return v.Title }).(pulumi.StringOutput)\n}", "func toUpperCaseFirst(s string) string {\n\tretStr := []string{}\n\tstrs := strings.Split(s, \" \")\n\tfor _, str := range strs {\n\t\tif len(str) > 1 {\n\t\t\tstr = strings.ToUpper(string(str[0])) + str[1:]\n\t\t}\n\t\tretStr = append(retStr, str)\n\t}\n\n\treturn strings.Join(retStr, \" \")\n}", "func Abbreviate(s string) string {\n\tre := regexp.MustCompile(\"\\\\b[a-zA-Z]\")\n\tres := re.FindAllString(s, -1)\n\tacronym := strings.Join(res, \"\")\n\treturn strings.ToUpper(acronym)\n}", "func main() {\n\ts := \"Sarah\" // This thing (type) is a string, with name of s, and value of Sarah\n\tc := capitalise(s)\n\n\tfmt.Println(\"Hello \" + c)\n}", "func (o EditingProjectOutput) Title() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *EditingProject) pulumi.StringOutput { return v.Title }).(pulumi.StringOutput)\n}", "func BaselineClassTitle(v string) predicate.BaselineClass {\n\treturn predicate.BaselineClass(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldBaselineClassTitle), v))\n\t})\n}", "func OutcomeOverviewTitle(v string) predicate.OutcomeOverview {\n\treturn predicate.OutcomeOverview(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldOutcomeOverviewTitle), v))\n\t})\n}", "func Uncapitalize(str string) string {\n\tif str == \"\" {\n\t\treturn str\n\t}\n\tr := []rune(str)\n\tif unicode.IsUpper(r[0]) {\n\t\treturn string(unicode.ToLower(r[0])) + string(r[1:])\n\t}\n\treturn str\n}", "func Abbreviate(s string) string {\n\t// re := regexp.MustCompile(\"[ \\\\-_]+\")\n\t// words := re.Split(s, -1)\n\ts = strings.ReplaceAll(s, \"-\", \" \")\n\ts = strings.ReplaceAll(s, \"_\", \"\")\n\twords := strings.Fields(s)\n\n\tvar ret = []string{}\n\tfor i := range words {\n\t\tret = append(ret, strings.ToUpper(string(words[i][0])))\n\t}\n\treturn strings.Join(ret, \"\")\n}", "func Abbreviate(s string) string {\n\tvar words []string\n\tvar result strings.Builder\n\n\tf := func(r rune) bool {\n\t\treturn unicode.IsSpace(r) || r == '-'\n\t}\n\n\twords = strings.FieldsFunc(s, f)\n\tfor _, w := range words {\n\t\tfor _, l := range w {\n\t\t\tif !unicode.IsLetter(l) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tresult.WriteString(strings.ToUpper(string(l)))\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn result.String()\n}", "func Title(a interfaces.AssumeCredentialProcess, emoji string, prefix string, message string) {\n\ts := a.GetDestination()\n\tf := a.GetFlags()\n\tif f.Verbose {\n\t\tformatted := format(a, textColorTitle, emoji, prefix, message)\n\t\tfmt.Fprint(s, formatted)\n\t}\n}", "func (o ExprPtrOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Expr) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Title\n\t}).(pulumi.StringPtrOutput)\n}", "func (o ExprPtrOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Expr) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Title\n\t}).(pulumi.StringPtrOutput)\n}", "func (o ExprPtrOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Expr) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Title\n\t}).(pulumi.StringPtrOutput)\n}", "func (o ExprPtrOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Expr) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Title\n\t}).(pulumi.StringPtrOutput)\n}", "func (o ExprPtrOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Expr) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Title\n\t}).(pulumi.StringPtrOutput)\n}", "func (o ExprPtrOutput) Title() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Expr) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Title\n\t}).(pulumi.StringPtrOutput)\n}", "func Abbreviate(s string) string {\n\tupperPhrase := strings.ToUpper(s)\n\n\tacronym := \"\"\n\n\tfor _, word := range phraseBreaker(upperPhrase) {\n\t\tacronym += firstLetter(word)\n\n\t}\n\n\treturn acronym\n}", "func UpperFirst(value string) string {\n\tfor i, v := range value {\n\t\treturn string(unicode.ToUpper(v)) + value[i+1:]\n\t}\n\treturn \"\"\n}", "func Capitalize(str string) string {\n\tvar upperStr string\n\tsrunes := []rune(str)\n\n\tfor idx := range srunes {\n\t\tif idx == 0 {\n\t\t\tsrunes[idx] = unicode.ToUpper(srunes[idx])\n\t\t}\n\t\tupperStr += string(srunes[idx])\n\t}\n\n\treturn upperStr\n}" ]
[ "0.75362426", "0.7420402", "0.7298345", "0.7221202", "0.71784824", "0.7118454", "0.70131034", "0.6994498", "0.6810526", "0.6740117", "0.6604884", "0.65940887", "0.65800273", "0.6561685", "0.65249354", "0.65127033", "0.6468561", "0.64599764", "0.64533174", "0.63843113", "0.63210255", "0.6313571", "0.6307369", "0.6277279", "0.6198531", "0.6198531", "0.6198531", "0.6198531", "0.6198531", "0.6198531", "0.6166403", "0.6088027", "0.60773355", "0.6041094", "0.602871", "0.60072976", "0.5981719", "0.59793943", "0.5968582", "0.59674096", "0.59674096", "0.59674096", "0.59674096", "0.59674096", "0.59674096", "0.5962194", "0.59618735", "0.5924552", "0.5922347", "0.59172744", "0.5904539", "0.5903078", "0.5896916", "0.58899975", "0.5879995", "0.5869884", "0.58622444", "0.58603704", "0.58479124", "0.58449405", "0.5839055", "0.5832898", "0.58326733", "0.5832659", "0.5830749", "0.58128345", "0.5804281", "0.5802702", "0.5788614", "0.57877827", "0.57799494", "0.57765883", "0.57709724", "0.5770014", "0.57673854", "0.57506245", "0.5747068", "0.5741144", "0.5736653", "0.57345563", "0.5728875", "0.5720966", "0.5716282", "0.571344", "0.5702873", "0.57012117", "0.569017", "0.56816727", "0.5677951", "0.56756383", "0.5673928", "0.5667195", "0.5667195", "0.5667195", "0.5667195", "0.5667195", "0.5667195", "0.5642047", "0.5640491", "0.5622883" ]
0.7397537
2
ToLower uses strings.ToLower to return operand with all unicode codepoints converted to lowercase.
func ToLower(operand string) string { return strings.ToLower(operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ToLower(r rune) rune", "func SpecialCaseToLower(special unicode.SpecialCase, r rune) rune", "func toLower(s string) string {\n\treturn strings.ToLower(s)\n}", "func toLower(tk token.Token) token.Token {\n\ts := strings.ToLower(tk.Text())\n\treturn token.UpdateText(tk, s)\n}", "func ToLower(str string) string {\n\treturn strings.ToLower(str)\n}", "func ToLower(s string) string {\n\treturn strings.ToLower(s)\n}", "func filterLower(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.ToLower(stick.CoerceString(val))\n}", "func ToLowerSpecial(c unicode.SpecialCase) MapFunc {\n\treturn func(s string) string { return strings.ToLowerSpecial(c, s) }\n}", "func toLowerCase(str string) string {\n\tvar strBuilder strings.Builder\n\tvar r rune\n\tvar ch string // Holds the character to add to the builder\n\n\tfor _, c := range str {\n\t\tif 'A' <= c && c <= 'Z' {\n\t\t\tr = c - 'A' + 'a'\n\t\t\tch = string(r)\n\t\t} else {\n\t\t\tch = string(c)\n\t\t}\n\t\tstrBuilder.WriteString(ch)\n\t}\n\tlowerStr := strBuilder.String()\n\treturn lowerStr\n}", "func ToLower() (desc string, f predicate.TransformFunc) {\n\tdesc = \"ToLower({})\"\n\tf = func(v interface{}) (r interface{}, ctx []predicate.ContextValue, err error) {\n\t\ts, ok := v.(string)\n\t\tif !ok {\n\t\t\treturn nil, nil, fmt.Errorf(\n\t\t\t\t\"value of type '%T' cannot be transformed to lowercase\", v)\n\t\t}\n\t\ts = strings.ToLower(s)\n\t\treturn s, []predicate.ContextValue{\n\t\t\t{Name: \"lower\", Value: s, Pre: false},\n\t\t}, nil\n\t}\n\treturn\n}", "func ToLower(rune int) int {\n\tif rune < 0x80 {\t// quick ASCII check\n\t\tif 'A' <= rune && rune <= 'Z' {\n\t\t\trune += 'a' - 'A'\n\t\t}\n\t\treturn rune;\n\t}\n\treturn To(LowerCase, rune);\n}", "func ToLowercase(str []byte) []byte {\n\tfor i, s := range str {\n\t\tif s > 64 && s < 91 {\n\t\t\tstr[i] = s + 32\n\t\t}\n\t}\n\treturn str\n}", "func toLower(word string) (string, error) {\n\t// Builder pattern was taken from `strings.ToLower()`\n\tvar b strings.Builder\n\tb.Grow(len(word))\n\n\tfor _, r := range word {\n\t\tif !alphabetic(r) {\n\t\t\treturn \"\", fmt.Errorf(\"word `%s`: %w\", word, ErrUnexpectedCharacters)\n\t\t}\n\t\tb.WriteByte(byte(unicode.ToLower(r)))\n\t}\n\n\treturn b.String(), nil\n}", "func lowerPrefix(s string) (lower string) {\n\tfor pos, char := range s {\n\t\tif unicode.IsUpper(char) {\n\t\t\tlower = lower + string(unicode.ToLower(char))\n\t\t} else {\n\t\t\tif pos > 1 {\n\t\t\t\tlower = lower[:len(lower)-1] + s[pos-1:]\n\t\t\t} else {\n\t\t\t\tlower = lower + s[pos:]\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\t}\n\treturn\n}", "func LowerFirst(v string) string {\n\tfor i, r := range v { // run loop to get first rune.\n\t\treturn string(unicode.ToLower(r)) + v[i+1:]\n\t}\n\n\treturn \"\"\n}", "func (g *Generator) toLowerCase() {\n\tfor _, runeValue := range g.Input {\n\t\tg.lowerCased = g.lowerCased + string(unicode.ToLower(runeValue))\n\t}\n\n\tg.lowerCased = helpers.CleanString(g.lowerCased)\n}", "func toLowerASCII(s string) string {\n\tvar b []byte\n\tfor i := 0; i < len(s); i++ {\n\t\tif c := s[i]; 'A' <= c && c <= 'Z' {\n\t\t\tif b == nil {\n\t\t\t\tb = make([]byte, len(s))\n\t\t\t\tcopy(b, s)\n\t\t\t}\n\t\t\tb[i] = s[i] + ('a' - 'A')\n\t\t}\n\t}\n\n\tif b == nil {\n\t\treturn s\n\t}\n\n\treturn string(b)\n}", "func (s *Stringish) ToLower() *Stringish {\n\ts.str = strings.ToLower(s.str)\n\treturn s\n}", "func (s *Str) ToLower() *Str {\n\ts.val = strings.ToLower(s.val)\n\treturn s\n}", "func TokenizeLowerCase(s string) []string {\n\ttokens := Tokenize(s)\n\tfor i := range tokens {\n\t\ttokens[i] = strings.ToLower(tokens[i])\n\t}\n\treturn tokens\n}", "func filterToLower(b byte) byte {\n\n\tif b >= 'A' && b <= 'Z' {\n\t\treturn 'a' + (b - 'A')\n\t} else if b >= 'a' && b <= 'z' {\n\t\treturn b\n\t} else {\n\t\treturn ' ' // binary etc converted to space\n\t}\n}", "func StrLower(s string) string {\n\treturn strings.ToLower(s)\n}", "func Lowercase(text string) string {\n\treturn strings.ToLower(text)\n}", "func FirstToLower(in string) string {\n\tif in == \"\" {\n\t\treturn in\n\t}\n\tr, size := utf8.DecodeRuneInString(in)\n\tif r == utf8.RuneError {\n\t\treturn in\n\t}\n\treturn string(unicode.ToLower(r)) + in[size:]\n}", "func IsLower(r rune) bool", "func lowerInitial(str string) string {\n\tfor i, v := range str {\n\t\treturn string(unicode.ToLower(v)) + str[i+1:]\n\t}\n\treturn \"\"\n}", "func Strtolower(str string) string {\n\treturn strings.ToLower(str)\n}", "func ToLowerFirst(str string) string {\n\tfor i, v := range str {\n\t\treturn string(unicode.ToLower(v)) + str[i+1:]\n\t}\n\treturn str\n}", "func LowerFirstChar(str string) string {\n\tfor i, v := range str {\n\t\treturn string(unicode.ToLower(v)) + str[i+1:]\n\t}\n\treturn str\n}", "func LowerFirst(value string) string {\n\tfor i, v := range value {\n\t\treturn string(unicode.ToLower(v)) + value[i+1:]\n\t}\n\treturn \"\"\n}", "func FirstLower(s string) string {\n\treturn strings.ToLower(string(s[0])) + s[1:]\n}", "func ToLower(b []byte) []byte {\n\tconst decr = 'a' - 'A'\n\tfor i, c := range b {\n\t\tif c >= 'A' && c <= 'Z' {\n\t\t\tb[i] += decr\n\t\t}\n\t}\n\treturn b\n}", "func (s String) GoLowerCase() string {\n\treturn gocase.To(strcase.ToLowerCamel(string(s)))\n}", "func ToLowerFirstRune(s string) string {\n\tif utf8.RuneCountInString(s) == 0 {\n\t\treturn s\n\t}\n\trunes := []rune(s)\n\trunes[0] = unicode.ToLower(runes[0])\n\treturn string(runes)\n}", "func StringLower(scope *Scope, input tf.Output, optional ...StringLowerAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"StringLower\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func toLowerCase(value string) (interface{}, bool) {\n\treturn strings.ToLower(value), false\n}", "func (f *StringSetFilter) ToLower() *StringSetFilter {\r\n\tf.strcase = STRING_LOWERCASE\r\n\treturn f\r\n}", "func (t *StringDataType) ToLower() *StringDataType {\n\treturn t.Formatter(func(s string) string {\n\t\treturn strings.ToLower(s)\n\t})\n}", "func LowerFirst(s string) string {\n\tif len(s) == 0 {\n\t\treturn \"\"\n\t}\n\tc := s[0]\n\tif ('A' <= c) && (c <= 'Z') {\n\t\treturn string(rune(int(c)+32)) + SubString(s, 1, len(s)-1)\n\t}\n\treturn s\n}", "func LowercaseFirstChar(s string) string {\n\tfor i, v := range s {\n\t\treturn string(unicode.ToLower(v)) + s[i+1:]\n\t}\n\treturn \"\"\n}", "func LowerFirst(s string) string {\n\tif len(s) == 0 {\n\t\treturn s\n\t}\n\tr, n := utf8.DecodeRuneInString(s)\n\treturn string(unicode.ToLower(r)) + s[n:]\n}", "func firstLowercase(s string) string {\n\ta := []rune(s)\n\ta[0] = unicode.ToLower(a[0])\n\treturn string(a)\n}", "func LowerSnakeCase(s string) string {\n\tparts := []string{}\n\tfor _, part := range SplitSymbol(s) {\n\t\tif part == \"\" {\n\t\t\tparts = append(parts, \"_\")\n\t\t\tcontinue\n\t\t}\n\t\tparts = append(parts, strings.ToLower(part))\n\t}\n\treturn strings.Join(parts, \"_\")\n}", "func (t Type) GoLowerCase() string {\n\treturn gocase.To(strcase.ToLowerCamel(string(t)))\n}", "func (fn *formulaFuncs) LOWER(argsList *list.List) formulaArg {\n\tif argsList.Len() != 1 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"LOWER requires 1 argument\")\n\t}\n\treturn newStringFormulaArg(strings.ToLower(argsList.Front().Value.(formulaArg).String))\n}", "func LowercaseFirstLetter(s string) string {\n\trunes := []rune(s)\n\trunes[0] = unicode.ToLower(runes[0])\n\treturn string(runes)\n}", "func templateFunctionLower(value string) string {\n\treturn strings.ToLower(value)\n}", "func LowerFirstRune(s string) string {\n\tif s == \"\" {\n\t\treturn \"\"\n\t}\n\tr, n := utf8.DecodeRuneInString(s)\n\treturn string(unicode.ToLower(r)) + s[n:]\n}", "func ToLower(v interface{}) {\n\tif !mustbePtr(v) {\n\t\tfmt.Printf(\"ToLower param(%s) not type of pointer\\n\",\n\t\t\treflect.ValueOf(v).Kind().String())\n\t\treturn\n\t}\n\t// must be pointer\n\tvalue := reflect.ValueOf(v).Elem()\n\tif !typeEqual(value, reflect.Struct) {\n\t\tfmt.Println(\"ToLower param not type of struct\")\n\t\treturn\n\t}\n\n\t// range and toLower\n\tfor i := 0; i < value.NumField(); i++ {\n\t\tfield := value.Field(i)\n\t\tswitch field.Type().Kind() {\n\t\tcase reflect.String:\n\t\t\tfield.SetString(\n\t\t\t\tstrings.ToLower(field.String()),\n\t\t\t)\n\t\tcase reflect.Ptr:\n\t\t\tToLower(field.Interface())\n\t\t}\n\t}\n}", "func fnLower(ctx Context, doc *JDoc, params []string) interface{} {\n\tstats := ctx.Value(EelTotalStats).(*ServiceStats)\n\tif params == nil || len(params) != 1 {\n\t\tctx.Log().Error(\"error_type\", \"func_lower\", \"op\", \"lower\", \"cause\", \"wrong_number_of_parameters\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, SyntaxError{fmt.Sprintf(\"wrong number of parameters in call to lower function\"), \"lower\", params})\n\t\treturn \"\"\n\t}\n\treturn strings.ToLower(extractStringParam(params[0]))\n}", "func isLower(c byte) bool {\n return c >= 97 && c <= 122\n}", "func ToLowerCamelCase(in string) string {\n\tout := toCamelCase([]rune(in))\n\tlength := len(out)\n\tfor i := 0; i < length; i++ {\n\t\tisUpper := unicode.IsUpper(out[i])\n\t\tif isUpper && (i == 0 || i+1 == length || unicode.IsUpper(out[i+1])) {\n\t\t\tout[i] -= 'A' - 'a'\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn string(out)\n}", "func (a AssertableString) IsLowerCase() AssertableString {\n\ta.t.Helper()\n\tif !a.actual.IsLowerCase() {\n\t\ta.t.Error(shouldBeLowerCase(a.actual))\n\t}\n\treturn a\n}", "func (b *Builder) ToLower() *Builder {\n\tb.p.RegisterTransformation(impl.ToLower())\n\treturn b\n}", "func (p Pair) Lower() Pair {\n\tp.Base = p.Base.Lower()\n\tp.Quote = p.Quote.Lower()\n\treturn p\n}", "func toLowerCamelCase(s string) string {\n\treturn toCamelInitCase(s, false)\n}", "func ToLowerSnakeCase(s string) string {\n\treturn strings.ToLower(ToSnakeCase(s))\n}", "func IsLower(b byte) bool { return lookup[b]&lowerMask != 0 }", "func IsLower(rune int) bool {\n\tif rune < 0x80 {\t// quick ASCII check\n\t\treturn 'a' <= rune && rune <= 'z'\n\t}\n\treturn Is(Lower, rune);\n}", "func ToLowerStrings(in []string) []string {\n\tfor i := range in {\n\t\tin[i] = strings.ToLower(in[i])\n\t}\n\treturn in\n}", "func (t Level) Lowercase() string {\n\tswitch t {\n\tcase LevelTrace:\n\t\treturn \"trace\"\n\tcase LevelDebug:\n\t\treturn \"debug\"\n\tcase LevelInfo:\n\t\treturn \"info\"\n\tcase LevelWarn:\n\t\treturn \"warn\"\n\tcase LevelError:\n\t\treturn \"error\"\n\tcase LevelOut:\n\t\treturn \"out\"\n\tdefault:\n\t\treturn \"<unknown>\"\n\t}\n}", "func IsLowerCase(str string) bool {\n\tif IsNull(str) {\n\t\treturn true\n\t}\n\treturn str == strings.ToLower(str)\n}", "func (ent *Entity) GetLowerCasePrefixLetter() string {\n\tif len(ent.Header.Value) > 0 {\n\t\treturn string(ent.Header.Value[0])\n\t}\n\treturn \"e\"\n}", "func bytesToLowerCase(inputBytes []byte) []byte {\r\n\r\n\tmixedcaseBytes := make([]byte, len(inputBytes))\r\n\r\n\tfor i := 0; i < len(inputBytes); i++ {\r\n\t\tif uppercase := isUppercase(inputBytes[i]); uppercase {\r\n\t\t\t// Convert to lowercase\r\n\t\t\tmixedcaseBytes[i] = inputBytes[i] + 0x20\r\n\t\t}\r\n\t}\r\n\treturn mixedcaseBytes\r\n}", "func lowercaseFilter(tokens []string) []string {\n\tr := make([]string, len(tokens))\n\tfor i, token := range tokens {\n\t\tr[i] = strings.ToLower(token)\n\t}\n\treturn r\n}", "func LowerCamelCase(s string) string {\n\tfirst := true\n\tparts := []string{}\n\tfor _, part := range SplitSymbol(s) {\n\t\tif part == \"\" {\n\t\t\tparts = append(parts, \"_\")\n\t\t\tcontinue\n\t\t}\n\t\tif first {\n\t\t\tparts = append(parts, strings.ToLower(part))\n\t\t\tfirst = false\n\t\t} else {\n\t\t\t// Merge trailing s\n\t\t\tif part == \"s\" && len(parts) > 0 {\n\t\t\t\tparts[len(parts)-1] += part\n\t\t\t} else {\n\t\t\t\tif commonInitialisms[strings.ToUpper(part)] {\n\t\t\t\t\tpart = strings.ToUpper(part)\n\t\t\t\t} else {\n\t\t\t\t\tpart = title(part)\n\t\t\t\t}\n\t\t\t\tparts = append(parts, part)\n\t\t\t}\n\t\t}\n\t}\n\treturn strings.Join(parts, \"\")\n}", "func TestPositiveEqualsIgnoreCaseFunc(t *testing.T) {\n\tfirstInput := \"Bola\"\n\tsecondInput := \"bola\"\n\n\tresult := EqualsIgnoreCase(firstInput, secondInput)\n\n\tassert.Equal(t, true, result, \"Result should return true, but actual get \"+result)\n}", "func ToLowerCamel(s string) string {\n\treturn snaker(s, rune(0), unicode.ToLower, unicode.ToUpper, noop)\n}", "func lowerASCIIBytes(x []byte) {\n\tfor i, b := range x {\n\t\tif 'A' <= b && b <= 'Z' {\n\t\t\tx[i] += 'a' - 'A'\n\t\t}\n\t}\n}", "func (g *generator) lowerExpression(expr model.Expression, typ model.Type) model.Expression {\n\trewriteApplies := true\n\treturn g.rewriteExpression(expr, typ, rewriteApplies)\n}", "func ToLowerCamelWithInitialisms(s string, initialisms map[string]bool) string {\n\tif s == \"\" {\n\t\treturn s\n\t}\n\tif initialisms == nil {\n\t\tinitialisms = map[string]bool{}\n\t}\n\tss := SplitIntoWordsWithInitialisms(s, initialisms)\n\tfor i, s := range ss {\n\t\tif i == 0 {\n\t\t\tss[i] = strings.ToLower(s)\n\t\t\tcontinue\n\t\t}\n\t\tif initialisms[strings.ToLower(s)] {\n\t\t\tss[i] = strings.ToUpper(s)\n\t\t\tcontinue\n\t\t}\n\t\tif strings.ToLower(s[len(s)-1:]) == \"s\" && initialisms[strings.ToLower(s[:len(s)-1])] {\n\t\t\tss[i] = strings.ToUpper(s[:len(s)-1]) + \"s\"\n\t\t\tcontinue\n\t\t}\n\t\tss[i] = strings.ToUpper(s[:1]) + strings.ToLower(s[1:])\n\t}\n\treturn strings.Join(ss, \"\")\n}", "func LowerCase(str string) bool {\n\tif len(str) == 0 {\n\t\treturn true\n\t}\n\treturn str == strings.ToLower(str)\n}", "func LowerCamelCase(s string) string {\n\tif s == \"\" {\n\t\treturn \"\"\n\t}\n\trunes := []rune(s)\n\treturn string(append([]rune{unicode.ToLower(runes[0])}, runes[1:]...))\n}", "func ToLowerSnake(s string) (string, bool) {\n\tsnake := strcase.ToSnake(strings.ToLower(s))\n\treturn snake, s == snake\n}", "func ToLowerCamel(s string) (string, bool) {\n\tcamel := strcase.ToLowerCamel(s)\n\tsnake := strcase.ToSnake(camel)\n\treturn camel, s == snake\n}", "func lowerSlice(input []string) []string {\n\tvar output []string\n\tfor _, v := range input {\n\t\toutput = append(output, strings.ToLower(v))\n\t}\n\treturn output\n}", "func (p Pair) Lower() Pair {\n\treturn Pair{\n\t\tDelimiter: p.Delimiter,\n\t\tBase: p.Base.Lower(),\n\t\tQuote: p.Quote.Lower(),\n\t}\n}", "func TrainCase(t string) string {\n\treturn KebabCase(t, true, false)\n}", "func ToLowerCamel(s string) string {\n\treturn ToLowerCamelWithInitialisms(s, CommonInitialisms)\n}", "func lowCamelName(s string) string {\n\ts = gogen.CamelCase(s)\n\tnew := []rune(s)\n\tif len(new) < 1 {\n\t\treturn s\n\t}\n\trv := []rune{}\n\trv = append(rv, unicode.ToLower(new[0]))\n\trv = append(rv, new[1:]...)\n\treturn string(rv)\n}", "func LowerAlphaASCII(loweredCol *chunk.Column, rowNum int) {\n\tfor i := 0; i < rowNum; i++ {\n\t\tstr := loweredCol.GetString(i)\n\t\tstrBytes := hack.Slice(str)\n\n\t\tstringutil.LowerOneString(strBytes)\n\t}\n}", "func toLowerAndFormat(args []string) string {\n\tvar endStr []string\n\tfor _, word := range args {\n\t\tendStr = append(endStr, strings.ToLower(word))\n\t}\n\tif len(args) > 1 {\n\t\treturn strings.Join(endStr, \"_\")\n\t}\n\treturn endStr[0]\n}", "func LowerRune(r rune) rune {\n\tstr := strings.ToLower(string(r))\n\treturn []rune(str)[0]\n}", "func normalize(s string) string {\n\tvar sb strings.Builder\n\tfor _, c := range s {\n\t\tif !unicode.IsLetter(c) && !unicode.IsNumber(c) {\n\t\t\tcontinue\n\t\t}\n\t\tsb.WriteRune(unicode.ToLower(c))\n\t}\n\treturn sb.String()\n}", "func ProcessCode(code string) string {\n\tcode = strings.Replace(code, \" \", \"-\", -1)\n\treturn strings.ToLower(code)\n}", "func StartsWithIgnoreCase(str string, prefix string) bool {\n\treturn internalStartsWith(str, prefix, true)\n}", "func LcFirst(str string) string {\n\tfor _, v := range str {\n\t\tu := string(unicode.ToLower(v))\n\t\treturn u + str[len(u):]\n\t}\n\treturn \"\"\n}", "func Lowercase(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tr.URL.Path = strings.ToLower(r.URL.Path)\n\t\tnext.ServeHTTP(w, r)\n\t})\n}", "func SnakeToLowerCamel(original string) string {\n\treturn toCamel(original, snakeDelimiter, false)\n}", "func isLowercase(fl FieldLevel) bool {\n\tfield := fl.Field()\n\n\tif field.Kind() == reflect.String {\n\t\tif field.String() == \"\" {\n\t\t\treturn false\n\t\t}\n\t\treturn field.String() == strings.ToLower(field.String())\n\t}\n\n\tpanic(fmt.Sprintf(\"Bad field type %T\", field.Interface()))\n}", "func caseInsensitiveLess(a, b string) bool {\n\tfor i := 0; i < len(a) && i < len(b); i++ {\n\t\tif a[i] >= 'A' && a[i] <= 'Z' {\n\t\t\tif b[i] >= 'A' && b[i] <= 'Z' {\n\t\t\t\t// both are uppercase, do nothing\n\t\t\t\tif a[i] < b[i] {\n\t\t\t\t\treturn true\n\t\t\t\t} else if a[i] > b[i] {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// a is uppercase, convert a to lowercase\n\t\t\t\tif a[i]+32 < b[i] {\n\t\t\t\t\treturn true\n\t\t\t\t} else if a[i]+32 > b[i] {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\t} else if b[i] >= 'A' && b[i] <= 'Z' {\n\t\t\t// b is uppercase, convert b to lowercase\n\t\t\tif a[i] < b[i]+32 {\n\t\t\t\treturn true\n\t\t\t} else if a[i] > b[i]+32 {\n\t\t\t\treturn false\n\t\t\t}\n\t\t} else {\n\t\t\t// neither are uppercase\n\t\t\tif a[i] < b[i] {\n\t\t\t\treturn true\n\t\t\t} else if a[i] > b[i] {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\treturn len(a) < len(b)\n}", "func AllLower(embed *discordgo.MessageEmbed) *discordgo.MessageEmbed {\n\tembed.Author.Name = \"Command: lower\"\n\tembed.Description = \"`lower <text>` will create the text given into lower case.\"\n\tembed.Fields = []*discordgo.MessageEmbedField{\n\t\t{\n\t\t\tName: \"<text>\",\n\t\t\tValue: \"The text to completely lower case.\",\n\t\t\tInline: true,\n\t\t},\n\t\t{\n\t\t\tName: \"Related commands:\",\n\t\t\tValue: \"`caps`, `randomcaps`, `swap`, `title`\",\n\t\t},\n\t}\n\treturn embed\n}", "func transformizer(replacer *strings.Replacer, str string) (result string, err error) {\n\tt := transform.Chain(norm.NFD, runes.Remove(runes.In(unicode.Mn)), norm.NFC)\n\tpreResult, _, err := transform.String(t, str)\n\n\tif err != nil {\n\t\treturn result, err\n\t}\n\n\tresult = strings.ToLower(replacer.Replace(preResult))\n\treturn\n}", "func (lc LowerCaseConvention) Convert(fieldName string) string {\n\treturn strings.ToLower(fieldName)\n}", "func isOnlyLowerCase(text string) bool {\n\tconst chars = \"0123456789abcdefghijklmnopqrstuvwxyz.+-*/%&!# _,;:()[]{}\"\n\tfor _, c := range text {\n\t\tif !strings.Contains(chars, string(c)) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (l Level) LowerName() string {\n\treturn strings.ToLower(LevelName(l))\n}", "func IgnoringCase() StringOpt {\n\treturn func(c *AssertableString) {\n\t\tc.actual = c.actual.AddDecorator(strings.ToLower)\n\t}\n}", "func IsLowercase(s string) bool {\n\tif s == \"\" {\n\t\treturn false\n\t}\n\treturn strings.ToLower(s) == s\n}", "func ToLowerCamel(s string) string {\n\tif s == \"\" {\n\t\treturn s\n\t}\n\tif uppercaseAcronym[s] {\n\t\ts = strings.ToLower(s)\n\t}\n\tif r := rune(s[0]); r == '_' {\n\t\ts = s[1:]\n\t}\n\tif r := rune(s[0]); r >= 'A' && r <= 'Z' {\n\t\ts = strings.ToLower(string(r)) + s[1:]\n\t}\n\treturn toCamelInitCase(s, false)\n}", "func inplaceASCIILower(s []byte) {\n\tfor i := 0; i < len(s); i++ {\n\t\tc := s[i]\n\t\tif 'A' <= c && c <= 'Z' {\n\t\t\tc += 'a' - 'A'\n\t\t}\n\t\ts[i] = c\n\t}\n}" ]
[ "0.7667526", "0.736745", "0.7208126", "0.7137268", "0.7069492", "0.705703", "0.705131", "0.7034256", "0.70304763", "0.6950598", "0.6929829", "0.6856248", "0.6828877", "0.6773903", "0.6757076", "0.673432", "0.67115104", "0.66437197", "0.6626718", "0.66254663", "0.6624735", "0.6601172", "0.65981877", "0.6580353", "0.6539085", "0.65339184", "0.650517", "0.6504496", "0.6473746", "0.64406896", "0.6433359", "0.6433133", "0.6396546", "0.6385746", "0.63755417", "0.6369667", "0.6365583", "0.63358927", "0.63127154", "0.6297414", "0.62845844", "0.6239979", "0.6234248", "0.6227224", "0.6176974", "0.6118055", "0.6116684", "0.60737425", "0.60569704", "0.60545903", "0.605188", "0.6039104", "0.6022573", "0.60112196", "0.6009337", "0.6003119", "0.599419", "0.59865505", "0.5969584", "0.5967293", "0.59562194", "0.59459704", "0.5931154", "0.5887373", "0.58335245", "0.57921964", "0.5775667", "0.5763398", "0.5702676", "0.56866854", "0.56852144", "0.5634466", "0.56144255", "0.55984765", "0.557399", "0.55693024", "0.5569109", "0.5563792", "0.55583864", "0.5557856", "0.55512357", "0.5546736", "0.5514679", "0.5513022", "0.55027556", "0.5499407", "0.5479563", "0.5476706", "0.54611176", "0.5436231", "0.5435668", "0.5432761", "0.5429231", "0.5421088", "0.5419005", "0.53938895", "0.5390049", "0.53784907", "0.5359348", "0.5352267" ]
0.8119913
0
ToUpper uses strings.ToUpper to return operand with all unicode codepoints converted to uppercase.
func ToUpper(operand string) string { return strings.ToUpper(operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func upper(in string) string {\n\tout := []rune(in)\n\tout[0] = unicode.ToUpper(out[0])\n\treturn string(out)\n}", "func ToUpper(r rune) rune", "func Upper(s string) string {\n\treturn strings.ToUpper(s)\n}", "func upperCase(v string) (string, error) {\n\treturn strings.ToUpper(v), nil\n}", "func ToUpper(s string) string {\n\treturn strings.ToUpper(s)\n}", "func (suc *StringUpperCaser) ToUpper() string {\n\treturn strings.ToUpper(suc.String())\n}", "func ToUpper(str string) string {\n\treturn strings.ToUpper(str)\n}", "func ToUpperCase(s string) string {\n\tr := []rune(s)\n\tfor i := 0; i < len(r); i = i + 1 {\n\t\tif r[i] >= 97 && r[i] <= 122 {\n\t\t\tr[i] -= 32\n\t\t}\n\t}\n\treturn string(r)\n}", "func ToUpper() (desc string, f predicate.TransformFunc) {\n\tdesc = \"ToUpper({})\"\n\tf = func(v interface{}) (r interface{}, ctx []predicate.ContextValue, err error) {\n\t\ts, ok := v.(string)\n\t\tif !ok {\n\t\t\treturn nil, nil, fmt.Errorf(\n\t\t\t\t\"value of type '%T' cannot be transformed to uppercase\", v)\n\t\t}\n\t\ts = strings.ToUpper(s)\n\t\treturn s, []predicate.ContextValue{\n\t\t\t{Name: \"upper\", Value: s, Pre: false},\n\t\t}, nil\n\t}\n\treturn\n}", "func SpecialCaseToUpper(special unicode.SpecialCase, r rune) rune", "func main() {\n\t//func ToUpper(s string) string\n\tfmt.Println(strings.ToUpper(\"test\"))\n\t// => TEST\n}", "func (s *Str) ToUpper() *Str {\n\ts.val = strings.ToUpper(s.val)\n\treturn s\n}", "func ToUpperSpecial(c unicode.SpecialCase) MapFunc {\n\treturn func(s string) string { return strings.ToUpperSpecial(c, s) }\n}", "func (s *Stringish) ToUpper() *Stringish {\n\ts.str = strings.ToUpper(s.str)\n\treturn s\n}", "func ToUpper(rune int) int {\n\tif rune < 0x80 {\t// quick ASCII check\n\t\tif 'a' <= rune && rune <= 'z' {\n\t\t\trune -= 'a' - 'A'\n\t\t}\n\t\treturn rune;\n\t}\n\treturn To(UpperCase, rune);\n}", "func filterUpper(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.ToUpper(stick.CoerceString(val))\n}", "func (StringService) Uppercase(s string) (string, error) {\n\tif s == \"\" {\n\t\treturn \"\", errEmpty\n\t}\n\treturn strings.ToUpper(s), nil\n}", "func UpperCase(str string) string {\n\treturn strings.ToUpper(str)\n}", "func VZEROUPPER() { ctx.VZEROUPPER() }", "func (s Set) Uppercase(ctx context.Context, a string) (str string, err error) {\n\tresp, err := s.UppercaseEndpoint(ctx, UppercaseRequest{\n\t\tS: a,\n\t})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tresponse := resp.(UppercaseResponse)\n\treturn response.V, str2err(response.Err)\n}", "func Strtoupper(str string) string {\n\treturn strings.ToUpper(str)\n}", "func (vl VerbLevel) Upper() string {\n\treturn strings.ToUpper(vl.Name())\n}", "func capitalize(s string) string {\n\treturn strings.ToUpper(s[:1]) + s[1:]\n}", "func UpperCamelCase(s string) string {\n\tparts := []string{}\n\tfor _, part := range SplitSymbol(s) {\n\t\tif part == \"\" {\n\t\t\tparts = append(parts, \"_\")\n\t\t\tcontinue\n\t\t}\n\t\tif part == \"s\" && len(parts) > 0 {\n\t\t\tparts[len(parts)-1] += part\n\t\t} else {\n\t\t\tif commonInitialisms[strings.ToUpper(part)] {\n\t\t\t\tpart = strings.ToUpper(part)\n\t\t\t} else {\n\t\t\t\tpart = title(part)\n\t\t\t}\n\t\t\tparts = append(parts, part)\n\t\t}\n\t}\n\treturn strings.Join(parts, \"\")\n}", "func (b *Builder) ToUpper() *Builder {\n\tb.p.RegisterTransformation(impl.ToUpper())\n\treturn b\n}", "func ToUpperCamel(s string) string {\n\treturn toCamelInitCase(s, true)\n}", "func capitalizedWord(word string) string {\n\tvar processedWord = removeAllSymbols(word)\n\treturn \"k\" + strings.ToUpper(processedWord)\n}", "func (t *StringDataType) ToUpper() *StringDataType {\n\treturn t.Formatter(func(s string) string {\n\t\treturn strings.ToUpper(s)\n\t})\n}", "func capitalise(text string) string {\n\tresult := strings.ToLower(text)\n\tresult = strings.Title(result)\n\treturn result\n}", "func ToUppercase(str []byte) []byte {\n\tfor i, s := range str {\n\t\tif s > 96 && s < 123 {\n\t\t\tstr[i] = s - 32\n\t\t}\n\t}\n\treturn str\n}", "func MakeUpperCase(str string) string {\n\treturn (strings.ToUpper(str))\n}", "func MapToUppercase(vs []string) []string {\n\tvso := make([]string, len(vs))\n\tfor i, v := range vs {\n\t\tvso[i] = strings.ToUpper(v)\n\t}\n\treturn vso\n}", "func templateFunctionUpper(value string) string {\n\treturn strings.ToUpper(value)\n}", "func Capitalize(str string) string {\n\tvar upperStr string\n\tsrunes := []rune(str)\n\n\tfor idx := range srunes {\n\t\tif idx == 0 {\n\t\t\tsrunes[idx] = unicode.ToUpper(srunes[idx])\n\t\t}\n\t\tupperStr += string(srunes[idx])\n\t}\n\n\treturn upperStr\n}", "func ConvertUpper(content string) <-chan string {\n\tuc := make(chan string)\n\tgo func() {\n\t\tuc <- strings.ToUpper(content)\n\t\tclose(uc)\n\t}()\n\treturn uc\n}", "func main() {\n\ts := \"Sarah\" // This thing (type) is a string, with name of s, and value of Sarah\n\tc := capitalise(s)\n\n\tfmt.Println(\"Hello \" + c)\n}", "func fnUpper(ctx Context, doc *JDoc, params []string) interface{} {\n\tstats := ctx.Value(EelTotalStats).(*ServiceStats)\n\tif params == nil || len(params) != 1 {\n\t\tctx.Log().Error(\"error_type\", \"func_upper\", \"op\", \"upper\", \"cause\", \"wrong_number_of_parameters\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, SyntaxError{fmt.Sprintf(\"wrong number of parameters in call to upper function\"), \"upper\", params})\n\t\treturn \"\"\n\t}\n\treturn strings.ToUpper(extractStringParam(params[0]))\n}", "func ToUpperCamelCase(name string) string {\n\tparts := nameParts(name)\n\tfor i := range parts {\n\t\tparts[i] = strings.Title(strings.ToLower(parts[i]))\n\t\tif parts[i] == \"\" {\n\t\t\tparts[i] = \"_\"\n\t\t}\n\t}\n\treturn strings.Join(parts, \"\")\n}", "func ToUpperSnakeCase(s string) string {\n\treturn strings.ToUpper(ToSnakeCase(s))\n}", "func capitalizeString(s string) string {\n\tr := []rune(s)\n\tr[0] = unicode.ToUpper(r[0])\n\n\treturn string(r)\n}", "func capitalize(s string) string {\n\tif s == \"\" {\n\t\treturn s\n\t}\n\tr, n := utf8.DecodeRuneInString(s)\n\treturn string(unicode.ToTitle(r)) + s[n:]\n}", "func (pair *StringPair) UpperCase() {\n\tpair.first = strings.ToUpper(pair.first)\n\tpair.second = strings.ToUpper(pair.second)\n}", "func UpperSnakeCase(s string) string {\n\tparts := []string{}\n\tfor _, part := range SplitSymbol(s) {\n\t\tif part == \"\" {\n\t\t\tparts = append(parts, \"_\")\n\t\t\tcontinue\n\t\t}\n\t\tparts = append(parts, strings.ToUpper(part))\n\t}\n\treturn strings.Join(parts, \"_\")\n}", "func (p Pair) Upper() Pair {\n\tp.Base = p.Base.Upper()\n\tp.Quote = p.Quote.Upper()\n\treturn p\n}", "func StringUpper(scope *Scope, input tf.Output, optional ...StringUpperAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"StringUpper\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func ToUpperCamel(s string) (string, bool) {\n\tcamel := strcase.ToCamel(s)\n\tsnake := strcase.ToSnake(camel)\n\treturn camel, s == snake\n}", "func IsUpper(r rune) bool", "func ToUpper() opt {\n\treturn func(key string) string {\n\t\treturn strings.ToUpper(key)\n\t}\n}", "func FirstToUpper(in string) string {\n\tif in == \"\" {\n\t\treturn in\n\t}\n\tr, size := utf8.DecodeRuneInString(in)\n\tif r == utf8.RuneError {\n\t\treturn in\n\t}\n\treturn string(unicode.ToUpper(r)) + in[size:]\n}", "func (c *Context) VZEROUPPER() {\n\tc.addinstruction(x86.VZEROUPPER())\n}", "func (p RProc) Upper() RProc { return RProc{p.p.upper, p.mrb} }", "func (f *StringSetFilter) ToUpper() *StringSetFilter {\r\n\tf.strcase = STRING_UPPERCASE\r\n\treturn f\r\n}", "func filterCapitalize(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\ts := stick.CoerceString(val)\n\treturn strings.ToUpper(s[:1]) + s[1:]\n}", "func Capitalize(str string) string {\n\tfor i, c := range str {\n\t\treturn string(unicode.ToUpper(c)) + str[i+1:]\n\t}\n\treturn \"\"\n}", "func (fn *formulaFuncs) UPPER(argsList *list.List) formulaArg {\n\tif argsList.Len() != 1 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"UPPER requires 1 argument\")\n\t}\n\treturn newStringFormulaArg(strings.ToUpper(argsList.Front().Value.(formulaArg).String))\n}", "func IsUpper(b byte) bool { return lookup[b]&upperMask != 0 }", "func Uncapitalize(str string) string {\n\tif str == \"\" {\n\t\treturn str\n\t}\n\tr := []rune(str)\n\tif unicode.IsUpper(r[0]) {\n\t\treturn string(unicode.ToLower(r[0])) + string(r[1:])\n\t}\n\treturn str\n}", "func (e *Upper) Process() {\n\tfor val := range e.Val {\n\t\te.Res <- strings.ToUpper(val)\n\t}\n}", "func capitalize(label string) string {\n\tfirstLetter := strings.SplitN(label, \"\", 2)\n\tif len(firstLetter) < 1 {\n\t\treturn label\n\t}\n\treturn fmt.Sprintf(\"%v%v\", strings.ToUpper(firstLetter[0]),\n\t\tstrings.TrimPrefix(label, firstLetter[0]))\n}", "func Capitalize(in *pongo2.Value, param *pongo2.Value) (*pongo2.Value, *pongo2.Error) {\n\tif !in.IsString() {\n\t\treturn pongo2.AsValue(\"\"), nil\n\t}\n\treturn pongo2.AsValue(strings.Title(strings.ToLower(in.String()))), nil\n}", "func Abbreviate(in string) string {\n\tpos := 0\n\tout := \"\"\n\tfor i, v := range in {\n\t\tif v == ' ' || v == '-' || i == len(in)-1 {\n\t\t\tout += strings.ToUpper(in[pos : pos+1])\n\t\t\tpos = i + 1\n\t\t}\n\t}\n\treturn out\n}", "func UpperRune(r rune) rune {\n\tstr := strings.ToUpper(string(r))\n\treturn []rune(str)[0]\n}", "func Capitalize(str string) string {\n\tif len(str) <= 0 {\n\t\treturn \"\"\n\t}\n\n\tif len(str) == 1 {\n\t\treturn strings.ToUpper(str)\n\t}\n\n\treturn strings.ToUpper(str[:1]) + strings.ToLower(str[1:])\n}", "func MakeUppercaseEndpoint(svc StringSvc) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(UppercaseRequest)\n\t\tv, err := svc.Uppercase(ctx, req.S)\n\t\treturn UppercaseResponse{V: v, Err: err}, nil\n\t}\n}", "func upperOnly(s string) []byte {\n\tu := make([]byte, 0, len(s))\n\tfor i := 0; i < len(s); i++ {\n\t\tc := s[i]\n\t\tif c >= 'A' && c <= 'Z' {\n\t\t\tu = append(u, c)\n\t\t} else if c >= 'a' && c <= 'z' {\n\t\t\tu = append(u, c-32)\n\t\t}\n\t}\n\treturn u\n}", "func upperFirst(s string) string {\n\treturn strings.ToUpper(string([]rune(s)[0])) + string([]rune(s)[1:])\n}", "func TestHappyUpper(t *testing.T) {\n\tvar testCases = []struct {\n\t\tinput string\n\t\twant string\n\t}{\n\t\t{\"low\", \"LOW😊\"},\n\t\t{\"l\", \"L😊\"},\n\t\t{\"1234\", \"1234😊\"},\n\t\t{\"\", \"😊\"},\n\t}\n\n\tfor _, tc := range testCases {\n\t\thave := happyUpper(tc.input)\n\t\tif have != tc.want {\n\t\t\tt.Errorf(\"Error. Want %s have %s\", tc.want, have)\n\t\t}\n\t}\n}", "func SwitchFormatToUpper(form Format) {\n\tSwitchFormat(Format(strings.ToUpper(string(form))))\n}", "func Capital(s string) string {\n\treturn cases.Title(language.English, cases.NoLower).String(s)\n}", "func upperFirst(str string) string {\n\tfor i, v := range str {\n\t\treturn string(unicode.ToUpper(v)) + str[i+1:]\n\t}\n\treturn \"\"\n}", "func Abbreviate(s string) string {\n\tre := regexp.MustCompile(\"\\\\b[a-zA-Z]\")\n\tres := re.FindAllString(s, -1)\n\tacronym := strings.Join(res, \"\")\n\treturn strings.ToUpper(acronym)\n}", "func CamelCase(s string, upper bool) string {\n\tif s == \"\" {\n\t\treturn \"\"\n\t}\n\n\tout := make([]rune, 1, len(s)+5)\n\tfor i, r := range s {\n\t\tif i == 0 {\n\t\t\tif upper {\n\t\t\t\tr = unicode.ToUpper(r)\n\t\t\t}\n\t\t\tout[0] = r\n\t\t\tcontinue\n\t\t}\n\n\t\tif i == 1 {\n\t\t\tif !upper && unicode.Is(unicode.Lower, r) {\n\t\t\t\tout[0] = unicode.ToLower(out[0])\n\t\t\t}\n\n\t\t\tupper = false\n\t\t}\n\n\t\tswitch {\n\t\tcase unicode.IsLetter(r):\n\t\t\tif upper {\n\t\t\t\tr = unicode.ToUpper(r)\n\t\t\t}\n\n\t\t\tfallthrough\n\t\tcase unicode.IsNumber(r):\n\t\t\tupper = false\n\t\t\tout = append(out, r)\n\n\t\tdefault:\n\t\t\tupper = true\n\n\t\t}\n\t}\n\n\treturn string(out)\n}", "func (t DbStrCmpAsciiCI) Convert(input string) string {\n\treturn strings.ToUpper(input)\n}", "func ToUpperSnake(s string) string {\n\treturn snaker(s, '_', unicode.ToUpper, unicode.ToUpper, unicode.ToUpper)\n}", "func (pm *ProductMetadata) ProductNameUpper() string {\n\treturn strings.ToUpper(pm.ProductName)\n}", "func ToCamelCase(in string) string {\n\trunes := []rune(in)\n\tout := make([]rune, 0, len(runes))\n\tup := true\n\tfor i := 0; i < len(runes); i++ {\n\t\tr := runes[i]\n\t\tif r == '_' {\n\t\t\tup = true\n\t\t} else {\n\t\t\tif up {\n\t\t\t\tr = unicode.ToUpper(r)\n\t\t\t\tup = false\n\t\t\t}\n\t\t\tout = append(out, r)\n\t\t}\n\t}\n\treturn string(out)\n}", "func (p Pair) Upper() Pair {\n\treturn Pair{\n\t\tDelimiter: p.Delimiter,\n\t\tBase: p.Base.Upper(),\n\t\tQuote: p.Quote.Upper(),\n\t}\n}", "func ToUpperFirst(s string) string {\n\treturn strings.ToUpper(s[:1]) + s[1:]\n}", "func (p RProc) SetUpper(upper RProc) { p.p.upper = upper.p }", "func Abbreviate(s string) string {\n\ts = strings.TrimSpace(s)\n\tvar outPut string\n\tfor i, w := range s {\n\n\t\tif i == 0 {\n\t\t\tfmt.Println(string(w))\n\t\t\toutPut += string(w)\n\t\t} else if !unicode.IsLetter(rune(s[i-1])) && rune(s[i-1]) != 39 {\n\t\t\tif unicode.IsLetter(w) {\n\t\t\t\toutPut += strings.ToUpper(string(w))\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\treturn outPut\n}", "func changeCase(str string) string {\n\n\tswitch capitalize {\n\tcase 0:\n\t\treturn str\n\tcase 1:\n\t\treturn strings.ToUpper(str)\n\tcase 2:\n\t\treturn strings.ToLower(str)\n\tdefault:\n\t\treturn str\n\t}\n}", "func SnakeToUpperCamel(original string) string {\n\treturn toCamel(original, snakeDelimiter, true)\n}", "func Abbreviate(s string) (output string) {\n\tfor _, v := range regex.FindAllString(strings.ToUpper(s), -1) {\n\t\toutput += string(v[0])\n\t}\n\treturn output\n}", "func isUppercase(char byte) bool {\r\n\treturn ((0x41 <= char) && (char <= 0x5A))\r\n}", "func UpperFirstChar(str string) string {\n\tfor i, v := range str {\n\t\treturn string(unicode.ToUpper(v)) + str[i+1:]\n\t}\n\treturn str\n}", "func ToLower(operand string) string { return strings.ToLower(operand) }", "func KebabToUpperCamel(original string) string {\n\treturn toCamel(original, kebabDelimiter, true)\n}", "func CamelCase(s string) string {\n\n\tconst Delemeter = '_'\n\n\twriter := bytes.NewBuffer(make([]byte, 0, len(s)))\n\treader := strings.NewReader(s)\n\n\tvar prev rune\n\tfor {\n\t\tr, _, err := reader.ReadRune()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\n\t\tif prev == 0x00 && r == Delemeter {\n\t\t\twriter.WriteRune('X')\n\n\t\t} else if unicode.IsLower(r) && (prev == 0x00 || prev == Delemeter || (prev >= '0' && prev <= '9')) {\n\t\t\tif lastIndex := writer.Len() - 1; lastIndex >= 0 && writer.Bytes()[lastIndex] == Delemeter {\n\t\t\t\twriter.Truncate(lastIndex) // remove delemeter before symbol\n\t\t\t}\n\t\t\twriter.WriteRune(unicode.ToUpper(r))\n\n\t\t} else {\n\t\t\twriter.WriteRune(r)\n\n\t\t}\n\n\t\tprev = r\n\t}\n\n\treturn writer.String()\n}", "func UpperFirst(value string) string {\n\tfor i, v := range value {\n\t\treturn string(unicode.ToUpper(v)) + value[i+1:]\n\t}\n\treturn \"\"\n}", "func (a AssertableString) IsUpperCase() AssertableString {\n\ta.t.Helper()\n\tif !a.actual.IsUpperCase() {\n\t\ta.t.Error(shouldBeUpperCase(a.actual))\n\t}\n\treturn a\n}", "func FirstToUpper(str string) string {\n\truneStr := []rune(str)\n\tfor i := 0; i < len(runeStr); i++ {\n\t\tif unicode.IsLetter(runeStr[i]) {\n\t\t\tif unicode.IsLower(runeStr[i]) {\n\t\t\t\truneStr[i] = unicode.ToUpper(runeStr[i])\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\t}\n\treturn string(runeStr)\n}", "func CamelCase(s string) string {\n\tif s == \"\" {\n\t\treturn \"\"\n\t}\n\n\tt := make([]byte, 0, 32)\n\ti := 0\n\tif s[0] == '_' {\n\t\tt = append(t, 'X')\n\t\ti++\n\t}\n\tfor ; i < len(s); i++ {\n\t\tc := s[i]\n\t\tif c == '_' && i+1 < len(s) && isASCIILower(s[i+1]) {\n\t\t\tcontinue // Skip the underscore in s.\n\t\t}\n\n\t\tif isASCIIDigit(c) {\n\t\t\tt = append(t, c)\n\t\t\tcontinue\n\t\t}\n\n\t\tif isASCIILower(c) {\n\t\t\tc ^= ' ' // Make it a capital letter.\n\t\t}\n\n\t\tt = append(t, c) // Guaranteed not lower case.\n\t\tfor i+1 < len(s) && isASCIILower(s[i+1]) {\n\t\t\ti++\n\t\t\tt = append(t, s[i])\n\t\t}\n\t}\n\n\treturn string(t)\n}", "func HeadToUpper(str string) string {\n\tln := len(str)\n\tif ln <= 0 {\n\t\treturn str\n\t}\n\treturn strings.ToUpper(str[0:1]) + str[1:]\n}", "func Abbreviate(s string) string {\n\tvar a string\n\tfor _, w := range regexp.MustCompile(\"\\\\s|\\\\-\").Split(s, 5) {\n\t\ta = a + strings.ToUpper(string([]rune(w)[0]))\n\t}\n\treturn a\n}", "func Abbreviate(s string) string {\n\twordParse := regexp.MustCompile(`[[:alpha:]']+`)\n\twords := wordParse.FindAllString(s, -1)\n\tabbr := make([]byte, len(words))\n\tfor i, w := range words {\n\t\tabbr[i] += strings.ToUpper(w)[0]\n\t}\n\treturn string(abbr)\n}", "func Abbreviate(s string) (retVal string) {\n\ts = strings.ReplaceAll(s, \"-\", \" \")\n\ts = strings.ReplaceAll(s, \"_\", \" \")\n\n\tfor _, v := range strings.Fields(s) {\n\t\tretVal += string(v[0])\n\t}\n\treturn strings.ToUpper(retVal)\n}", "func WordCapitalize(word string) string {\n\tvar n []string\n\tfor i, r := range word {\n\t\tif i == 0 {\n\t\t\tn = append(n, strings.ToUpper(string(r)))\n\t\t} else {\n\t\t\tn = append(n, strings.ToLower(string(r)))\n\t\t}\n\t}\n\treturn strings.Join(n, \"\")\n}", "func ToCamelCase(in string) string {\n\treturn string(toCamelCase([]rune(in)))\n}", "func (u *Upper) Eval(\n\tctx *sql.Context,\n\trow sql.Row,\n) (interface{}, error) {\n\tv, err := u.Child.Eval(ctx, row)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif v == nil {\n\t\treturn nil, nil\n\t}\n\n\tv, err = sql.LongText.Convert(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn strings.ToUpper(v.(string)), nil\n}", "func allUpper(s string) bool {\n\treturn strings.ToUpper(s) == s\n}" ]
[ "0.7619067", "0.7483768", "0.7391097", "0.7280122", "0.72594786", "0.7113133", "0.71099466", "0.70730865", "0.7053634", "0.701086", "0.68747383", "0.6840409", "0.68256366", "0.6812164", "0.67067987", "0.66705513", "0.6659231", "0.66334844", "0.6598262", "0.6582776", "0.6562218", "0.6530775", "0.65123546", "0.6453987", "0.6436832", "0.64170504", "0.64092076", "0.6365288", "0.62594545", "0.6253234", "0.6244984", "0.61964804", "0.6188273", "0.6184813", "0.61754656", "0.6168621", "0.61554396", "0.611004", "0.6102723", "0.6098441", "0.60961556", "0.6060144", "0.6048016", "0.60279906", "0.60265535", "0.60166246", "0.6008231", "0.5986598", "0.59776855", "0.5919531", "0.591507", "0.59089494", "0.5908121", "0.59028375", "0.5883549", "0.5840499", "0.5801754", "0.5789766", "0.57834375", "0.5780625", "0.57805336", "0.5763376", "0.57411444", "0.5740739", "0.5726143", "0.57061917", "0.5679966", "0.5674299", "0.5668352", "0.5658817", "0.56455153", "0.56224203", "0.5574586", "0.55723965", "0.55580366", "0.5553767", "0.5552987", "0.55510354", "0.5535023", "0.55325675", "0.552682", "0.5512458", "0.5497557", "0.54793197", "0.54755664", "0.5457737", "0.5445452", "0.54384065", "0.54367936", "0.5425091", "0.5422041", "0.54030234", "0.5389203", "0.53788245", "0.5377744", "0.53769815", "0.5373368", "0.5357486", "0.53573334", "0.535463" ]
0.84688693
0
Trim uses strings.Trim to remove any occurrences of chars from the beginning and end of operand.
func Trim(chars string, operand string) string { return strings.Trim(operand, chars) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TrimSpace(operand string) string { return strings.TrimSpace(operand) }", "func TrimLeft(chars string, operand string) string { return strings.TrimLeft(operand, chars) }", "func TrimPrefix(prefix, operand string) string { return strings.TrimPrefix(operand, prefix) }", "func filterTrim(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.TrimSpace(stick.CoerceString(val))\n}", "func TrimRight(cutset string, operand string) string { return strings.TrimRight(operand, cutset) }", "func Trim(text string) string {\n\treturn trimRx.FindStringSubmatch(text)[1]\n}", "func trim(s string) string {\n\tfor len(s) > 0 {\n\t\tif s[0] <= ' ' {\n\t\t\ts = s[1:]\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tfor len(s) > 0 {\n\t\tif s[len(s)-1] <= ' ' {\n\t\t\ts = s[:len(s)-1]\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn s\n}", "func TrimSuffix(suffix, operand string) string { return strings.TrimSuffix(operand, suffix) }", "func Trim(str, chars string) string {\n\treturn LeftTrim(RightTrim(str, chars), chars)\n}", "func trim(strn string) string {\n\treturn strings.Trim(strn, \" \\t\\n\\r\")\n}", "func trim(name string) string {\n\treturn strings.TrimPrefix(name, Prefix)\n}", "func execTrimString(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := textproto.TrimString(args[0].(string))\n\tp.Ret(1, ret)\n}", "func Trim(str string) string {\n\treturn strings.TrimSpace(str)\n}", "func trimWhitespaces(value string) string {\n\treturn strings.Trim(value, \"\")\n}", "func Trim(str string) string {\n\treturn strings.Trim(str, \" \")\n}", "func ScanlineTrim() string {\n\treturn strings.TrimSpace(Scanline())\n}", "func trimS(s string) string {\r\n\treturn strings.TrimSpace(s)\r\n}", "func StrTrim(s string) string {\n\treturn strings.TrimSpace(s)\n}", "func TestTrim(t *testing.T) {\n\ttext := \"Hola Mundo TDA\"\n\tt.Logf(\"text:[%s]\", text)\n\tt.Logf(\"trim:[%s]\", utl.Trim(text))\n}", "func (s *Str) Trim(cutset string) *Str {\n\ts.val = strings.Trim(s.val, cutset)\n\treturn s\n}", "func Strip(str string) string {\n\treturn regexp.MustCompile(`^\\s+|\\s+$`).ReplaceAllString(str, \"\")\n}", "func (p *Parser) Trim(input string) string {\n\tt := strings.Replace(strings.Replace(input, \"<output>\", \"\", -1), \"</output>\", \"\", -1)\n\tt1 := strings.Replace(strings.Replace(t, \"<configuration-information>\", \"\", -1), \"</configuration-information>\", \"\", -1)\n\treturn strings.Replace(strings.Replace(t1, \"<configuration-output>\", \"\", -1), \"</configuration-output>\", \"\", -1)\n}", "func trim(s []byte) []byte {\n\ti := 0\n\tfor i < len(s) && (s[i] == ' ' || s[i] == '\\t') {\n\t\ti++\n\t}\n\tn := len(s)\n\tfor n > i && (s[n-1] == ' ' || s[n-1] == '\\t') {\n\t\tn--\n\t}\n\treturn s[i:n]\n}", "func trim(s []byte) []byte {\n\ti := 0\n\tfor i < len(s) && (s[i] == ' ' || s[i] == '\\t') {\n\t\ti++\n\t}\n\tn := len(s)\n\tfor n > i && (s[n-1] == ' ' || s[n-1] == '\\t') {\n\t\tn--\n\t}\n\treturn s[i:n]\n}", "func TrimSpaceAndQuotes(answer string) string {\n\ttext := strings.TrimSpace(answer)\n\tfor _, q := range quotes {\n\t\tif strings.HasPrefix(text, q) && strings.HasSuffix(text, q) {\n\t\t\treturn strings.TrimPrefix(strings.TrimSuffix(text, q), q)\n\t\t}\n\t}\n\treturn text\n}", "func Trim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_BOTH,\n\t}\n}", "func TrimSpace(ctx context.Context, t *mold.Transformer, v reflect.Value) error {\n\tv.Set(reflect.ValueOf(strings.TrimSpace(v.String())))\n\treturn nil\n}", "func TrimRedundantSpaces(text string) string {\n\ttext = spaceRegex.ReplaceAllString(text, \" \")\n\treturn strings.TrimSpace(text)\n}", "func RTrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_TRAILING,\n\t\tchars: chars,\n\t}\n}", "func RTrim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_TRAILING,\n\t}\n}", "func Trim(str string, characterMask ...string) string {\n\ttrimChars := DefaultTrimChars\n\tif len(characterMask) > 0 {\n\t\ttrimChars += characterMask[0]\n\t}\n\treturn strings.Trim(str, trimChars)\n}", "func (fn *formulaFuncs) TRIM(argsList *list.List) formulaArg {\n\tif argsList.Len() != 1 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"TRIM requires 1 argument\")\n\t}\n\treturn newStringFormulaArg(strings.TrimSpace(argsList.Front().Value.(formulaArg).Value()))\n}", "func TrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_BOTH,\n\t\tchars: chars,\n\t}\n}", "func trimRight(s string) string {\n\treturn strings.TrimRightFunc(s, unicode.IsSpace)\n}", "func RightTrim(str, chars string) string {\n\tif chars == \"\" {\n\t\treturn strings.TrimRightFunc(str, unicode.IsSpace)\n\t}\n\tr, _ := regexp.Compile(\"[\" + chars + \"]+$\")\n\treturn r.ReplaceAllString(str, \"\")\n}", "func (seg *Segmenter) CutTrim(str string, hmm ...bool) []string {\n\ts := seg.Cut(str, hmm...)\n\treturn seg.Trim(s)\n}", "func (seg *Segmenter) Trim(s []string) (r []string) {\n\tfor i := 0; i < len(s); i++ {\n\t\tsi := FilterSymbol(s[i])\n\t\tif !seg.NotStop && seg.IsStop(si) {\n\t\t\tsi = \"\"\n\t\t}\n\n\t\tif si != \"\" {\n\t\t\tr = append(r, si)\n\t\t}\n\t}\n\n\treturn\n}", "func trimDot(s string) string {\n\treturn strings.Trim(s, \".\")\n}", "func trim(s string, n int) string {\n\tif len(s) > n {\n\t\treturn s[:n]\n\t}\n\treturn s\n}", "func trim(s string, n int) string {\n\tif len(s) > n {\n\t\treturn s[:n]\n\t}\n\treturn s\n}", "func trimSpace(b []byte) string {\n\treturn string(bytes.TrimRight(b, \" \"))\n}", "func trimSpace(b []byte) string {\n\treturn string(bytes.TrimRight(b, \" \"))\n}", "func (g *GraphiteProvider) trimQuery(query string) string {\n\tspace := regexp.MustCompile(`\\s+`)\n\treturn space.ReplaceAllString(query, \" \")\n}", "func LeftTrim(str, chars string) string {\n\tif chars == \"\" {\n\t\treturn strings.TrimLeftFunc(str, unicode.IsSpace)\n\t}\n\tr, _ := regexp.Compile(\"^[\" + chars + \"]+\")\n\treturn r.ReplaceAllString(str, \"\")\n}", "func ToTrimmedString(in interface{}) string {\n\tif str, ok := in.(string); ok {\n\t\treturn strings.TrimSpace(str)\n\t}\n\treturn \"\"\n}", "func StringStrip(scope *Scope, input tf.Output) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"StringStrip\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func TrimSlash(str string) string {\n\treturn TrimByteSuffix(TrimBytePrefix(str, '/'), '/')\n}", "func Trim(cutset string) MapFunc {\n\treturn func(s string) string { return strings.Trim(s, cutset) }\n}", "func execTrimBytes(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := textproto.TrimBytes(args[0].([]byte))\n\tp.Ret(1, ret)\n}", "func trimSpaceAndJoin(args []string, sep string) string {\n\ttrimmedArgs := make([]string, len(args))\n\tfor i, arg := range args {\n\t\ttrimmedArgs[i] = strings.TrimSpace(arg)\n\t}\n\treturn strings.TrimSpace(strings.Join(trimmedArgs, sep))\n}", "func trimChar(s string, r byte) string {\n\tsz := len(s)\n\n\tif sz > 0 && s[sz-1] == r {\n\t\ts = s[:sz-1]\n\t}\n\tsz = len(s)\n\tif sz > 0 && s[0] == r {\n\t\ts = s[1:sz]\n\t}\n\treturn s\n}", "func trimChar(s string, r byte) string {\n\tsz := len(s)\n\n\tif sz > 0 && s[sz-1] == r {\n\t\ts = s[:sz-1]\n\t}\n\tsz = len(s)\n\tif sz > 0 && s[0] == r {\n\t\ts = s[1:sz]\n\t}\n\treturn s\n}", "func (seg *Segmenter) PosTrim(str string, search bool, pos ...string) []SegPos {\n\tp := seg.Pos(str, search)\n\tp = seg.TrimWithPos(p, pos...)\n\treturn seg.TrimPos(p)\n}", "func trimBetween(s string, start string, end string) string {\n\tif idx := strings.Index(s, start); idx != -1 {\n\t\ts = s[idx+len(start):]\n\t}\n\n\tif idx := strings.Index(s, end); idx != -1 {\n\t\ts = s[:idx]\n\t}\n\n\treturn s\n}", "func TrimAllSpaceAndNewline(input string) string {\n\toutput := input\n\tfor _, f := range []string{\"\\n\", \"\\t\", \" \"} {\n\t\toutput = strings.Replace(output, f, \"\", -1)\n\t}\n\n\treturn output\n}", "func findtrimmedexpression(s string) string {\r\n\tresult := \"\"\r\n\tfirst := false\r\n\tfirst_idx := 0\r\n\tidx := 0\r\n\tsecond_idx := 0\r\n\t//get correct indices\r\n\tfor idx < len(s) {\r\n\t\tif (string(s[idx]) == \"\\\"\" && (first == false)) || (string(s[idx]) == \"'\" && (first == false)) {\r\n\t\t\tfirst_idx = idx\r\n\t\t\tfirst = true\r\n\t\t} else if string(s[idx]) == \"\\\"\" || string(s[idx]) == \"'\" {\r\n\t\t\tsecond_idx = idx\r\n\t\t}\r\n\t\tidx = idx + 1\r\n\t}\r\n\tidx = first_idx + 1\r\n\tfor idx < second_idx {\r\n\t\tresult = result + string(s[idx])\r\n\t\tidx = idx + 1\r\n\t}\r\n\treturn result\r\n}", "func stripWhiteSpace(str string) string {\n\treturn strings.ReplaceAll(str, \" \", \"\")\n}", "func TrimAll(s string) string {\n\treturn strings.Join(strings.Fields(s), \" \")\n}", "func DeleteEmptyAndTrim(s []string) []string {\n\tvar r []string\n\tfor _, str := range s {\n\t\tstr = strings.TrimSpace(str)\n\t\tif str != \"\" {\n\t\t\tr = append(r, str)\n\t\t}\n\t}\n\treturn r\n}", "func removeWhitespaces(value string) string {\n\treturn strings.ReplaceAll(value, \" \", \"\")\n}", "func trimLeadingWhiteSpace(buf string) string {\n\treturn strings.TrimLeft(buf, \" \")\n}", "func trimQuotes(str string) string {\n\treturn strings.TrimFunc(str, func(char rune) bool {\n\t\tswitch char {\n\t\tcase '\"', '\\'', '`':\n\t\t\treturn true\n\t\t}\n\t\treturn false\n\t})\n}", "func (h *Helpers) Strip(s string) string {\n\tre1 := regexp.MustCompile(stripEmptyRegex)\n\tre2 := regexp.MustCompile(stripEdgesRegex)\n\ts = re1.ReplaceAllString(s, \"\")\n\ts = re2.ReplaceAllString(s, \"\")\n\treturn s\n}", "func TrimFunc(f func(rune) bool) MapFunc {\n\treturn func(s string) string { return strings.TrimFunc(s, f) }\n}", "func splitTrim(s string, sep string) []string {\n\tsplitItems := strings.Split(s, sep)\n\ttrimItems := make([]string, 0, len(splitItems))\n\tfor _, item := range splitItems {\n\t\tif item = strings.TrimSpace(item); item != \"\" {\n\t\t\ttrimItems = append(trimItems, item)\n\t\t}\n\t}\n\treturn trimItems\n}", "func trimLast(text string) string {\n\ttextLen := len(text)\n\tif textLen == 0 {\n\t\treturn text\n\t}\n\treturn text[:textLen-1]\n}", "func trimStringN(c *C.char, l C.int) string {\n\tvar rc string\n\ts := C.GoStringN(c, l)\n\ti := strings.IndexByte(s, 0)\n\tif i == -1 {\n\t\trc = s\n\t} else {\n\t\trc = s[0:i]\n\t}\n\treturn strings.TrimSpace(rc)\n}", "func trimWS(input string) string {\n\n\tvar out strings.Builder\n\tfor _, v := range input {\n\t\tif !(v == '\\u0009' || v == '\\u0020' || v == '\\u000A' || v == '\\u000D' || v == ',') {\n\t\t\tout.WriteRune(v)\n\t\t}\n\t}\n\treturn out.String()\n\n}", "func trimTrailingWhiteSpace(s []byte) []byte {\n\tfor (len(s) > 0) && ((s[len(s)-1] == ' ') || (s[len(s)-1] == '\\t')) {\n\t\ts = s[:len(s)-1]\n\t}\n\treturn s\n}", "func trimTrailingWhitespace(details string) string {\n\treturn strings.TrimSuffix(details, \" \")\n}", "func signV4TrimAll(input string) string {\n\t// Compress adjacent spaces (a space is determined by\n\t// unicode.IsSpace() internally here) to one space and return\n\treturn strings.Join(strings.Fields(input), \" \")\n}", "func (e EmbeddedString) Trim() string {\n\ts := string(e)\n\tif strings.HasPrefix(s, \"// Copyright \") {\n\t\tif i := strings.Index(s, \"\\n\\n\"); i >= 0 {\n\t\t\ts = s[i+2:]\n\t\t}\n\t}\n\treturn s\n}", "func LTrim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_LEADING,\n\t}\n}", "func (s *Str) TrimSpaces() *Str {\n\ts.val = strings.TrimSpace(s.val)\n\treturn s\n}", "func stripSlash(str string) string {\n\treturn strings.TrimRight(strings.TrimLeft(str, \"/\"), \"/\")\n}", "func IdentityTrim(ident string) string {\n\tif len(ident) > 0 {\n\t\tif ident[0] == '`' || ident[0] == '[' {\n\t\t\tident = ident[1:]\n\t\t}\n\t\tif len(ident) > 0 {\n\t\t\tif ident[len(ident)-1] == '`' || ident[len(ident)-1] == ']' {\n\t\t\t\tident = ident[0 : len(ident)-1]\n\t\t\t}\n\t\t}\n\t}\n\treturn ident\n}", "func (su StringUtility) TrimEndMultiple(targetStr string, trimChar rune) (rStr string, err error) {\n\n\tif targetStr == \"\" {\n\t\terr = errors.New(\"Empty targetStr\")\n\t\treturn\n\t}\n\n\tfStr := []rune(targetStr)\n\tlenTargetStr := len(fStr)\n\toutputStr := make([]rune, lenTargetStr)\n\tlenTargetStr--\n\tidx := lenTargetStr\n\tfoundFirstChar := false\n\n\tfor i := lenTargetStr; i >= 0; i-- {\n\n\t\tif !foundFirstChar && fStr[i] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tif i > 0 && fStr[i] == trimChar && fStr[i-1] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tif i == 0 && fStr[i] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tfoundFirstChar = true\n\t\toutputStr[idx] = fStr[i]\n\t\tidx--\n\t}\n\n\tif idx != lenTargetStr {\n\t\tidx++\n\t}\n\n\tif outputStr[idx] == trimChar {\n\t\tidx++\n\t}\n\n\tresult := string(outputStr[idx:])\n\n\treturn result, nil\n\n}", "func trimToPrefix(str, prefix string) string {\n\ti := strings.Index(str, prefix)\n\tif i < 0 {\n\t\treturn str\n\t}\n\treturn str[i:]\n}", "func (s *Stringish) TrimSpaces() *Stringish {\n\ts.str = strings.TrimSpace(s.str)\n\treturn s\n}", "func (seg *Segmenter) TrimSymbol(s []string) (r []string) {\n\tfor i := 0; i < len(s); i++ {\n\t\tsi := FilterSymbol(s[i])\n\t\tif si != \"\" {\n\t\t\tr = append(r, si)\n\t\t}\n\t}\n\n\treturn\n}", "func TrimTrailingSpaces(text string) string {\n\tparts := strings.Split(text, \"\\n\")\n\tfor i := range parts {\n\t\tparts[i] = strings.TrimRightFunc(parts[i], func(r rune) bool {\n\t\t\treturn unicode.IsSpace(r)\n\t\t})\n\n\t}\n\n\treturn strings.Join(parts, \"\\n\")\n}", "func (s *Str) TrimLeft(cutset string) *Str {\n\ts.val = strings.TrimLeft(s.val, cutset)\n\treturn s\n}", "func TrimLineSpaces(str string) string {\n\treturn TrimLineSpaces2(str, \"\")\n}", "func TrimTrailingSpaces(p []byte) []byte {\n\tfor i := len(p) - 1; i >= 0; i-- {\n\t\tif p[i] != 0x20 && p[i] != '\\n' && p[i] != '\\t' {\n\t\t\treturn p[:i+1]\n\t\t}\n\t}\n\t// it was all spaces\n\treturn p[:0]\n}", "func trimQuotes(buf string) string {\n\tbuflen := len(buf)\n\tif buflen == 0 {\n\t\treturn buf\n\t}\n\tif buf[0:1] == \"\\\"\" && buf[buflen-1:buflen] == \"\\\"\" {\n\t\treturn buf[1 : buflen-1]\n\t}\n\treturn buf\n}", "func stripTrailingSpace(str string) string {\n\tbuf := bytes.NewBuffer(nil)\n\n\tscan := bufio.NewScanner(strings.NewReader(str))\n\tfor scan.Scan() {\n\t\tbuf.WriteString(strings.TrimRight(scan.Text(), \" \\t\\r\\n\"))\n\t\tbuf.WriteString(\"\\n\")\n\t}\n\treturn buf.String()\n}", "func LTrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_LEADING,\n\t\tchars: chars,\n\t}\n}", "func (chars *Chars) TrimLength() uint16 {\n\tif chars.trimLengthKnown {\n\t\treturn chars.trimLength\n\t}\n\tchars.trimLengthKnown = true\n\tvar i int\n\tlen := chars.Length()\n\tfor i = len - 1; i >= 0; i-- {\n\t\tchar := chars.Get(i)\n\t\tif !unicode.IsSpace(char) {\n\t\t\tbreak\n\t\t}\n\t}\n\t// Completely empty\n\tif i < 0 {\n\t\treturn 0\n\t}\n\n\tvar j int\n\tfor j = 0; j < len; j++ {\n\t\tchar := chars.Get(j)\n\t\tif !unicode.IsSpace(char) {\n\t\t\tbreak\n\t\t}\n\t}\n\tchars.trimLength = AsUint16(i - j + 1)\n\treturn chars.trimLength\n}", "func TrimSubstrings(haystack string, needles []string) (result string) {\n\tfor _, item := range needles {\n\t\thaystack = strings.ReplaceAll(haystack, item, \"\")\n\t}\n\treturn haystack\n}", "func trimOutPrefix(str, prefix string) string {\n\ti := strings.Index(str, prefix)\n\tif i < 0 {\n\t\treturn str\n\t}\n\treturn str[i+len(prefix):]\n}", "func trimLeftSlash(name string) string {\n\treturn strings.TrimPrefix(name, \"/\")\n}", "func CutTrimSpace(s, sep string) (before, after string, found bool) {\n\tif i := strings.Index(s, sep); i >= 0 {\n\t\treturn strings.TrimSpace(s[:i]), strings.TrimSpace(s[i+len(sep):]), true\n\t}\n\treturn s, \"\", false\n}", "func (t *standard) trim() {\n\n\tt.trimIfAdded()\n\tt.trimIfSubtracted()\n\n\tif t.Day() < 0 {\n\t\tpanic(ErrNegativeTime)\n\t}\n\n}", "func trimLeadingWhiteSpace(s []byte) []byte {\n\tfor (len(s) > 0) && ((s[0] == ' ') || (s[0] == '\\t')) {\n\t\ts = s[1:]\n\t}\n\treturn s\n}", "func TrimString(s string) string {\n\treturn replacer.Replace(s)\n}", "func TrimLeadingSpaces(p []byte) []byte {\n\tfor i := 0; i < len(p); i++ {\n\t\tif p[i] != 0x20 && p[i] != '\\t' {\n\t\t\treturn p[i:]\n\t\t}\n\t}\n\t// it was all spaces\n\treturn p[:0]\n}", "func trimCut(label []byte, maxlen int, left, right string) []byte {\n\ttrim := bytes.TrimLeft(label, left)\n\tsize := min(len(trim), maxlen)\n\thead := bytes.TrimRight(trim[:size], right)\n\tif len(head) == size {\n\t\treturn head\n\t}\n\ttail := bytes.TrimLeft(trim[size:], right)\n\tif len(tail) > 0 {\n\t\treturn append(head, tail[:min(len(tail), size-len(head))]...)\n\t}\n\treturn head\n}", "func Strip(s string) string {\n\treturn strip.ReplaceAllString(s, \"\")\n}", "func Strip(s string) string {\n\treturn strip.ReplaceAllString(s, \"\")\n}", "func trimmed(bs []byte) []byte {\n\tfor i, b := range bs {\n\t\tif b == 0x00 {\n\t\t\treturn bs[:i]\n\t\t}\n\t}\n\treturn bs\n}" ]
[ "0.8349016", "0.7741901", "0.7373609", "0.7302248", "0.7040318", "0.69703877", "0.69395834", "0.6924436", "0.6859497", "0.6822096", "0.6818247", "0.67978364", "0.6779929", "0.67027116", "0.66982675", "0.6682328", "0.66658735", "0.6663513", "0.6622947", "0.6596055", "0.64982265", "0.64438576", "0.6440511", "0.6440511", "0.6433716", "0.633796", "0.6319466", "0.63175374", "0.6253703", "0.62513", "0.62408894", "0.62094015", "0.6198147", "0.6162491", "0.6160594", "0.61017865", "0.6090275", "0.6075992", "0.6072917", "0.6072917", "0.60724616", "0.60724616", "0.6057541", "0.5969238", "0.59650207", "0.59449774", "0.5894855", "0.5873381", "0.586691", "0.5862617", "0.58533037", "0.58533037", "0.5852135", "0.5852112", "0.5838989", "0.5833387", "0.5830001", "0.582888", "0.5821153", "0.5817075", "0.5816747", "0.5787124", "0.57690996", "0.5755905", "0.575267", "0.575228", "0.5736498", "0.57333237", "0.5709742", "0.5705321", "0.56979454", "0.5696091", "0.5691349", "0.5689949", "0.5686432", "0.56860334", "0.5671366", "0.5666808", "0.56638795", "0.56622213", "0.56619054", "0.56492436", "0.56473696", "0.56454605", "0.5640826", "0.56378895", "0.5630536", "0.56284356", "0.5621226", "0.5607978", "0.5591701", "0.55812806", "0.5576389", "0.5575078", "0.55714905", "0.5570915", "0.55695313", "0.556696", "0.556696", "0.552822" ]
0.89910746
0
TrimLeft uses strings.TrimLeft to remove any occurrences of chars from the beginning of operand.
func TrimLeft(chars string, operand string) string { return strings.TrimLeft(operand, chars) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Str) TrimLeft(cutset string) *Str {\n\ts.val = strings.TrimLeft(s.val, cutset)\n\treturn s\n}", "func LeftTrim(str, chars string) string {\n\tif chars == \"\" {\n\t\treturn strings.TrimLeftFunc(str, unicode.IsSpace)\n\t}\n\tr, _ := regexp.Compile(\"^[\" + chars + \"]+\")\n\treturn r.ReplaceAllString(str, \"\")\n}", "func Trim(chars string, operand string) string { return strings.Trim(operand, chars) }", "func TrimPrefix(prefix, operand string) string { return strings.TrimPrefix(operand, prefix) }", "func trimLeft(r *syntax.Regexp) (bool, *syntax.Regexp) {\n\tif eqPrefixAnyRegex(r, patDotStar, patNullBeginDotStar) {\n\t\ttmp := *r\n\t\ttmp.Sub = tmp.Sub[1:]\n\t\treturn true, &tmp\n\t}\n\n\treturn false, r\n}", "func TrimLeftFunc(f func(rune) bool) MapFunc {\n\treturn func(s string) string { return strings.TrimLeftFunc(s, f) }\n}", "func TrimLeft(cutset string) MapFunc {\n\treturn func(s string) string { return strings.TrimLeft(s, cutset) }\n}", "func TrimLeftChar(s string) string {\n\tfor i := range s {\n\t\tif i > 0 {\n\t\t\treturn s[i:]\n\t\t}\n\t}\n\treturn s[:0]\n}", "func TrimSpace(operand string) string { return strings.TrimSpace(operand) }", "func trimLeftSlash(name string) string {\n\treturn strings.TrimPrefix(name, \"/\")\n}", "func TrimRight(cutset string, operand string) string { return strings.TrimRight(operand, cutset) }", "func (n *nodeHeader) leftTrimPrefix(l uint16) {\n\tif l < 1 {\n\t\treturn\n\t}\n\tpLen, pBytes := n.prefixFields()\n\tif l > *pLen {\n\t\tl = *pLen\n\t}\n\tnewLen := *pLen - uint16(l)\n\tcopy(pBytes[0:newLen], pBytes[l:*pLen])\n\t*pLen = newLen\n}", "func filterTrim(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.TrimSpace(stick.CoerceString(val))\n}", "func Left(s string, n int) string {\n\tif n < 1 {\n\t\treturn \"\"\n\t}\n\trunes := []rune(s)\n\tif n >= len(runes) {\n\t\treturn s\n\t}\n\n\treturn string(runes[:n])\n\n}", "func trimLeadingWhiteSpace(buf string) string {\n\treturn strings.TrimLeft(buf, \" \")\n}", "func Left(str string, size int) string {\n\tif str == \"\" || size < 0 {\n\t\treturn \"\"\n\t}\n\tif len(str) <= size {\n\t\treturn str\n\t}\n\treturn str[0:size]\n}", "func LTrim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_LEADING,\n\t}\n}", "func StripStart(str string) string {\n\treturn regexp.MustCompile(`^\\s+`).ReplaceAllString(str, \"\")\n}", "func trim_string_before_s(s string, x string) (r string) {\n\tif idx := strings.LastIndex(s, x); idx != -1 {\n\t\treturn s[idx+1:]\n\t}\n\treturn s\n}", "func trimToPrefix(str, prefix string) string {\n\ti := strings.Index(str, prefix)\n\tif i < 0 {\n\t\treturn str\n\t}\n\treturn str[i:]\n}", "func trim(s string) string {\n\tfor len(s) > 0 {\n\t\tif s[0] <= ' ' {\n\t\t\ts = s[1:]\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tfor len(s) > 0 {\n\t\tif s[len(s)-1] <= ' ' {\n\t\t\ts = s[:len(s)-1]\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn s\n}", "func TestTrim(t *testing.T) {\n\ttext := \"Hola Mundo TDA\"\n\tt.Logf(\"text:[%s]\", text)\n\tt.Logf(\"trim:[%s]\", utl.Trim(text))\n}", "func trim(name string) string {\n\treturn strings.TrimPrefix(name, Prefix)\n}", "func execTrimString(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := textproto.TrimString(args[0].(string))\n\tp.Ret(1, ret)\n}", "func LTrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_LEADING,\n\t\tchars: chars,\n\t}\n}", "func Left(text string, size int) string {\n\tspaces := size - Length(text)\n\tif spaces <= 0 {\n\t\treturn text\n\t}\n\n\tvar buffer bytes.Buffer\n\tbuffer.WriteString(text)\n\n\tfor i := 0; i < spaces; i++ {\n\t\tbuffer.WriteString(space)\n\t}\n\treturn buffer.String()\n}", "func StripLeftTabs(s string) string { return cptNC.stripLeftTabs(s) }", "func leftTruncate(val string, n int) string {\n if ( len(val) > n ){\n return val[len(val)-n:]\n } else {\n return val\n }\n}", "func TrimSuffix(suffix, operand string) string { return strings.TrimSuffix(operand, suffix) }", "func Trim(str, chars string) string {\n\treturn LeftTrim(RightTrim(str, chars), chars)\n}", "func (s *Stringish) TrimPrefix(prefix string) *Stringish {\n\ts.str = strings.TrimPrefix(s.str, prefix)\n\treturn s\n}", "func trimS(s string) string {\r\n\treturn strings.TrimSpace(s)\r\n}", "func TrimLeadingSpaces(p []byte) []byte {\n\tfor i := 0; i < len(p); i++ {\n\t\tif p[i] != 0x20 && p[i] != '\\t' {\n\t\t\treturn p[i:]\n\t\t}\n\t}\n\t// it was all spaces\n\treturn p[:0]\n}", "func (b *Bar) TrimLeftSpace() *Bar {\n\tif isClosed(b.done) {\n\t\treturn b\n\t}\n\tb.trimLeftCh <- true\n\treturn b\n}", "func Trim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_BOTH,\n\t}\n}", "func StripLeftTabsOnly(s string) string { return cptNC.stripLeftTabsOnly(s) }", "func LeftRight(val string) (string, string, bool) {\n\tif len(val) < 2 {\n\t\treturn \"\", val, false\n\t}\n\tswitch by := val[0]; by {\n\tcase '`':\n\t\tvals := strings.Split(val, \"`.`\")\n\t\tif len(vals) == 1 {\n\t\t\treturn \"\", IdentityTrim(val), false\n\t\t} else if len(vals) == 2 {\n\t\t\treturn IdentityTrim(vals[0]), IdentityTrim(vals[1]), true\n\t\t}\n\t\t// wat, no idea what this is\n\t\treturn \"\", val, false\n\tcase '[':\n\t\tvals := strings.Split(val, \"].[\")\n\t\tif len(vals) == 1 {\n\t\t\treturn \"\", IdentityTrim(val), false\n\t\t} else if len(vals) == 2 {\n\t\t\treturn IdentityTrim(vals[0]), IdentityTrim(vals[1]), true\n\t\t}\n\t\t// wat, no idea what this is\n\t\treturn \"\", val, false\n\tdefault:\n\t\tvals := strings.SplitN(val, \".\", 2)\n\t\tif len(vals) == 1 {\n\t\t\treturn \"\", val, false\n\t\t} else if len(vals) == 2 {\n\t\t\treturn IdentityTrim(vals[0]), IdentityTrim(vals[1]), true\n\t\t}\n\t}\n\n\treturn \"\", val, false\n}", "func trimWhitespaces(value string) string {\n\treturn strings.Trim(value, \"\")\n}", "func RTrim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_TRAILING,\n\t}\n}", "func (fn *formulaFuncs) TRIM(argsList *list.List) formulaArg {\n\tif argsList.Len() != 1 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"TRIM requires 1 argument\")\n\t}\n\treturn newStringFormulaArg(strings.TrimSpace(argsList.Front().Value.(formulaArg).Value()))\n}", "func trimCut(label []byte, maxlen int, left, right string) []byte {\n\ttrim := bytes.TrimLeft(label, left)\n\tsize := min(len(trim), maxlen)\n\thead := bytes.TrimRight(trim[:size], right)\n\tif len(head) == size {\n\t\treturn head\n\t}\n\ttail := bytes.TrimLeft(trim[size:], right)\n\tif len(tail) > 0 {\n\t\treturn append(head, tail[:min(len(tail), size-len(head))]...)\n\t}\n\treturn head\n}", "func (s *Stringish) TrimPrefixSpaces() *Stringish {\n\treturn s.TrimPrefix(\" \")\n}", "func (s *Str) Trim(cutset string) *Str {\n\ts.val = strings.Trim(s.val, cutset)\n\treturn s\n}", "func StrTrim(s string) string {\n\treturn strings.TrimSpace(s)\n}", "func trimLeadingWhiteSpace(s []byte) []byte {\n\tfor (len(s) > 0) && ((s[0] == ' ') || (s[0] == '\\t')) {\n\t\ts = s[1:]\n\t}\n\treturn s\n}", "func PadLeft(str string, padStr string, padLen int) string {\n\treturn buildPadStr(str, padStr, padLen, true, false)\n}", "func MaskLeft(s string) string {\n\trs := []rune(s)\n\tfor i := 0; i < len(rs)-4; i++ {\n\t\trs[i] = 'X'\n\t}\n\treturn string(rs)\n}", "func Strip(str string) string {\n\treturn regexp.MustCompile(`^\\s+|\\s+$`).ReplaceAllString(str, \"\")\n}", "func ScanlineTrim() string {\n\treturn strings.TrimSpace(Scanline())\n}", "func Trim(text string) string {\n\treturn trimRx.FindStringSubmatch(text)[1]\n}", "func trimRight(s string) string {\n\treturn strings.TrimRightFunc(s, unicode.IsSpace)\n}", "func PadLeft(s string, padStr string, totalLen int) string {\n\tvar padCountInt int\n\tpadCountInt = 1 + ((totalLen - len(padStr)) / len(padStr))\n\tvar retStr = strings.Repeat(padStr, padCountInt) + s\n\treturn retStr[(len(retStr) - totalLen):]\n}", "func TrimPrefix(prefix string) MapFunc {\n\treturn func(s string) string { return strings.TrimPrefix(s, prefix) }\n}", "func (fn *formulaFuncs) LEFT(argsList *list.List) formulaArg {\n\treturn fn.leftRight(\"LEFT\", argsList)\n}", "func trimOutPrefix(str, prefix string) string {\n\ti := strings.Index(str, prefix)\n\tif i < 0 {\n\t\treturn str\n\t}\n\treturn str[i+len(prefix):]\n}", "func (sopsTxtJustify TextJustify) Left() TextJustify {\n\n\tlockStrOpsTextJustify.Lock()\n\n\tdefer lockStrOpsTextJustify.Unlock()\n\n\treturn TextJustify(1)\n}", "func RTrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_TRAILING,\n\t\tchars: chars,\n\t}\n}", "func RightTrim(str, chars string) string {\n\tif chars == \"\" {\n\t\treturn strings.TrimRightFunc(str, unicode.IsSpace)\n\t}\n\tr, _ := regexp.Compile(\"[\" + chars + \"]+$\")\n\treturn r.ReplaceAllString(str, \"\")\n}", "func Trim(str string) string {\n\treturn strings.TrimSpace(str)\n}", "func TrimRedundantSpaces(text string) string {\n\ttext = spaceRegex.ReplaceAllString(text, \" \")\n\treturn strings.TrimSpace(text)\n}", "func PadLeft(s string, padStr string, lenStr int) string {\n\tvar padCount int\n\tpadCount = I.MaxOf(lenStr-len(s), 0)\n\treturn strings.Repeat(padStr, padCount) + s\n}", "func trimBetween(s string, start string, end string) string {\n\tif idx := strings.Index(s, start); idx != -1 {\n\t\ts = s[idx+len(start):]\n\t}\n\n\tif idx := strings.Index(s, end); idx != -1 {\n\t\ts = s[:idx]\n\t}\n\n\treturn s\n}", "func TrimFunc(f func(rune) bool) MapFunc {\n\treturn func(s string) string { return strings.TrimFunc(s, f) }\n}", "func (p Prefix) LeftShift(word string) string {\r\n\telem := p[len(p)-1]\r\n\tcopy(p[1:], p[:len(p)-1])\r\n\tp[0] = word\r\n\treturn elem\r\n}", "func (seg *Segmenter) PosTrim(str string, search bool, pos ...string) []SegPos {\n\tp := seg.Pos(str, search)\n\tp = seg.TrimWithPos(p, pos...)\n\treturn seg.TrimPos(p)\n}", "func trimLeading(s string, char string) (suffix string, count int) {\n\tcount = 0\n\tprevChar := char\n\tfor i := 0; i < len(s) && prevChar == char; i++ {\n\t\tprevChar = string(s[i])\n\t\tif prevChar == char {\n\t\t\tcount++\n\t\t}\n\t}\n\tsuffix = strings.TrimLeft(s, char)\n\treturn suffix, count\n}", "func leftPad(s string, padStr string, pLen int) string {\n\tr := pLen - len(s)\n\tif r > 0 {\n\t\treturn strings.Repeat(padStr, pLen-len(s)) + s\n\t}\n\treturn s\n}", "func TrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_BOTH,\n\t\tchars: chars,\n\t}\n}", "func trimStringN(c *C.char, l C.int) string {\n\tvar rc string\n\ts := C.GoStringN(c, l)\n\ti := strings.IndexByte(s, 0)\n\tif i == -1 {\n\t\trc = s\n\t} else {\n\t\trc = s[0:i]\n\t}\n\treturn strings.TrimSpace(rc)\n}", "func padLeft(str string, max int) string {\n\tpad := max - len(str)\n\tb := strings.Builder{}\n\tb.WriteString(strings.Repeat(\" \", pad))\n\tb.WriteString(str)\n\treturn b.String()\n}", "func trim(strn string) string {\n\treturn strings.Trim(strn, \" \\t\\n\\r\")\n}", "func stripPrefix(s string) (string, bool) {\n\ts = strings.TrimSpace(s)\n\tres := strings.TrimPrefix(s, \"What is\")\n\tif res == s {\n\t\treturn \"\", false\n\t}\n\treturn res, true\n}", "func RemoveStart(str string, remove string) string {\n\tif IsEmpty(str) || IsEmpty(remove) {\n\t\treturn str\n\t}\n\tif StartsWith(str, remove) {\n\t\treturn str[len(remove)+1:]\n\t}\n\treturn str\n}", "func (s *Str) TrimRight(cutset string) *Str {\n\ts.val = strings.TrimRight(s.val, cutset)\n\treturn s\n}", "func Trim(str string) string {\n\treturn strings.Trim(str, \" \")\n}", "func PadLeft(src, padding string, length int) (string, error) {\n\n\tif length <= len(src) {\n\t\treturn \"\", errors.New(\"Target length must be greater than the length\" +\n\t\t\t\" of the original string.\")\n\t}\n\n\tif len(padding) != 1 {\n\t\treturn \"\", errors.New(\"Padding must be a single character.\")\n\t}\n\n\treturn strings.Repeat(padding, length-len(src)) + src, nil\n\n}", "func Trim(cutset string) MapFunc {\n\treturn func(s string) string { return strings.Trim(s, cutset) }\n}", "func trim(s string, n int) string {\n\tif len(s) > n {\n\t\treturn s[:n]\n\t}\n\treturn s\n}", "func trim(s string, n int) string {\n\tif len(s) > n {\n\t\treturn s[:n]\n\t}\n\treturn s\n}", "func (o DashboardSpacingOutput) Left() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v DashboardSpacing) *string { return v.Left }).(pulumi.StringPtrOutput)\n}", "func trim(s []byte) []byte {\n\ti := 0\n\tfor i < len(s) && (s[i] == ' ' || s[i] == '\\t') {\n\t\ti++\n\t}\n\tn := len(s)\n\tfor n > i && (s[n-1] == ' ' || s[n-1] == '\\t') {\n\t\tn--\n\t}\n\treturn s[i:n]\n}", "func trim(s []byte) []byte {\n\ti := 0\n\tfor i < len(s) && (s[i] == ' ' || s[i] == '\\t') {\n\t\ti++\n\t}\n\tn := len(s)\n\tfor n > i && (s[n-1] == ' ' || s[n-1] == '\\t') {\n\t\tn--\n\t}\n\treturn s[i:n]\n}", "func (a Amount) StringBefore() string {\n\tparts := strings.Split(a.String(), \" \")\n\tif len(parts) > 1 {\n\t\treturn parts[1] + parts[0]\n\t}\n\treturn parts[0]\n}", "func DeleteEmptyAndTrim(s []string) []string {\n\tvar r []string\n\tfor _, str := range s {\n\t\tstr = strings.TrimSpace(str)\n\t\tif str != \"\" {\n\t\t\tr = append(r, str)\n\t\t}\n\t}\n\treturn r\n}", "func stripSlash(str string) string {\n\treturn strings.TrimRight(strings.TrimLeft(str, \"/\"), \"/\")\n}", "func TrimSubstrings(haystack string, needles []string) (result string) {\n\tfor _, item := range needles {\n\t\thaystack = strings.ReplaceAll(haystack, item, \"\")\n\t}\n\treturn haystack\n}", "func LeftPad(s string, padStr string, overallLen int) string {\n\tvar padCountInt = 1 + ((overallLen - len(padStr)) / len(padStr))\n\tvar retStr = strings.Repeat(padStr, padCountInt) + s\n\treturn retStr[(len(retStr) - overallLen):]\n}", "func IdentityTrim(ident string) string {\n\tif len(ident) > 0 {\n\t\tif ident[0] == '`' || ident[0] == '[' {\n\t\t\tident = ident[1:]\n\t\t}\n\t\tif len(ident) > 0 {\n\t\t\tif ident[len(ident)-1] == '`' || ident[len(ident)-1] == ']' {\n\t\t\t\tident = ident[0 : len(ident)-1]\n\t\t\t}\n\t\t}\n\t}\n\treturn ident\n}", "func trimSpaceAndJoin(args []string, sep string) string {\n\ttrimmedArgs := make([]string, len(args))\n\tfor i, arg := range args {\n\t\ttrimmedArgs[i] = strings.TrimSpace(arg)\n\t}\n\treturn strings.TrimSpace(strings.Join(trimmedArgs, sep))\n}", "func (o DashboardSpacingPtrOutput) Left() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *DashboardSpacing) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Left\n\t}).(pulumi.StringPtrOutput)\n}", "func CutTrimSpace(s, sep string) (before, after string, found bool) {\n\tif i := strings.Index(s, sep); i >= 0 {\n\t\treturn strings.TrimSpace(s[:i]), strings.TrimSpace(s[i+len(sep):]), true\n\t}\n\treturn s, \"\", false\n}", "func execTrimBytes(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := textproto.TrimBytes(args[0].([]byte))\n\tp.Ret(1, ret)\n}", "func Trim(str string, characterMask ...string) string {\n\ttrimChars := DefaultTrimChars\n\tif len(characterMask) > 0 {\n\t\ttrimChars += characterMask[0]\n\t}\n\treturn strings.Trim(str, trimChars)\n}", "func trimTitle(title string) string {\n\treturn strings.TrimSpace(title)\n}", "func trimPrefix(command, prefix string) (string, bool) {\n\tvar c string\n\tif prefix == \"\" && command != \"\" {\n\t\treturn command, true\n\t}\n\tif command == \"\" {\n\t\treturn command, false\n\t}\n\tif c = strings.TrimPrefix(command, prefix); c == command {\n\t\treturn \"\", false\n\t}\n\n\treturn c, len(c) != 0 // if command was \"[prefix]\" and it's trimmed into \"\" it should be false\n\n}", "func trimLeadingWhiteSpaceAndNewLines(s []byte) []byte {\n\tfor (len(s) > 0) && ((s[0] == ' ') || (s[0] == '\\t') || (s[0] == '\\n')) {\n\t\ts = s[1:]\n\t}\n\treturn s\n}", "func trimLast(text string) string {\n\ttextLen := len(text)\n\tif textLen == 0 {\n\t\treturn text\n\t}\n\treturn text[:textLen-1]\n}", "func trimTreePrefix(name string, n int) string {\n\ti := 0\n\tfor ; i < len(name) && n > 0; i++ {\n\t\tif name[i] == '/' {\n\t\t\tn--\n\t\t}\n\t}\n\treturn name[i:]\n}", "func SplitAndTrim(s string, separator string) ([]string, error) {\n\t// Trims the spaces and then splits\n\ttrimmed := strings.TrimSpace(s)\n\tsplit := strings.Split(trimmed, separator)\n\tcleanSplit := make([]string, len(split))\n\tfor i, val := range split {\n\t\tcleanSplit[i] = strings.TrimSpace(val)\n\t}\n\n\treturn cleanSplit, nil\n}", "func (seg *Segmenter) PosTrimStr(str string, search bool, pos ...string) string {\n\tpa := seg.PosTrimArr(str, search, pos...)\n\treturn seg.CutStr(pa)\n}" ]
[ "0.7660647", "0.7622056", "0.7534656", "0.74435884", "0.7333051", "0.72390336", "0.72058535", "0.6937747", "0.6934556", "0.67538977", "0.666173", "0.63456863", "0.6317367", "0.6309685", "0.62236136", "0.6174375", "0.6161742", "0.6141191", "0.6040537", "0.599031", "0.5936853", "0.59277225", "0.59178746", "0.5863252", "0.58588856", "0.58431154", "0.58064663", "0.58021706", "0.5793688", "0.57870245", "0.57688683", "0.5735736", "0.56965905", "0.56898", "0.56641376", "0.5647645", "0.5640076", "0.5627152", "0.56239533", "0.56214505", "0.5619416", "0.56125885", "0.56022084", "0.56005096", "0.559849", "0.55628556", "0.5549894", "0.55482125", "0.5546481", "0.55446655", "0.55397886", "0.553095", "0.55091226", "0.54973567", "0.5488336", "0.54883254", "0.5461102", "0.5449867", "0.54439193", "0.5435526", "0.54101807", "0.5389301", "0.53873163", "0.53812194", "0.53746235", "0.536415", "0.536301", "0.5362173", "0.5351415", "0.5325329", "0.5322123", "0.53216743", "0.53147805", "0.530874", "0.5296009", "0.5294241", "0.52775764", "0.52693397", "0.52693397", "0.5264473", "0.52579206", "0.52579206", "0.5257295", "0.5246897", "0.5241644", "0.5234891", "0.52214897", "0.52187747", "0.5218229", "0.52127117", "0.52020955", "0.51999277", "0.51808804", "0.5179603", "0.517909", "0.5170886", "0.51635605", "0.5139012", "0.51386505", "0.5133911" ]
0.90295964
0
TrimPrefix uses strings.TrimPrefix to remove prefix from the beginning of operand.
func TrimPrefix(prefix, operand string) string { return strings.TrimPrefix(operand, prefix) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func trimToPrefix(str, prefix string) string {\n\ti := strings.Index(str, prefix)\n\tif i < 0 {\n\t\treturn str\n\t}\n\treturn str[i:]\n}", "func trimOutPrefix(str, prefix string) string {\n\ti := strings.Index(str, prefix)\n\tif i < 0 {\n\t\treturn str\n\t}\n\treturn str[i+len(prefix):]\n}", "func (s *Stringish) TrimPrefix(prefix string) *Stringish {\n\ts.str = strings.TrimPrefix(s.str, prefix)\n\treturn s\n}", "func TrimPrefix(prefix string) MapFunc {\n\treturn func(s string) string { return strings.TrimPrefix(s, prefix) }\n}", "func trim(name string) string {\n\treturn strings.TrimPrefix(name, Prefix)\n}", "func RemovePrefix(prefix, str string) (string, bool) {\n\thad := strings.HasPrefix(str, prefix)\n\treturn strings.TrimPrefix(str, prefix), had\n}", "func TrimLeft(chars string, operand string) string { return strings.TrimLeft(operand, chars) }", "func trimPrefix(command, prefix string) (string, bool) {\n\tvar c string\n\tif prefix == \"\" && command != \"\" {\n\t\treturn command, true\n\t}\n\tif command == \"\" {\n\t\treturn command, false\n\t}\n\tif c = strings.TrimPrefix(command, prefix); c == command {\n\t\treturn \"\", false\n\t}\n\n\treturn c, len(c) != 0 // if command was \"[prefix]\" and it's trimmed into \"\" it should be false\n\n}", "func TrimSuffix(suffix, operand string) string { return strings.TrimSuffix(operand, suffix) }", "func TrimPrefix(objects []string, prefix string) []string {\n\tvar results []string\n\tfor _, object := range objects {\n\t\tresults = append(results, strings.TrimPrefix(object, prefix))\n\t}\n\treturn results\n}", "func CleanPrefix(prefix string) string {\n\trunes := []rune(prefix)\n\tif len(runes) == 0 {\n\t\treturn prefix\n\t}\n\t// Trim duplicate leading slash(es).\n\tif runes[0] == '/' {\n\t\tfor len(runes) > 1 && runes[1] == '/' {\n\t\t\trunes = runes[1:]\n\t\t}\n\t}\n\t// Trim trailing slash(es).\n\tfor len(runes) > 0 && runes[len(runes)-1] == '/' {\n\t\trunes = runes[:len(runes)-1]\n\t}\n\t// Were there nothing but slashes?\n\tif len(runes) == 0 {\n\t\treturn \"\"\n\t}\n\t// Build a properly formatted string.\n\tvar b strings.Builder\n\tif runes[0] != '/' {\n\t\tb.Grow(len(runes) + 1)\n\t\tb.WriteRune('/')\n\t} else {\n\t\tb.Grow(len(runes))\n\t}\n\tfor _, r := range runes {\n\t\tb.WriteRune(r)\n\t}\n\treturn b.String()\n}", "func (s *Stringish) TrimPrefixSpaces() *Stringish {\n\treturn s.TrimPrefix(\" \")\n}", "func StripPrefix(prefix string, handler func(*Request, *Response)) func(*Request, *Response) {\n\t// check for nil prefix\n\tif prefix == \"\" {\n\t\treturn handler\n\t}\n\t// create handler to trip the prefix before running handler\n\treturn func(req *Request, resp *Response) {\n\t\treq.RequestURI = strings.TrimPrefix(req.RequestURI, prefix)\n\t\thandler(req, resp)\n\t}\n}", "func Trim(chars string, operand string) string { return strings.Trim(operand, chars) }", "func removePrefix(s string) string {\n\tif !strings.ContainsRune(s, ':') {\n\t\treturn s\n\t}\n\treturn strings.Split(s, \":\")[1]\n}", "func stripPrefix(s string) (string, bool) {\n\ts = strings.TrimSpace(s)\n\tres := strings.TrimPrefix(s, \"What is\")\n\tif res == s {\n\t\treturn \"\", false\n\t}\n\treturn res, true\n}", "func (o BucketReplicationConfigRuleFilterAndOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigRuleFilterAnd) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func stripPrefixes(subj string) string {\n\tredo := true\n\tfor redo {\n\t\tredo = false\n\t\tfor _, prefix := range _BAD_PREFIXES {\n\t\t\tif strings.HasPrefix(strings.ToLower(subj), prefix) {\n\t\t\t\tsubj = subj[len(prefix):]\n\t\t\t\tredo = true\n\t\t\t}\n\t\t}\n\t}\n\n\treturn subj\n}", "func (o *Orders) TrimPrefixOrderId(prefix_id string) (int, error) {\n\ttrimmed_id := bytes.TrimPrefix([]byte(prefix_id), []byte(\"order-\"))\n\torder_id, err := strconv.Atoi(string(trimmed_id))\n\treturn order_id, err\n}", "func (o BucketReplicationConfigurationRuleFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigurationRuleFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func (o BucketMetricFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketMetricFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func (o BucketReplicationConfigRuleFilterAndPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketReplicationConfigRuleFilterAnd) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func HasPrefix(prefix, operand string) bool { return strings.HasPrefix(operand, prefix) }", "func TrimSpace(operand string) string { return strings.TrimSpace(operand) }", "func (o BucketReplicationConfigRuleFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigRuleFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func TrimPrefixPath(prefix string, path string) string {\n\treturn strings.TrimPrefix(path, prefix)\n}", "func (o BucketMetricFilterPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketMetricFilter) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func (g *gnmiPath) StripPrefix(pfx *gnmiPath) (*gnmiPath, error) {\n\tif !g.isSameType(pfx) {\n\t\treturn nil, fmt.Errorf(\"mismatched path formats in prefix and path, isElementPath: %v != %v\", g.isStringSlicePath(), pfx.isStringSlicePath())\n\t}\n\n\tif !g.isValid() || !pfx.isValid() {\n\t\treturn nil, fmt.Errorf(\"invalid paths supplied for stripPrefix: %v, %v\", g, pfx)\n\t}\n\n\tif pfx.isStringSlicePath() {\n\t\tfor i, e := range pfx.stringSlicePath {\n\t\t\tif g.stringSlicePath[i] != e {\n\t\t\t\treturn nil, fmt.Errorf(\"prefix is not a prefix of the supplied path, %v is not a subset of %v\", pfx, g)\n\t\t\t}\n\t\t}\n\t\treturn newStringSliceGNMIPath(g.stringSlicePath[len(pfx.stringSlicePath):]), nil\n\t}\n\n\tfor i, e := range pfx.pathElemPath {\n\t\tif !util.PathElemsEqual(g.pathElemPath[i], e) {\n\t\t\treturn nil, fmt.Errorf(\"prefix is not a prefix of the supplied path, %v is not a subset of %v\", pfx, g)\n\t\t}\n\t}\n\treturn newPathElemGNMIPath(g.pathElemPath[len(pfx.pathElemPath):]), nil\n}", "func TrimBytePrefix(str string, prefix byte) string {\n\tif len(str) > 0 && str[0] == prefix {\n\t\tstr = str[1:]\n\t}\n\treturn str\n}", "func (id ID) trimPrefix() ID {\n\treturn ID(strings.TrimPrefix(string(id), \"sha256:\"))\n}", "func (n *nodeHeader) leftTrimPrefix(l uint16) {\n\tif l < 1 {\n\t\treturn\n\t}\n\tpLen, pBytes := n.prefixFields()\n\tif l > *pLen {\n\t\tl = *pLen\n\t}\n\tnewLen := *pLen - uint16(l)\n\tcopy(pBytes[0:newLen], pBytes[l:*pLen])\n\t*pLen = newLen\n}", "func (f FunctionSelector) WithoutPrefix() string { return f.String()[2:] }", "func (o BucketReplicationConfigurationRuleFilterPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketReplicationConfigurationRuleFilter) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func (o BucketLifecycleConfigurationV2RuleFilterAndOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketLifecycleConfigurationV2RuleFilterAnd) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func trimTreePrefix(name string, n int) string {\n\ti := 0\n\tfor ; i < len(name) && n > 0; i++ {\n\t\tif name[i] == '/' {\n\t\t\tn--\n\t\t}\n\t}\n\treturn name[i:]\n}", "func (o BucketReplicationConfigRuleFilterPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketReplicationConfigRuleFilter) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func (b DeleteBuilder) Prefix(sql string, args ...interface{}) DeleteCondition {\n\treturn builder.Append(b, \"Prefixes\", Expr(sql, args...)).(DeleteBuilder)\n}", "func trimLeftSlash(name string) string {\n\treturn strings.TrimPrefix(name, \"/\")\n}", "func (o BucketV2ReplicationConfigurationRuleFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketV2ReplicationConfigurationRuleFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func FilterPrefix(stringSet sets.String, prefix string, ignoreCase bool) sets.String {\n\tif prefix == \"\" {\n\t\treturn stringSet\n\t}\n\treturn filterSet(stringSet, prefix, ignoreCase, strings.HasPrefix)\n}", "func (o InventoryFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InventoryFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func (o InventoryFilterPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *InventoryFilter) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func (o BucketLifecycleConfigurationV2RuleFilterAndPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketLifecycleConfigurationV2RuleFilterAnd) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func (o BucketReplicationConfigurationRuleOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigurationRule) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func (n Name) Prefix(suffix Name) Name {\n\toffset := len(n) - len(suffix)\n\tif offset < 0 {\n\t\treturn n\n\t}\n\tif n[offset:].Equal(suffix) {\n\t\treturn n[:offset]\n\t}\n\treturn n\n}", "func (o BucketNotificationQueueOutput) FilterPrefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketNotificationQueue) *string { return v.FilterPrefix }).(pulumi.StringPtrOutput)\n}", "func trimLeft(r *syntax.Regexp) (bool, *syntax.Regexp) {\n\tif eqPrefixAnyRegex(r, patDotStar, patNullBeginDotStar) {\n\t\ttmp := *r\n\t\ttmp.Sub = tmp.Sub[1:]\n\t\treturn true, &tmp\n\t}\n\n\treturn false, r\n}", "func (o BucketNotificationLambdaFunctionOutput) FilterPrefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketNotificationLambdaFunction) *string { return v.FilterPrefix }).(pulumi.StringPtrOutput)\n}", "func (o AnalyticsConfigurationFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AnalyticsConfigurationFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func EnsurePrefix(s, prefix string) string {\n\tif strings.HasPrefix(s, prefix) {\n\t\treturn s\n\t}\n\treturn prefix + s\n}", "func (s Settings) TrimPrefix() bool {\n\treturn s.trimPrefix\n}", "func ensurePrefix(s, prefix string) string {\n\tif !strings.HasPrefix(s, prefix) {\n\t\ts = prefix + s\n\t}\n\treturn s\n}", "func (o AnalyticsConfigurationFilterPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AnalyticsConfigurationFilter) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func (o BucketIntelligentTieringConfigurationFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketIntelligentTieringConfigurationFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func (o BucketReplicationConfigRuleOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigRule) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func (o BucketNotificationTopicOutput) FilterPrefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketNotificationTopic) *string { return v.FilterPrefix }).(pulumi.StringPtrOutput)\n}", "func CorrectionHasPrefix(v string) predicate.TradeCorrection {\n\treturn predicate.TradeCorrection(sql.FieldHasPrefix(FieldCorrection, v))\n}", "func (o BucketLifecycleRuleOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketLifecycleRule) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func StripPrefix(prefix string, h Handler) Handler {\n\tif prefix == \"\" {\n\t\treturn h\n\t}\n\treturn HandlerFunc(func(w ResponseWriter, r *Request) {\n\t\tp := strings.TrimPrefix(r.URL.Path, prefix)\n\t\trp := strings.TrimPrefix(r.URL.RawPath, prefix)\n\t\tif len(p) < len(r.URL.Path) && (r.URL.RawPath == \"\" || len(rp) < len(r.URL.RawPath)) {\n\t\t\tr2 := new(Request)\n\t\t\t*r2 = *r\n\t\t\tr2.URL = new(url.URL)\n\t\t\t*r2.URL = *r.URL\n\t\t\tr2.URL.Path = p\n\t\t\tr2.URL.RawPath = rp\n\t\t\th.ServeHTTP(w, r2)\n\t\t} else {\n\t\t\tNotFound(w, r)\n\t\t}\n\t})\n}", "func (tr *NormalizingTarReader) Strip(prefix string) {\n\ttr.headerOpts = append(tr.headerOpts, func(header *tar.Header) *tar.Header {\n\t\theader.Name = strings.TrimPrefix(header.Name, prefix)\n\t\treturn header\n\t})\n}", "func (o BucketLifecycleConfigurationV2RuleFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketLifecycleConfigurationV2RuleFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func (s *IPSet) RemovePrefix(p IPPrefix) { s.RemoveRange(p.Range()) }", "func (s *IPSet) RemovePrefix(p IPPrefix) { s.RemoveRange(p.Range()) }", "func PrefixFilter(prefix string) FilterFunc {\n\treturn func(m *MountInfo) (bool, bool) {\n\t\tskip := !strings.HasPrefix(m.MountPoint, prefix)\n\t\treturn skip, false\n\t}\n}", "func StripStart(str string) string {\n\treturn regexp.MustCompile(`^\\s+`).ReplaceAllString(str, \"\")\n}", "func compactPrefix() []byte { return []byte{0, 1, 0, 0} }", "func (o BucketIntelligentTieringConfigurationFilterPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketIntelligentTieringConfigurationFilter) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func (p *Parser) parsePrefixExpression() ast.Expression {\n\texpression := &ast.PrefixExpression{\n\t\tToken: p.curToken,\n\t\tOperator: p.curToken.Literal}\n\tp.nextToken()\n\texpression.Right = p.parseExpression(PREFIX)\n\treturn expression\n}", "func (f *filter) Prefix(prefix string) Filter {\n\tif f.prefixed || prefix == \"\" {\n\t\treturn f\n\t}\n\n\tif len(f.allowed) > 0 {\n\t\tfor _, r := range f.allowed {\n\t\t\tr.prefix(prefix)\n\t\t}\n\n\t\tf.prefixed = true\n\t}\n\n\treturn f\n}", "func (pars *Parser) parsePrefixExpression() tree.Expression {\n\texpression := &tree.PrefixExpression{\n\t\tToken: pars.thisToken,\n\t\tOperator: pars.thisToken.Val,\n\t}\n\n\tpars.nextToken()\n\n\texpression.Right = pars.parseExpression(PREFIX)\n\treturn expression\n}", "func TrimLeft(cutset string) MapFunc {\n\treturn func(s string) string { return strings.TrimLeft(s, cutset) }\n}", "func PhoneHasPrefix(v string) predicate.Patient {\n\treturn predicate.Patient(func(s *sql.Selector) {\n\t\ts.Where(sql.HasPrefix(s.C(FieldPhone), v))\n\t})\n}", "func (o BucketLifecycleConfigurationV2RuleFilterPtrOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketLifecycleConfigurationV2RuleFilter) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefix\n\t}).(pulumi.StringPtrOutput)\n}", "func (b UpdateBuilder) Prefix(sql string, args ...interface{}) UpdateCondition {\n\treturn builder.Append(b, \"Prefixes\", Expr(sql, args...)).(UpdateBuilder)\n}", "func compactPrefix() []byte { return []byte{0, 3, 0, 0} }", "func (o BucketV2ReplicationConfigurationRuleOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketV2ReplicationConfigurationRule) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func SwapPrefix(group, prefix, swap string) string {\n\tprefix = strings.ToLower(prefix)\n\tgroup = strings.ToLower(group)\n\tif strings.HasPrefix(group, prefix) {\n\t\ts := swap + strings.TrimPrefix(group, prefix)\n\t\treturn Format(s)\n\t}\n\treturn \"\"\n}", "func PhoneHasPrefix(v string) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.HasPrefix(s.C(FieldPhone), v))\n\t})\n}", "func PhoneHasPrefix(v string) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.HasPrefix(s.C(FieldPhone), v))\n\t})\n}", "func PhoneHasPrefix(v string) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.HasPrefix(s.C(FieldPhone), v))\n\t})\n}", "func PhoneHasPrefix(v string) predicate.Surgeryappointment {\n\treturn predicate.Surgeryappointment(func(s *sql.Selector) {\n\t\ts.Where(sql.HasPrefix(s.C(FieldPhone), v))\n\t})\n}", "func (p *Parser) parsePrefixExpression() asti.ExpressionI {\n\texpression := &ast.PrefixExpression{\n\t\tToken: p.curToken,\n\t\tOperator: p.curToken.Type,\n\t}\n\tp.nextToken()\n\texpression.Right = p.parseExpression(precedence.PREFIX)\n\treturn expression\n}", "func PhoneHasPrefix(v string) predicate.User {\n\treturn predicate.User(sql.FieldHasPrefix(FieldPhone, v))\n}", "func TrimLeftFunc(f func(rune) bool) MapFunc {\n\treturn func(s string) string { return strings.TrimLeftFunc(s, f) }\n}", "func StringEnsurePrefix(s string, cut string) string {\n\tif strings.HasPrefix(s, cut) {\n\t\treturn s\n\t}\n\treturn cut + s\n}", "func RemoveStart(str string, remove string) string {\n\tif IsEmpty(str) || IsEmpty(remove) {\n\t\treturn str\n\t}\n\tif StartsWith(str, remove) {\n\t\treturn str[len(remove)+1:]\n\t}\n\treturn str\n}", "func Prefix() string {\n\treturn std.Prefix()\n}", "func DeletePrefix(ctx context.Context, prefix string) error {\n\terr := Client().DeletePrefix(ctx, prefix)\n\tTrace(\"DeletePrefix\", err, logrus.Fields{fieldPrefix: prefix})\n\treturn err\n}", "func filterTrim(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.TrimSpace(stick.CoerceString(val))\n}", "func (o *SignalPersonName) UnsetPrefix() {\n\to.Prefix.Unset()\n}", "func (s *Str) TrimLeft(cutset string) *Str {\n\ts.val = strings.TrimLeft(s.val, cutset)\n\treturn s\n}", "func cleansePrefixes(ss []string) []string {\n\n\tret := []string{}\n\tfor _, s := range ss {\n\t\tstripped := \"\"\n\t\tfor i := len(ss) - 1; i > -1; i-- { // reversely\n\t\t\tpref := ss[i]\n\t\t\tif s != pref && strings.HasPrefix(s, pref) {\n\n\t\t\t\tstripped = strings.TrimPrefix(s, pref)\n\n\t\t\t\tstripped = strings.TrimSpace(stripped)\n\t\t\t\tstripped = strings.TrimPrefix(stripped, \"-- \")\n\t\t\t\tstripped = strings.TrimSuffix(stripped, \" --\")\n\n\t\t\t\t// log.Printf(\"stripped off\\n\\t%q \\n\\t%q \\n\\t%q\", s, pref, stripped)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif stripped == \"\" {\n\t\t\tret = append(ret, s)\n\t\t} else {\n\t\t\tret = append(ret, stripped)\n\t\t}\n\t}\n\n\treturn ret\n\n}", "func PrefixString(prefix, key string) string {\n\treturn GetString(Prefix(prefix, key))\n}", "func (o InventoryDestinationBucketOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InventoryDestinationBucket) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func trimWhitespaces(value string) string {\n\treturn strings.Trim(value, \"\")\n}", "func (db *DB) DeletePrefix(prefix interface{}, value interface{}) error {\n\treturn db.bolt.Update(func(tx *bolt.Tx) error {\n\t\treturn db.DeleteTx(tx, prefix, value)\n\t})\n}", "func LTrim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_LEADING,\n\t}\n}", "func (o BucketLifecycleConfigurationV2RuleOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketLifecycleConfigurationV2Rule) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func removeSemverPrefix(s string) string {\n\ts = strings.TrimPrefix(s, \"v\")\n\ts = strings.TrimPrefix(s, \"go\")\n\treturn s\n}", "func (q *queryImpl) Prefix(sql string, args ...interface{}) Query {\n\tq.prefixes = append(q.suffixes, ExprString{sql, args})\n\treturn q\n}" ]
[ "0.79802936", "0.76670146", "0.7575979", "0.7347845", "0.7151964", "0.6931681", "0.6902632", "0.6860264", "0.6841062", "0.68230784", "0.6656401", "0.65976673", "0.65567815", "0.64447945", "0.642312", "0.64137286", "0.6408851", "0.64047384", "0.63290054", "0.63274646", "0.63249403", "0.63130045", "0.6304696", "0.629505", "0.6259701", "0.625187", "0.6250502", "0.621472", "0.62122416", "0.61933464", "0.6191105", "0.6188272", "0.6181847", "0.6151284", "0.6139785", "0.6129984", "0.6124559", "0.60681885", "0.6060039", "0.60590506", "0.60401547", "0.60106236", "0.60020936", "0.5995301", "0.5989006", "0.59603846", "0.59549034", "0.59416187", "0.5938841", "0.5921516", "0.59147644", "0.5904205", "0.5880739", "0.58793443", "0.5879019", "0.58649653", "0.5864534", "0.5861943", "0.58583236", "0.5845484", "0.5841424", "0.58393216", "0.58393216", "0.5832796", "0.5826596", "0.5796546", "0.5796288", "0.5780446", "0.5769881", "0.5756505", "0.57236034", "0.5716412", "0.5713775", "0.570975", "0.5706058", "0.570401", "0.56863105", "0.56749046", "0.56749046", "0.56749046", "0.5655737", "0.56364864", "0.5608826", "0.5602078", "0.55970824", "0.5573222", "0.55719006", "0.5571717", "0.5544241", "0.5541167", "0.5512706", "0.5502435", "0.5502336", "0.54959434", "0.5492789", "0.5481292", "0.5477081", "0.5470325", "0.5470227", "0.5463645" ]
0.9364741
0
TrimRight uses strings.TrimRight to remove any occurrences of chars from the end of operand.
func TrimRight(cutset string, operand string) string { return strings.TrimRight(operand, cutset) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func trimRight(s string) string {\n\treturn strings.TrimRightFunc(s, unicode.IsSpace)\n}", "func RightTrim(str, chars string) string {\n\tif chars == \"\" {\n\t\treturn strings.TrimRightFunc(str, unicode.IsSpace)\n\t}\n\tr, _ := regexp.Compile(\"[\" + chars + \"]+$\")\n\treturn r.ReplaceAllString(str, \"\")\n}", "func (s *Str) TrimRight(cutset string) *Str {\n\ts.val = strings.TrimRight(s.val, cutset)\n\treturn s\n}", "func Trim(chars string, operand string) string { return strings.Trim(operand, chars) }", "func trimRight(r *syntax.Regexp) (bool, *syntax.Regexp) {\n\tif eqSuffixAnyRegex(r, patDotStar, patNullEndDotStar) {\n\t\ti := len(r.Sub) - 1\n\t\ttmp := *r\n\t\ttmp.Sub = tmp.Sub[0:i]\n\t\treturn true, &tmp\n\t}\n\n\treturn false, r\n}", "func TrimSpace(operand string) string { return strings.TrimSpace(operand) }", "func TrimLeft(chars string, operand string) string { return strings.TrimLeft(operand, chars) }", "func TrimRight(cutset string) MapFunc {\n\treturn func(s string) string { return strings.TrimRight(s, cutset) }\n}", "func TrimRightFunc(f func(rune) bool) MapFunc {\n\treturn func(s string) string { return strings.TrimRightFunc(s, f) }\n}", "func TrimSuffix(suffix, operand string) string { return strings.TrimSuffix(operand, suffix) }", "func filterTrim(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.TrimSpace(stick.CoerceString(val))\n}", "func RTrim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_TRAILING,\n\t}\n}", "func TrimPrefix(prefix, operand string) string { return strings.TrimPrefix(operand, prefix) }", "func RTrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_TRAILING,\n\t\tchars: chars,\n\t}\n}", "func Right(str string, size int) string {\n\tif str == \"\" || size < 0 {\n\t\treturn \"\"\n\t}\n\tif len(str) <= size {\n\t\treturn str\n\t}\n\treturn str[len(str)-size:]\n}", "func (b *Bar) TrimRightSpace() *Bar {\n\tif isClosed(b.done) {\n\t\treturn b\n\t}\n\tb.trimRightCh <- true\n\treturn b\n}", "func Right(s string, n int) string {\n\trunes := []rune(s)\n\tif n < 1 {\n\t\treturn \"\"\n\t}\n\tif n >= len(runes) {\n\t\treturn s\n\t}\n\n\treturn string(runes[len(runes)-n:])\n}", "func Trim(str, chars string) string {\n\treturn LeftTrim(RightTrim(str, chars), chars)\n}", "func Trim(text string) string {\n\treturn trimRx.FindStringSubmatch(text)[1]\n}", "func trimSpace(b []byte) string {\n\treturn string(bytes.TrimRight(b, \" \"))\n}", "func trimSpace(b []byte) string {\n\treturn string(bytes.TrimRight(b, \" \"))\n}", "func trimLast(text string) string {\n\ttextLen := len(text)\n\tif textLen == 0 {\n\t\treturn text\n\t}\n\treturn text[:textLen-1]\n}", "func TrimTrailingSpaces(text string) string {\n\tparts := strings.Split(text, \"\\n\")\n\tfor i := range parts {\n\t\tparts[i] = strings.TrimRightFunc(parts[i], func(r rune) bool {\n\t\t\treturn unicode.IsSpace(r)\n\t\t})\n\n\t}\n\n\treturn strings.Join(parts, \"\\n\")\n}", "func (su StringUtility) TrimEndMultiple(targetStr string, trimChar rune) (rStr string, err error) {\n\n\tif targetStr == \"\" {\n\t\terr = errors.New(\"Empty targetStr\")\n\t\treturn\n\t}\n\n\tfStr := []rune(targetStr)\n\tlenTargetStr := len(fStr)\n\toutputStr := make([]rune, lenTargetStr)\n\tlenTargetStr--\n\tidx := lenTargetStr\n\tfoundFirstChar := false\n\n\tfor i := lenTargetStr; i >= 0; i-- {\n\n\t\tif !foundFirstChar && fStr[i] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tif i > 0 && fStr[i] == trimChar && fStr[i-1] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tif i == 0 && fStr[i] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tfoundFirstChar = true\n\t\toutputStr[idx] = fStr[i]\n\t\tidx--\n\t}\n\n\tif idx != lenTargetStr {\n\t\tidx++\n\t}\n\n\tif outputStr[idx] == trimChar {\n\t\tidx++\n\t}\n\n\tresult := string(outputStr[idx:])\n\n\treturn result, nil\n\n}", "func (fn *formulaFuncs) TRIM(argsList *list.List) formulaArg {\n\tif argsList.Len() != 1 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"TRIM requires 1 argument\")\n\t}\n\treturn newStringFormulaArg(strings.TrimSpace(argsList.Front().Value.(formulaArg).Value()))\n}", "func execTrimString(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := textproto.TrimString(args[0].(string))\n\tp.Ret(1, ret)\n}", "func Right(str string, length int, pad string) string {\n\treturn str + times(pad, length-len(str))\n}", "func trim(strn string) string {\n\treturn strings.Trim(strn, \" \\t\\n\\r\")\n}", "func stripTrailingSpace(str string) string {\n\tbuf := bytes.NewBuffer(nil)\n\n\tscan := bufio.NewScanner(strings.NewReader(str))\n\tfor scan.Scan() {\n\t\tbuf.WriteString(strings.TrimRight(scan.Text(), \" \\t\\r\\n\"))\n\t\tbuf.WriteString(\"\\n\")\n\t}\n\treturn buf.String()\n}", "func StrTrim(s string) string {\n\treturn strings.TrimSpace(s)\n}", "func TrimSlash(str string) string {\n\treturn TrimByteSuffix(TrimBytePrefix(str, '/'), '/')\n}", "func RightJustify(s string, w int) string {\n\tsw := len(s)\n\tdiff := w - sw\n\tif diff > 0 {\n\t\ts = strings.Repeat(\" \", diff) + s\n\t} else if diff < 0 {\n\t\ts = s[:w]\n\t}\n\treturn s\n}", "func Right(text string, size int) string {\n\tspaces := size - Length(text)\n\tif spaces <= 0 {\n\t\treturn text\n\t}\n\n\tvar buffer bytes.Buffer\n\tfor i := 0; i < spaces; i++ {\n\t\tbuffer.WriteString(space)\n\t}\n\n\tbuffer.WriteString(text)\n\treturn buffer.String()\n}", "func Strip(str string) string {\n\treturn regexp.MustCompile(`^\\s+|\\s+$`).ReplaceAllString(str, \"\")\n}", "func LeftRight(val string) (string, string, bool) {\n\tif len(val) < 2 {\n\t\treturn \"\", val, false\n\t}\n\tswitch by := val[0]; by {\n\tcase '`':\n\t\tvals := strings.Split(val, \"`.`\")\n\t\tif len(vals) == 1 {\n\t\t\treturn \"\", IdentityTrim(val), false\n\t\t} else if len(vals) == 2 {\n\t\t\treturn IdentityTrim(vals[0]), IdentityTrim(vals[1]), true\n\t\t}\n\t\t// wat, no idea what this is\n\t\treturn \"\", val, false\n\tcase '[':\n\t\tvals := strings.Split(val, \"].[\")\n\t\tif len(vals) == 1 {\n\t\t\treturn \"\", IdentityTrim(val), false\n\t\t} else if len(vals) == 2 {\n\t\t\treturn IdentityTrim(vals[0]), IdentityTrim(vals[1]), true\n\t\t}\n\t\t// wat, no idea what this is\n\t\treturn \"\", val, false\n\tdefault:\n\t\tvals := strings.SplitN(val, \".\", 2)\n\t\tif len(vals) == 1 {\n\t\t\treturn \"\", val, false\n\t\t} else if len(vals) == 2 {\n\t\t\treturn IdentityTrim(vals[0]), IdentityTrim(vals[1]), true\n\t\t}\n\t}\n\n\treturn \"\", val, false\n}", "func chopRight(expr string) (left string, tok rune, right string) {\n\t// XXX implementation redacted for CHALLENGE1.\n\t// TODO restore implementation and replace '~'\n\tparts := strings.Split(expr, \"~\")\n\tif len(parts) != 4 {\n\t\treturn\n\t}\n\tleft = parts[0]\n\ttok = rune(parts[1][0])\n\tright = parts[2]\n\t// close = parts[3]\n\treturn\n}", "func ScanlineTrim() string {\n\treturn strings.TrimSpace(Scanline())\n}", "func stripSlash(str string) string {\n\treturn strings.TrimRight(strings.TrimLeft(str, \"/\"), \"/\")\n}", "func trimTrailingWhitespace(details string) string {\n\treturn strings.TrimSuffix(details, \" \")\n}", "func (chars *Chars) TrimLength() uint16 {\n\tif chars.trimLengthKnown {\n\t\treturn chars.trimLength\n\t}\n\tchars.trimLengthKnown = true\n\tvar i int\n\tlen := chars.Length()\n\tfor i = len - 1; i >= 0; i-- {\n\t\tchar := chars.Get(i)\n\t\tif !unicode.IsSpace(char) {\n\t\t\tbreak\n\t\t}\n\t}\n\t// Completely empty\n\tif i < 0 {\n\t\treturn 0\n\t}\n\n\tvar j int\n\tfor j = 0; j < len; j++ {\n\t\tchar := chars.Get(j)\n\t\tif !unicode.IsSpace(char) {\n\t\t\tbreak\n\t\t}\n\t}\n\tchars.trimLength = AsUint16(i - j + 1)\n\treturn chars.trimLength\n}", "func (fn *formulaFuncs) RIGHT(argsList *list.List) formulaArg {\n\treturn fn.leftRight(\"RIGHT\", argsList)\n}", "func TrimTrailingSpaces(p []byte) []byte {\n\tfor i := len(p) - 1; i >= 0; i-- {\n\t\tif p[i] != 0x20 && p[i] != '\\n' && p[i] != '\\t' {\n\t\t\treturn p[:i+1]\n\t\t}\n\t}\n\t// it was all spaces\n\treturn p[:0]\n}", "func TestTrim(t *testing.T) {\n\ttext := \"Hola Mundo TDA\"\n\tt.Logf(\"text:[%s]\", text)\n\tt.Logf(\"trim:[%s]\", utl.Trim(text))\n}", "func trim(s string) string {\n\tfor len(s) > 0 {\n\t\tif s[0] <= ' ' {\n\t\t\ts = s[1:]\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tfor len(s) > 0 {\n\t\tif s[len(s)-1] <= ' ' {\n\t\t\ts = s[:len(s)-1]\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn s\n}", "func (s *Str) Trim(cutset string) *Str {\n\ts.val = strings.Trim(s.val, cutset)\n\treturn s\n}", "func Trim(str string) string {\n\treturn strings.TrimSpace(str)\n}", "func PadRight(str string, padStr string, padLen int) string {\n\treturn buildPadStr(str, padStr, padLen, false, true)\n}", "func StripEnd(str string) string {\n\treturn regexp.MustCompile(`\\s+$`).ReplaceAllString(str, \"\")\n}", "func Trim(str string) string {\n\treturn strings.Trim(str, \" \")\n}", "func (s *Str) TrimLeft(cutset string) *Str {\n\ts.val = strings.TrimLeft(s.val, cutset)\n\treturn s\n}", "func trimPathAfterLastSlash(path string) string {\n\tif index := strings.LastIndex(path, \"/\"); index != -1 {\n\t\t// fmt.Println(path, \" Trimmed =\", path[:index])\n\t\treturn path[:index] //remove including the last /\n\t}\n\tfmt.Println(\"Failed to trim strings after last '/'\")\n\treturn path\n}", "func (p *Parser) Trim(input string) string {\n\tt := strings.Replace(strings.Replace(input, \"<output>\", \"\", -1), \"</output>\", \"\", -1)\n\tt1 := strings.Replace(strings.Replace(t, \"<configuration-information>\", \"\", -1), \"</configuration-information>\", \"\", -1)\n\treturn strings.Replace(strings.Replace(t1, \"<configuration-output>\", \"\", -1), \"</configuration-output>\", \"\", -1)\n}", "func (o DashboardSpacingOutput) Right() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v DashboardSpacing) *string { return v.Right }).(pulumi.StringPtrOutput)\n}", "func trimS(s string) string {\r\n\treturn strings.TrimSpace(s)\r\n}", "func PadRight(s string, padStr string, lenStr int) string {\n\tvar padCount int\n\tpadCount = I.MaxOf(lenStr-len(s), 0)\n\treturn s + strings.Repeat(padStr, padCount)\n}", "func BytesStripRight(s []byte, pad byte) []byte {\n\tn := len(s)\n\tfor n > 0 && s[n-1] == pad {\n\t\tn--\n\t}\n\treturn s[:n]\n}", "func TrimSpaceAndQuotes(answer string) string {\n\ttext := strings.TrimSpace(answer)\n\tfor _, q := range quotes {\n\t\tif strings.HasPrefix(text, q) && strings.HasSuffix(text, q) {\n\t\t\treturn strings.TrimPrefix(strings.TrimSuffix(text, q), q)\n\t\t}\n\t}\n\treturn text\n}", "func TrimSpace(ctx context.Context, t *mold.Transformer, v reflect.Value) error {\n\tv.Set(reflect.ValueOf(strings.TrimSpace(v.String())))\n\treturn nil\n}", "func (ctx *RequestContext) RemoveRight(right string) {\n\tif !ctx.IsAuthenticated() {\n\t\treturn\n\t}\n\n\trights := ctx.principal.Rights\n\n\ti := sort.Search(len(rights), func(i int) bool {\n\t\treturn rights[i] == right\n\t})\n\n\tif i >= len(rights) {\n\t\treturn\n\t}\n\n\tctx.principal.Rights = append(rights[:i], rights[i+1:]...)\n}", "func TrimRedundantSpaces(text string) string {\n\ttext = spaceRegex.ReplaceAllString(text, \" \")\n\treturn strings.TrimSpace(text)\n}", "func padRight(str string, max int) string {\n\tpad := max - len(str)\n\tb := strings.Builder{}\n\tb.WriteString(str)\n\tb.WriteString(strings.Repeat(\" \", pad))\n\treturn b.String()\n}", "func trimTrailingWhiteSpace(s []byte) []byte {\n\tfor (len(s) > 0) && ((s[len(s)-1] == ' ') || (s[len(s)-1] == '\\t')) {\n\t\ts = s[:len(s)-1]\n\t}\n\treturn s\n}", "func trimLeftSlash(name string) string {\n\treturn strings.TrimPrefix(name, \"/\")\n}", "func trimChar(s string, r byte) string {\n\tsz := len(s)\n\n\tif sz > 0 && s[sz-1] == r {\n\t\ts = s[:sz-1]\n\t}\n\tsz = len(s)\n\tif sz > 0 && s[0] == r {\n\t\ts = s[1:sz]\n\t}\n\treturn s\n}", "func trimChar(s string, r byte) string {\n\tsz := len(s)\n\n\tif sz > 0 && s[sz-1] == r {\n\t\ts = s[:sz-1]\n\t}\n\tsz = len(s)\n\tif sz > 0 && s[0] == r {\n\t\ts = s[1:sz]\n\t}\n\treturn s\n}", "func removeTrailingSlash(path string) string {\n\treturn strings.TrimRight(path, \"/\")\n}", "func stripEnds(tk token.Token) token.Token {\n\ts := tk.Text()[1 : len(tk.Text())-1]\n\treturn token.UpdateText(tk, s)\n}", "func (o DashboardSpacingPtrOutput) Right() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *DashboardSpacing) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Right\n\t}).(pulumi.StringPtrOutput)\n}", "func LeftTrim(str, chars string) string {\n\tif chars == \"\" {\n\t\treturn strings.TrimLeftFunc(str, unicode.IsSpace)\n\t}\n\tr, _ := regexp.Compile(\"^[\" + chars + \"]+\")\n\treturn r.ReplaceAllString(str, \"\")\n}", "func padRight(str string, length int, pad string) string {\n\treturn str + times(pad, length-len(str))\n}", "func RemoveBothEndChars(data string, front int, last int) string {\n\t\n\tif len(data) > 0 {\n\t\treturn data[front:len(data)-last]\n\t}\n\t\n\treturn data\n}", "func trimLeft(r *syntax.Regexp) (bool, *syntax.Regexp) {\n\tif eqPrefixAnyRegex(r, patDotStar, patNullBeginDotStar) {\n\t\ttmp := *r\n\t\ttmp.Sub = tmp.Sub[1:]\n\t\treturn true, &tmp\n\t}\n\n\treturn false, r\n}", "func trim(s []byte) []byte {\n\ti := 0\n\tfor i < len(s) && (s[i] == ' ' || s[i] == '\\t') {\n\t\ti++\n\t}\n\tn := len(s)\n\tfor n > i && (s[n-1] == ' ' || s[n-1] == '\\t') {\n\t\tn--\n\t}\n\treturn s[i:n]\n}", "func trim(s []byte) []byte {\n\ti := 0\n\tfor i < len(s) && (s[i] == ' ' || s[i] == '\\t') {\n\t\ti++\n\t}\n\tn := len(s)\n\tfor n > i && (s[n-1] == ' ' || s[n-1] == '\\t') {\n\t\tn--\n\t}\n\treturn s[i:n]\n}", "func (sopsTxtJustify TextJustify) Right() TextJustify {\n\n\tlockStrOpsTextJustify.Lock()\n\n\tdefer lockStrOpsTextJustify.Unlock()\n\n\treturn TextJustify(2)\n}", "func Trim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_BOTH,\n\t}\n}", "func trimCut(label []byte, maxlen int, left, right string) []byte {\n\ttrim := bytes.TrimLeft(label, left)\n\tsize := min(len(trim), maxlen)\n\thead := bytes.TrimRight(trim[:size], right)\n\tif len(head) == size {\n\t\treturn head\n\t}\n\ttail := bytes.TrimLeft(trim[size:], right)\n\tif len(tail) > 0 {\n\t\treturn append(head, tail[:min(len(tail), size-len(head))]...)\n\t}\n\treturn head\n}", "func RemoveTrailingSlash(s string) string {\n\tre := regexp.MustCompile(`/$`)\n\treturn re.ReplaceAllString(s, \"\")\n}", "func LTrim(p projection) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_LEADING,\n\t}\n}", "func trim(s string, n int) string {\n\tif len(s) > n {\n\t\treturn s[:n]\n\t}\n\treturn s\n}", "func trim(s string, n int) string {\n\tif len(s) > n {\n\t\treturn s[:n]\n\t}\n\treturn s\n}", "func StringStrip(scope *Scope, input tf.Output) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"StringStrip\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (seg *Segmenter) CutTrim(str string, hmm ...bool) []string {\n\ts := seg.Cut(str, hmm...)\n\treturn seg.Trim(s)\n}", "func TrimLeft(cutset string) MapFunc {\n\treturn func(s string) string { return strings.TrimLeft(s, cutset) }\n}", "func TrimFunc(f func(rune) bool) MapFunc {\n\treturn func(s string) string { return strings.TrimFunc(s, f) }\n}", "func rightPad(s string, padStr string, pLen int) string {\n\treturn s + strings.Repeat(padStr, pLen)\n}", "func trim(name string) string {\n\treturn strings.TrimPrefix(name, Prefix)\n}", "func RemoveTrailingSlash(dir string) string {\n\treturn strings.TrimRight(dir, string(os.PathSeparator))\n}", "func TrimLeftChar(s string) string {\n\tfor i := range s {\n\t\tif i > 0 {\n\t\t\treturn s[i:]\n\t\t}\n\t}\n\treturn s[:0]\n}", "func TrimLeftFunc(f func(rune) bool) MapFunc {\n\treturn func(s string) string { return strings.TrimLeftFunc(s, f) }\n}", "func RightPad(s string, padStr string, overallLen int) string {\n\tvar padCountInt = 1 + ((overallLen - len(padStr)) / len(padStr))\n\tvar retStr = s + strings.Repeat(padStr, padCountInt)\n\treturn retStr[:overallLen]\n}", "func (s Stream) DropRight(input interface{}) Stream {\n\ts.operations = append(s.operations, &streamDropRight{\n\t\tItemsValue: s.itemsValue,\n\t\tItemsType: s.itemsType,\n\t\tItem: input,\n\t\tOption: drop.Right,\n\t})\n\treturn s\n}", "func trimSpaceAndJoin(args []string, sep string) string {\n\ttrimmedArgs := make([]string, len(args))\n\tfor i, arg := range args {\n\t\ttrimmedArgs[i] = strings.TrimSpace(arg)\n\t}\n\treturn strings.TrimSpace(strings.Join(trimmedArgs, sep))\n}", "func findtrimmedexpression(s string) string {\r\n\tresult := \"\"\r\n\tfirst := false\r\n\tfirst_idx := 0\r\n\tidx := 0\r\n\tsecond_idx := 0\r\n\t//get correct indices\r\n\tfor idx < len(s) {\r\n\t\tif (string(s[idx]) == \"\\\"\" && (first == false)) || (string(s[idx]) == \"'\" && (first == false)) {\r\n\t\t\tfirst_idx = idx\r\n\t\t\tfirst = true\r\n\t\t} else if string(s[idx]) == \"\\\"\" || string(s[idx]) == \"'\" {\r\n\t\t\tsecond_idx = idx\r\n\t\t}\r\n\t\tidx = idx + 1\r\n\t}\r\n\tidx = first_idx + 1\r\n\tfor idx < second_idx {\r\n\t\tresult = result + string(s[idx])\r\n\t\tidx = idx + 1\r\n\t}\r\n\treturn result\r\n}", "func trim_string_after_s(s string, x string) (r string) {\n\tif idx := strings.Index(s, x); idx != -1 {\n\t\treturn s[:idx]\n\t}\n\treturn s\n}", "func (h *Helpers) Strip(s string) string {\n\tre1 := regexp.MustCompile(stripEmptyRegex)\n\tre2 := regexp.MustCompile(stripEdgesRegex)\n\ts = re1.ReplaceAllString(s, \"\")\n\ts = re2.ReplaceAllString(s, \"\")\n\treturn s\n}", "func TrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_BOTH,\n\t\tchars: chars,\n\t}\n}", "func Trim(cutset string) MapFunc {\n\treturn func(s string) string { return strings.Trim(s, cutset) }\n}", "func (seg *Segmenter) Trim(s []string) (r []string) {\n\tfor i := 0; i < len(s); i++ {\n\t\tsi := FilterSymbol(s[i])\n\t\tif !seg.NotStop && seg.IsStop(si) {\n\t\t\tsi = \"\"\n\t\t}\n\n\t\tif si != \"\" {\n\t\t\tr = append(r, si)\n\t\t}\n\t}\n\n\treturn\n}", "func Trim(str string, characterMask ...string) string {\n\ttrimChars := DefaultTrimChars\n\tif len(characterMask) > 0 {\n\t\ttrimChars += characterMask[0]\n\t}\n\treturn strings.Trim(str, trimChars)\n}" ]
[ "0.7941458", "0.7652819", "0.7532694", "0.74803454", "0.70558035", "0.6945625", "0.69270116", "0.6880997", "0.6880915", "0.6215482", "0.61146474", "0.60969275", "0.6040601", "0.589196", "0.5886939", "0.58238745", "0.57857376", "0.57648623", "0.57059145", "0.5652308", "0.5652308", "0.56168103", "0.55878055", "0.55846214", "0.5583379", "0.5579458", "0.5571917", "0.55547476", "0.55458987", "0.5528638", "0.5520801", "0.5515922", "0.55147254", "0.5504656", "0.54886174", "0.5434617", "0.54313844", "0.5427758", "0.5421728", "0.54056716", "0.54021406", "0.5391942", "0.5386882", "0.5331598", "0.5327009", "0.5323671", "0.5288556", "0.527587", "0.52733845", "0.52668923", "0.52615714", "0.5249284", "0.52478236", "0.5246093", "0.52449584", "0.5220424", "0.51953816", "0.517923", "0.51791036", "0.51607764", "0.5157695", "0.5141877", "0.5123625", "0.5119498", "0.5119498", "0.51115847", "0.5103538", "0.5099622", "0.5096853", "0.50725865", "0.5072389", "0.5061976", "0.5049717", "0.5049717", "0.50404906", "0.5037236", "0.5037062", "0.5029856", "0.50166184", "0.4988645", "0.4988645", "0.49640408", "0.49585363", "0.49575603", "0.49541196", "0.4938871", "0.4937749", "0.4936077", "0.49265915", "0.49184036", "0.49152234", "0.49075902", "0.49056682", "0.48947746", "0.48944566", "0.4887274", "0.48858187", "0.48809233", "0.48761132", "0.48739186" ]
0.83351445
0
TrimSpace uses strings.TrimSpace to remove all unicode whitespace codepoints from the beginning and end of operand.
func TrimSpace(operand string) string { return strings.TrimSpace(operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Trim(chars string, operand string) string { return strings.Trim(operand, chars) }", "func TrimLeft(chars string, operand string) string { return strings.TrimLeft(operand, chars) }", "func trimS(s string) string {\r\n\treturn strings.TrimSpace(s)\r\n}", "func ShouldEqualTrimSpace(actual interface{}, expected ...interface{}) error {\n\tactualS, err := cast.ToStringE(actual)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ShouldEqual(strings.TrimSpace(actualS), expected...)\n}", "func trimWhitespaces(value string) string {\n\treturn strings.Trim(value, \"\")\n}", "func trim(s string) string {\n\tfor len(s) > 0 {\n\t\tif s[0] <= ' ' {\n\t\t\ts = s[1:]\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tfor len(s) > 0 {\n\t\tif s[len(s)-1] <= ' ' {\n\t\t\ts = s[:len(s)-1]\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn s\n}", "func Strip(str string) string {\n\treturn regexp.MustCompile(`^\\s+|\\s+$`).ReplaceAllString(str, \"\")\n}", "func stripWhiteSpace(str string) string {\n\treturn strings.ReplaceAll(str, \" \", \"\")\n}", "func execTrimString(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := textproto.TrimString(args[0].(string))\n\tp.Ret(1, ret)\n}", "func Trim(str string) string {\n\treturn strings.TrimSpace(str)\n}", "func filterTrim(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.TrimSpace(stick.CoerceString(val))\n}", "func TrimRedundantSpaces(text string) string {\n\ttext = spaceRegex.ReplaceAllString(text, \" \")\n\treturn strings.TrimSpace(text)\n}", "func removeWhiteSpace(input string) string {\n\tif runtime.GOOS == \"windows\" {\n\t\tinput = strings.TrimRight(input, \"\\r\\n\")\n\t} else {\n\t\tinput = strings.TrimRight(input, \"\\n\")\n\t}\n\treturn input\n}", "func TrimTrailingSpaces(p []byte) []byte {\n\tfor i := len(p) - 1; i >= 0; i-- {\n\t\tif p[i] != 0x20 && p[i] != '\\n' && p[i] != '\\t' {\n\t\t\treturn p[:i+1]\n\t\t}\n\t}\n\t// it was all spaces\n\treturn p[:0]\n}", "func TrimAllSpaceAndNewline(input string) string {\n\toutput := input\n\tfor _, f := range []string{\"\\n\", \"\\t\", \" \"} {\n\t\toutput = strings.Replace(output, f, \"\", -1)\n\t}\n\n\treturn output\n}", "func TrimSpace(ctx context.Context, t *mold.Transformer, v reflect.Value) error {\n\tv.Set(reflect.ValueOf(strings.TrimSpace(v.String())))\n\treturn nil\n}", "func removeWhitespaces(value string) string {\n\treturn strings.ReplaceAll(value, \" \", \"\")\n}", "func TrimLeadingSpaces(p []byte) []byte {\n\tfor i := 0; i < len(p); i++ {\n\t\tif p[i] != 0x20 && p[i] != '\\t' {\n\t\t\treturn p[i:]\n\t\t}\n\t}\n\t// it was all spaces\n\treturn p[:0]\n}", "func trimTrailingWhiteSpace(s []byte) []byte {\n\tfor (len(s) > 0) && ((s[len(s)-1] == ' ') || (s[len(s)-1] == '\\t')) {\n\t\ts = s[:len(s)-1]\n\t}\n\treturn s\n}", "func StrTrim(s string) string {\n\treturn strings.TrimSpace(s)\n}", "func trim(strn string) string {\n\treturn strings.Trim(strn, \" \\t\\n\\r\")\n}", "func simpleTextStrip(s string) string {\n\ttext := fmt.Sprintf(\"%s\", s)\n\n\treplaces := map[string]string{\n\t\t\"\\n\": \" \",\n\t\t\"\\t\": \" \",\n\t\t\"\\\"\": \"'\",\n\t\t`\\u003c`: \"<\",\n\t\t`\\u003e`: \">\",\n\t}\n\n\tfor k, v := range replaces {\n\t\ttext = strings.Replace(text, k, v, -1)\n\t}\n\n\treturn strings.TrimSpace(text)\n}", "func trimLeadingWhiteSpace(s []byte) []byte {\n\tfor (len(s) > 0) && ((s[0] == ' ') || (s[0] == '\\t')) {\n\t\ts = s[1:]\n\t}\n\treturn s\n}", "func normalizeStr(v string) string {\n\tv = strings.TrimSpace(v)\n\tv = regexp.MustCompile(`[^\\S\\r\\n]+`).ReplaceAllString(v, \" \")\n\tv = regexp.MustCompile(`[\\r\\n]+`).ReplaceAllString(v, \"\\n\")\n\n\treturn v\n}", "func trim(s []byte) []byte {\n\ti := 0\n\tfor i < len(s) && (s[i] == ' ' || s[i] == '\\t') {\n\t\ti++\n\t}\n\tn := len(s)\n\tfor n > i && (s[n-1] == ' ' || s[n-1] == '\\t') {\n\t\tn--\n\t}\n\treturn s[i:n]\n}", "func trim(s []byte) []byte {\n\ti := 0\n\tfor i < len(s) && (s[i] == ' ' || s[i] == '\\t') {\n\t\ti++\n\t}\n\tn := len(s)\n\tfor n > i && (s[n-1] == ' ' || s[n-1] == '\\t') {\n\t\tn--\n\t}\n\treturn s[i:n]\n}", "func trimWS(input string) string {\n\n\tvar out strings.Builder\n\tfor _, v := range input {\n\t\tif !(v == '\\u0009' || v == '\\u0020' || v == '\\u000A' || v == '\\u000D' || v == ',') {\n\t\t\tout.WriteRune(v)\n\t\t}\n\t}\n\treturn out.String()\n\n}", "func (h *Helpers) Strip(s string) string {\n\tre1 := regexp.MustCompile(stripEmptyRegex)\n\tre2 := regexp.MustCompile(stripEdgesRegex)\n\ts = re1.ReplaceAllString(s, \"\")\n\ts = re2.ReplaceAllString(s, \"\")\n\treturn s\n}", "func TrimPrefix(prefix, operand string) string { return strings.TrimPrefix(operand, prefix) }", "func stripTrailingSpace(str string) string {\n\tbuf := bytes.NewBuffer(nil)\n\n\tscan := bufio.NewScanner(strings.NewReader(str))\n\tfor scan.Scan() {\n\t\tbuf.WriteString(strings.TrimRight(scan.Text(), \" \\t\\r\\n\"))\n\t\tbuf.WriteString(\"\\n\")\n\t}\n\treturn buf.String()\n}", "func TestTrim(t *testing.T) {\n\ttext := \"Hola Mundo TDA\"\n\tt.Logf(\"text:[%s]\", text)\n\tt.Logf(\"trim:[%s]\", utl.Trim(text))\n}", "func trimLeadingWhiteSpace(buf string) string {\n\treturn strings.TrimLeft(buf, \" \")\n}", "func (s *Stringish) TrimSpaces() *Stringish {\n\ts.str = strings.TrimSpace(s.str)\n\treturn s\n}", "func trimStringN(c *C.char, l C.int) string {\n\tvar rc string\n\ts := C.GoStringN(c, l)\n\ti := strings.IndexByte(s, 0)\n\tif i == -1 {\n\t\trc = s\n\t} else {\n\t\trc = s[0:i]\n\t}\n\treturn strings.TrimSpace(rc)\n}", "func TrimEleSpaceAndRemoveEmpty(vs []string) (r []string) {\n\tfor _, v := range vs {\n\t\tv = strings.TrimSpace(v)\n\t\tif v != \"\" {\n\t\t\tr = append(r, v)\n\t\t}\n\t}\n\n\treturn\n}", "func (s *Str) TrimSpaces() *Str {\n\ts.val = strings.TrimSpace(s.val)\n\treturn s\n}", "func execTrimBytes(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := textproto.TrimBytes(args[0].([]byte))\n\tp.Ret(1, ret)\n}", "func TrimSpaceAndQuotes(answer string) string {\n\ttext := strings.TrimSpace(answer)\n\tfor _, q := range quotes {\n\t\tif strings.HasPrefix(text, q) && strings.HasSuffix(text, q) {\n\t\t\treturn strings.TrimPrefix(strings.TrimSuffix(text, q), q)\n\t\t}\n\t}\n\treturn text\n}", "func Trim(str, chars string) string {\n\treturn LeftTrim(RightTrim(str, chars), chars)\n}", "func StripAll(s string) string {\n\treturn strings.Map(func(r rune) rune {\n\t\tif r < ' ' || r == zwsp {\n\t\t\treturn -1\n\t\t}\n\n\t\treturn r\n\t}, s)\n}", "func TrimTrailingSpaces(text string) string {\n\tparts := strings.Split(text, \"\\n\")\n\tfor i := range parts {\n\t\tparts[i] = strings.TrimRightFunc(parts[i], func(r rune) bool {\n\t\t\treturn unicode.IsSpace(r)\n\t\t})\n\n\t}\n\n\treturn strings.Join(parts, \"\\n\")\n}", "func Trim(str string) string {\n\treturn strings.Trim(str, \" \")\n}", "func CutTrimSpace(s, sep string) (before, after string, found bool) {\n\tif i := strings.Index(s, sep); i >= 0 {\n\t\treturn strings.TrimSpace(s[:i]), strings.TrimSpace(s[i+len(sep):]), true\n\t}\n\treturn s, \"\", false\n}", "func removeSpaces(s string) string {\n\tvar ss = regSpaces.ReplaceAllString(s, \" \")\n\treturn strings.Trim(ss, \" \")\n}", "func StringStrip(scope *Scope, input tf.Output) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"StringStrip\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func trimSpace(b []byte) string {\n\treturn string(bytes.TrimRight(b, \" \"))\n}", "func trimSpace(b []byte) string {\n\treturn string(bytes.TrimRight(b, \" \"))\n}", "func trimTitle(title string) string {\n\treturn strings.TrimSpace(title)\n}", "func normalizeWhitespace(x string) string {\n\tx = strings.Join(strings.Fields(x), \" \")\n\tx = strings.Replace(x, \"( \", \"(\", -1)\n\tx = strings.Replace(x, \" )\", \")\", -1)\n\tx = strings.Replace(x, \")->\", \") ->\", -1)\n\treturn x\n}", "func stripString(s string) (o string) {\n\to = RegularExp.ReplaceAllString(s, \"\")\n\tlog.Printf(\" String: %s\", o)\n\to = strings.ToLower(o)\n\treturn\n}", "func TrimLeftChar(s string) string {\n\tfor i := range s {\n\t\tif i > 0 {\n\t\t\treturn s[i:]\n\t\t}\n\t}\n\treturn s[:0]\n}", "func TrimNL(s string) string {\n\ts = strings.TrimSuffix(s, \"\\n\")\n\ts = strings.TrimSuffix(s, \"\\r\")\n\n\treturn s\n}", "func trimLeadingWhiteSpaceAndNewLines(s []byte) []byte {\n\tfor (len(s) > 0) && ((s[0] == ' ') || (s[0] == '\\t') || (s[0] == '\\n')) {\n\t\ts = s[1:]\n\t}\n\treturn s\n}", "func CPUClean(str string, args ...string) string {\n\tfor _, arg := range args {\n\t\tstr = strings.Replace(str, arg, \"\", -1)\n\t}\n\tstr = strings.TrimSpace(str)\n\treturn str\n}", "func ToTrimmedString(in interface{}) string {\n\tif str, ok := in.(string); ok {\n\t\treturn strings.TrimSpace(str)\n\t}\n\treturn \"\"\n}", "func TrimAll(s string) string {\n\treturn strings.Join(strings.Fields(s), \" \")\n}", "func stripSpaces(str string) string {\n\treturn strings.Map(func(r rune) rune {\n\t\tif unicode.IsSpace(r) {\n\t\t\treturn -1\n\t\t}\n\t\treturn r\n\t}, str)\n}", "func stripSpaces(str string) string {\n\treturn strings.Map(func(r rune) rune {\n\t\tif unicode.IsSpace(r) {\n\t\t\t// if the character is a space, drop it\n\t\t\treturn -1\n\t\t}\n\t\t// else keep it in the string\n\t\treturn r\n\t}, str)\n}", "func RemoveSpaces(in []byte) (out []byte) {\n\tvar c byte\n\tout = make([]byte, 0, len(in))\n\tfor _, c = range in {\n\t\tif ascii.IsSpace(c) {\n\t\t\tcontinue\n\t\t}\n\t\tout = append(out, c)\n\t}\n\treturn out\n}", "func Clean_up(input_str string) string {\n \n\t// compile regular expression\n\t// to match 2-or-more-than-2 whitespaces\n\tvar validID = regexp.MustCompile(`\\s{2,}`)\n\tvar temp_str string = validID.ReplaceAllString(input_str, \" \")\n\tvar temp_arr []string = strings.Split(temp_str, \" \")\n\n\t// pop off the last unnecessary character\n\tif temp_arr[0] == \"\" || temp_arr[0] == \" \" {\n\t\ttemp_arr = temp_arr[1:]\n\t}\n\n\t// delete the first unnessary character\n\tif temp_arr[len(temp_arr)-1] == \"\" || temp_arr[len(temp_arr)-1] == \" \" {\n\t\ttemp_arr = temp_arr[:len(temp_arr)-1]\n\t}\n\n\t// return in string format\n\treturn strings.Join(temp_arr, \" \")\n}", "func Trim(text string) string {\n\treturn trimRx.FindStringSubmatch(text)[1]\n}", "func DeleteWhiteSpace(str string) string {\n\tif str == \"\" {\n\t\treturn str\n\t}\n\tsz := len(str)\n\tvar chs bytes.Buffer\n\tcount := 0\n\tfor i := 0; i < sz; i++ {\n\t\tch := rune(str[i])\n\t\tif !unicode.IsSpace(ch) {\n\t\t\tchs.WriteRune(ch)\n\t\t\tcount++\n\t\t}\n\t}\n\tif count == sz {\n\t\treturn str\n\t}\n\treturn chs.String()\n}", "func TrimRight(cutset string, operand string) string { return strings.TrimRight(operand, cutset) }", "func TrimExtraSpaces(s string) string {\n\tspace := regexp.MustCompile(`\\s+`)\n\ts = space.ReplaceAllString(s, \" \")\n\ts = strings.TrimPrefix(s, \" \")\n\ts = strings.TrimSuffix(s, \" \")\n\treturn s\n}", "func TrimSpaceNewline(str string) string {\n\tstr = strings.TrimSpace(str)\n\treturn strings.Trim(str, \"\\r\\n\")\n}", "func removeBlanks(s string) string {\n\tregex, err := regexp.Compile(\"\\n$\")\n\tif err != nil {\n\t\tFatalf(\"Failure in removing blank lines, err: %v\", err)\n\t}\n\ts = strings.TrimSpace(regex.ReplaceAllString(s, \"\"))\n\treturn s\n}", "func trimChar(s string, r byte) string {\n\tsz := len(s)\n\n\tif sz > 0 && s[sz-1] == r {\n\t\ts = s[:sz-1]\n\t}\n\tsz = len(s)\n\tif sz > 0 && s[0] == r {\n\t\ts = s[1:sz]\n\t}\n\treturn s\n}", "func trimChar(s string, r byte) string {\n\tsz := len(s)\n\n\tif sz > 0 && s[sz-1] == r {\n\t\ts = s[:sz-1]\n\t}\n\tsz = len(s)\n\tif sz > 0 && s[0] == r {\n\t\ts = s[1:sz]\n\t}\n\treturn s\n}", "func trimRight(s string) string {\n\treturn strings.TrimRightFunc(s, unicode.IsSpace)\n}", "func removeSpace(s string) string {\n\trunes := []rune(s)\n\tout := runes[:0] // zero-length slice of original\n\tfor i := 0; i < len(runes); i++ {\n\t\tif i == 0 {\n\t\t\tout = append(out, runes[i])\n\t\t} else if !unicode.IsSpace(runes[i]) || !unicode.IsSpace(runes[i-1]) {\n\t\t\tout = append(out, runes[i])\n\t\t}\n\t}\n\treturn string(out)\n}", "func Strip(s string) string {\n\treturn strip.ReplaceAllString(s, \"\")\n}", "func Strip(s string) string {\n\treturn strip.ReplaceAllString(s, \"\")\n}", "func cleancompstr(cleanup string) string {\n\tcleanup = strings.TrimSpace(cleanup)\n\tcleanup = strings.ToLower(cleanup)\n\treturn cleanup\n}", "func signV4TrimAll(input string) string {\n\t// Compress adjacent spaces (a space is determined by\n\t// unicode.IsSpace() internally here) to one space and return\n\treturn strings.Join(strings.Fields(input), \" \")\n}", "func trimAdjacentSpaces(bs []byte) []byte {\n\tout := bs[:0] // A new output slice based upon the original.\n\trunes := bytes.Runes(bs) // Turning the byte slice into runes guarantees things will work with any characters (e.g. emojis).\n\tindex := 0 // Keep a running index so we know exactly how many characters we've inserted.\n\n\tfor _, char := range runes {\n\t\tif unicode.IsSpace(char) {\n\t\t\tif index > 0 && unicode.IsSpace(runes[index-1]) {\n\t\t\t\t// We're in multiple spaces here, so we don't need to append anything.\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tout = append(out, ' ') // Append a single space to the output instead of multiple.\n\t\t\t\tindex++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\t// Here we turn the rune into a string, and then the string into a byte slice so we can append it to the output.\n\t\tfor _, b := range []byte(string(char)) {\n\t\t\tout = append(out, b)\n\t\t\tindex++\n\t\t}\n\t}\n\n\treturn out[:index]\n}", "func DeleteEmptyAndTrim(s []string) []string {\n\tvar r []string\n\tfor _, str := range s {\n\t\tstr = strings.TrimSpace(str)\n\t\tif str != \"\" {\n\t\t\tr = append(r, str)\n\t\t}\n\t}\n\treturn r\n}", "func trimmed(bs []byte) []byte {\n\tfor i, b := range bs {\n\t\tif b == 0x00 {\n\t\t\treturn bs[:i]\n\t\t}\n\t}\n\treturn bs\n}", "func unescapeValueAfterGet(value string) string {\n\tif len(value) == 0 {\n\t\treturn value\n\t}\n\n\trunes := []rune(value)\n\tfor _, ru := range runes {\n\t\tif ru != rune(' ') {\n\t\t\treturn value\n\t\t}\n\t}\n\treturn string(runes[0: len(runes)-1])\n}", "func CleanString(str string, unescape bool) string {\n\tif unescape {\n\t\tstr = html.UnescapeString(str)\n\t}\n\tstr = strings.Replace(str, \"\\n\", \"\", -1)\n\tstr = strings.Replace(str, \"\\r\", \"\", -1)\n\treturn strings.TrimFunc(str, unicode.IsSpace)\n}", "func stripMotd(motd string) string {\n\treturn regexp.MustCompile(\"\\\\s+\").ReplaceAllString(regexp.MustCompile(\"`+.\").ReplaceAllString(\n\t\tregexp.MustCompile(\"[^\\x20-\\x7f]\").ReplaceAllString(\n\t\t\tmotd, \"`\"),\n\t\t\"\"),\n\t\t\" \")\n}", "func (reader *Reader) IsTrimSpace() bool {\n\treturn reader.TrimSpace\n}", "func RemoveUnwantedSpaces(str string) string {\n\treturn spaceRemover.ReplaceAllString(str, \" \")\n}", "func (s *Stringish) TrimPrefixSpaces() *Stringish {\n\treturn s.TrimPrefix(\" \")\n}", "func LeftTrim(str, chars string) string {\n\tif chars == \"\" {\n\t\treturn strings.TrimLeftFunc(str, unicode.IsSpace)\n\t}\n\tr, _ := regexp.Compile(\"^[\" + chars + \"]+\")\n\treturn r.ReplaceAllString(str, \"\")\n}", "func (p *Parser) Trim(input string) string {\n\tt := strings.Replace(strings.Replace(input, \"<output>\", \"\", -1), \"</output>\", \"\", -1)\n\tt1 := strings.Replace(strings.Replace(t, \"<configuration-information>\", \"\", -1), \"</configuration-information>\", \"\", -1)\n\treturn strings.Replace(strings.Replace(t1, \"<configuration-output>\", \"\", -1), \"</configuration-output>\", \"\", -1)\n}", "func trim(name string) string {\n\treturn strings.TrimPrefix(name, Prefix)\n}", "func SpaceFilter(input string) string {\n\treturn strings.TrimSpace(input)\n}", "func UnicodeSanitize(s string) string {\n\tsource := []rune(s)\n\ttarget := make([]rune, 0, len(source))\n\n\tfor i, r := range source {\n\t\tif r == '%' && i+2 < len(source) && ishex(source[i+1]) && ishex(source[i+2]) {\n\t\t\ttarget = append(target, r)\n\t\t} else if unicode.IsLetter(r) || unicode.IsDigit(r) || unicode.IsMark(r) || r == '.' || r == '/' || r == '\\\\' || r == '_' || r == '-' || r == '#' || r == '+' || r == '~' {\n\t\t\ttarget = append(target, r)\n\t\t}\n\t}\n\n\tvar result string\n\n\tif true {\n\t\t// remove accents - see https://blog.golang.org/normalization\n\t\tt := transform.Chain(norm.NFD, transform.RemoveFunc(isMn), norm.NFC)\n\t\tresult, _, _ = transform.String(t, string(target))\n\t} else {\n\t\tresult = string(target)\n\t}\n\n\treturn result\n}", "func trimTrailingWhitespace(details string) string {\n\treturn strings.TrimSuffix(details, \" \")\n}", "func TrimChars(p projection, chars string) *trimFunc {\n\treturn &trimFunc{\n\t\tsubject: p.(element),\n\t\tsel: p.from(),\n\t\tlocation: TRIM_BOTH,\n\t\tchars: chars,\n\t}\n}", "func CleanString(str string) string {\n\treturn strings.Join(strings.Fields(strings.TrimSpace(str)), \" \")\n}", "func CleanString(str string) string {\n\treturn strings.Join(strings.Fields(strings.TrimSpace(str)), \" \")\n}", "func CleanString(str string) string {\n\treturn strings.Join(strings.Fields(strings.TrimSpace(str)), \" \")\n}", "func CleanString(str string) string {\n\treturn strings.Join(strings.Fields(strings.TrimSpace(str)), \" \")\n}", "func stripSlash(str string) string {\n\treturn strings.TrimRight(strings.TrimLeft(str, \"/\"), \"/\")\n}", "func Clean (current string) string {\n\n // Remove unicode characters.\n formatted := html2text.HTML2Text(current)\n // Remove specific characters.\n formatted = strings.Replace(formatted, \"\\n\", \" \", -1)\n formatted = strings.Replace(formatted, \"\\r\", \" \", -1)\n\n // Final unicode removal (specific to code sections).\n return html2text.HTML2Text(formatted)\n}", "func (chars *Chars) TrimLength() uint16 {\n\tif chars.trimLengthKnown {\n\t\treturn chars.trimLength\n\t}\n\tchars.trimLengthKnown = true\n\tvar i int\n\tlen := chars.Length()\n\tfor i = len - 1; i >= 0; i-- {\n\t\tchar := chars.Get(i)\n\t\tif !unicode.IsSpace(char) {\n\t\t\tbreak\n\t\t}\n\t}\n\t// Completely empty\n\tif i < 0 {\n\t\treturn 0\n\t}\n\n\tvar j int\n\tfor j = 0; j < len; j++ {\n\t\tchar := chars.Get(j)\n\t\tif !unicode.IsSpace(char) {\n\t\t\tbreak\n\t\t}\n\t}\n\tchars.trimLength = AsUint16(i - j + 1)\n\treturn chars.trimLength\n}", "func StringOutCharset(a, b string) bool { return len(StringTrim(a, b)) > 0 }", "func ScanlineTrim() string {\n\treturn strings.TrimSpace(Scanline())\n}", "func removeExtraSpaces(val string) string {\n\treturn strings.Join(strings.Fields(val), \" \")\n}" ]
[ "0.7410103", "0.68665975", "0.6548592", "0.6355161", "0.62889844", "0.62683105", "0.6243955", "0.6243547", "0.6224234", "0.61773384", "0.617549", "0.61590344", "0.61192745", "0.60075575", "0.60062546", "0.59572494", "0.59486145", "0.5935191", "0.5934588", "0.5899888", "0.5875586", "0.5845168", "0.5835582", "0.5827005", "0.5815971", "0.5815971", "0.581582", "0.5793755", "0.57628816", "0.57153296", "0.57133645", "0.57112736", "0.5699708", "0.5681141", "0.56499064", "0.5639581", "0.56395215", "0.5633957", "0.56239784", "0.5620285", "0.5608411", "0.558599", "0.55858094", "0.55851984", "0.5574701", "0.5570182", "0.5570182", "0.5557171", "0.55499965", "0.55186373", "0.549539", "0.5492238", "0.5481448", "0.5478939", "0.54559803", "0.5437421", "0.54360217", "0.5431989", "0.5431873", "0.54219013", "0.5415518", "0.5414349", "0.54102975", "0.5410087", "0.5397602", "0.5393139", "0.5390324", "0.5390324", "0.53378725", "0.53333145", "0.53289336", "0.53289336", "0.5314002", "0.53103817", "0.5304947", "0.5301658", "0.5300821", "0.5300003", "0.5295169", "0.52914643", "0.52826023", "0.528062", "0.52754074", "0.52728754", "0.5263752", "0.52571595", "0.5254173", "0.52406895", "0.5235649", "0.5234545", "0.5228557", "0.5228557", "0.5228557", "0.5228557", "0.52079123", "0.5207094", "0.52006525", "0.5194233", "0.5191025", "0.51703477" ]
0.75475335
0
TrimSuffix uses strings.TrimSuffix to remove suffix from the end of operand.
func TrimSuffix(suffix, operand string) string { return strings.TrimSuffix(operand, suffix) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Stringish) TrimSuffix(suffix string) *Stringish {\n\ts.str = strings.TrimSuffix(s.str, suffix)\n\treturn s\n}", "func TrimPrefix(prefix, operand string) string { return strings.TrimPrefix(operand, prefix) }", "func removeSuffix(value, suffix string) string {\n\treturn value[0 : len(value)-len(suffix)]\n}", "func Trim(chars string, operand string) string { return strings.Trim(operand, chars) }", "func TrimSuffix(suffix string) MapFunc {\n\treturn func(s string) string { return strings.TrimSuffix(s, suffix) }\n}", "func (s *Stringish) TrimSuffixSpaces() *Stringish {\n\treturn s.TrimSuffix(\" \")\n}", "func trim(name string) string {\n\treturn strings.TrimPrefix(name, Prefix)\n}", "func TrimSpace(operand string) string { return strings.TrimSpace(operand) }", "func TrimRight(cutset string, operand string) string { return strings.TrimRight(operand, cutset) }", "func TrimByteSuffix(str string, suffix byte) string {\n\tif len(str) > 0 && str[len(str)-1] == suffix {\n\t\tstr = str[:len(str)-1]\n\t}\n\treturn str\n}", "func TestTrimKafkaChannelServiceNameSuffix(t *testing.T) {\n\n\t// Test Data\n\tchannelName := \"TestChannelName\"\n\tchannelServiceName := fmt.Sprintf(\"%s-%s\", channelName, constants.KafkaChannelServiceNameSuffix)\n\n\t// Perform The Test\n\tactualResult := TrimKafkaChannelServiceNameSuffix(channelServiceName)\n\n\t// Verify The Results\n\texpectedResult := channelName\n\tassert.Equal(t, expectedResult, actualResult)\n}", "func HasSuffix(suffix, operand string) bool { return strings.HasSuffix(operand, suffix) }", "func trimDot(s string) string {\n\treturn strings.Trim(s, \".\")\n}", "func Trim(text string) string {\n\treturn trimRx.FindStringSubmatch(text)[1]\n}", "func trimLast(text string) string {\n\ttextLen := len(text)\n\tif textLen == 0 {\n\t\treturn text\n\t}\n\treturn text[:textLen-1]\n}", "func trimTrailingWhitespace(details string) string {\n\treturn strings.TrimSuffix(details, \" \")\n}", "func trimOutPrefix(str, prefix string) string {\n\ti := strings.Index(str, prefix)\n\tif i < 0 {\n\t\treturn str\n\t}\n\treturn str[i+len(prefix):]\n}", "func WithoutSuffix(filename string) string {\n\tsuffix := Suffix(filename)\n\tif suffix == \"\" {\n\t\treturn filename\n\t}\n\treturn filename[0 : len(filename)-len(suffix)-1]\n}", "func (su StringUtility) TrimEndMultiple(targetStr string, trimChar rune) (rStr string, err error) {\n\n\tif targetStr == \"\" {\n\t\terr = errors.New(\"Empty targetStr\")\n\t\treturn\n\t}\n\n\tfStr := []rune(targetStr)\n\tlenTargetStr := len(fStr)\n\toutputStr := make([]rune, lenTargetStr)\n\tlenTargetStr--\n\tidx := lenTargetStr\n\tfoundFirstChar := false\n\n\tfor i := lenTargetStr; i >= 0; i-- {\n\n\t\tif !foundFirstChar && fStr[i] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tif i > 0 && fStr[i] == trimChar && fStr[i-1] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tif i == 0 && fStr[i] == trimChar {\n\t\t\tcontinue\n\t\t}\n\n\t\tfoundFirstChar = true\n\t\toutputStr[idx] = fStr[i]\n\t\tidx--\n\t}\n\n\tif idx != lenTargetStr {\n\t\tidx++\n\t}\n\n\tif outputStr[idx] == trimChar {\n\t\tidx++\n\t}\n\n\tresult := string(outputStr[idx:])\n\n\treturn result, nil\n\n}", "func truncateSuffix(value string, suffix string, max int) (string, error) {\n\tvalLen := len(value)\n\tsufLen := len(suffix)\n\tif valLen+sufLen <= max {\n\t\treturn value, nil\n\t}\n\tif sufLen+1 >= max {\n\t\treturn errorMessage[0].val, fmt.Errorf(\"sufixo [%s] nao pode ser aplicado porque estoura \"+\n\t\t\t\"o tamanho maximo [%d] no elemento [%s]\", suffix, max, value)\n\t}\n\tr := []rune(value)\n\tif l := len(r); max > l {\n\t\tmax = l\n\t}\n\tfor l := max - 1; l >= 0 && r[l] == '_'; l-- {\n\t\tmax--\n\t}\n\tsafeSubstring := string(r[0:max])\n\treturn testInvalidChars(safeSubstring)\n}", "func Suffix() string {\n\treturn std.Suffix()\n}", "func Suffix(filename string) string {\n\tname := path.Base(filename) // Make sure we only have filename, not full path\n\n\t// Many files in aspera_test have suffixes like \"mxf.xml\" or \"mov.aspx\" or similar.\n\t// We need to check the full siffix, not only the last part.\n\tindex := strings.Index(name, \".\")\n\tif index < 0 || index == len(name)-1 {\n\t\treturn \"\"\n\t}\n\treturn strings.ToLower(name[index+1:])\n}", "func (n Name) Suffix() Name {\n\tif len(n) > 0 {\n\t\treturn n[1:]\n\t}\n\treturn nil\n}", "func trimTrailingPeriod(v interface{}) string {\n\tvar str string\n\tswitch value := v.(type) {\n\tcase *string:\n\t\tstr = aws.StringValue(value)\n\tcase string:\n\t\tstr = value\n\tdefault:\n\t\treturn \"\"\n\t}\n\n\tif str == \".\" {\n\t\treturn str\n\t}\n\n\treturn strings.TrimSuffix(str, \".\")\n}", "func trimToPrefix(str, prefix string) string {\n\ti := strings.Index(str, prefix)\n\tif i < 0 {\n\t\treturn str\n\t}\n\treturn str[i:]\n}", "func (o BucketNotificationQueueOutput) FilterSuffix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketNotificationQueue) *string { return v.FilterSuffix }).(pulumi.StringPtrOutput)\n}", "func (o BucketNotificationLambdaFunctionOutput) FilterSuffix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketNotificationLambdaFunction) *string { return v.FilterSuffix }).(pulumi.StringPtrOutput)\n}", "func TrimLeft(chars string, operand string) string { return strings.TrimLeft(operand, chars) }", "func trimRight(r *syntax.Regexp) (bool, *syntax.Regexp) {\n\tif eqSuffixAnyRegex(r, patDotStar, patNullEndDotStar) {\n\t\ti := len(r.Sub) - 1\n\t\ttmp := *r\n\t\ttmp.Sub = tmp.Sub[0:i]\n\t\treturn true, &tmp\n\t}\n\n\treturn false, r\n}", "func filterTrim(ctx stick.Context, val stick.Value, args ...stick.Value) stick.Value {\n\treturn strings.TrimSpace(stick.CoerceString(val))\n}", "func TrimSpaceAndQuotes(answer string) string {\n\ttext := strings.TrimSpace(answer)\n\tfor _, q := range quotes {\n\t\tif strings.HasPrefix(text, q) && strings.HasSuffix(text, q) {\n\t\t\treturn strings.TrimPrefix(strings.TrimSuffix(text, q), q)\n\t\t}\n\t}\n\treturn text\n}", "func trim(strn string) string {\n\treturn strings.Trim(strn, \" \\t\\n\\r\")\n}", "func (b DeleteBuilder) Suffix(sql string, args ...interface{}) WhereConditions {\n\treturn builder.Append(b, \"Suffixes\", Expr(sql, args...)).(DeleteBuilder)\n}", "func parseSuffix(s string) (string, int64, int64) {\n\tfor _, v := range uomSuffixes {\n\t\tif strings.HasSuffix(s, v.unit) {\n\t\t\treturn s[0 : len(s)-len(v.unit)], v.mul, v.div\n\t\t}\n\t}\n\n\treturn s, 1, 1\n}", "func trim_string_after_s(s string, x string) (r string) {\n\tif idx := strings.Index(s, x); idx != -1 {\n\t\treturn s[:idx]\n\t}\n\treturn s\n}", "func trimRight(s string) string {\n\treturn strings.TrimRightFunc(s, unicode.IsSpace)\n}", "func stripZone(zone string) string {\n\treturn strings.TrimRight(zone, \"abc\")\n}", "func split(n string) (string, string) {\n\tfor _, s := range allSuffixes {\n\t\tif strings.HasSuffix(n, s) {\n\t\t\tp := strings.TrimSuffix(n, s)\n\t\t\treturn p, s\n\t\t}\n\t}\n\treturn n, \"\"\n}", "func suffixReplacer(suffix, replace string) NameMapper {\n\treturn func(s string) string {\n\t\ttrimmed := strings.TrimSuffix(s, suffix)\n\n\t\tif trimmed != s {\n\t\t\treturn trimmed + replace\n\t\t}\n\n\t\treturn s\n\t}\n}", "func (o BucketNotificationTopicOutput) FilterSuffix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketNotificationTopic) *string { return v.FilterSuffix }).(pulumi.StringPtrOutput)\n}", "func TestTrim(t *testing.T) {\n\ttext := \"Hola Mundo TDA\"\n\tt.Logf(\"text:[%s]\", text)\n\tt.Logf(\"trim:[%s]\", utl.Trim(text))\n}", "func (n *Name) Suffix() string {\n\t// TODO handle gender\n\treturn n.pick(namePrefix + \"/suffix\")\n}", "func stringSuffixFunc(s string, f func(c rune) bool) (suff string) {\n if i := stringSuffixIndexFunc(s, f); i >= 0 {\n suff = s[i:]\n }\n return\n}", "func (b *Builder) HasSuffix(rhs interface{}) *predicate.Predicate {\n\tb.p.RegisterPredicate(impl.HasSuffix(rhs))\n\tif b.t != nil {\n\t\tb.t.Helper()\n\t\tEvaluate(b)\n\t}\n\treturn &b.p\n}", "func execTrimString(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := textproto.TrimString(args[0].(string))\n\tp.Ret(1, ret)\n}", "func SuffixTransform(b []byte) []byte { return append([]byte(\".\"), b...) }", "func RightTrim(str, chars string) string {\n\tif chars == \"\" {\n\t\treturn strings.TrimRightFunc(str, unicode.IsSpace)\n\t}\n\tr, _ := regexp.Compile(\"[\" + chars + \"]+$\")\n\treturn r.ReplaceAllString(str, \"\")\n}", "func Suffix() (string, error) {\n\tslice, err := loader(\"name_suffixes\")\n\tcheckErr(err)\n\treturn random(slice), nil\n}", "func (q *queryImpl) Suffix(sql string, args ...interface{}) Query {\n\tq.suffixes = append(q.suffixes, ExprString{sql, args})\n\treturn q\n}", "func (o *SignalPersonName) UnsetSuffix() {\n\to.Suffix.Unset()\n}", "func stripEnds(tk token.Token) token.Token {\n\ts := tk.Text()[1 : len(tk.Text())-1]\n\treturn token.UpdateText(tk, s)\n}", "func RemoveEnd(str string, remove string) string {\n\tif IsEmpty(str) || IsEmpty(remove) {\n\t\treturn str\n\t}\n\tif EndsWith(str, remove) {\n\t\treturn str[:len(str)-len(remove)]\n\t}\n\treturn str\n}", "func StripEnd(str string) string {\n\treturn regexp.MustCompile(`\\s+$`).ReplaceAllString(str, \"\")\n}", "func trim(s string, n int) string {\n\tif len(s) > n {\n\t\treturn s[:n]\n\t}\n\treturn s\n}", "func trim(s string, n int) string {\n\tif len(s) > n {\n\t\treturn s[:n]\n\t}\n\treturn s\n}", "func (p *Parser) Trim(input string) string {\n\tt := strings.Replace(strings.Replace(input, \"<output>\", \"\", -1), \"</output>\", \"\", -1)\n\tt1 := strings.Replace(strings.Replace(t, \"<configuration-information>\", \"\", -1), \"</configuration-information>\", \"\", -1)\n\treturn strings.Replace(strings.Replace(t1, \"<configuration-output>\", \"\", -1), \"</configuration-output>\", \"\", -1)\n}", "func (seg *Segmenter) CutTrim(str string, hmm ...bool) []string {\n\ts := seg.Cut(str, hmm...)\n\treturn seg.Trim(s)\n}", "func extractSuffix() string {\n\tc := obtainConfig()\n\tprefix := ExtractDNSPrefix()\n\turl := c.Clusters[prefix].Server\n\tsuffix := url[strings.Index(url, \".\"):]\n\tif strings.Contains(suffix, \":\") {\n\t\tsuffix = suffix[:strings.Index(suffix, \":\")]\n\t}\n\treturn suffix\n}", "func hasSuffixDemo(a string, b string) bool {\n\treturn strings.HasSuffix(a, b)\n}", "func BaselineMeasureDenomUnitsHasSuffix(v string) predicate.BaselineMeasureDenom {\n\treturn predicate.BaselineMeasureDenom(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldBaselineMeasureDenomUnits), v))\n\t})\n}", "func (m *Immutable) Suffix() string {\n\treturn m.contained.Suffix()\n}", "func (ub *UpdateBuilder) Suffix(\n\tsql string,\n\targs ...interface{},\n) *UpdateBuilder {\n\tub.sql = ub.sql.Suffix(sql, args...)\n\treturn ub\n}", "func (q *Query) Suffix(sql string, args ...interface{}) *Query {\n\tq.suffixes = append(q.suffixes, Expr(sql, args...))\n\treturn q\n}", "func fileNameWOExt(filePath string) string {\n\tfileName := filepath.Base(filePath)\n\treturn strings.TrimSuffix(fileName, filepath.Ext(fileName))\n}", "func (b UpdateBuilder) Suffix(sql string, args ...interface{}) WhereConditions {\n\treturn builder.Append(b, \"Suffixes\", Expr(sql, args...)).(UpdateBuilder)\n}", "func (s *Str) Trim(cutset string) *Str {\n\ts.val = strings.Trim(s.val, cutset)\n\treturn s\n}", "func (o BucketWebsiteConfigurationV2IndexDocumentOutput) Suffix() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketWebsiteConfigurationV2IndexDocument) string { return v.Suffix }).(pulumi.StringOutput)\n}", "func LastNameHasSuffix(v string) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldLastName), v))\n\t})\n}", "func (fn *FakeName) Suffix() string {\n\treturn random.PickString(fn.Suffixs)\n}", "func trimCut(label []byte, maxlen int, left, right string) []byte {\n\ttrim := bytes.TrimLeft(label, left)\n\tsize := min(len(trim), maxlen)\n\thead := bytes.TrimRight(trim[:size], right)\n\tif len(head) == size {\n\t\treturn head\n\t}\n\ttail := bytes.TrimLeft(trim[size:], right)\n\tif len(tail) > 0 {\n\t\treturn append(head, tail[:min(len(tail), size-len(head))]...)\n\t}\n\treturn head\n}", "func TrimTrailingSpaces(text string) string {\n\tparts := strings.Split(text, \"\\n\")\n\tfor i := range parts {\n\t\tparts[i] = strings.TrimRightFunc(parts[i], func(r rune) bool {\n\t\t\treturn unicode.IsSpace(r)\n\t\t})\n\n\t}\n\n\treturn strings.Join(parts, \"\\n\")\n}", "func LastnameHasSuffix(v string) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldLastname), v))\n\t})\n}", "func (o BucketLifecycleRuleItemConditionPtrOutput) MatchesSuffix() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *BucketLifecycleRuleItemCondition) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.MatchesSuffix\n\t}).(pulumi.StringArrayOutput)\n}", "func (o *SignalPersonName) SetSuffix(v string) {\n\to.Suffix.Set(&v)\n}", "func TrimExtension(file FileEntry, extensions []string) (new FileEntry) {\n\tnew = file\n\tfor _, ext := range extensions {\n\t\tif file.Extension == ext {\n\t\t\tnew.Name = strings.TrimSuffix(file.Name, ext)\n\t\t\tnew.Extension = filepath.Ext(new.Name)\n\t\t}\n\t}\n\treturn\n}", "func MannerNameHasSuffix(v string) predicate.Manner {\n\treturn predicate.Manner(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldMannerName), v))\n\t})\n}", "func (o BucketWebsiteConfigurationV2IndexDocumentPtrOutput) Suffix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketWebsiteConfigurationV2IndexDocument) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Suffix\n\t}).(pulumi.StringPtrOutput)\n}", "func Trim(str string) string {\n\treturn strings.Trim(str, \" \")\n}", "func Trim(str string, characterMask ...string) string {\n\ttrimChars := DefaultTrimChars\n\tif len(characterMask) > 0 {\n\t\ttrimChars += characterMask[0]\n\t}\n\treturn strings.Trim(str, trimChars)\n}", "func Trim(s string, l int, ellipsis string) string {\n\tif len(s) < l {\n\t\treturn s\n\t}\n\treturn s[:l] + ellipsis\n}", "func MednoHasSuffix(v string) predicate.Medicalfile {\n\treturn predicate.Medicalfile(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldMedno), v))\n\t})\n}", "func DropOptionalHasSuffix(v string) predicate.User {\n\treturn predicate.User(sql.FieldHasSuffix(FieldDropOptional, v))\n}", "func trimSpace(b []byte) string {\n\treturn string(bytes.TrimRight(b, \" \"))\n}", "func trimSpace(b []byte) string {\n\treturn string(bytes.TrimRight(b, \" \"))\n}", "func fnHasSuffix(ctx Context, doc *JDoc, params []string) interface{} {\n\tstats := ctx.Value(EelTotalStats).(*ServiceStats)\n\tif params == nil || len(params) != 2 {\n\t\tctx.Log().Error(\"error_type\", \"func_hassuffix\", \"op\", \"match\", \"cause\", \"wrong_number_of_parameters\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, SyntaxError{fmt.Sprintf(\"wrong number of parameters in call to hassuffix function\"), \"hassuffix\", params})\n\t\treturn nil\n\t}\n\treturn strings.HasSuffix(extractStringParam(params[0]), extractStringParam(params[1]))\n}", "func (o BucketLifecycleRuleItemConditionOutput) MatchesSuffix() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v BucketLifecycleRuleItemCondition) []string { return v.MatchesSuffix }).(pulumi.StringArrayOutput)\n}", "func Trim(str, chars string) string {\n\treturn LeftTrim(RightTrim(str, chars), chars)\n}", "func trimNIDs(s string) string {\n\treturn strings.Trim(s, \" .-\")\n}", "func trimPathAfterLastSlash(path string) string {\n\tif index := strings.LastIndex(path, \"/\"); index != -1 {\n\t\t// fmt.Println(path, \" Trimmed =\", path[:index])\n\t\treturn path[:index] //remove including the last /\n\t}\n\tfmt.Println(\"Failed to trim strings after last '/'\")\n\treturn path\n}", "func TrimSlash(str string) string {\n\treturn TrimByteSuffix(TrimBytePrefix(str, '/'), '/')\n}", "func stripExt(s string) string {\n\text := filepath.Ext(s)\n\tif len(ext) > 0 {\n\t\treturn s[:len(s)-len(ext)]\n\t}\n\treturn s\n}", "func (f FunctionSelector) WithoutPrefix() string { return f.String()[2:] }", "func (s *Stringish) TrimPrefix(prefix string) *Stringish {\n\ts.str = strings.TrimPrefix(s.str, prefix)\n\treturn s\n}", "func StrTrim(s string) string {\n\treturn strings.TrimSpace(s)\n}", "func NameHasSuffix(v string) predicate.Conversion {\n\treturn predicate.Conversion(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldName), v))\n\t})\n}", "func StringValHasSuffix(v string) predicate.Property {\n\treturn predicate.Property(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldStringVal), v))\n\t})\n}", "func SubHasSuffix(v string) predicate.Account {\n\treturn predicate.Account(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldSub), v))\n\t})\n}", "func BenefitsHasSuffix(v string) predicate.Job {\n\treturn predicate.Job(func(s *sql.Selector) {\n\t\ts.Where(sql.HasSuffix(s.C(FieldBenefits), v))\n\t})\n}", "func Trim(str string) string {\n\treturn strings.TrimSpace(str)\n}", "func getPatchVersionTrimmed(version string) string {\n\tendIndex := strings.LastIndex(version, \".\")\n\tif endIndex != -1 {\n\t\tversion = version[:endIndex]\n\t}\n\treturn version\n}" ]
[ "0.7361788", "0.71717155", "0.7046378", "0.68964106", "0.6881872", "0.66473514", "0.63875514", "0.62565345", "0.6204681", "0.61894584", "0.61191064", "0.60459554", "0.6032065", "0.5949064", "0.59068537", "0.5882754", "0.5873037", "0.57579243", "0.5750609", "0.574808", "0.5739919", "0.56416774", "0.5628701", "0.5622454", "0.5616339", "0.56150734", "0.56020194", "0.55960447", "0.5590597", "0.55777854", "0.5572178", "0.5554436", "0.5552236", "0.5498452", "0.5468744", "0.5439631", "0.54286146", "0.54211324", "0.5420454", "0.53981984", "0.5386082", "0.53858227", "0.5363233", "0.5357665", "0.5351558", "0.53017163", "0.53010803", "0.5288662", "0.5282833", "0.5281313", "0.5274254", "0.5272474", "0.5271351", "0.5265909", "0.5265909", "0.5264114", "0.5258515", "0.52354693", "0.5227949", "0.5226082", "0.5215546", "0.52105206", "0.51957786", "0.5194392", "0.51913565", "0.51900125", "0.51883686", "0.51834786", "0.51660585", "0.51565695", "0.51469105", "0.514458", "0.51409847", "0.5139247", "0.513709", "0.51315814", "0.5126807", "0.51180947", "0.5117742", "0.51034236", "0.50964785", "0.5090279", "0.508978", "0.508978", "0.5089635", "0.50835174", "0.50806916", "0.5080647", "0.5077023", "0.50691986", "0.50683546", "0.5066323", "0.5064817", "0.50496906", "0.50481856", "0.5030488", "0.5018659", "0.50167876", "0.5015929", "0.50154185" ]
0.91758645
0
Unquote uses strconv.Unquote to the underlying, unquoted string value of operand.
func Unquote(operand string) (unquoted string, err error) { return strconv.Unquote(operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func unquote(s string) string {\n\tif strings.HasPrefix(s, `\"`) && strings.HasSuffix(s, `\"`) {\n\t\ts, err := strconv.Unquote(s)\n\t\tif err != nil {\n\t\t\tpanic(fmt.Errorf(\"unable to unquote %q; %v\", s, err))\n\t\t}\n\t\treturn s\n\t}\n\treturn s\n}", "func unquote(s []byte) (t string, ok bool) {\n\ts, ok = unquoteBytes(s)\n\tt = string(s)\n\treturn\n}", "func Quote(operand string) string { return strconv.Quote(operand) }", "func StringUnEscape(quote rune, val string) (string, bool) {\n\tvar buf bytes.Buffer\n\tprevEscape, hasEscape := false, false\n\tfor _, r := range val {\n\t\tif r == quote && prevEscape {\n\t\t\thasEscape = true\n\t\t\tbuf.WriteByte(byte(r))\n\t\t\tprevEscape = false\n\t\t} else if r == quote {\n\t\t\tprevEscape = true\n\t\t} else if r == '\\\\' {\n\t\t\tprevEscape = true\n\t\t} else {\n\t\t\tbuf.WriteByte(byte(r))\n\t\t}\n\t}\n\n\treturn buf.String(), hasEscape\n}", "func UnescapeValue(s string) (string, error) {\n\t// if there's no escaping or special characters, just return to avoid allocation\n\tif !strings.ContainsAny(s, `\\,=`) {\n\t\treturn s, nil\n\t}\n\n\tv := bytes.NewBuffer(make([]byte, 0, len(s)))\n\tinSlash := false\n\tfor _, c := range s {\n\t\tif inSlash {\n\t\t\tswitch c {\n\t\t\tcase '\\\\', ',', '=':\n\t\t\t\t// omit the \\ for recognized escape sequences\n\t\t\t\tv.WriteRune(c)\n\t\t\tdefault:\n\t\t\t\t// error on unrecognized escape sequences\n\t\t\t\treturn \"\", InvalidEscapeSequence{sequence: string([]rune{'\\\\', c})}\n\t\t\t}\n\t\t\tinSlash = false\n\t\t\tcontinue\n\t\t}\n\n\t\tswitch c {\n\t\tcase '\\\\':\n\t\t\tinSlash = true\n\t\tcase ',', '=':\n\t\t\t// unescaped , and = characters are not allowed in field selector values\n\t\t\treturn \"\", UnescapedRune{r: c}\n\t\tdefault:\n\t\t\tv.WriteRune(c)\n\t\t}\n\t}\n\n\t// Ending with a single backslash is an invalid sequence\n\tif inSlash {\n\t\treturn \"\", InvalidEscapeSequence{sequence: \"\\\\\"}\n\t}\n\n\treturn v.String(), nil\n}", "func unquote(s string) string {\n\tu, q, esc := make([]rune, 0, len(s)), false, false\n\tfor _, c := range s {\n\t\tif esc {\n\t\t\tuc, ok := unescape[c]\n\t\t\tswitch {\n\t\t\tcase ok:\n\t\t\t\tu = append(u, uc)\n\t\t\t\tfallthrough\n\t\t\tcase !q && c == '\\n':\n\t\t\t\tesc = false\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tpanic(\"invalid escape sequence\")\n\t\t}\n\t\tswitch c {\n\t\tcase '\"':\n\t\t\tq = !q\n\t\tcase '\\\\':\n\t\t\tesc = true\n\t\tdefault:\n\t\t\tu = append(u, c)\n\t\t}\n\t}\n\tif q {\n\t\tpanic(\"missing end quote\")\n\t}\n\tif esc {\n\t\tpanic(\"invalid escape sequence\")\n\t}\n\treturn string(u)\n}", "func UnescapeArg(param interface{}) interface{} {\n\tif str, ok := param.(string); ok {\n\t\tend := len(str) - 1\n\n\t\tswitch GetTerminal() {\n\t\tcase TermBash:\n\t\t\tif str[0] == '\\'' && str[end] == '\\'' {\n\t\t\t\treturn strings.Replace(str[1:end], \"'\\\\''\", \"'\", -1)\n\t\t\t}\n\t\tcase TermCmd, TermPowershell:\n\t\t\tif str[0] == '\\'' && str[end] == '\\'' {\n\t\t\t\treturn strings.Replace(str[1:end], \"''\", \"'\", -1)\n\t\t\t}\n\t\t}\n\t}\n\treturn param\n}", "func unquote(s string) (string, error) {\n\n\t// check whether the string contains the expected embracing quotes\n\tif len(s) < 2 || s[0] != '\"' || s[len(s)-1] != '\"' {\n\t\treturn \"\", fmt.Errorf(\"String is not properly quoted\")\n\t}\n\n\t// strip embracing quotes\n\ts = s[1 : len(s)-1]\n\n\tunquoted := strings.Builder{}\n\tunquotePending := false\n\tfor _, c := range s {\n\t\tif !unquotePending && c == '\\\\' {\n\t\t\tunquotePending = true\n\t\t\tcontinue\n\t\t}\n\n\t\tif unquotePending && !strings.ContainsRune(charsToQuote, c) {\n\t\t\tunquoted.WriteRune('\\\\')\n\t\t}\n\n\t\tunquoted.WriteRune(c)\n\t\tunquotePending = false\n\t}\n\n\tif unquotePending {\n\t\treturn \"\", fmt.Errorf(\"Unpaired backslash found when unquoting string (%s)\", s)\n\t}\n\n\treturn unquoted.String(), nil\n}", "func bashUnescape(s string) string {\n\tif len(s) <= 1 {\n\t\treturn s\n\t}\n\tif s[0] == '\\'' && s[len(s)-1] == '\\'' {\n\t\treturn strings.ReplaceAll(s[1:len(s)-1], `'\\''`, `'`)\n\t}\n\tif s[0] == '\"' && s[len(s)-1] == '\"' {\n\t\ts = s[1 : len(s)-1]\n\t\t// Unescape \\\\, \\$, \\\", \\`, and \\!.\n\t\tre := regexp.MustCompile(\"\\\\\\\\([\\\\$\\\"`!\\\\\\\\])\")\n\t\treturn re.ReplaceAllString(s, \"$1\")\n\t}\n\tre := regexp.MustCompile(`\\\\(.)`)\n\treturn re.ReplaceAllString(s, \"$1\")\n}", "func UnquoteImportPath(s *ast.ImportSpec) ImportPath {\n\tpath, err := strconv.Unquote(s.Path.Value)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn ImportPath(path)\n}", "func Unescape(in []byte) []byte {\n\tif bytes.IndexByte(in, '\\\\') == -1 {\n\t\treturn in\n\t}\n\n\tinLen := len(in)\n\ti := 0\n\n\t// The output will be no more than inLen. Preallocating the\n\t// capacity here is faster and uses less memory than letting\n\t// append allocate.\n\tout := make([]byte, inLen)\n\tj := 0\n\n\tfor {\n\t\tif i >= inLen {\n\t\t\tbreak\n\t\t}\n\t\tii := i + 1\n\t\tif in[i] == '\\\\' && ii < inLen {\n\t\t\tswitch in[ii] {\n\t\t\tcase ',', '\"', ' ', '=':\n\t\t\t\tout[j] = in[ii]\n\t\t\t\ti, j = i+2, j+1\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tout[j] = in[i]\n\t\ti, j = ii, j+1\n\t}\n\treturn out[:j]\n}", "func unquote(pat string) upspin.PathName {\n\tif !strings.Contains(pat, \"\\\\\") {\n\t\treturn upspin.PathName(pat)\n\t}\n\tb := make([]byte, 0, len(pat))\n\tesc := false\n\tfor _, c := range []byte(pat) {\n\t\tif !esc && c == '\\\\' {\n\t\t\tesc = true\n\t\t\tcontinue\n\t\t}\n\t\tesc = false\n\t\tb = append(b, c)\n\t}\n\tif esc {\n\t\tb = append(b, '\\\\')\n\t}\n\treturn upspin.PathName(b)\n}", "func Unquote(s string) (t string, err error) {\n\tn := len(s)\n\tif n < 2 {\n\t\treturn \"\", ErrSyntax\n\t}\n\tquote := s[0]\n\tif quote != s[n-1] {\n\t\treturn \"\", ErrSyntax\n\t}\n\ts = s[1 : n-1]\n\n\tif quote != '\"' {\n\t\treturn \"\", ErrSyntax\n\t}\n\tif !contains(s, '$') && !contains(s, '{') && contains(s, '\\n') {\n\t\treturn \"\", ErrSyntax\n\t}\n\n\t// Is it trivial? Avoid allocation.\n\tif !contains(s, '\\\\') && !contains(s, quote) && !contains(s, '$') {\n\t\tswitch quote {\n\t\tcase '\"':\n\t\t\treturn s, nil\n\t\tcase '\\'':\n\t\t\tr, size := utf8.DecodeRuneInString(s)\n\t\t\tif size == len(s) && (r != utf8.RuneError || size != 1) {\n\t\t\t\treturn s, nil\n\t\t\t}\n\t\t}\n\t}\n\n\tvar runeTmp [utf8.UTFMax]byte\n\tbuf := make([]byte, 0, 3*len(s)/2) // Try to avoid more allocations.\n\tfor len(s) > 0 {\n\t\t// If we're starting a '${}' then let it through un-unquoted.\n\t\t// Specifically: we don't unquote any characters within the `${}`\n\t\t// section.\n\t\tif s[0] == '$' && len(s) > 1 && s[1] == '{' {\n\t\t\tbuf = append(buf, '$', '{')\n\t\t\ts = s[2:]\n\n\t\t\t// Continue reading until we find the closing brace, copying as-is\n\t\t\tbraces := 1\n\t\t\tfor len(s) > 0 && braces > 0 {\n\t\t\t\tr, size := utf8.DecodeRuneInString(s)\n\t\t\t\tif r == utf8.RuneError {\n\t\t\t\t\treturn \"\", ErrSyntax\n\t\t\t\t}\n\n\t\t\t\ts = s[size:]\n\n\t\t\t\tn := utf8.EncodeRune(runeTmp[:], r)\n\t\t\t\tbuf = append(buf, runeTmp[:n]...)\n\n\t\t\t\tswitch r {\n\t\t\t\tcase '{':\n\t\t\t\t\tbraces++\n\t\t\t\tcase '}':\n\t\t\t\t\tbraces--\n\t\t\t\t}\n\t\t\t}\n\t\t\tif braces != 0 {\n\t\t\t\treturn \"\", ErrSyntax\n\t\t\t}\n\t\t\tif len(s) == 0 {\n\t\t\t\t// If there's no string left, we're done!\n\t\t\t\tbreak\n\t\t\t} else {\n\t\t\t\t// If there's more left, we need to pop back up to the top of the loop\n\t\t\t\t// in case there's another interpolation in this string.\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\tif s[0] == '\\n' {\n\t\t\treturn \"\", ErrSyntax\n\t\t}\n\n\t\tc, multibyte, ss, err := unquoteChar(s, quote)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\ts = ss\n\t\tif c < utf8.RuneSelf || !multibyte {\n\t\t\tbuf = append(buf, byte(c))\n\t\t} else {\n\t\t\tn := utf8.EncodeRune(runeTmp[:], c)\n\t\t\tbuf = append(buf, runeTmp[:n]...)\n\t\t}\n\t\tif quote == '\\'' && len(s) != 0 {\n\t\t\t// single-quoted must be single character\n\t\t\treturn \"\", ErrSyntax\n\t\t}\n\t}\n\treturn string(buf), nil\n}", "func unescape(str string) (string, error) {\n\tbuf := make([]byte, len(str))\n\toffset := 0\n\tfor len(str) > 0 {\n\t\tval, _, tail, err := strconv.UnquoteChar(str, 0)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\toffset += utf8.EncodeRune(buf[offset:], val)\n\t\tstr = tail\n\t}\n\treturn string(buf[0:offset]), nil\n}", "func unquoteMetadataValue(s string) string {\n\tif strings.HasPrefix(s, \"'\") {\n\t\treturn strings.TrimPrefix(strings.TrimSuffix(s, \"'\"), \"'\")\n\t}\n\treturn s\n}", "func (v *Value) GetUnEscapeStr() *string {\n\tret := C.zj_GetUnEscapeStr(v.V)\n\tif ret == nil {\n\t\treturn nil\n\t}\n\tretStr := C.GoString(ret)\n\treturn &retStr\n}", "func unescape(escaped string) string {\n unescaped, e := url.QueryUnescape(escaped)\n if e != nil {\n log.Fatal(e)\n }\n return unescaped\n}", "func UnEscape(s string) string {\n\treturn strings.Replace(strings.Replace(s, specialString1, \"<\", -1), specialString2, \">\", -1)\n}", "func lexOpQuoted(l *lexer) lexFn {\n\tomitSpaces(l)\n\n\tl.acceptRun(_OpValueRunes)\n\tl.emit(OPV_QUOTED)\n\n\treturn lexText\n}", "func lexDubQuotedString(lx *lexer) stateFn {\r\n\tr := lx.next()\r\n\tswitch {\r\n\tcase r == '\\\\':\r\n\t\tlx.addCurrentStringPart(1)\r\n\t\treturn lexStringEscape\r\n\tcase r == dqStringEnd:\r\n\t\tlx.backup()\r\n\t\tlx.emitString()\r\n\t\tlx.next()\r\n\t\tlx.ignore()\r\n\t\treturn lx.pop()\r\n\tcase r == eof:\r\n\t\tif lx.pos > lx.start {\r\n\t\t\treturn lx.errorf(\"Unexpected EOF.\")\r\n\t\t}\r\n\t\tlx.emit(itemEOF)\r\n\t\treturn nil\r\n\t}\r\n\treturn lexDubQuotedString\r\n}", "func interpretString(s string, quote byte) (string, error) {\n\tswitch quote {\n\tcase '\\'', '\"':\n\t\t// OK\n\tdefault:\n\t\tpanic(\"invalid quote type\")\n\t}\n\n\tif !strings.Contains(s, `\\`) {\n\t\t// Fast path: nothing to replace.\n\t\treturn s, nil\n\t}\n\n\t// To understand what's going on, consult the manual:\n\t// https://www.php.net/manual/en/language.types.string.php#language.types.string.syntax.double\n\n\tif quote == '\"' {\n\t\treturn interpretStringQ2(s)\n\t}\n\treturn interpretStringQ1(s)\n}", "func (p *Parser) operand(tok scan.Token, indexOK bool) value.Expr {\n\tvar expr value.Expr\n\tswitch tok.Type {\n\tcase scan.Operator:\n\t\texpr = &unary{\n\t\t\top: tok.Text,\n\t\t\tright: p.expr(),\n\t\t}\n\tcase scan.Identifier:\n\t\tif p.context.DefinedUnary(tok.Text) {\n\t\t\texpr = &unary{\n\t\t\t\top: tok.Text,\n\t\t\t\tright: p.expr(),\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\t\tfallthrough\n\tcase scan.Number, scan.Rational, scan.Complex, scan.String, scan.LeftParen:\n\t\texpr = p.numberOrVector(tok)\n\tdefault:\n\t\tp.errorf(\"unexpected %s\", tok)\n\t}\n\tif indexOK {\n\t\texpr = p.index(expr)\n\t}\n\treturn expr\n}", "func unquoteSASSString(q []byte) string {\n\t// Adjust quotes so that Unquote is happy. We need a double quoted string\n\t// without unescaped double quote characters inside.\n\tnoQuotes := bytes.Split(q[1:len(q)-1], []byte{'\"'})\n\tif len(noQuotes) != 1 {\n\t\tfor i := 0; i < len(noQuotes)-1; i++ {\n\t\t\tif len(noQuotes[i]) == 0 || noQuotes[i][len(noQuotes[i])-1] != '\\\\' {\n\t\t\t\tnoQuotes[i] = append(noQuotes[i], '\\\\')\n\t\t\t}\n\t\t}\n\t\tq = append([]byte{'\"'}, bytes.Join(noQuotes, []byte{'\"'})...)\n\t\tq = append(q, '\"')\n\t}\n\tif q[0] == '\\'' {\n\t\tq[0] = '\"'\n\t\tq[len(q)-1] = '\"'\n\t}\n\n\ts, err := strconv.Unquote(string(q))\n\tif err != nil {\n\t\tlog.Panicf(\"unquoting string literal %s from sass: %v\", q, err)\n\t}\n\treturn s\n}", "func Unquote(unquote bool) Option {\n\treturn func(p *PathOptions) {\n\t\tp.Unquote = unquote\n\t}\n}", "func unparseString(v string) string {\n\tvar buf bytes.Buffer\n\tbuf.WriteString(\"\\\"\")\n\tfor _, c := range v {\n\t\tswitch c {\n\t\tcase '\"':\n\t\t\tbuf.WriteString(\"\\\\\\\"\")\n\t\tcase '\\\\':\n\t\t\tbuf.WriteString(\"\\\\\\\\\")\n\t\tcase '\\b':\n\t\t\tbuf.WriteString(\"\\\\b\")\n\t\tcase '\\f':\n\t\t\tbuf.WriteString(\"\\\\f\")\n\t\tcase '\\n':\n\t\t\tbuf.WriteString(\"\\\\n\")\n\t\tcase '\\r':\n\t\t\tbuf.WriteString(\"\\\\r\")\n\t\tcase '\\t':\n\t\t\tbuf.WriteString(\"\\\\t\")\n\t\tcase 0:\n\t\t\tbuf.WriteString(\"\\\\u0000\")\n\t\tdefault:\n\t\t\tif c < 0x20 || (c >= 0x7f && c <= 0x9f) {\n\t\t\t\tbuf.WriteString(fmt.Sprintf(\"\\\\u%04x\", int(c)))\n\t\t\t} else {\n\t\t\t\tbuf.WriteRune(c)\n\t\t\t}\n\t\t}\n\t}\n\tbuf.WriteString(\"\\\"\")\n\treturn buf.String()\n}", "func QuoteRegex(operand string) string { return regexp.QuoteMeta(operand) }", "func evalCommandSafely(v string) string {\n\tcmdWithArgs, err := shellquote.Split(v)\n\tif err != nil {\n\t\tfmt.Println(\"error processing command, will return empty string, ignored: \", err.Error())\n\t\treturn \"\"\n\t}\n\tif output, err := exec.Command(cmdWithArgs[0], cmdWithArgs[1:]...).Output(); err == nil {\n\t\treturn strings.TrimSpace(string(output))\n\t}\n\tfmt.Println(\"error executing command, will return empty string, ignored: \", err.Error())\n\treturn \"\"\n}", "func Unescape(s string) string {\n\tb := bytes.NewBuffer(make([]byte, 0, len(s)))\n\ti := 0\n\tfor _, v := range re.FindAllStringSubmatchIndex(s, -1) {\n\t\tj := v[0]\n\t\tb.WriteString(s[i:j])\n\t\tif k := v[3]; k >= 0 {\n\t\t\tb.WriteString(unescape_uXXXX(s, j, k))\n\t\t\ti = k\n\t\t} else {\n\t\t\tk := v[5]\n\t\t\tb.WriteString(unescape_XX(s, j, k))\n\t\t\ti = k\n\t\t}\n\t}\n\tb.WriteString(s[i:])\n\treturn b.String()\n}", "func FormatUnquoted(arg interface{}) string {\n\tswitch v := arg.(type) {\n\tcase string:\n\t\treturn v\n\tdefault:\n\t\treturn Format(v)\n\t}\n}", "func unescapeString(str string) string {\n\tvar sb strings.Builder\n\tfor i := 0; i < len(str); i++ {\n\t\tb := str[i]\n\t\tif b == '\\\\' && i+3 < len(str) {\n\t\t\tif parsed, err := strconv.ParseInt(str[i+1:i+4], 8, 8); err == nil {\n\t\t\t\tb = uint8(parsed)\n\t\t\t\ti += 3\n\t\t\t}\n\t\t}\n\t\tsb.WriteByte(b)\n\t}\n\treturn sb.String()\n}", "func (p *Parser) operand(tok token.Token, indexOK bool) value.Expr {\n\tvar expr value.Expr\n\tswitch tok.Type {\n\tcase token.Identifier:\n\t\t// TODO\n\t\tfallthrough\n\tcase token.Number, token.Rational, token.String, token.LeftParen:\n\t\texpr = p.numberOrVector(tok)\n\tdefault:\n\t\tp.errorf(\"unexpected %s\", tok)\n\t}\n\tif indexOK {\n\t\texpr = p.index(expr)\n\t}\n\treturn expr\n}", "func (t *Int64) UnmarshalGraphQL(input interface{}) error {\n\tswitch input := input.(type) {\n\tcase int:\n\t\t*t = Int64(input)\n\tcase int32:\n\t\t*t = Int64(input)\n\tcase int64:\n\t\t*t = Int64(input)\n\tcase float64:\n\t\t*t = Int64(input)\n\tcase string:\n\t\tval, err := strconv.Atoi(strings.Trim(input, `\"`))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t*t = Int64(val)\n\tdefault:\n\t\treturn errors.Errorf(convertErrorFormat, reflect.TypeOf(input), reflect.TypeOf(*t))\n\t}\n\treturn nil\n}", "func stringValue(e ast.Expr) (string, error) {\n\tlit, ok := e.(*ast.BasicLit)\n\tif !ok {\n\t\treturn \"\", fmt.Errorf(\"not a string, but %T\", e)\n\t}\n\tif lit.Kind != token.STRING {\n\t\treturn \"\", fmt.Errorf(\"not a string, but %v\", lit.Kind)\n\t}\n\treturn strconv.Unquote(lit.Value)\n}", "func (this *JsonWriter) Unquoted(v []byte) *JsonWriter { // {{{\n\treturn this.write(v)\n}", "func TestEval(t *testing.T) {\n\tany := `.+`\n\ttestCases := []struct {\n\t\tname string\n\t\tquery string\n\t\twantErr string\n\t\twant []values.Value\n\t}{\n\t\t{\n\t\t\tname: \"string interpolation\",\n\t\t\tquery: `\n\t\t\t\tstr = \"str\"\n\t\t\t\ting = \"ing\"\n\t\t\t\t\"str + ing = ${str+ing}\"`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewString(\"str + ing = string\"),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"string interpolation missing field\",\n\t\t\tquery: `\n\t\t\t\tr = makeRecord(o: {a: \"foo\", b: 42})\n\t\t\t\t\"r._value = ${r._value}\"`,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"string interpolation non-string type\",\n\t\t\tquery: `\n\t\t\t\tr = makeRecord(o: {a: \"foo\", b: 42})\n\t\t\t\t\"r._value = ${r.b}\"`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewString(\"r._value = 42\"),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"string interpolation wrong type\",\n\t\t\tquery: `\n\t\t\t\tr = makeRecord(o: {a: \"foo\", b: 42})\n\t\t\t\t\"r = ${r}\"`,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"call builtin function\",\n\t\t\tquery: \"six()\",\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewFloat(6.0),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"call function with fail\",\n\t\t\tquery: \"fail()\",\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"call function with duplicate args\",\n\t\t\tquery: \"plusOne(x:1.0, x:2.0)\",\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"binary expressions\",\n\t\t\tquery: `\n\t\t\tsix_value = six()\n\t\t\tnine_value = nine()\n\n\t\t\tfortyTwo() == six_value * nine_value\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"logical expressions short circuit\",\n\t\t\tquery: `\n six_value = six()\n nine_value = nine()\n\n not (fortyTwo() == six_value * nine_value) or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"function\",\n\t\t\tquery: `\n plusSix = (r) => r + six()\n plusSix(r:1.0) == 7.0 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"function block\",\n\t\t\tquery: `\n f = (r) => {\n r1 = 1.0 + r\n return (r + r1) / r\n }\n f(r:1.0) == 3.0 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"function block polymorphic\",\n\t\t\tquery: `\n f = (r) => {\n r2 = r * r\n return r2 / r\n }\n f(r:2.0) == 2.0 or fail()\n f(r:2) == 2 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"function with default param\",\n\t\t\tquery: `\n addN = (r,n=4) => r + n\n addN(r:2) == 6 or fail()\n addN(r:3,n:1) == 4 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"scope closing\",\n\t\t\tquery: `\n\t\t\tx = 5\n plusX = (r) => r + x\n plusX(r:2) == 7 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"nested scope mutations not visible outside\",\n\t\t\tquery: `\n\t\t\tx = 5\n xinc = () => {\n x = x + 1\n return x\n }\n xinc() == 6 or fail()\n x == 5 or fail()\n\t\t\t`,\n\t\t},\n\t\t// TODO(jsternberg): This test seems to not\n\t\t// infer the type constraints correctly for m.a,\n\t\t// but it doesn't fail.\n\t\t{\n\t\t\tname: \"return map from func\",\n\t\t\tquery: `\n toMap = (a,b) => ({\n a: a,\n b: b,\n })\n m = toMap(a:1, b:false)\n m.a == 1 or fail()\n not m.b or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"pipe expression\",\n\t\t\tquery: `\n\t\t\tadd = (a=<-,b) => a + b\n\t\t\tone = 1\n\t\t\tone |> add(b:2) == 3 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"ignore pipe default\",\n\t\t\tquery: `\n\t\t\tadd = (a=<-,b) => a + b\n\t\t\tadd(a:1, b:2) == 3 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"pipe expression function\",\n\t\t\tquery: `\n\t\t\tadd = (a=<-,b) => a + b\n\t\t\tsix() |> add(b:2.0) == 8.0 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"pipe builtin function\",\n\t\t\tquery: `\n\t\t\tsix() |> plusOne() == 7.0 or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"regex match\",\n\t\t\tquery: `\n\t\t\t\"abba\" =~ /^a.*a$/ or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"regex not match\",\n\t\t\tquery: `\n\t\t\t\"abc\" =~ /^a.*a$/ and fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"not regex match\",\n\t\t\tquery: `\n\t\t\t\"abc\" !~ /^a.*a$/ or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"not regex not match\",\n\t\t\tquery: `\n\t\t\t\"abba\" !~ /^a.*a$/ and fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"options metadata\",\n\t\t\tquery: `\n\t\t\toption task = {\n\t\t\t\tname: \"foo\",\n\t\t\t\trepeat: 100,\n\t\t\t}\n\t\t\ttask.name == \"foo\" or fail()\n\t\t\ttask.repeat == 100 or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"query with side effects\",\n\t\t\tquery: `sideEffect() == 0 or fail()`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewInt(0),\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"array index expression\",\n\t\t\tquery: `\n\t\t\t\ta = [1, 2, 3]\n\t\t\t\tx = a[1]\n\t\t\t\tx == 2 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"dict expression\",\n\t\t\tquery: `\n\t\t\t\tm = [\"a\" + \"b\": 0, \"c\" + \"d\": 1]\n\t\t\t\tx = get(dict: m, key: \"ab\", default: 2)\n\t\t\t\ty = get(dict: m, key: \"cd\", default: 2)\n\t\t\t\tz = get(dict: m, key: \"ef\", default: 2)\n\t\t\t\tx == 0 and y == 1 and z == 2 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"empy dictionary\",\n\t\t\tquery: `\n\t\t\t\tm0 = [:]\n\t\t\t\tm1 = insert(dict: m0, key: \"a\", value: 0)\n\t\t\t\tm2 = insert(dict: m0, key: 0, value: \"a\")\n\t\t\t\tv1 = get(dict: m1, key: \"a\", default: -1)\n\t\t\t\tv2 = get(dict: m2, key: 0, default: \"b\")\n\t\t\t\tv1 == 0 and v2 == \"a\" or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"array index expression out of bounds low\",\n\t\t\tquery: `\n\t\t\t\ta = [1, 2, 3]\n\t\t\t\ti = -1\n\t\t\t\tx = a[i]\n\t\t\t`,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"array index expression out of bounds high\",\n\t\t\tquery: `\n\t\t\t\ta = [1, 2, 3]\n\t\t\t\ti = 3\n\t\t\t\tx = a[i]\n\t\t\t`,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"array with complex index expression\",\n\t\t\tquery: `\n\t\t\t\tf = () => ({l: 0, m: 1, n: 2})\n\t\t\t\ta = [1, 2, 3]\n\t\t\t\tx = a[f().l]\n\t\t\t\ty = a[f().m]\n\t\t\t\tz = a[f().n]\n\t\t\t\tx == 1 or fail()\n\t\t\t\ty == 2 or fail()\n\t\t\t\tz == 3 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"short circuit logical and\",\n\t\t\tquery: `\n false and fail()\n `,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"short circuit logical or\",\n\t\t\tquery: `\n true or fail()\n `,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"no short circuit logical and\",\n\t\t\tquery: `\n true and fail()\n `,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"no short circuit logical or\",\n\t\t\tquery: `\n false or fail()\n `,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"conditional true\",\n\t\t\tquery: `\n\t\t\t\tif 1 != 0 then 10 else 100\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewInt(10),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"conditional false\",\n\t\t\tquery: `\n\t\t\t\tif 1 == 0 then 10 else 100\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewInt(100),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"conditional in function\",\n\t\t\tquery: `\n\t\t\t\tf = (t, c, a) => if t then c else a\n\t\t\t\t{\n\t\t\t\t\tv1: f(t: false, c: 30, a: 300),\n\t\t\t\t\tv2: f(t: true, c: \"cats\", a: \"dogs\"),\n\t\t\t\t}\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewObjectWithValues(map[string]values.Value{\n\t\t\t\t\t\"v1\": values.NewInt(300),\n\t\t\t\t\t\"v2\": values.NewString(\"cats\"),\n\t\t\t\t}),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"exists\",\n\t\t\tquery: `hasValue(o: makeRecord(o: {value: 1}))`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"exists null\",\n\t\t\tquery: `hasValue(o: makeRecord(o: {val: 2}))`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"invalid function parameter\",\n\t\t\tquery: `from(bucket: \"telegraf\") |> window(every: 0s)`,\n\t\t\twantErr: `error calling function \"window\" @\\d+:\\d+-\\d+:\\d+: window function requires at least one of \"every\" or \"period\" to be set and non-zero`,\n\t\t},\n\t\t{\n\t\t\t// tests that we don't nest error messages when\n\t\t\t// a function call fails and gets piped into another\n\t\t\t// function.\n\t\t\tname: \"nested function error\",\n\t\t\tquery: `from(bucket: \"telegraf\") |> window(every: 0s) |> mean()`,\n\t\t\twantErr: `error calling function \"window\" @\\d+:\\d+-\\d+:\\d+: window function requires at least one of \"every\" or \"period\" to be set and non-zero`,\n\t\t},\n\t}\n\n\tfor _, tc := range testCases {\n\t\ttc := tc\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\tsrc := prelude + tc.query\n\n\t\t\tctx, deps := dependency.Inject(context.Background(), dependenciestest.Default())\n\t\t\tdefer deps.Finish()\n\n\t\t\tsideEffects, _, err := runtime.Eval(ctx, src)\n\t\t\tif err != nil {\n\t\t\t\tif tc.wantErr == \"\" {\n\t\t\t\t\tt.Fatalf(\"unexpected error: %s\", err)\n\t\t\t\t}\n\n\t\t\t\t// We expect an error, so it should be a non-internal Flux error.\n\t\t\t\tif code := flux.ErrorCode(err); code == codes.Internal || code == codes.Unknown {\n\t\t\t\t\tt.Errorf(\"expected non-internal error code, got %s\", code)\n\t\t\t\t}\n\n\t\t\t\tre := regexp.MustCompile(tc.wantErr)\n\t\t\t\tif got := err.Error(); !re.MatchString(got) {\n\t\t\t\t\tt.Errorf(\"expected error to match pattern %q, but error was %q\", tc.wantErr, got)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t} else if tc.wantErr != \"\" {\n\t\t\t\tt.Fatal(\"expected error\")\n\t\t\t}\n\n\t\t\tvs := getSideEffectsValues(sideEffects)\n\t\t\tif tc.want != nil && !cmp.Equal(tc.want, vs, semantictest.CmpOptions...) {\n\t\t\t\tt.Fatalf(\"unexpected side effect values -want/+got: \\n%s\", cmp.Diff(tc.want, vs, semantictest.CmpOptions...))\n\t\t\t}\n\t\t})\n\t}\n}", "func Htmlunquote(text string) string {\n\t//实体符号解释为HTML\n\t/*\n\t Decodes `text` that's HTML quoted.\n\t >>> htmlunquote('&lt;&#39;&amp;&quot;&gt;')\n\t '<\\\\'&\">'\n\t*/\n\n\ttext = html.UnescapeString(text)\n\n\treturn strings.TrimSpace(text)\n}", "func UnescapeTagValue(inString string) string {\n\t// buf.Len() == 0 is the fastpath where we have not needed to unescape any chars\n\tvar buf strings.Builder\n\tremainder := inString\n\tfor {\n\t\tbackslashPos := strings.IndexByte(remainder, '\\\\')\n\n\t\tif backslashPos == -1 {\n\t\t\tif buf.Len() == 0 {\n\t\t\t\treturn inString\n\t\t\t} else {\n\t\t\t\tbuf.WriteString(remainder)\n\t\t\t\tbreak\n\t\t\t}\n\t\t} else if backslashPos == len(remainder)-1 {\n\t\t\t// trailing backslash, which we strip\n\t\t\tif buf.Len() == 0 {\n\t\t\t\treturn inString[:len(inString)-1]\n\t\t\t} else {\n\t\t\t\tbuf.WriteString(remainder[:len(remainder)-1])\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\t// non-trailing backslash detected; we're now on the slowpath\n\t\t// where we modify the string\n\t\tif buf.Len() == 0 {\n\t\t\tbuf.Grow(len(inString)) // just an optimization\n\t\t}\n\t\tbuf.WriteString(remainder[:backslashPos])\n\t\tbuf.WriteByte(escapedCharLookupTable[remainder[backslashPos+1]])\n\t\tremainder = remainder[backslashPos+2:]\n\t}\n\n\treturn buf.String()\n}", "func (rt *operatorRuntime) strOp(op func(string, string) interface{},\n\tvs parser.Scope, is map[string]interface{}, tid uint64) (interface{}, error) {\n\n\tvar ret interface{}\n\n\terrorutil.AssertTrue(len(rt.node.Children) == 2,\n\t\tfmt.Sprint(\"Operation requires 2 operands\", rt.node))\n\n\tres1, err := rt.node.Children[0].Runtime.Eval(vs, is, tid)\n\tif err == nil {\n\t\tvar res2 interface{}\n\n\t\tif res2, err = rt.node.Children[1].Runtime.Eval(vs, is, tid); err == nil {\n\t\t\tret = op(fmt.Sprint(res1), fmt.Sprint(res2))\n\t\t}\n\t}\n\n\treturn ret, err\n}", "func lexOpValue(l *lexer) lexFn {\n\tomitSpaces(l)\n\t// handle quoted values\n\tif l.peek() == '\\'' {\n\t\treturn lexOpQuoted\n\t}\n\n\treturn lexOpNumber\n}", "func MarshalUnescaped(v interface{}) string {\n\tm := Marshal(v)\n\ts, err := url.QueryUnescape(m)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn m\n\t}\n\treturn s\n\n}", "func unescapeLabel(escaped string) string {\n\tvar out []byte\n\tvar last rune\n\tvar value byte\n\tstate := 0\n\n\tfor _, r := range escaped {\n\t\tswitch state {\n\t\tcase 0:\n\t\t\tif r == '\\\\' {\n\t\t\t\tstate++\n\t\t\t\tcontinue\n\t\t\t}\n\t\tcase 1:\n\t\t\tif r == 'x' {\n\t\t\t\tstate++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tout = append(out, '\\\\')\n\t\tcase 2:\n\t\t\tif v, ok := isHex(r); ok {\n\t\t\t\tvalue = v\n\t\t\t\tlast = r\n\t\t\t\tstate++\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tout = append(out, '\\\\', 'x')\n\t\t\t}\n\t\tcase 3:\n\t\t\tif v, ok := isHex(r); ok {\n\t\t\t\tvalue <<= 4\n\t\t\t\tvalue += v\n\t\t\t\tout = append(out, byte(value))\n\t\t\t\tstate = 0\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tout = append(out, '\\\\', 'x')\n\t\t\t\tout = appendRune(out, last)\n\t\t\t}\n\t\t}\n\t\tstate = 0\n\t\tout = appendRune(out, r)\n\t}\n\n\tswitch state {\n\tcase 3:\n\t\tout = append(out, '\\\\', 'x')\n\t\tout = appendRune(out, last)\n\tcase 2:\n\t\tout = append(out, '\\\\', 'x')\n\tcase 1:\n\t\tout = append(out, '\\\\')\n\t}\n\n\treturn string(out)\n}", "func unary(typ int, op string, od1 *expr) *expr {\n\treturn &expr{\n\t\tsexp: append(exprlist{atomic(typ, op)}, od1),\n\t}\n}", "func unintern(x string) string {\n\tret := \"\"\n\tfor _, t := range x {\n\t\tret = ret + string(t)\n\t}\n\treturn ret\n}", "func ParseFloat(operand string) (f float64, err error) { return strconv.ParseFloat(operand, 64) }", "func Unpack(str string) string {\n\tvar builder strings.Builder\n\treader := strings.NewReader(str)\n\tprevChar, _, _ := reader.ReadRune()\n\tif unicode.IsDigit(prevChar) {\n\t\treturn \"\"\n\t}\n\tfor {\n\t\tcurrChar, _, readErr := reader.ReadRune()\n\t\tif readErr != nil {\n\t\t\tbuilder.WriteRune(prevChar)\n\t\t\tbreak\n\t\t}\n\n\t\tdigit, atoiErr := strconv.Atoi(string(currChar))\n\t\tif atoiErr == nil {\n\t\t\tbuilder.WriteString(strings.Repeat(string(prevChar), digit))\n\t\t} else {\n\t\t\tbuilder.WriteRune(prevChar)\n\t\t}\n\n\t\tif currChar == '\\\\' || atoiErr == nil {\n\t\t\tprevChar, _, readErr = reader.ReadRune()\n\t\t\tif readErr != nil {\n\t\t\t\tbreak\n\t\t\t}\n\t\t} else {\n\t\t\tprevChar = currChar\n\t\t}\n\t}\n\treturn builder.String()\n}", "func (v *Value) GetUnEscapeKey() *string {\n\tret := C.zj_GetUnEscapeKey(v.V)\n\tif ret == nil {\n\t\treturn nil\n\t}\n\tretStr := C.GoString(ret)\n\treturn &retStr\n}", "func unparse(v, pow int64, mult []int64) string {\n\ttype term struct {\n\t\tn int64\n\t\tu string\n\t}\n\tvar terms []term\n\tadd := func(n int64, u, prev string, v int64) int64 {\n\t\t// If the remaining value is zero and there is a previous term one place\n\t\t// higher, lower the previous term by one place and combine them.\n\t\t// For example, 1G+1M = 1025M with pow == 1024.\n\n\t\tif p := len(terms) - 1; p >= 0 && v == 0 && terms[p].u == prev {\n\t\t\tterms[p].n = terms[p].n*pow + n\n\t\t\tterms[p].u = u\n\t\t} else {\n\t\t\tterms = append(terms, term{n, u})\n\t\t}\n\t\treturn v\n\t}\n\n\tz := v\n\tfor i, div := range mult {\n\t\tif n := z / div; n > 0 {\n\t\t\tz = add(n, labels[i+1], labels[i], z%div)\n\t\t}\n\t}\n\tif len(terms) == 0 || z > 0 {\n\t\tadd(z, \"\", \"K\", 0)\n\t}\n\n\tparts := make([]string, len(terms))\n\tfor i, t := range terms {\n\t\tparts[i] = fmt.Sprintf(\"%d%s\", t.n, t.u)\n\t}\n\treturn strings.Join(parts, \" \")\n}", "func HandleUnquotedStrings(value []byte, dt jsonparser.ValueType) []byte {\n\tif dt == jsonparser.String {\n\t\t// bookend() is destructive to underlying slice, need to copy.\n\t\t// extra capacity saves an allocation and copy during bookend.\n\t\ttmp := make([]byte, len(value), len(value)+2)\n\t\tcopy(tmp, value)\n\t\tvalue = bookend(tmp, '\"', '\"')\n\t}\n\treturn value\n}", "func unfakeBackquote(s string) string {\n\treturn strings.ReplaceAll(s, \"‵\", \"`\")\n}", "func TestCheckBinaryExprStringQuoString(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `\"abc\" / \"abc\"`, env,\n\t\t`invalid operation: \"abc\" / \"abc\" (operator / not defined on string)`,\n\t)\n\n}", "func SimpleSubtract(exp string) string {\n var index int = strings.Index(exp, \"-\")\n var arg1 string = exp[:index]\n var arg2 string = exp[index + 1:]\n val1 := NotateToDouble(arg1)\n val2 := NotateToDouble(arg2)\n return NegativeNotate(fmt.Sprintf(\"%f\", val1 - val2))\n}", "func (p *Parser) lexQuote(l *lex.Lexer) lex.StateFn {\n\t// lexQuote is called for ', \", and `.\n\tif l.Next() != '\"' {\n\t\treturn l.Errorf(\"only support double-quoted strings\")\n\t}\n\tl.Ignore()\n\nloop:\n\tfor {\n\t\tswitch l.Next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.Next(); r != lex.EOF && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase lex.EOF, '\\n':\n\t\t\treturn l.Errorf(\"unterminated quoted string\")\n\t\tcase '\"':\n\t\t\tbreak loop\n\t\t}\n\t}\n\tl.Dec(1)\n\tl.Emit(typeString)\n\tl.Inc(1)\n\tl.Ignore()\n\treturn p.lexInsideAction\n}", "func (ns *Namespace) HTMLUnescape(s interface{}) (string, error) {\n\tss, err := cast.ToStringE(s)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn html.UnescapeString(ss), nil\n}", "func OpString(i syntax.Op,) string", "func yarnLiteralToToken(str string) Token {\n\t// String literal must end with '\"' and have length at least 2\n\tif l := len(str); l < 2 || str[l-1] != '\"' {\n\t\treturn Token{Err, \"Invalid string literal: \" + str}\n\t}\n\t// chop off the start and end quotes\n\treturn Token{Literal, str[1 : len(str)-1]}\n}", "func (t *Binary) UnmarshalGraphQL(input interface{}) error {\n\tswitch input := input.(type) {\n\tcase string:\n\t\t*t = Binary(input)\n\t\treturn nil\n\tdefault:\n\t\treturn errors.Errorf(convertErrorFormat, reflect.TypeOf(input), reflect.TypeOf(*t))\n\t}\n}", "func UnaryOp(op token.Token, y Value, prec uint) Value {\n\tif v, ok := y.(*ratVal); ok {\n\t\tret := constant.UnaryOp(gotoken.Token(op), v.Value, prec)\n\t\treturn &ratVal{ret}\n\t}\n\treturn constant.UnaryOp(gotoken.Token(op), y, prec)\n}", "func decodeOp(s string) (op, error) {\n\tparts := strings.SplitN(s, \",\", 2)\n\tt := parts[0]\n\tswitch t {\n\tcase \"ci\":\n\t\tparts = strings.SplitN(s, \",\", 4)\n\t\tif len(parts) < 4 {\n\t\t\treturn nil, newParseError(s)\n\t\t}\n\t\tvar prevPid, nextPid *pid\n\t\tvar err error\n\t\tif parts[1] != \"\" {\n\t\t\tif prevPid, err = decodePid(parts[1]); err != nil {\n\t\t\t\treturn nil, newParseError(s)\n\t\t\t}\n\t\t}\n\t\tif parts[2] != \"\" {\n\t\t\tif nextPid, err = decodePid(parts[2]); err != nil {\n\t\t\t\treturn nil, newParseError(s)\n\t\t\t}\n\t\t}\n\t\tif err != nil {\n\t\t\treturn nil, newParseError(s)\n\t\t}\n\t\treturn &clientInsert{prevPid, nextPid, parts[3]}, nil\n\tcase \"i\":\n\t\tparts = strings.SplitN(s, \",\", 3)\n\t\tif len(parts) < 3 {\n\t\t\treturn nil, newParseError(s)\n\t\t}\n\t\tpid, err := decodePid(parts[1])\n\t\tif err != nil {\n\t\t\treturn nil, newParseError(s)\n\t\t}\n\t\treturn &insert{pid, parts[2]}, nil\n\tcase \"d\":\n\t\tparts = strings.SplitN(s, \",\", 2)\n\t\tif len(parts) < 2 {\n\t\t\treturn nil, newParseError(s)\n\t\t}\n\t\tpid, err := decodePid(parts[1])\n\t\tif err != nil {\n\t\t\treturn nil, newParseError(s)\n\t\t}\n\t\treturn &delete{pid}, nil\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unknown op type: %s\", t)\n\t}\n}", "func (iop *internalOp) Quote(el Element, env *Environment) Element {\n\t// TODO is env needed for internal ops?\n\tif iop.quoter == nil {\n\t\tif el.IsAtom() {\n\t\t\treturn Elem(Cons(Atomize(iop), Cons(el.AsAtom(), nil)))\n\t\t}\n\t\treturn Elem(Cons(Atomize(iop), el.AsList()))\n\t}\n\treturn iop.quoter(el)\n}", "func (op *Operation) Unmarshal(raw []byte) error {\n\treturn json.Unmarshal(raw, &op)\n}", "func UnescapeToString(s string) (string, error) {\n\tb, err := Unescape(s)\n\treturn string(b), err\n}", "func Base64Decode(operand string) (string, error) {\n\tbytes, err := base64.StdEncoding.DecodeString(operand)\n\treturn string(bytes), err\n}", "func lexRawQuote(l *lexer) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated raw quoted string\")\n\t\tcase '`':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(itemRawString)\n\treturn lexInsideAction\n}", "func TestStrings(t *testing.T){\n\tsrc := \"a\\\\b\\nc\\n\"\n\tdst := \"a\\\\\\\\b\\\\nc\\\\n\"\n\tif StringEscape(src) != dst {\n\t\tt.Fatal(dst, StringEscape(src))\n\t}\n\t// rev := StringUnescape(dst)\n\trev := dst\n\tfmt.Println(src)\n\tfmt.Println(dst)\n\tfmt.Println(rev)\n}", "func lexDoubleQuote(l *lexer) stateFn {\r\nLoop:\r\n\tfor {\r\n\t\tswitch l.next() {\r\n\t\tcase '\\\\':\r\n\t\t\tif r := l.next(); r != eof && r != '\\n' {\r\n\t\t\t\tbreak\r\n\t\t\t}\r\n\t\t\tfallthrough\r\n\t\tcase eof, '\\n':\r\n\t\t\treturn l.errorf(\"未闭合的字符串。\")\r\n\t\tcase '\"':\r\n\t\t\tbreak Loop\r\n\t\t}\r\n\t}\r\n\tl.emit(TokDoubleQuot)\r\n\r\n\treturn lexAny\r\n}", "func (s *String) Eval(_, _ *Scope) (Value, error) {\n\treturn s, nil\n}", "func evalBinaryStringExpr(ctx *Ctx, x reflect.Value, op token.Token, y reflect.Value) (reflect.Value, error) {\n\tvar err error\n\tvar r string\n\tvar b bool\n\tis_bool := false\n\n\txx, yy := x.String(), y.String()\n\tswitch op {\n\tcase token.ADD:\tr = xx + yy\n\tcase token.EQL: b = xx == yy; is_bool = true\n\tcase token.NEQ: b = xx != yy; is_bool = true\n\tcase token.LEQ: b = xx <= yy; is_bool = true\n\tcase token.GEQ: b = xx >= yy; is_bool = true\n\tcase token.LSS: b = xx < yy; is_bool = true\n\tcase token.GTR: b = xx > yy; is_bool = true\n\tdefault: err = ErrInvalidOperands{x, op, y}\n\t}\n\tif is_bool {\n\t\treturn reflect.ValueOf(b), err\n\t} else {\n\t\treturn reflect.ValueOf(r).Convert(x.Type()), err\n\t}\n}", "func (p *parser) parseOperand(lhs bool) ast.Expr {\n\n\tswitch p.tok {\n\tcase token.IDENT:\n\t\tx := p.parseIdent()\n\t\t//\t\tif !lhs {\n\t\t//\t\t\tp.resolve(x)\n\t\t//\t\t}\n\t\treturn x\n\n\tcase token.INT, token.FLOAT, token.IMAG, token.CHAR, token.STRING:\n\t\tx := &ast.BasicLit{ValuePos: p.pos, Kind: p.tok, Value: p.lit}\n\t\tp.next()\n\t\treturn x\n\n\tcase token.LPAREN:\n\t\tlparen := p.pos\n\t\tp.next()\n\t\tp.exprLev++\n\t\tx := p.parseOperand(false)\n\t\tp.exprLev--\n\t\trparen := p.expect(token.RPAREN)\n\t\treturn &ast.ParenExpr{Lparen: lparen, X: x, Rparen: rparen}\n\n\tdefault:\n\t\tp.errorExpected(p.pos, \"operand\")\n\t\treturn &ast.BadExpr{From: p.pos - 10, To: p.pos + 10}\n\t}\n\n}", "func ResolveQuotedArgs(template string, args []v1alpha1.Argument) (string, error) {\n\tquotedArgs := make([]v1alpha1.Argument, len(args))\n\tfor i, arg := range args {\n\t\tquotedArg := v1alpha1.Argument{\n\t\t\tName: arg.Name,\n\t\t}\n\t\tif arg.Value != nil {\n\t\t\t// The following escapes any special characters (e.g. newlines, tabs, etc...)\n\t\t\t// in preparation for substitution\n\t\t\treplacement := strconv.Quote(*arg.Value)\n\t\t\treplacement = replacement[1 : len(replacement)-1]\n\t\t\tquotedArg.Value = &replacement\n\t\t}\n\t\tquotedArgs[i] = quotedArg\n\t}\n\treturn ResolveArgs(template, quotedArgs)\n}", "func lvalProcessStr(str string) string {\n\treturn strings.Replace(str, \"\\\"\", \"\", -1)\n}", "func UTokenFromTokenDenom(tokenDenom string) string {\n\treturn UTokenPrefix + tokenDenom\n}", "func indexUnescaped(s string, term []byte) (int, string, error) {\n\tvar i int\n\tunesc := make([]byte, 0)\n\tfor i = 0; i < len(s); i++ {\n\t\tb := s[i]\n\t\t// A terminator byte?\n\t\tif bytes.IndexByte(term, b) != -1 {\n\t\t\tbreak\n\t\t}\n\t\tif b == '\\\\' {\n\t\t\ti++\n\t\t\tif i >= len(s) {\n\t\t\t\treturn 0, \"\", fmt.Errorf(\"nothing following final escape in %q\", s)\n\t\t\t}\n\t\t\tb = s[i]\n\t\t}\n\t\tunesc = append(unesc, b)\n\t}\n\treturn i, string(unesc), nil\n}", "func (o *Operator) UnmarshalJSON(data []byte) error {\n\tu := jsonpb.Unmarshaler{}\n\tbuf := bytes.NewBuffer(data)\n\n\treturn u.Unmarshal(buf, &*o)\n}", "func lexQuote(l *lexer) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.next(); r != eof && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated quoted string\")\n\t\tcase '\"':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(itemString)\n\treturn lexInsideAction\n}", "func (t *HashType) Unpack(v string) error {\n\t*t = HashType(v)\n\treturn nil\n}", "func AsString(input starlark.Value) string {\n\tvalue, err := strconv.Unquote(input.String())\n\tif nil != err {\n\t\treturn \"\"\n\t}\n\treturn value\n}", "func lexDubQuotedKey(lx *lexer) stateFn {\r\n\tr := lx.peek()\r\n\tif r == dqStringEnd {\r\n\t\tlx.emit(itemKey)\r\n\t\tlx.next()\r\n\t\treturn lexSkip(lx, lexKeyEnd)\r\n\t} else if r == eof {\r\n\t\tif lx.pos > lx.start {\r\n\t\t\treturn lx.errorf(\"Unexpected EOF.\")\r\n\t\t}\r\n\t\tlx.emit(itemEOF)\r\n\t\treturn nil\r\n\t}\r\n\tlx.next()\r\n\treturn lexDubQuotedKey\r\n}", "func lexRawQuote(l *lexer) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase token.Eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated raw quoted string\")\n\t\tcase '`':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(token.ItemRawString)\n\treturn lexInsideList\n}", "func _op(op string) token.Token {\n\tswitch op {\n\tcase \"+\":\n\t\treturn token.ADD\n\tcase \"-\":\n\t\treturn token.SUB\n\tcase \"*\":\n\t\treturn token.MUL\n\tcase \"/\":\n\t\treturn token.QUO\n\tcase \"%\":\n\t\treturn token.REM\n\tcase \"&\":\n\t\treturn token.AND\n\tcase \"|\":\n\t\treturn token.OR\n\tcase \"^\":\n\t\treturn token.XOR\n\tcase \"<<\":\n\t\treturn token.SHL\n\tcase \">>\":\n\t\treturn token.SHR\n\tcase \"&^\":\n\t\treturn token.AND_NOT\n\tcase \"&&\":\n\t\treturn token.LAND\n\tcase \"||\":\n\t\treturn token.LOR\n\tcase \"<-\":\n\t\treturn token.ARROW\n\tcase \"++\":\n\t\treturn token.INC\n\tcase \"--\":\n\t\treturn token.DEC\n\tcase \"==\":\n\t\treturn token.EQL\n\tcase \"<\":\n\t\treturn token.LSS\n\tcase \">\":\n\t\treturn token.GTR\n\tcase \"!\":\n\t\treturn token.NOT\n\tcase \"!=\":\n\t\treturn token.NEQ\n\tcase \"<=\":\n\t\treturn token.LEQ\n\tcase \">=\":\n\t\treturn token.GEQ\n\tdefault:\n\t\tpanic(\"unrecognized binary/unary operator \" + op)\n\t}\n}", "func Tokenize(s string) (e shared.Equation, err error) {\n e.Arguments = extractArguements(s)\n e.Operators = extractOperators(s)\n return\n}", "func (m *EqOp) UnmarshalJSON(raw []byte) error {\n\tvar data struct {\n\t\tArgs json.RawMessage `json:\"args\"`\n\t}\n\tbuf := bytes.NewBuffer(raw)\n\tdec := json.NewDecoder(buf)\n\tdec.UseNumber()\n\n\tif err := dec.Decode(&data); err != nil {\n\t\treturn err\n\t}\n\n\tvar base struct {\n\t\t/* Just the base type fields. Used for unmashalling polymorphic types.*/\n\n\t\tType string `json:\"type\"`\n\t}\n\tbuf = bytes.NewBuffer(raw)\n\tdec = json.NewDecoder(buf)\n\tdec.UseNumber()\n\n\tif err := dec.Decode(&base); err != nil {\n\t\treturn err\n\t}\n\n\tallOfArgs, err := UnmarshalQuerySlice(bytes.NewBuffer(data.Args), runtime.JSONConsumer())\n\tif err != nil && err != io.EOF {\n\t\treturn err\n\t}\n\n\tvar result EqOp\n\n\tif base.Type != result.Type() {\n\t\t/* Not the type we're looking for. */\n\t\treturn errors.New(422, \"invalid type value: %q\", base.Type)\n\t}\n\n\tresult.argsField = allOfArgs\n\n\t*m = result\n\n\treturn nil\n}", "func lexEvaluatedString(lexer *Lexer) stateModifier {\n\trune := lexer.Peek()\n\tswitch rune {\n\tcase '\"':\n\t\tlexer.Emit(ItemStringPart)\n\t\tlexer.Next()\n\t\tlexer.Emit(ItemDoubleQuote)\n\t\treturn lexRoot\n\tcase '\\\\':\n\t\tlexer.Next()\n\t\tescaped := lexer.Next()\n\t\tswitch escaped {\n\t\tcase '\"', '\\\\', '$':\n\t\t\t// Do nothing\n\t\tdefault:\n\t\t\tlexer.errorf(\"unexpected escape sequence '%c'\", rune)\n\t\t\treturn nil\n\t\t}\n\tcase '$':\n\t\tlexer.Next()\n\t\tif rune := lexer.Peek(); rune == '(' {\n\t\t\tlexer.Backtrack()\n\t\t\tlexer.Emit(ItemStringPart)\n\t\t\tlexer.Next()\n\t\t\tlexer.Next()\n\t\t\tlexer.Emit(ItemSubstitutionStart)\n\t\t\tlexer.Mode = ModeEvaluatedString\n\t\t\tlexer.substitutionDepth++\n\t\t\treturn lexRoot\n\t\t} else {\n\t\t\tlexer.Next()\n\t\t}\n\tdefault:\n\t\tlexer.Next()\n\t}\n\n\treturn lexEvaluatedString\n}", "func EvaluatePostfix(input string) (float64, error) {\n\n\tevaluateOperators := map[string]func(float64, float64) float64{\n\t\t\"+\": func(arg1, arg2 float64) float64 {\n\t\t\treturn arg1 + arg2\n\t\t},\n\t\t\"-\": func(arg1, arg2 float64) float64 {\n\t\t\treturn arg1 - arg2\n\t\t},\n\t\t\"*\": func(arg1, arg2 float64) float64 {\n\t\t\treturn arg1 * arg2\n\t\t},\n\t\t\"/\": func(arg1, arg2 float64) float64 {\n\t\t\treturn arg1 / arg2\n\t\t},\n\t\t\"^\": func(arg1, arg2 float64) float64 {\n\t\t\treturn math.Pow(arg1, arg2)\n\t\t},\n\t}\n\n\tif input == \"\" {\n\t\treturn 0, errors.New(\"Input can not be empty\")\n\t}\n\n\tstringArray := strings.Fields(input)\n\tvar stack []float64\n\n\tfor _, token := range stringArray {\n\t\tif operators, consist := evaluateOperators[token]; consist {\n\t\t\tif len(stack) < 2 {\n\t\t\t\treturn 0, errors.New(\"Missing operand\")\n\t\t\t}\n\t\t\targ1, arg2 := stack[len(stack)-2], stack[len(stack)-1]\n\t\t\tstack = stack[:len(stack)-2]\n\t\t\tstack = append(stack, operators(arg1, arg2))\n\t\t} else {\n\t\t\tval, err := strconv.ParseFloat(token, 64)\n\t\t\tif err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t\tstack = append(stack, val)\n\t\t}\n\t}\n\n\tif len(stack) != 1 {\n\t\treturn 0, errors.New(\"Stack corrupted\")\n\t}\n\n\treturn stack[len(stack)-1], nil\n}", "func UnquoteUsage(flag *Flag) (name string, usage string) {\n\t// Look for a back-quoted name, but avoid the strings package.\n\tusage = flag.Usage\n\tfor i := 0; i < len(usage); i++ {\n\t\tif usage[i] == '`' {\n\t\t\tfor j := i + 1; j < len(usage); j++ {\n\t\t\t\tif usage[j] == '`' {\n\t\t\t\t\tname = usage[i+1 : j]\n\t\t\t\t\tusage = usage[:i] + name + usage[j+1:]\n\t\t\t\t\treturn name, usage\n\t\t\t\t}\n\t\t\t}\n\t\t\tbreak // Only one back quote; use type name.\n\t\t}\n\t}\n\n\t// No explicit name, so use type if we can find one.\n\tname = \"value\"\n\tswitch fv := flag.Value.(type) {\n\tcase boolFlag:\n\t\tif fv.IsBoolFlag() {\n\t\t\tname = \"\"\n\t\t}\n\tcase *durationValue:\n\t\tname = \"duration\"\n\tcase *float64Value:\n\t\tname = \"float\"\n\tcase *intValue, *int64Value:\n\t\tname = \"int\"\n\tcase *stringValue:\n\t\tname = \"string\"\n\tcase *uintValue, *uint64Value:\n\t\tname = \"uint\"\n\t}\n\treturn\n}", "func InstOpString(i syntax.InstOp,) string", "func (p *Parser) parseOperator() (op Operator, err error) {\n\ttok, lit := p.consume(KeyAndOperator)\n\tswitch tok {\n\t// DoesNotExistToken shouldn't be here because it's a unary operator, not a binary operator\n\tcase InToken:\n\t\top = Operator_in\n\tcase EqualsToken:\n\t\top = Operator_equals\n\tcase NotInToken:\n\t\top = Operator_notIn\n\tcase NotEqualsToken:\n\t\top = Operator_notEquals\n\tdefault:\n\t\treturn -1, fmt.Errorf(\"found '%s', expected: '=', '!=', 'in', notin'\", lit)\n\t}\n\treturn op, nil\n}", "func (t *Tokeniser) tokeniseDoubleQuotes(b byte) bool {\n\tswitch b {\n\tcase '\"':\n\t\tt.currentQuoteType = none\n\tcase '\\\\':\n\t\tt.escapeNextChar = true\n\tdefault:\n\t\tt.put(b)\n\t}\n\n\treturn false\n}", "func parseExpression(l string, existing expression) (expression, error) {\n\tif len(l) == 0 {\n\t\treturn nil, errors.New(\"error parsing empty string as expression\")\n\t}\n\te := existing\n\tvar opStack []binaryOperator\n\tfor i := 0; i < len(l); i++ {\n\t\tif l[i] == ' ' {\n\t\t\tcontinue\n\t\t}\n\t\tif l[i] == '(' {\n\t\t\tvar right int\n\t\t\tparens := 1\n\t\t\tfor right = i + 1; right < len(l); right++ {\n\t\t\t\tif l[right] == '(' {\n\t\t\t\t\tparens++\n\t\t\t\t}\n\t\t\t\tif l[right] == ')' {\n\t\t\t\t\tparens--\n\t\t\t\t}\n\t\t\t\tif parens == 0 {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif parens != 0 {\n\t\t\t\treturn nil, fmt.Errorf(\"missing end paren at %d after %q\", i, l[i:])\n\t\t\t}\n\t\t\tsubE, err := parseExpression(l[i+1:right], nil)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"error at %d parsing subexpression %q: %v\", i, l[i+1:right], err)\n\t\t\t}\n\t\t\ti = right + 1\n\t\t\tif e == nil {\n\t\t\t\te = subE\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\te = binaryExpression{left: e, right: subE, op: opStack[len(opStack)-1]}\n\t\t\topStack = opStack[:len(opStack)-1]\n\t\t\tcontinue\n\t\t}\n\t\tif l[i] == '+' {\n\t\t\topStack = append(opStack, add)\n\t\t\tcontinue\n\t\t}\n\t\tif l[i] == '*' {\n\t\t\topStack = append(opStack, mul)\n\t\t\tcontinue\n\t\t}\n\t\tn, err := strconv.Atoi(l[i : i+1])\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error parsing int at %d: %v\", i, err)\n\t\t}\n\t\tif e == nil {\n\t\t\te = intValue(n)\n\t\t\tcontinue\n\t\t}\n\t\tif len(opStack) == 0 {\n\t\t\t// is len ever != 1?\n\t\t\treturn nil, fmt.Errorf(\"opStack is empty at %d, expression so far: %v\", i, e)\n\t\t}\n\t\te = binaryExpression{left: e, right: intValue(n), op: opStack[len(opStack)-1]}\n\t\topStack = opStack[:len(opStack)-1]\n\t}\n\treturn e, nil\n}", "func tokenToFormulaArg(token efp.Token) formulaArg {\n\tswitch token.TSubType {\n\tcase efp.TokenSubTypeLogical:\n\t\treturn newBoolFormulaArg(strings.EqualFold(token.TValue, \"TRUE\"))\n\tcase efp.TokenSubTypeNumber:\n\t\tnum, _ := strconv.ParseFloat(token.TValue, 64)\n\t\treturn newNumberFormulaArg(num)\n\tdefault:\n\t\treturn newStringFormulaArg(token.TValue)\n\t}\n}", "func UnquoteUsage(flag *Flag) (name string, usage string) {\n\t// Look for a back-quoted name, but avoid the strings package.\n\tusage = flag.Usage\n\tfor i := 0; i < len(usage); i++ {\n\t\tif usage[i] == '`' {\n\t\t\tfor j := i + 1; j < len(usage); j++ {\n\t\t\t\tif usage[j] == '`' {\n\t\t\t\t\tname = usage[i+1 : j]\n\t\t\t\t\tusage = usage[:i] + name + usage[j+1:]\n\t\t\t\t\treturn name, usage\n\t\t\t\t}\n\t\t\t}\n\t\t\tbreak // Only one back quote; use type name.\n\t\t}\n\t}\n\n\tname = flag.Value.Type()\n\tswitch name {\n\tcase \"bool\":\n\t\tname = \"\"\n\tcase \"float64\":\n\t\tname = \"float\"\n\tcase \"int64\":\n\t\tname = \"int\"\n\tcase \"uint64\":\n\t\tname = \"uint\"\n\tcase \"stringSlice\":\n\t\tname = \"strings\"\n\tcase \"intSlice\":\n\t\tname = \"ints\"\n\tcase \"uintSlice\":\n\t\tname = \"uints\"\n\tcase \"boolSlice\":\n\t\tname = \"bools\"\n\t}\n\n\treturn\n}", "func UnquoteUsage(flag *Flag) (name string, usage string) {\n\t// Look for a back-quoted name, but avoid the strings package.\n\tusage = flag.Usage\n\tfor i := 0; i < len(usage); i++ {\n\t\tif usage[i] == '`' {\n\t\t\tfor j := i + 1; j < len(usage); j++ {\n\t\t\t\tif usage[j] == '`' {\n\t\t\t\t\tname = usage[i+1 : j]\n\t\t\t\t\tusage = usage[:i] + name + usage[j+1:]\n\t\t\t\t\treturn name, usage\n\t\t\t\t}\n\t\t\t}\n\t\t\tbreak // Only one back quote; use type name.\n\t\t}\n\t}\n\t// No explicit name, so use type if we can find one.\n\tname = \"value\"\n\tswitch flag.Value.(type) {\n\tcase boolFlag:\n\t\tname = \"\"\n\tcase *durationValue:\n\t\tname = \"duration\"\n\tcase *float64Value:\n\t\tname = \"float\"\n\tcase *intValue, *int64Value:\n\t\tname = \"int\"\n\tcase *stringValue:\n\t\tname = \"string\"\n\tcase *uintValue, *uint64Value:\n\t\tname = \"uint\"\n\t}\n\treturn\n}", "func Unescape(s string) ([]byte, error) {\n\tvar buf = new(bytes.Buffer)\n\treader := strings.NewReader(s)\n\n\tfor {\n\t\tr, size, err := reader.ReadRune()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif size > 1 {\n\t\t\treturn nil, fmt.Errorf(\"rfc2396: non-ASCII char detected\")\n\t\t}\n\n\t\tswitch r {\n\t\tcase '%':\n\t\t\teb1, err := reader.ReadByte()\n\t\t\tif err == io.EOF {\n\t\t\t\treturn nil, fmt.Errorf(\"rfc2396: unexpected end of unescape sequence\")\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif !isHex(eb1) {\n\t\t\t\treturn nil, fmt.Errorf(\"rfc2396: invalid char 0x%x in unescape sequence\", r)\n\t\t\t}\n\t\t\teb0, err := reader.ReadByte()\n\t\t\tif err == io.EOF {\n\t\t\t\treturn nil, fmt.Errorf(\"rfc2396: unexpected end of unescape sequence\")\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif !isHex(eb0) {\n\t\t\t\treturn nil, fmt.Errorf(\"rfc2396: invalid char 0x%x in unescape sequence\", r)\n\t\t\t}\n\t\t\tbuf.WriteByte(unhex(eb0) + unhex(eb1)*16)\n\t\tdefault:\n\t\t\tbuf.WriteByte(byte(r))\n\t\t}\n\t}\n\treturn buf.Bytes(), nil\n}", "func lexRawQuote(l *reader) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.next(); r != eof && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated quoted string\")\n\t\tcase '`':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(itemString)\n\treturn lexPunctuation\n}", "func TestPerformArithmeticOperationMissingOperand(t *testing.T) {\n\t// operand stack (also known as data stack)\n\tstack := evaluator.Stack{}\n\n\t// push just one value onto the stack\n\tstack.Push(1)\n\n\t// any token that is not token.QUO or token.REM\n\ttok := token.ADD\n\n\t// perform the selected arithmetic operation\n\taddOperation := func(x int, y int) int { return x + y }\n\n\t// perform the selected arithmetic operation\n\terr := evaluator.PerformArithmeticOperation(&stack, addOperation, tok)\n\tassert.Error(t, err)\n}", "func ParseBool(operand string) (value bool, err error) { return strconv.ParseBool(operand) }", "func lexQuote(l *lexer) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.next(); r != token.Eof && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase token.Eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated quoted string\")\n\t\tcase '\"':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(token.ItemString)\n\treturn lexInsideList\n}", "func stringExpr(expr ast.Expr) string {\n\targ, ok := expr.(*ast.BasicLit)\n\tif !ok {\n\t\treturn \"\"\n\t}\n\treturn strings.Trim(arg.Value, `\"`)\n}", "func AsString(x skylark.Value) (string, error) {\n\treturn strconv.Unquote(x.String())\n}", "func (self *codeGen) cgUnopExp(exp *UnopExp, a int) {\n\tallocator := self.newTmpAllocator(a)\n\tb, _ := self.exp2OpArg(exp.Exp, ARG_REG, allocator)\n\tself.emitUnaryOp(exp.Line, exp.Op, a, b)\n\tallocator.freeAll()\n}", "func (e *Evaluator) Evaluate(expression string) (*string, error) {\n\tinfixExpression, err := e.tknzr.Tokenize(expression)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// a bit of syntax sugar: if expression contains only atoms\n\t// consider it as just a string literal\n\tif e.onlyAtoms(infixExpression) {\n\t\treturn &expression, nil\n\t}\n\n\tpostfixExpression, err := e.cnvtr.Convert(infixExpression)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn e.evaluateExpression(postfixExpression)\n}" ]
[ "0.69857085", "0.6599664", "0.62009734", "0.5994443", "0.595505", "0.5912972", "0.5821383", "0.5570849", "0.55220246", "0.5440596", "0.5429673", "0.54185534", "0.54095936", "0.5297827", "0.52768856", "0.52635616", "0.52546376", "0.52247566", "0.51736724", "0.51180476", "0.51095134", "0.50663936", "0.5054962", "0.5027655", "0.49483216", "0.4887109", "0.48705432", "0.48494488", "0.48493725", "0.4779849", "0.47779864", "0.47777998", "0.47503114", "0.47418672", "0.469726", "0.46937487", "0.46407282", "0.46076715", "0.4572506", "0.45576817", "0.45501265", "0.45437574", "0.45437002", "0.45396644", "0.45170268", "0.45137766", "0.45078692", "0.45046043", "0.45010805", "0.44886678", "0.44852072", "0.44827873", "0.44749042", "0.44688493", "0.44681105", "0.4441661", "0.4433756", "0.44295064", "0.4422407", "0.44027984", "0.44000843", "0.43993852", "0.43952587", "0.43942398", "0.43917754", "0.43894434", "0.43668073", "0.43593287", "0.43562117", "0.43465498", "0.43379375", "0.43265778", "0.4323383", "0.431807", "0.43148366", "0.43096608", "0.43014792", "0.4287227", "0.4279916", "0.42756015", "0.4272341", "0.42620105", "0.42588288", "0.42398003", "0.42393872", "0.42344606", "0.42283836", "0.42240444", "0.42210272", "0.4217189", "0.42162523", "0.4215973", "0.42132106", "0.42070207", "0.4195913", "0.41814426", "0.4174662", "0.417269", "0.416475", "0.41553995" ]
0.8678551
0
/ Slice Manipulation Grep filters operand according to pattern, returning a slice of matching elements. Pattern is treated as a regexp.
func Grep(pattern string, operand []string) ([]string, error) { rex, err := regexp.Compile(pattern) if err != nil { return nil, err } var matching []string for _, elem := range operand { if rex.MatchString(elem) { matching = append(matching, elem) } } return matching, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Filter(slice []string, match MatchFunc) []string {\n\tout := make([]string, 0, len(slice))\n\tfor _, s := range slice {\n\t\tif match(s) {\n\t\t\tout = append(out, s)\n\t\t}\n\t}\n\treturn out\n}", "func FilterSliceWithRegexps(sl []string, excludeRegexps []*regexp.Regexp) []string {\n\tres := make([]string, len(sl))\n\ti := 0\n\tfor k, s := range sl {\n\t\tif IsStringExcluded(s, excludeRegexps) {\n\t\t\tcontinue\n\t\t}\n\t\tres[k] = s\n\t\ti++\n\t}\n\treturn res[:i]\n}", "func (re *Regexp) MatchSlices(b []byte) (a [][]byte) {\n\tr := re.doExecute(\"\", b, 0);\n\tif r == nil {\n\t\treturn nil\n\t}\n\ta = make([][]byte, len(r)/2);\n\tfor i := 0; i < len(r); i += 2 {\n\t\tif r[i] != -1 {\t// -1 means no match for this subexpression\n\t\t\ta[i/2] = b[r[i]:r[i+1]]\n\t\t}\n\t}\n\treturn;\n}", "func Grep(b []byte, search ...string) []string {\n\tvar lines []string\n\ttext_lines := strings.Split(string(b), \"\\n\")\ncheck_line:\n\tfor _, text_line := range text_lines {\n\t\tfor _, word := range search {\n\t\t\tif match, _ := regexp.MatchString(word, text_line); match == false {\n\t\t\t\tcontinue check_line\n\t\t\t}\n\t\t}\n\t\tlines = append(lines, text_line)\n\t}\n\treturn lines\n}", "func Grepping(data, regex string) []string {\n\tbyteData := []byte(data)\n\tvar bodySlice []string\n\tvar pattern = regexp.MustCompile(regex)\n\tresult := pattern.FindAllIndex(byteData, -1)\n\t_ = result\n\n\tfor _, v := range result {\n\t\tbodySlice = append(bodySlice, data[v[0]:v[1]])\n\t}\n\treturn bodySlice\n}", "func Grep(exp string, src string) string {\n\tre, err := regexp.Compile(exp)\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"%q is not a valid, %s\\n\", exp, err)\n\t\treturn \"\"\n\t}\n\tlines := strings.Split(src, \"\\n\")\n\tfor _, line := range lines {\n\t\ts := re.FindString(line)\n\t\tif len(s) > 0 {\n\t\t\treturn s\n\t\t}\n\t}\n\treturn \"\"\n}", "func reExtract(regex, target string, matchGroup ...int) ([]string, error) {\n\tresults := make([]string, 0)\n\tmatches := regexp.MustCompile(regex).FindStringSubmatch(target)\n\tfor _, mg := range matchGroup {\n\t\tif mg > len(matches)-1 {\n\t\t\treturn nil, fmt.Errorf(\"not enough matches\")\n\t\t}\n\t\tresults = append(results, matches[mg])\n\t}\n\treturn results, nil\n}", "func Grep(in chan string, pattern string) chan string {\n\tout := make(chan string)\n\tgo func() {\n\t\tr := regexp.MustCompile(pattern)\n\t\tfor l := range in {\n\t\t\tif r.Find([]byte(l)) != nil {\n\t\t\t\tout <- l\n\t\t\t}\n\t\t}\n\t\tclose(out)\n\t}()\n\treturn out\n}", "func (re *RegexpStd) FindAllSubmatch(b []byte, n int) [][][]byte {\n\t// if n < 0 {\n\t// \tn = len(b) + 1\n\t// }\n\t// var result [][][]byte\n\t// re.allMatches(\"\", b, n, func(match []int) {\n\t// \tif result == nil {\n\t// \t\tresult = make([][][]byte, 0, startSize)\n\t// \t}\n\t// \tslice := make([][]byte, len(match)/2)\n\t// \tfor j := range slice {\n\t// \t\tif match[2*j] >= 0 {\n\t// \t\t\tslice[j] = b[match[2*j]:match[2*j+1]:match[2*j+1]]\n\t// \t\t}\n\t// \t}\n\t// \tresult = append(result, slice)\n\t// })\n\t// return result\n\tpanic(\"\")\n}", "func (re *RegexpStd) FindSubmatch(b []byte) [][]byte {\n\t// var dstCap [4]int\n\t// a := re.doExecute(nil, b, \"\", 0, re.prog.NumCap, dstCap[:0])\n\t// if a == nil {\n\t// \treturn nil\n\t// }\n\t// ret := make([][]byte, 1+re.numSubexp)\n\t// for i := range ret {\n\t// \tif 2*i < len(a) && a[2*i] >= 0 {\n\t// \t\tret[i] = b[a[2*i]:a[2*i+1]:a[2*i+1]]\n\t// \t}\n\t// }\n\t// return ret\n\tpanic(\"\")\n}", "func grep(sel *Selection, predicate func(i int, s *Selection) bool) (result []*html.Node) {\n\tfor i, n := range sel.Nodes {\n\t\tif predicate(i, newSingleSelection(n, sel.document)) {\n\t\t\tresult = append(result, n)\n\t\t}\n\t}\n\treturn result\n}", "func RegexpFilter(pattern string) Filter {\n\tp := regexp.MustCompile(pattern)\n\treturn regexpFilter{p}\n}", "func RegexpParse(content, pattern string) ([]string, error) {\n\tre, err := pool.Compile(pattern)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"re:[%s] error:%+v\", pattern, err)\n\t}\n\tvar ret []string\n\tres := re.FindAllStringSubmatch(content, -1)\n\tfor i := range res {\n\t\tswitch {\n\t\tcase len(res[i]) == 1:\n\t\t\tret = append(ret, res[i][0])\n\t\tcase len(res[i]) > 1:\n\t\t\tret = append(ret, res[i][1:]...)\n\t\t}\n\t}\n\treturn ret, nil\n}", "func find(text []byte, exp string) []byte {\n\treturn regexp.MustCompile(exp).FindSubmatch(text)[1]\n}", "func Egrep(b []byte, search ...string) []string {\n\tvar lines []string\n\ttext_lines := strings.Split(string(b), \"\\n\")\n\n\tfor _, text_line := range text_lines {\n\t\tfor _, word := range search {\n\t\t\tif match, _ := regexp.MatchString(word, text_line); match {\n\t\t\t\tlines = append(lines, text_line)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn lines\n}", "func (s *Action) Grep(c *cli.Context) error {\n\tctx := ctxutil.WithGlobalFlags(c)\n\tif !c.Args().Present() {\n\t\treturn exit.Error(exit.Usage, nil, \"Usage: %s grep arg\", s.Name)\n\t}\n\n\t// get the search term.\n\tneedle := c.Args().First()\n\n\thaystack, err := s.Store.List(ctx, tree.INF)\n\tif err != nil {\n\t\treturn exit.Error(exit.List, err, \"failed to list store: %s\", err)\n\t}\n\n\tmatchFn := func(haystack string) bool {\n\t\treturn strings.Contains(haystack, needle)\n\t}\n\n\tif c.Bool(\"regexp\") {\n\t\tre, err := regexp.Compile(needle)\n\t\tif err != nil {\n\t\t\treturn exit.Error(exit.Usage, err, \"failed to compile regexp %q: %s\", needle, err)\n\t\t}\n\t\tmatchFn = re.MatchString\n\t}\n\n\tvar matches int\n\tvar errors int\n\tfor _, v := range haystack {\n\t\tsec, err := s.Store.Get(ctx, v)\n\t\tif err != nil {\n\t\t\tout.Errorf(ctx, \"failed to decrypt %s: %v\", v, err)\n\n\t\t\tcontinue\n\t\t}\n\n\t\tif matchFn(string(sec.Bytes())) {\n\t\t\tout.Printf(ctx, \"%s matches\", color.BlueString(v))\n\t\t}\n\t}\n\n\tif errors > 0 {\n\t\tout.Warningf(ctx, \"%d secrets failed to decrypt\", errors)\n\t}\n\tout.Printf(ctx, \"\\nScanned %d secrets. %d matches, %d errors\", len(haystack), matches, errors)\n\n\treturn nil\n}", "func RegexpExtract(content, pattern string) (map[string]string, error) {\n\tre, err := pool.Compile(pattern)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"re:[%s] error:%+v\", pattern, err)\n\t}\n\tmatch := re.FindStringSubmatch(content)\n\tif len(match) == 0 {\n\t\treturn nil, nil\n\t}\n\tret := make(map[string]string)\n\tfor i, name := range re.SubexpNames() {\n\t\tif i != 0 {\n\t\t\tif name == \"\" {\n\t\t\t\tname = fmt.Sprintf(\"%d\", i)\n\t\t\t}\n\t\t\tret[name] = strings.TrimSpace(match[i])\n\t\t}\n\t}\n\treturn ret, nil\n}", "func (s *GrepEngine) Search(terms ...string) ([]*Result, error) {\n\texpr := \"\"\n\tif len(terms) > 0 {\n\t\texpr = terms[0]\n\t}\n\n\trx, err := compileRx(expr, !s.CaseSensitive)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmatches := func(s string) []string {\n\t\treturn rx.FindAllString(s, -1)\n\t}\n\n\treturn searchInNotes(s.notes, matches, s.OnlyNames)\n}", "func grepLines(data string, like string) []string {\n\tvar result = []string{}\n\tif like == \"\" {\n\t\tlog.Printf(\"ERROR: unable to look for empty pattern\")\n\t\treturn result\n\t}\n\tlike_bytes := []byte(like)\n\n\tscanner := bufio.NewScanner(strings.NewReader(data))\n\tfor scanner.Scan() {\n\t\tif bytes.Contains(scanner.Bytes(), like_bytes) {\n\t\t\tresult = append(result, scanner.Text())\n\t\t}\n\t}\n\tif err := scanner.Err(); err != nil {\n\t\tlog.Printf(\"WARN: error scanning string for %s: %s\", like, err)\n\t}\n\n\treturn result\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func Regexp(pattern string) (*regexp.Regexp, error) {\n\treturn pool.Compile(pattern)\n}", "func (r Report) FilterRegexp(pattern ...string) (result Report) {\n\tif len(pattern) == 0 {\n\t\treturn r\n\t}\n\n\tregexps := make([]*regexp.Regexp, len(pattern))\n\tfor i := range pattern {\n\t\tregexps[i] = regexp.MustCompile(pattern[i])\n\t}\n\n\treturn r.filter(func(filterPath *ytbx.Path) bool {\n\t\tfor _, regexp := range regexps {\n\t\t\tif filterPath != nil && regexp.MatchString(filterPath.String()) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t\treturn false\n\t})\n}", "func Regexp(expr *regexp.Regexp) Pattern {\n\treturn regexpMatch{expr}\n}", "func (re *RegexpStd) FindAll(b []byte, n int) [][]byte {\n\t// if n < 0 {\n\t// \tn = len(b) + 1\n\t// }\n\t// var result [][]byte\n\t// re.allMatches(\"\", b, n, func(match []int) {\n\t// \tif result == nil {\n\t// \t\tresult = make([][]byte, 0, startSize)\n\t// \t}\n\t// \tresult = append(result, b[match[0]:match[1]:match[1]])\n\t// })\n\t// return result\n\tpanic(\"\")\n}", "func FilterSlice[S any](s []S, keep func(S) bool) []S {\n\tvar result []S\n\tfor _, e := range s {\n\t\tif keep(e) {\n\t\t\tresult = append(result, e)\n\t\t}\n\t}\n\treturn result\n}", "func FilterRegex(regex string, metrics []string) ([]string, error) {\n\tr, err := regexp.Compile(regex)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresult := make([]string, 0)\n\n\tfor _, v := range metrics {\n\t\tif r.MatchString(v) {\n\t\t\tresult = append(result, v)\n\t\t}\n\t}\n\n\treturn result, nil\n}", "func r(pattern string) *regexp.Regexp { return regexp.MustCompile(pattern) }", "func regexpFindAllString(s string, regExp string, n int) []string {\n\treturn regexp.MustCompile(regExp).FindAllString(s, n)\n}", "func group(res ...*regexp.Regexp) *regexp.Regexp {\n\treturn match(`(?:` + expression(res...).String() + `)`)\n}", "func group(res ...*regexp.Regexp) *regexp.Regexp {\n\treturn match(`(?:` + expression(res...).String() + `)`)\n}", "func group(res ...*regexp.Regexp) *regexp.Regexp {\n\treturn match(`(?:` + expression(res...).String() + `)`)\n}", "func FindAllMatch(toMatch string, regx string, selecter int) []string {\n\tre := regexp.MustCompile(regx)\n\tout := re.FindAllStringSubmatch(string(toMatch), -1)\n\ttoReturn := []string{}\n\tfor _, value := range out {\n\t\ttoReturn = append(toReturn, value[selecter])\n\t}\n\treturn toReturn\n}", "func (s String) Filter(modifier func(index int, val string) bool) String {\n\tif len(s) == 0 {\n\t\treturn s\n\t}\n\n\tres := make([]string, 0)\n\tfor index, value := range s {\n\t\tif modifier(index, value) {\n\t\t\tres = append(res, value)\n\t\t}\n\t}\n\n\treturn res\n}", "func Filter(arr interface{}, cond func(interface{}) bool) interface{} {\n\tcontentType := reflect.TypeOf(arr)\n\tcontentValue := reflect.ValueOf(arr)\n\n\tnewContent := reflect.MakeSlice(contentType, 0, 0)\n\tfor i := 0; i < contentValue.Len(); i++ {\n\t\tif content := contentValue.Index(i); cond(content.Interface()) {\n\t\t\tnewContent = reflect.Append(newContent, content)\n\t\t}\n\t}\n\treturn newContent.Interface()\n}", "func (re *RegexpStd) FindStringSubmatch(s string) []string {\n\t// var dstCap [4]int\n\t// a := re.doExecute(nil, nil, s, 0, re.prog.NumCap, dstCap[:0])\n\t// if a == nil {\n\t// \treturn nil\n\t// }\n\t// ret := make([]string, 1+re.numSubexp)\n\t// for i := range ret {\n\t// \tif 2*i < len(a) && a[2*i] >= 0 {\n\t// \t\tret[i] = s[a[2*i]:a[2*i+1]]\n\t// \t}\n\t// }\n\t// return ret\n\tpanic(\"\")\n}", "func regexpFilter(regexps []string, keep bool) (func(string) bool, error) {\n\tif len(regexps) == 0 {\n\t\treturn func(name string) bool {\n\t\t\treturn false\n\t\t}, nil\n\t}\n\tfilterNames := []*regexp.Regexp{}\n\tfor _, n := range regexps {\n\t\tre, err := regexp.Compile(n)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"filter pattern: %s\", n)\n\t\t}\n\t\tfilterNames = append(filterNames, re)\n\t}\n\n\treturn func(name string) bool {\n\t\tfor _, r := range filterNames {\n\t\t\tif r.MatchString(name) {\n\t\t\t\treturn !keep\n\t\t\t}\n\t\t}\n\t\treturn keep\n\t}, nil\n}", "func Regex(path string, r string) ([]string, error) {\n\tclearVars()\n\n\tregex, err := regexp.Compile(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = Find(path, func(s string, i os.FileInfo) {\n\t\tif regex.FindString(i.Name()) != \"\" {\n\t\t\tresults = append(results, s)\n\t\t}\n\t})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn results, nil\n}", "func (re *RegexpStd) FindAllStringSubmatch(s string, n int) [][]string {\n\tm, err := re.p.FindStringMatch(s)\n\tif err != nil {\n\t\tprintln(err.Error())\n\t\treturn nil\n\t}\n\n\tm.populateOtherGroups()\n\tsubs := make([]string, 0, len(m.otherGroups)+1)\n\tsubs = append(subs, m.Group.String())\n\tfor i := 0; i < len(m.otherGroups); i++ {\n\t\tsubs = append(subs, (&m.otherGroups[i]).String())\n\t}\n\n\treturn [][]string{subs}\n}", "func (n *LibPcreRegexp) FindStringSubmatch(s string) []string {\n\tn.currentMatcher = n.re.MatcherString(s, 0)\n\tm := n.currentMatcher\n\n\tif !m.Matches() {\n\t\treturn []string{}\n\t}\n\n\tgroupsCnt := m.Groups() + 1\n\n\tmatches := make([]string, groupsCnt)\n\n\tfor i := 0; i < groupsCnt; i++ {\n\t\tmatches[i] = m.GroupString(i)\n\t}\n\n\treturn matches\n}", "func execmStringSliceSearch(_ int, p *gop.Context) {\n\targs := p.GetArgs(2)\n\tret := args[0].(sort.StringSlice).Search(args[1].(string))\n\tp.Ret(2, ret)\n}", "func subSlice(out, a, b []float64)", "func grep(path string, searchWords, excludeDirs []string) ([]GrepResult, error) {\n\targs := grepExcludeDirStr(excludeDirs)\n\targs = append(args, searchWordsStr(searchWords)...)\n\targs = append(args, \"--recursive\", \"--ignore-case\", \"--only-matching\", path)\n\n\tgrepCmd := exec.Command(\"grep\", args...)\n\tlog.Println(\"running command: \" + strings.Join(grepCmd.Args, \" \"))\n\tgrepOut, err := grepCmd.Output()\n\tif err != nil {\n\t\tvar exitError *exec.ExitError\n\t\tif errors.As(err, &exitError) {\n\t\t\tif exitError.ExitCode() == GrepErrorCodeNoMatches {\n\t\t\t\treturn []GrepResult{}, nil\n\t\t\t}\n\t\t\treturn nil, fmt.Errorf(\"unable to execute grep command: %s\", string(exitError.Stderr))\n\t\t}\n\t\treturn nil, fmt.Errorf(\"unable to execute grep command: %w\", err)\n\t}\n\treturn parseGrepOutput(string(grepOut), path), nil\n}", "func RemoveFromArray(slice []string, input string) []string {\n\tvar output []string\n\tfor i, item := range slice {\n\t\tif item == input {\n\t\t\toutput = append(slice[:i], slice[i+1:]...)\n\t\t\tbreak\n\t\t}\n\t}\n\treturn output\n}", "func MatchAll(text, pattern string) [][]string {\n\tre := regexp.MustCompile(pattern)\n\tvalue := re.FindAllStringSubmatch(text, -1)\n\treturn value\n}", "func (re *RegexpStd) FindAllSubmatchIndex(b []byte, n int) [][]int {\n\t// if n < 0 {\n\t// \tn = len(b) + 1\n\t// }\n\t// var result [][]int\n\t// re.allMatches(\"\", b, n, func(match []int) {\n\t// \tif result == nil {\n\t// \t\tresult = make([][]int, 0, startSize)\n\t// \t}\n\t// \tresult = append(result, match)\n\t// })\n\t// return result\n\tpanic(\"\")\n}", "func sliceSubset(a, b []string) []string {\n\tresults := []string{}\n\n\tfor _, aValue := range a {\n\t\tif !existsInList(b, aValue) {\n\t\t\tresults = append(results, aValue)\n\t\t}\n\t}\n\n\treturn results\n}", "func SelectPrefixInStringSlice(prefix string, items []string) []string {\n\n\tl := len(prefix)\n\n\tvar results []string\n\n\t// iterate through the slice of items\n\tfor _, item := range items {\n\n\t\t// check the item length is geater than or equal to the prefix length\n\t\t// this ensures no out of bounds memory errors will occur\n\t\tif len(item) >= l {\n\t\t\tif prefix == item[:l] {\n\t\t\t\tresults = append(results, item)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn results\n}", "func (s *BuildahTestSession) GrepString(term string) (bool, []string) {\n\tvar (\n\t\tgreps []string\n\t\tmatches bool\n\t)\n\n\tfor _, line := range strings.Split(s.OutputToString(), \"\\n\") {\n\t\tif strings.Contains(line, term) {\n\t\t\tmatches = true\n\t\t\tgreps = append(greps, line)\n\t\t}\n\t}\n\treturn matches, greps\n}", "func IncludeRegexp(expr string) (Filter, error) {\n\treturn regexpFilter(expr, false)\n}", "func extractAmavisRegexpArrayParameter(content string) []string{\n reString := `^new_RE\\((?P<result>.*)\\);`\n re := regexp.MustCompile(reString)\n reResult := []byte{}\n for _, submatches := range re.FindAllStringSubmatchIndex(content, -1) {\n // Apply the captured submatches to the template and append the output\n // to the result.\n reResult = re.ExpandString(reResult, \"$result\", content, submatches)\n }\n resultRaw := strings.Split(string(reResult), \", \")\n result := []string{}\n for pcreExpression := range resultRaw{\n result = append(result, extractAmavisPCREExpression(resultRaw[pcreExpression]))\n }\n return result\n}", "func wildCardToRegexp(pattern string) string {\n\tvar result strings.Builder\n\tfor i, literal := range strings.Split(pattern, \"*\") {\n\n\t\t// Replace * with .*\n\t\tif i > 0 {\n\t\t\tresult.WriteString(\".*\")\n\t\t}\n\n\t\t// Quote any regular expression meta characters in the\n\t\t// literal text.\n\t\tresult.WriteString(regexp.QuoteMeta(literal))\n\t}\n\treturn result.String()\n}", "func CopyFiltered(slice interface{}, funcs ...FilterFunc) interface{} {\n\trv := reflect.ValueOf(slice)\n\tif rv.Kind() != reflect.Slice {\n\t\tpanic(\"not a slice\")\n\t}\n\n\tlength := rv.Len()\n\n\tptrfiltered := reflect.New(rv.Type())\n\tptrfiltered.Elem().Set(\n\t\t//reflect.MakeSlice(rv.Type(), 0, length))\n\t\treflect.MakeSlice(rv.Type(), length, length)) // copy is done by dest[j] = src[i], so it's allocated in advance\n\tfiltered := ptrfiltered.Elem()\n\n\treflect.Copy(filtered, rv)\n\n\tFilter(ptrfiltered.Interface(), funcs...)\n\n\treturn filtered.Interface()\n}", "func GetMatches(queryS string, rCount uint64, sliceData *[]string) *[]string {\n\tvar tempSlice []string\n\tfor i := 0; ( i < len(*sliceData) )&&( uint64(len(tempSlice)) < rCount ); i++ {\n\t\tif strings.HasPrefix(strings.ToLower((*sliceData)[i]), queryS) {\n\t\t\ttempSlice = append(tempSlice,(*sliceData)[i])\n\t\t}\n\t}\n\treturn &tempSlice\n}", "func (iobuf *buf) slice(free, base, bound uint) *Slice {\n\tatomic.AddInt32(&iobuf.refcount, 1)\n\treturn &Slice{iobuf: iobuf, free: free, base: base, Contents: iobuf.Contents[base:bound]}\n}", "func (s *Series) RegexExtract(re *regexp.Regexp, index int) *Series {\n\tif s.column.Dtype != base.Object {\n\t\tpanic(errors.IncorrectDataType(base.Object))\n\t}\n\n\textractedData := make([]interface{}, 0, s.Len())\n\n\tfor i, val := range s.Data {\n\t\tif val == nil {\n\t\t\textractedData = append(extractedData, \"\")\n\t\t\tcontinue\n\t\t}\n\t\tstrVal, ok := val.(string)\n\t\tif !ok {\n\t\t\tpanic(errors.InvalidSeriesValError(val, i, s.column.Name))\n\t\t}\n\n\t\tmatches := re.FindStringSubmatch(strVal)\n\t\tif matches == nil {\n\t\t\textractedData = append(extractedData, \"\")\n\t\t\tcontinue\n\t\t}\n\n\t\tif index > len(matches) {\n\t\t\tpanic(errors.CustomError(\"index is out of range\"))\n\t\t}\n\n\t\textractedData = append(extractedData, matches[index])\n\t}\n\n\tnewS := s.ShallowCopy()\n\tnewS.Data = extractedData\n\tnewS.column.Name = helpers.FunctionNameWrapper(\"regex_extract\", s.column.Name)\n\n\treturn newS\n}", "func parsePattern(pattern string) []string{\n\tvs := strings.Split(pattern, \"/\")\n\tparts := make([]string, 0)\n\tfor _, item := range vs{\n\t\tif item != \"\"{\n\t\t\tparts = append(parts, item)\n\t\t\tif item[0] == '*'{\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn parts\n}", "func MatchAndFindAll(pattern, target string) ([][]string, error) {\n\tr, err := regexp.Compile(pattern)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"[WARN] %s\", err)\n\t}\n\n\tif !r.MatchString(target) {\n\t\treturn nil, fmt.Errorf(\"[WARN] Target '%s' did not match against '%s'\", target, pattern)\n\t}\n\n\treturn r.FindAllStringSubmatch(target, -1), nil\n}", "func (re *RegexpStd) FindAllStringSubmatchIndex(s string, n int) [][]int {\n\t// if n < 0 {\n\t// \tn = len(s) + 1\n\t// }\n\t// var result [][]int\n\t// re.allMatches(s, nil, n, func(match []int) {\n\t// \tif result == nil {\n\t// \t\tresult = make([][]int, 0, startSize)\n\t// \t}\n\t// \tresult = append(result, match)\n\t// })\n\t// return result\n\tpanic(\"\")\n}", "func parsePattern(pattern string) []string {\n\tvs := strings.Split(pattern, \"/\")\n\n\tparts := make([]string, 0)\n\tfor _, item := range vs {\n\t\tif item != \"\" {\n\t\t\tparts = append(parts, item)\n\t\t\tif item[0] == '*' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn parts\n}", "func groupOnlyMatches(group *syntax.Regexp, s string) bool {\n\tswitch group.Op {\n\tcase syntax.OpLiteral:\n\t\tfor _, r := range group.Rune {\n\t\t\tif !strings.ContainsRune(s, r) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t\treturn true\n\n\tcase syntax.OpCharClass:\n\t\tfor i := 0; i < len(group.Rune); i += 2 {\n\t\t\tlo, hi := group.Rune[i], group.Rune[i+1]\n\t\t\tfor r := lo; r <= hi; r++ {\n\t\t\t\tif !strings.ContainsRune(s, r) {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn true\n\n\tcase syntax.OpStar, syntax.OpPlus, syntax.OpRepeat, syntax.OpQuest, syntax.OpCapture:\n\t\treturn groupOnlyMatches(group.Sub[0], s)\n\n\tcase syntax.OpConcat, syntax.OpAlternate:\n\t\tfor _, sub := range group.Sub {\n\t\t\tif !groupOnlyMatches(sub, s) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\n\tdefault:\n\t\treturn false\n\t}\n\treturn true\n}", "func ExtractSubmatch(re *regexp.Regexp, src []byte, values ...interface{}) error {\n\tif re.NumSubexp() != len(values) {\n\t\treturn ErrCount\n\t}\n\tsm := re.FindSubmatch(src)\n\tif sm == nil {\n\t\treturn ErrMatch\n\t}\n\tfor i, val := range values {\n\t\tif err := extractTo(val, sm[i+1]); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func GetPatternMatches(source, pattern string) ([]*PatternMatch, error) {\n\tif source == \"\" {\n\t\treturn nil, errors.New(\"source glob is required\")\n\t}\n\n\tif pattern == \"\" {\n\t\treturn nil, errors.New(\"pattern is required\")\n\t}\n\n\trgx, err := regexp.Compile(pattern)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to compile regex pattern: %s\", pattern)\n\t}\n\n\tmatches, err := filepath.Glob(source)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to find files matching glob (%s): %s\", source, err.Error())\n\t}\n\n\tmLen := len(matches)\n\n\tif mLen == 0 {\n\t\treturn nil, fmt.Errorf(\"no files were found for glob (%s)\", source)\n\t}\n\n\tpms := make([]*PatternMatch, mLen, mLen)\n\n\tfor i, path := range matches {\n\t\tpathMatches := rgx.FindAllStringSubmatch(path, -1)\n\n\t\tpm := PatternMatch{\n\t\t\tPath: path,\n\t\t\tPathMatches: pathMatches,\n\t\t}\n\n\t\tpms[i] = &pm\n\t}\n\n\treturn pms, nil\n}", "func (i *Iter) returnSlice(a, b int) []byte {\n\tif i.rb.src.bytes == nil {\n\t\treturn i.buf[:copy(i.buf[:], i.rb.src.str[a:b])]\n\t}\n\treturn i.rb.src.bytes[a:b]\n}", "func Search(txt, pat string) (offsets []int) {\n\treturn Find(txt, pat, FailT(pat))\n}", "func GenRangeByPattern(p string) (shSlice []string) {\n\tfor _, shId := range shRange {\n\t\tshSlice = append(shSlice, strings.Replace(p, \"%s\", shId, -1))\n\t}\n\treturn\n}", "func extensionPattern(pattern *regexp.Regexp) *regexp.Regexp {\n\treturn suffixPattern(regexp.MustCompile(\"(^|/)[^/]+.\" + pattern.String()))\n}", "func segmentPattern(pattern *regexp.Regexp) *regexp.Regexp {\n\treturn regexp.MustCompile(\"(^|/)\" + pattern.String() + \"($|/)\")\n}", "func suffixPattern(pattern *regexp.Regexp) *regexp.Regexp {\n\treturn regexp.MustCompile(pattern.String() + \"$\")\n}", "func Regex() {\n\n\t// This tests whether a pattern matches a string.\n\tmatch, _ := regexp.MatchString(\"p([a-z]+)ch\", \"peach\")\n\tfmt.Println(match)\n\n\t// Above we used a string pattern directly,\n\t// but for other regexp tasks you’ll need to Compile an optimized\n\t// Regexp struct.\n\tr, _ := regexp.Compile(\"p([a-z]+)ch\")\n\n\t// Many methods are available on these structs.\n\t// Here’s a match test like we saw earlier.\n\tfmt.Println(r.MatchString(\"peach\"))\n\n\t// This finds the match for the regexp.\n\tfmt.Println(r.FindString(\"peach punch\"))\n\n\t// This also finds the first match but returns the start\n\t// and end indexes for the match instead of the matching text.\n\tfmt.Println(r.FindStringIndex(\"peach punch\"))\n\n\t// The Submatch variants include information about both\n\t// the whole-pattern matches and the submatches within those matches.\n\t// For example this will return information for both p([a-z]+)ch and ([a-z]+).\n\tfmt.Println(r.FindStringSubmatch(\"peach punch\"))\n\n\t// Similarly this will return information about the indexes of matches and submatches.\n\tfmt.Println(r.FindStringSubmatchIndex(\"peach punch\"))\n\n\t// The All variants of these functions apply to all matches in the input,\n\t// not just the first. For example to find all matches for a regexp.\n\tfmt.Println(r.FindAllString(\"peach punch pinch azerty\", -1))\n\n\t// These All variants are available for the other functions we saw above as well.\n\tfmt.Println(r.FindAllStringSubmatchIndex(\"peach punch pinch azert\", -1))\n\n\t// Providing a non-negative integer as the second argument\n\t// to these functions will limit the number of matches.\n\tfmt.Println(r.FindAllString(\"peach punch pinch azert\", 2))\n\n\t// Our examples above had string arguments and used names like MatchString.\n\t// We can also provide []byte arguments and drop String from the function name.\n\tfmt.Println(r.Match([]byte(\"peach\")))\n\n\t// When creating constants with regular expressions\n\t// you can use the MustCompile variation of Compile.\n\t// A plain Compile won’t work for constants because it has 2 return values.\n\tr = regexp.MustCompile(\"p([a-z]+)ch\")\n\tfmt.Println(r)\n\n\t// The regexp package can also be used to replace subsets of strings with other values.\n\tfmt.Println(r.ReplaceAllString(\"a peach\", \"<fruit>\"))\n\n\t// The Func variant allows you to transform matched text with a given function.\n\tin := []byte(\"a peach\")\n\tout := r.ReplaceAllFunc(in, bytes.ToUpper)\n\tfmt.Println(string(out))\n}", "func GroupBy(p func(rune, rune) bool, s string) []string {\n\tss := []string{}\n\tfor len(s) > 0 {\n\t\tr0, n := utf8.DecodeRuneInString(s)\n\t\tt := TakeWhile(func(r rune) bool {\n\t\t\treturn p(r0, r)\n\t\t}, s[n:])\n\t\tn += len(t)\n\t\tss = append(ss, s[0:n])\n\t\ts = s[n:]\n\t}\n\treturn ss\n}", "func ArrayInStringToRegularExpression(arrayString string) string {\n\tif arrayString == \"x\" || len(arrayString) == 0 {\n\t\treturn allThingsRegex\n\t}\n\tarray := strings.Split(arrayString, \",\")\n\tif len(array) == 1 {\n\t\treturn fmt.Sprint(\"[0-9,]*\", array[0], \"[0-9,]*\")\n\t}\n\treturn fmt.Sprint(\"[0-9,]*(\", strings.Join(array, \"|\"), \")[0-9,]*\")\n}", "func (items Float64Slice) SubSlice(i, j int) Interface { return items[i:j] }", "func (s strings) Filter(in []string, keep func(item string) bool) []string {\n\tvar out []string\n\tfor _, item := range in {\n\t\tif keep(item) {\n\t\t\tout = append(out, item)\n\t\t}\n\t}\n\n\treturn out\n}", "func Regexp(expr string) (func(string) bool, error) {\n\tif expr == \"\" {\n\t\treturn nil, fmt.Errorf(\"empty regex expression\")\n\t}\n\n\t// add the last $ if missing (and not wildcard(?))\n\tif i := expr[len(expr)-1]; i != '$' && i != '*' {\n\t\texpr += \"$\"\n\t}\n\n\tr, err := regexp.Compile(expr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn r.MatchString, nil\n}", "func builtinGrep(data string) map[string][]string {\n\t// \"pattern name\":[\"list of grep\"]\n\tresult := make(map[string][]string)\n\t// \"pattern name\":\"regex\"\n\tpattern := map[string]string{\n\t\t\"dalfox-ssti\": \"2958816\",\n\t\t\"dalfox-rsa-key\": \"-----BEGIN RSA PRIVATE KEY-----|-----END RSA PRIVATE KEY-----\",\n\t\t\"dalfox-priv-key\": \"-----BEGIN PRIVATE KEY-----|-----END PRIVATE KEY-----\",\n\t\t\"dalfox-aws-s3\": \"s3\\\\.amazonaws.com[/]+|[a-zA-Z0-9_-]*\\\\.s3\\\\.amazonaws.com\",\n\t\t\"dalfox-slack-webhook\": \"https://hooks.slack.com/services/T[a-zA-Z0-9_]{8}/B[a-zA-Z0-9_]{8}/[a-zA-Z0-9_]{24}\",\n\t\t\"dalfox-slack-token\": \"(xox[p|b|o|a]-[0-9]{12}-[0-9]{12}-[0-9]{12}-[a-z0-9]{32})\",\n\t\t\"dalfox-facebook-oauth\": \"[f|F][a|A][c|C][e|E][b|B][o|O][o|O][k|K].{0,30}['\\\"\\\\s][0-9a-f]{32}['\\\"\\\\s]\",\n\t\t\"dalfox-twitter-oauth\": \"[t|T][w|W][i|I][t|T][t|T][e|E][r|R].{0,30}['\\\"\\\\s][0-9a-zA-Z]{35,44}['\\\"\\\\s]\",\n\t\t\"dalfox-heroku-api\": \"[h|H][e|E][r|R][o|O][k|K][u|U].{0,30}[0-9A-F]{8}-[0-9A-F]{4}-[0-9A-F]{4}-[0-9A-F]{4}-[0-9A-F]{12}\",\n\t\t\"dalfox-mailgun-api\": \"key-[0-9a-zA-Z]{32}\",\n\t\t\"dalfox-mailchamp-api\": \"[0-9a-f]{32}-us[0-9]{1,2}\",\n\t\t\"dalfox-picatic-api\": \"sk_live_[0-9a-z]{32}\",\n\t\t\"dalfox-google-oauth-id\": \"[0-9(+-[0-9A-Za-z_]{32}.apps.qooqleusercontent.com\",\n\t\t\"dalfox-google-api\": \"AIza[0-9A-Za-z-_]{35}\",\n\t\t\"dalfox-google-captcha\": \"6L[0-9A-Za-z-_]{38}\",\n\t\t\"dalfox-google-oauth\": \"ya29\\\\.[0-9A-Za-z\\\\-_]+\",\n\t\t\"dalfox-aws-access-key\": \"AKIA[0-9A-Z]{16}\",\n\t\t\"dalfox-amazon-mws-auth-token\": \"amzn\\\\.mws\\\\.[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}\",\n\t\t\"dalfox-facebook-access-token\": \"EAACEdEose0cBA[0-9A-Za-z]+\",\n\t\t\"dalfox-authorization-basic\": \"basic [a-zA-Z0-9_\\\\-:\\\\.]+\",\n\t\t\"dalfox-authorization-beare\": \"bearer [a-zA-Z0-9_\\\\-\\\\.]+\",\n\t\t\"dalfox-github-access-token\": \"[a-zA-Z0-9_-]*:[a-zA-Z0-9_\\\\-]+@github\\\\.com*\",\n\t\t\"dalfox-azure-storage\": \"[a-zA-Z0-9_-]*\\\\.file.core.windows.net\",\n\t\t\"dalfox-access-key\": \"[Aa](ccess|CCESS)_?[Kk](ey|EY)|[Aa](ccess|CCESS)_?[sS](ecret|ECRET)|[Aa](ccess|CCESS)_?(id|ID|Id)\",\n\t\t\"dalfox-secret-key\": \"[Ss](ecret|ECRET)_?[Kk](ey|EY)\",\n\t\t\"dalfox-error-mysql\":\"(SQL syntax.*MySQL|Warning.*mysql_.*|MySqlException \\\\(0x|valid MySQL result|check the manual that corresponds to your (MySQL|MariaDB) server version|MySqlClient\\\\.|com\\\\.mysql\\\\.jdbc\\\\.exceptions)\",\n\t\t\"dalfox-error-postgresql\":\"(PostgreSQL.*ERROR|Warning.*\\\\Wpg_.*|valid PostgreSQL result|Npgsql\\\\.|PG::SyntaxError:|org\\\\.postgresql\\\\.util\\\\.PSQLException|ERROR:\\\\s\\\\ssyntax error at or near)\",\n\t\t\"dalfox-error-mssql\":\"(Driver.* SQL[\\\\-\\\\_\\\\ ]*Server|OLE DB.* SQL Server|\\bSQL Server.*Driver|Warning.*mssql_.*|\\bSQL Server.*[0-9a-fA-F]{8}|[\\\\s\\\\S]Exception.*\\\\WSystem\\\\.Data\\\\.SqlClient\\\\.|[\\\\s\\\\S]Exception.*\\\\WRoadhouse\\\\.Cms\\\\.|Microsoft SQL Native Client.*[0-9a-fA-F]{8})\",\n\t\t\"dalfox-error-msaccess\":\"(Microsoft Access (\\\\d+ )?Driver|JET Database Engine|Access Database Engine|ODBC Microsoft Access)\",\n\t\t\"dalfox-error-oracle\":\"(\\\\bORA-\\\\d{5}|Oracle error|Oracle.*Driver|Warning.*\\\\Woci_.*|Warning.*\\\\Wora_.*)\",\n\t\t\"dalfox-error-ibmdb2\":\"(CLI Driver.*DB2|DB2 SQL error|\\\\bdb2_\\\\w+\\\\(|SQLSTATE.+SQLCODE)\",\n\t\t\"dalfox-error-informix\":\"(Exception.*Informix)\",\n\t\t\"dalfox-error-firebird\":\"(Dynamic SQL Error|Warning.*ibase_.*)\",\n\t\t\"dalfox-error-sqlite\":\"(SQLite\\\\/JDBCDriver|SQLite.Exception|System.Data.SQLite.SQLiteException|Warning.*sqlite_.*|Warning.*SQLite3::|\\\\[SQLITE_ERROR\\\\])\",\n\t\t\"dalfox-error-sapdb\":\"(SQL error.*POS([0-9]+).*|Warning.*maxdb.*)\",\n\t\t\"dalfox-error-sybase\":\"(Warning.*sybase.*|Sybase message|Sybase.*Server message.*|SybSQLException|com\\\\.sybase\\\\.jdbc)\",\n\t\t\"dalfox-error-ingress\":\"(Warning.*ingres_|Ingres SQLSTATE|Ingres\\\\W.*Driver)\",\n\t\t\"dalfox-error-frontbase\":\"(Exception (condition )?\\\\d+. Transaction rollback.)\",\n\t\t\"dalfox-error-hsqldb\":\"(org\\\\.hsqldb\\\\.jdbc|Unexpected end of command in statement \\\\[|Unexpected token.*in statement \\\\[)\",\n\t}\n\tfor k, v := range pattern {\n\t\tresultArr := Grepping(data, v)\n\t\tif len(resultArr) > 0 {\n\t\t\tresult[k] = resultArr\n\t\t}\n\t}\n\n\treturn result\n}", "func (x *Index) FindAllIndex(r *regexp.Regexp, n int) (result [][]int)", "func NasGrep(nases []Nas, s string) []Nas {\n\tvar n []Nas\n\tfor i := range nases {\n\t\ts = strings.ToLower(s)\n\t\taddr := strings.ToLower(nases[i].Street + \" \" + nases[i].Build)\n\t\tif strings.Index(addr, s) >= 0 ||\n\t\t\tstrings.Index(nases[i].IP.String(), s) >= 0 ||\n\t\t\tstrings.Index(nases[i].MAC.String(), s) >= 0 {\n\t\t\tn = append(n, nases[i])\n\t\t}\n\t}\n\treturn n\n}", "func (a myArray) splice(start int, data ...string) myArray {\n\tcopy(a[start:], data)\n\treturn a\n}", "func (re *Regexp) MatchStrings(s string) (a []string) {\n\tr := re.doExecute(s, nil, 0);\n\tif r == nil {\n\t\treturn nil\n\t}\n\ta = make([]string, len(r)/2);\n\tfor i := 0; i < len(r); i += 2 {\n\t\tif r[i] != -1 {\t// -1 means no match for this subexpression\n\t\t\ta[i/2] = s[r[i]:r[i+1]]\n\t\t}\n\t}\n\treturn;\n}", "func ExtractFromRegex(regex string, data string) ([]string, error) {\n\tif regex == \"\" {\n\t\treturn nil, errors.New(\"regex must be provied\")\n\t}\n\n\tr, err := regexp.Compile(regex)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmatch := r.FindStringSubmatch(data)\n\treturn match, nil\n}", "func Horspool(pattern string, text *string) ([]int) {\n matches := []int{}\n\n if len(pattern) == 0 || len(*text) == 0 {\n return matches\n }\n\n shiftMap := horspoolShiftMap(pattern)\n\n // find match\n for textI := 0; textI < len(*text) - len(pattern) + 1; {\n\n // loop over pattern\n match := false\n for pPos, tPos := 0, textI; pPos < len(pattern); pPos++ {\n if pattern[pPos] == (*text)[tPos] {\n match = true\n } else {\n match = false\n pPos = len(pattern)\n }\n tPos++\n }\n\n if match {\n matches = append(matches, textI)\n }\n\n // shift\n shiftAmount, found := shiftMap[(*text)[textI + len(pattern) - 1]]\n if found {\n textI += shiftAmount\n } else {\n textI += len(pattern)\n }\n }\n\n return matches\n}", "func TestFindStringSubmatch(t *testing.T) {\n\tregex := regexp.MustCompile(\"Hello.*(world)\")\n\tsubject := \"Hello brave new world\"\n\tmatches := regex.FindStringSubmatch(subject)\n\tAssert(\"world\", matches[1], t)\n}", "func filter(s []string, fn func(string) bool) []string {\n\tvar p []string\n\tfor _, v := range s {\n\t\tif fn(v) {\n\t\t\tp = append(p, v)\n\t\t}\n\t}\n\treturn p\n}", "func getMatchingText(s string, re *regexp.Regexp) []string {\n\tvar matches []string\n\tfor _, m := range re.FindAllStringSubmatch(s, -1) {\n\t\tmatches = append(matches, m[0])\n\t}\n\treturn matches\n}", "func (me Tokens) FindSub(beginsWith Tokens, endsWith Tokens) (slice Tokens) {\n\tbeginner, ender := beginsWith.First(nil), endsWith.Last(nil)\n\tif slice = me.FromUntil(beginner, ender, true); slice == nil {\n\t\tvar db, de bool\n\t\tfor i := range me {\n\t\t\tif (!db) && me[i].Pos.Off0 == beginner.Pos.Off0 {\n\t\t\t\tdb, beginner = true, &me[i]\n\t\t\t} else if db && me[i].Pos.Off0 == ender.Pos.Off0 {\n\t\t\t\tde, ender = true, &me[i]\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif de {\n\t\t\tslice = me.FromUntil(beginner, ender, true)\n\t\t}\n\t}\n\treturn\n}", "func (r Report) ExcludeRegexp(pattern ...string) (result Report) {\n\tif len(pattern) == 0 {\n\t\treturn r\n\t}\n\n\tregexps := make([]*regexp.Regexp, len(pattern))\n\tfor i := range pattern {\n\t\tregexps[i] = regexp.MustCompile(pattern[i])\n\t}\n\n\treturn r.filter(func(filterPath *ytbx.Path) bool {\n\t\tfor _, regexp := range regexps {\n\t\t\tif filterPath != nil && regexp.MatchString(filterPath.String()) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t\treturn true\n\t})\n}", "func wildCardToRegexp(pattern string) string {\n\tvar result strings.Builder\n\tresult.WriteString(\"(?i)\")\n\n\trpattern := strings.Replace(pattern, \"%\", \".*\", -1)\n\trpattern = strings.Replace(rpattern, \"_\", \".+\", -1)\n\tresult.WriteString(rpattern)\n\n\treturn result.String()\n}", "func subtractSlice(x, y []string) []string {\n\tm := make(map[string]bool)\n\n\tfor _, y := range y {\n\t\tm[y] = true\n\t}\n\n\tvar ret []string\n\tfor _, x := range x {\n\t\tif m[x] {\n\t\t\tcontinue\n\t\t}\n\t\tret = append(ret, x)\n\t}\n\n\treturn ret\n}", "func (items IntSlice) SubSlice(i, j int) Interface { return items[i:j] }", "func delete(slice []string, el string) (a []string) {\n\ti := -1\n\tfor j, s := range slice {\n\t\tif s == el {\n\t\t\ti = j\n\t\t}\n\t}\n\ta = append(slice[:i], slice[i+1:]...)\n\treturn a\n}", "func (s *Series) RegexContains(re *regexp.Regexp) *Series {\n\tif s.column.Dtype != base.Object {\n\t\tpanic(errors.IncorrectDataType(base.Object))\n\t}\n\n\textractedData := make([]interface{}, 0, s.Len())\n\n\tfor i, val := range s.Data {\n\t\tif val == nil {\n\t\t\textractedData = append(extractedData, false)\n\t\t\tcontinue\n\t\t}\n\t\tstrVal, ok := val.(string)\n\t\tif !ok {\n\t\t\tpanic(errors.InvalidSeriesValError(val, i, s.column.Name))\n\t\t}\n\n\t\tmatch := re.MatchString(strVal)\n\t\textractedData = append(extractedData, match)\n\t}\n\n\tnewS := s.ShallowCopy()\n\tnewS.Data = extractedData\n\tnewS.column.Name = helpers.FunctionNameWrapper(\"regex_contains\", s.column.Name)\n\tnewS.column.Dtype = base.Bool\n\n\treturn newS\n}", "func R(pattern string) *regexp.Regexp {\n\treturn regexp.MustCompile(pattern)\n}", "func RegexpBuilder(spec FilterSpec) (Filter, error) {\n\tswitch strings.ToLower(spec.Type) {\n\tcase \"regexp\", \"regex\", \"re\":\n\tdefault:\n\t\treturn nil, nil\n\t}\n\n\tswitch len(spec.Args) {\n\tcase 0, 1:\n\t\treturn nil, errors.New(\"regular expression filter requires a subject and an expression\")\n\tcase 2:\n\t\tsubject, expression := spec.Args[0], spec.Args[1]\n\n\t\t// Force case-insensitive matching\n\t\tif !strings.HasPrefix(expression, \"(?i)\") {\n\t\t\texpression = \"(?i)\" + expression\n\t\t}\n\n\t\tre, err := regexp.Compile(expression)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"invalid expression \\\"%s\\\": %v\", expression, err)\n\t\t}\n\n\t\treturn regexpFilter{\n\t\t\tsubject: subject,\n\t\t\tre: re,\n\t\t}.Filter, nil\n\tdefault:\n\t\treturn nil, errors.New(\"regular expression filter has %d arguments when two are needed\")\n\t}\n}", "func extractTags(pattern string) []string {\n\tif len(pattern) == 0 {\n\t\treturn nil // empty pattern is not allowed.\n\t}\n\tsplitted := strings.Split(pattern, \"%\")\n\tif len(splitted)%2 == 0 {\n\t\treturn nil // invalid tags.\n\t}\n\n\tresult := make([]string, len(splitted)/2, len(splitted)/2)\n\tfor index, token := range splitted {\n\t\tif isTagPortion(index) {\n\t\t\tif len(token) == 0 {\n\t\t\t\treturn nil // no empty tag\n\t\t\t}\n\t\t\tresult[index/2] = token\n\t\t}\n\t}\n\n\t// check for duplicates.\n\texists := make(map[string]bool)\n\tfor _, token := range result {\n\t\tif exists[token] {\n\t\t\treturn nil // no duplicate\n\t\t}\n\t\texists[token] = true\n\t}\n\treturn result\n}", "func (ix *Search) filterFileIndicesForRegexpMatch(post []uint32, re *regexp.Regexp, fnames []uint32) []uint32 {\n\t// This loop could conceivably be over all of the filenames. This could\n\t// be large. Keeping the body efficient has large impact.\n\tfor i := 0; len(fnames) < MaximumMatches && i < len(post); i++ {\n\t\tfileid := post[i]\n\n\t\tname := ix.NameBytes(fileid)\n\t\tsname := ix.trimmer(name)\n\n\t\tif re.Match(sname, true, true) >= 0 {\n\t\t\tfnames = append(fnames, fileid)\n\t\t\tcontinue\n\t\t}\n\t}\n\treturn fnames\n}", "func (arr *Array) Slice(i, j int) *Array {\n\tvar elems []*Term\n\tvar hashs []int\n\tif j == -1 {\n\t\telems = arr.elems[i:]\n\t\thashs = arr.hashs[i:]\n\t} else {\n\t\telems = arr.elems[i:j]\n\t\thashs = arr.hashs[i:j]\n\t}\n\t// If arr is ground, the slice is, too.\n\t// If it's not, the slice could still be.\n\tgr := arr.ground || termSliceIsGround(elems)\n\n\ts := &Array{elems: elems, hashs: hashs, ground: gr}\n\ts.rehash()\n\treturn s\n}", "func (ti *Index) Filter(textTags, metrics []string) []string {\n\tmatches := []string{}\n\tintersectionCounts := make([]int, len(metrics))\n\tfor _, tag := range textTags {\n\t\tsearch := strings.TrimPrefix(tag, ti.textMatchPrefix)\n\t\t// broken pin -> no possible matches -> empty intersection\n\t\tif search[0] == '$' || search[len(search)-1] == '^' {\n\t\t\treturn []string{}\n\t\t}\n\t\tcaret := search[0] == '^'\n\t\tdollar := search[len(search)-1] == '$'\n\t\tnonpositional := strings.Trim(search, \"^$\")\n\n\t\t// TODO: maybe this is slow? map based intersect and such\n\t\t// this case is a little silly, since you should probably just query\n\t\t// graphite for that metric directly\n\t\tif caret && dollar {\n\t\t\tfor i, rawMetric := range metrics {\n\t\t\t\tif rawMetric == nonpositional {\n\t\t\t\t\tintersectionCounts[i]++\n\t\t\t\t\tif intersectionCounts[i] == len(textTags) {\n\t\t\t\t\t\tmatches = append(matches, rawMetric)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t} else if caret {\n\t\t\tfor i, rawMetric := range metrics {\n\t\t\t\tif strings.HasPrefix(rawMetric, nonpositional) {\n\t\t\t\t\tintersectionCounts[i]++\n\t\t\t\t\tif intersectionCounts[i] == len(textTags) {\n\t\t\t\t\t\tmatches = append(matches, rawMetric)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t} else if dollar {\n\t\t\tfor i, rawMetric := range metrics {\n\t\t\t\tif strings.HasSuffix(rawMetric, nonpositional) {\n\t\t\t\t\tintersectionCounts[i]++\n\t\t\t\t\tif intersectionCounts[i] == len(textTags) {\n\t\t\t\t\t\tmatches = append(matches, rawMetric)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tfor i, rawMetric := range metrics {\n\t\t\t\tif strings.Contains(rawMetric, nonpositional) {\n\t\t\t\t\tintersectionCounts[i]++\n\t\t\t\t\tif intersectionCounts[i] == len(textTags) {\n\t\t\t\t\t\tmatches = append(matches, rawMetric)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn matches\n}", "func FilterEpdLines(lines []EpdLine, regex string) []EpdLine {\n\tmatchingLines := make([]EpdLine, 0)\n\tfor _, line := range lines {\n\t\tres, err := regexp.MatchString(regex, line.name)\n\t\tif err != nil {\n\t\t\t// shouldn't be possible (?)\n\t\t\tpanic(err)\n\t\t}\n\n\t\tif res {\n\t\t\tmatchingLines = append(matchingLines, line)\n\t\t}\n\t}\n\n\treturn matchingLines\n}" ]
[ "0.6164099", "0.59633416", "0.58574915", "0.5823077", "0.5570078", "0.54958916", "0.5297229", "0.528321", "0.5275069", "0.5222033", "0.52046746", "0.51630586", "0.50914747", "0.502559", "0.49922454", "0.49668378", "0.49371526", "0.492161", "0.49108174", "0.4897063", "0.4897063", "0.4897063", "0.48961708", "0.48942924", "0.48890144", "0.48470396", "0.48372212", "0.48336363", "0.48247525", "0.47824833", "0.4772238", "0.4772238", "0.4772238", "0.476954", "0.47670868", "0.4759186", "0.47395483", "0.4738491", "0.47026506", "0.4700094", "0.46797657", "0.4644594", "0.46393052", "0.46328366", "0.4618057", "0.46136633", "0.45595688", "0.45564654", "0.45448583", "0.45358798", "0.45209736", "0.45120308", "0.45075068", "0.45031705", "0.4499505", "0.44857988", "0.4481614", "0.447101", "0.44640815", "0.44599798", "0.4445018", "0.44416496", "0.44085017", "0.43935543", "0.4393284", "0.43914053", "0.4391215", "0.43904817", "0.43898544", "0.43882832", "0.43865982", "0.43865877", "0.43844035", "0.43739307", "0.43714806", "0.4371453", "0.43661284", "0.43623608", "0.43451363", "0.43385628", "0.4336437", "0.43269038", "0.43133196", "0.43078455", "0.43056017", "0.4304273", "0.4303546", "0.42948487", "0.42922175", "0.42850274", "0.4273102", "0.42727852", "0.4264671", "0.4260488", "0.42472425", "0.42461714", "0.4245453", "0.42436022", "0.42419595", "0.4240101" ]
0.68440825
0
Head returns the first n elements of operand. If less than n elements are in operand, it returns all of operand.
func Head(n int, operand []string) []string { if len(operand) < n { return operand } return operand[:n] }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s VectOp) Head(n int) VectOp {\n\treturn On(s[:n])\n}", "func Tail(n int, operand []string) []string {\n\tif len(operand) < n {\n\t\treturn operand\n\t}\n\treturn operand[len(operand)-n:]\n}", "func (s Stream) Head(n int) Stream {\n\tvar mod Modifer\n\tif n >= 0 {\n\t\th := head(n)\n\t\tmod = &h\n\t} else {\n\t\tt := make(tail, 0, -n)\n\t\tmod = &t\n\t}\n\treturn s.Modify(mod)\n}", "func Head(l List) interface{} {\n\tif l == nil {\n\t\tl = Mzero()\n\t}\n\tif _, ok := l.(unit); ok {\n\t\treturn unit{}\n\t}\n\tlf := l.([2]interface{})[0].(func() interface{})\n\treturn lf()\n}", "func NthElement[T any](first, last iterator.RandomAccessIterator[T], n int, cmp comparator.Comparator[T]) {\n\tif first.Position() < 0 || last.Position()-first.Position() < n {\n\t\treturn\n\t}\n\tlen := last.Position() - first.Position()\n\tnthElement(first, last.IteratorAt(first.Position()+len-1), n, cmp)\n}", "func (n Nodes) First() *Node", "func Head(arr interface{}) interface{} {\n\tvalue := redirectValue(reflect.ValueOf(arr))\n\tvalueType := value.Type()\n\n\tkind := value.Kind()\n\n\tif kind == reflect.Array || kind == reflect.Slice {\n\t\tif value.Len() == 0 {\n\t\t\treturn nil\n\t\t}\n\n\t\treturn value.Index(0).Interface()\n\t}\n\n\tpanic(fmt.Sprintf(\"Type %s is not supported by Head\", valueType.String()))\n}", "func Head(in chan string, n int) chan string {\n\tout := make(chan string)\n\tgo func() {\n\t\tfor l := range in {\n\t\t\tif n > 0 {\n\t\t\t\tout <- l\n\t\t\t\tn--\n\t\t\t}\n\t\t}\n\t\tclose(out)\n\t}()\n\treturn out\n}", "func Head(t Term) Callable {\n\treturn t.(*Compound).Arguments()[0].(Callable)\n}", "func nth_element(buildData []bvhPrimitiveInfo, first, nth, last int, comp func(info0, info1 *bvhPrimitiveInfo) bool) {\n\tBy(comp).Sort(buildData[first:last])\n}", "func (q *OperationQueue) Top(n int) []*SignedOperation {\n\tanswer := []*SignedOperation{}\n\tfor _, item := range q.set.Values() {\n\t\tanswer = append(answer, item.(*SignedOperation))\n\t\tif len(answer) == n {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn answer\n}", "func (s Sequence) FirstN(n int) []interface{} {\n\tr := make([]interface{}, n)\n\tx := 0\n\ts.Find(func(el El)bool{\n\t\tr[x] = el\n\t\tx++\n\t\treturn x == n\n\t})\n\treturn r\n}", "func (v Value) Head(subquery bool) bool {\n\tif v.Kind != Literal {\n\t\treturn false\n\t}\n\tlits := headLiterals[:]\n\tif subquery {\n\t\tlits = sqLiterals[:]\n\t}\n\tfor _, h := range lits {\n\t\tif v.Canon == h {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (x *Lazy) Take(n int, ar AnyValue) {\n\toar := reflect.ValueOf(ar)\n\tfor i := 0; i < n; i++ {\n\t\tif v, ok := x.omap[i]; ok {\n\t\t\toar.Index(i).Set(v)\n\t\t\tcontinue\n\t\t}\n\t\tvar v = []reflect.Value{x.iar.Index(i)}\n\t\tfor j := 0; j < len(x.fns); j++ {\n\t\t\tv = x.fns[j].Call(v)\n\t\t}\n\t\toar.Index(i).Set(v[0])\n\t\tx.omap[i] = oar.Index(i)\n\t}\n}", "func (pq *askPQueue) Head() *models.Ask {\n\tpq.RLock()\n\tdefer pq.RUnlock()\n\n\tif pq.size() < 1 {\n\t\treturn nil\n\t}\n\n\theadValue := pq.items[1].value\n\n\treturn headValue\n}", "func Head(data interface{}) (interface{}, error) {\n\treturn First(data)\n}", "func (s *Script) getNthOp(pcStart, n int) (OpCode, Operand, int /* pc */, error) {\n\topCode, operand, newPc, err := OpCode(0), Operand(nil), 0, error(nil)\n\n\tfor pc, i := pcStart, 0; i <= n; i++ {\n\t\topCode, operand, newPc, err = s.parseNextOp(pc)\n\t\tif err != nil {\n\t\t\treturn 0, nil, 0, err\n\t\t}\n\t\tpc = newPc\n\t}\n\treturn opCode, operand, newPc, err\n}", "func TestNth(t *T) {\n\t// Normal case, in bounds\n\tintl := []interface{}{0, 2, 4, 6, 8}\n\tl := NewList(intl...)\n\tr, ok := l.Nth(3)\n\tassertSaneList(l, t)\n\tassert.Equal(t, intl, ToSlice(l))\n\tassert.Equal(t, 6, r)\n\tassert.Equal(t, true, ok)\n\n\t// Normal case, out of bounds\n\tr, ok = l.Nth(8)\n\tassertSaneList(l, t)\n\tassert.Equal(t, intl, ToSlice(l))\n\tassert.Equal(t, nil, r)\n\tassert.Equal(t, false, ok)\n\n\t// Degenerate case\n\tl = NewList()\n\tr, ok = l.Nth(0)\n\tassert.Equal(t, 0, Size(l))\n\tassert.Equal(t, nil, r)\n\tassert.Equal(t, false, ok)\n}", "func (lst List) Nth(n Number) LangType {\n\tnode := lst.head\n\tN := int(n)\n\tfor i := 0; i < N; i++ {\n\t\tnode = node.next\n\t}\n\treturn node.value\n}", "func first_n_of_five(arr [5]int, n int) []int {\n // If n is out of range, return a slice of full array\n if n > 5 || n < 0 {\n fmt.Println(\"n must be between 0 and 5. Full array:\")\n return arr[:]\n }\n // Look at slice of array\n return arr[0:n]\n}", "func (s *summary) HeadSum(idx int) (sum float64) {\n\treturn float64(sumUntilIndex(s.counts, idx))\n}", "func TestNth(t *T) {\n\t// Normal case, in bounds\n\tintl := []interface{}{0, 2, 4, 6, 8}\n\tl := NewList(intl...)\n\tr, ok := l.Nth(3)\n\tassertSaneList(l, t)\n\tassertSeqContents(l, intl, t)\n\tassertValue(r, 6, t)\n\tassertValue(ok, true, t)\n\n\t// Normal case, out of bounds\n\tr, ok = l.Nth(8)\n\tassertSaneList(l, t)\n\tassertSeqContents(l, intl, t)\n\tassertValue(r, nil, t)\n\tassertValue(ok, false, t)\n\n\t// Degenerate case\n\tl = NewList()\n\tr, ok = l.Nth(0)\n\tassertEmpty(l, t)\n\tassertValue(r, nil, t)\n\tassertValue(ok, false, t)\n}", "func NthElement(v []int, border int) {\n\tleft := 0\n\tright := len(v) - 1\n\tfor {\n\t\tif left >= border || right <= border {\n\t\t\tbreak\n\t\t}\n\t\tnth := quickSelect(v, left, right, border)\n\t\tif nth == border { // left <= nth, border <= right\n\t\t\tbreak\n\t\t}\n\t\tif nth < border {\n\t\t\tleft = nth\n\t\t} else {\n\t\t\tright = nth\n\t\t}\n\t}\n}", "func (this *Tuple) Left(n int) *Tuple {\n\treturn this.Slice(0, n)\n}", "func (ts TrickSlice) First(n int) TrickSlice {\n\tv := reflect.Value(ts)\n\tif n > v.Len() {\n\t\tn = v.Len()\n\t}\n\treturn TrickSlice(v.Slice3(0, n, n))\n}", "func Nth(data interface{}, i int) (interface{}, error) {\n\tvar err error\n\n\tresult := func(err *error) interface{} {\n\t\tdefer catch(err)\n\n\t\tif !isNonNilData(err, \"data\", data) {\n\t\t\treturn nil\n\t\t}\n\n\t\tdataValue, _, _, dataValueLen := inspectData(data)\n\n\t\tif !isSlice(err, \"data\", dataValue) {\n\t\t\treturn nil\n\t\t}\n\n\t\tif dataValueLen == 0 {\n\t\t\treturn nil\n\t\t}\n\n\t\tif i < 0 {\n\t\t\ti = dataValueLen + i\n\t\t}\n\n\t\tif i < dataValueLen {\n\t\t\treturn dataValue.Index(i).Interface()\n\t\t}\n\n\t\treturn nil\n\t}(&err)\n\n\treturn result, err\n}", "func (m *sparse) Head() []*sparseRow {\n\titerator := m.Rows()\n\tcount := int(math.Min(10, m.Size()))\n\trows := make([]*sparseRow, count)\n\n\tfor i := 0; i <= count; i++ {\n\t\trow := iterator()\n\t\tif row == nil {\n\t\t\tbreak\n\t\t}\n\t\trows[i] = row\n\t}\n\n\treturn rows\n}", "func NthElement(scope *Scope, input tf.Output, n tf.Output, optional ...NthElementAttr) (values tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"NthElement\",\n\t\tInput: []tf.Input{\n\t\t\tinput, n,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (vec Vector) Nth(n Number) LangType {\n\tindex := int(n)\n\treturn vec[index]\n}", "func (l LinkedList) GetNth(n int) (val datatype, found bool) {\n\tval = errorVal\n\tfound = false\n\tif l.head != nil {\n\t\tcur := l.head\n\t\ti := 0\n\t\tfor {\n\t\t\tif i == n {\n\t\t\t\tval = cur.data\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t} else {\n\t\t\t\tif cur.next != nil {\n\t\t\t\t\tcur = cur.next\n\t\t\t\t\ti++\n\t\t\t\t} else {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn val, found\n}", "func main() {\n\tfmt.Println(getKthFromEnd(&ListNode{\n\t\tVal: 1,\n\t\tNext: &ListNode{\n\t\t\tVal: 2,\n\t\t\tNext: &ListNode{\n\t\t\t\tVal: 3,\n\t\t\t\tNext: &ListNode{\n\t\t\t\t\tVal: 4,\n\t\t\t\t\tNext: &ListNode{\n\t\t\t\t\t\tVal: 5,\n\t\t\t\t\t\tNext: nil,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}, 2))\n}", "func (s *List) GetHead() (interface{}, error) {\n\tif s.IsEmpty() {\n\t\treturn 0, ErrEmpty\n\t}\n\n\treturn s.Head.Data, nil\n}", "func middleNodeBrute(head *ListNode) *ListNode {\n\t// brute force\n\tnode := head\n\tslice := []*ListNode{}\n\n\tfor node != nil {\n\t\tslice = append(slice, node)\n\n\t\tnode = node.Next\n\t}\n\n\treturn slice[len(slice)/2]\n}", "func (pq *bidPQueue) Head() *models.Bid {\n\tpq.RLock()\n\tdefer pq.RUnlock()\n\n\tif pq.size() < 1 {\n\t\treturn nil\n\t}\n\n\theadValue := pq.items[1].value\n\n\treturn headValue\n}", "func (snapshots EBSSnapshots) TrimHead(n int) EBSSnapshots {\n\tif n > len(snapshots) {\n\t\treturn EBSSnapshots{}\n\t}\n\treturn snapshots[n:]\n}", "func (n Nodes) Slice() []*Node", "func (q *Queue) Head() uint64 { return q.head }", "func (p *Permutator) NextN(count int) interface{} { \n\tif count <= 0 || p.left() == 0 {\n\t\treturn reflect.MakeSlice(reflect.SliceOf(p.value.Type()), 0, 0).Interface()\n\t}\n \n cap := p.left()\n\tif cap > count {\n\t\tcap = count\n\t}\n\n result := reflect.MakeSlice(reflect.SliceOf(p.value.Type()), cap, cap)\n\n length := 0 \n for index := 0; index < cap; index++ { \n if _, ok := p.Next(); ok {\n length++\n list := p.copySliceValue()\n result.Index(index).Set(list)\n }\n }\n\n list := reflect.MakeSlice(result.Type(), length, length)\n reflect.Copy(list, result)\n \n return list.Interface()\n}", "func (h headTailIndex) head() uint32 {\n\treturn uint32(h >> 32)\n}", "func (t *Tensor) Min(along ...int) (retVal *Tensor, err error) {\n\tmonotonic, incr1 := types.IsMonotonicInts(along) // if both are true, then it means all axes are accounted for, then it'll return a scalar value\n\tif (monotonic && incr1 && len(along) == t.Dims()) || len(along) == 0 {\n\t\tret := sliceMin(t.data)\n\t\tretVal = NewTensor(AsScalar(ret))\n\t\treturn\n\t}\n\tretVal = t\n\tprev := -1\n\tdims := len(retVal.Shape())\n\n\tfor _, axis := range along {\n\t\tif prev == -1 {\n\t\t\tprev = axis\n\t\t}\n\t\tif axis > prev {\n\t\t\taxis--\n\t\t}\n\n\t\tif axis >= dims {\n\t\t\terr = types.DimMismatchErr(axis, retVal.Dims())\n\t\t\treturn\n\t\t}\n\n\t\tretVal = retVal.min(axis)\n\t}\n\treturn\n}", "func (root *TreeNode) topN(n int) (res []interface{}) {\n\tif root == nil {\n\t\treturn res\n\t}\n\n\tif root.right != nil {\n\t\ttmp := root.right.topN(n)\n\t\tfor _, k := range tmp {\n\t\t\tif len(res) == n {\n\t\t\t\treturn res\n\t\t\t}\n\n\t\t\tres = append(res, k)\n\t\t}\n\t}\n\n\tif len(res) < n {\n\t\tres = append(res, root.Values...)\n\t}\n\n\tif len(res) == n {\n\t\treturn res\n\t}\n\n\tif root.left != nil {\n\t\ttmp := root.left.topN(n - len(res))\n\t\tfor _, k := range tmp {\n\t\t\tif len(res) == n {\n\t\t\t\treturn res\n\t\t\t}\n\t\t\tres = append(res, k)\n\t\t}\n\t}\n\treturn res\n}", "func Nth(seq Seq, n uint) interface{} {\n\treturn First(NthRest(seq, n))\n}", "func Element(data sort.Interface, n int) {\n\tl := data.Len()\n\tif n < 0 || n >= l {\n\t\treturn\n\t}\n\tquickSelectAdaptive(data, n, 0, l)\n}", "func (list *TList) Head() *list.Element {\n\treturn list.list.Front()\n}", "func removeNthFromEndSlice(head *ListNode, n int) *ListNode {\n\tif head == nil || head.Next == nil {\n\t\treturn nil\n\t}\n\ts := []*ListNode{}\n\tfor {\n\t\ts = append(s, head)\n\t\thead = head.Next\n\t\tif head == nil {\n\t\t\tbreak\n\t\t}\n\t}\n\tx := len(s) - n\n\tif x > 0 {\n\t\ts[x-1].Next = s[x].Next\n\t\treturn s[0]\n\t}\n\treturn s[1]\n}", "func (list *List) Head() *Node {\n return list.head\n}", "func frontier(n uint) []uint {\n\tif n == 0 {\n\t\treturn []uint{}\n\t}\n\n\tr := root(n)\n\ts := subtreeSize(r, n)\n\tf := []uint{}\n\tfor s != (1 << log2(s)) {\n\t\tl := left(r)\n\t\tr = right(r, n)\n\t\ts = subtreeSize(r, n)\n\t\tf = append(f, l)\n\t}\n\tf = append(f, r)\n\treturn f\n}", "func GetLeftIndex(n int) int {\n\treturn 2*n + 1\n}", "func removeNthFromEnd(head *ListNode, n int) *ListNode {\n\tdummyHead := ListNode{Next: head}\n\n\tvar length int\n\tfor cursor := &dummyHead; cursor.Next != nil; cursor = cursor.Next {\n\t\tlength++\n\t}\n\n\tif n <= 0 || n > length {\n\t\treturn dummyHead.Next\n\t}\n\n\tpreIdx := length - n\n\tpreNode := &dummyHead\n\tfor i := 0; i < preIdx; i++ {\n\t\tpreNode = preNode.Next\n\t}\n\n\tdelNode := preNode.Next\n\tpreNode.Next = delNode.Next\n\tdelNode.Next = nil // avoid memory leaks\n\n\treturn dummyHead.Next\n}", "func GetNthByMinHeapFloat64(data []float64, k int) float64 {\n\th := make(minHeapFloat64, k + 1)\n\tfor i := 0; i < k + 1; i++ {\n\t\th[i] = data[i]\n\t\tj := i\n\n\t\tfor j != 0 && h[heapParent(j)] > h[j] {\n\t\t\th[heapParent(j)], h[j] = h[j], h[heapParent(j)]\n\t\t\tj = heapParent(j)\n\t\t}\n\t}\n\tfor i := k + 1; i < len(data); i++ {\n\t\tif data[i] > h[0] {\n\t\t\th[0] = data[i]\n\t\t\th.minHeapify(0)\n\t\t}\n\t}\n\treturn h[0]\n}", "func (a Slice[T]) Firsts(count int) Slice[T] {\n\treturn a[0:count]\n}", "func (r *ObjectsListingXact) peekN(n uint) (result []*cmn.BucketEntry, err error) {\n\tif len(r.buff) >= int(n) && n != 0 {\n\t\treturn r.buff[:n], nil\n\t}\n\n\tfor len(r.buff) < int(n) || n == 0 {\n\t\tres, ok := <-r.resultCh\n\t\tif !ok {\n\t\t\terr = io.EOF\n\t\t\tbreak\n\t\t}\n\t\tif res.err != nil {\n\t\t\terr = res.err\n\t\t\tbreak\n\t\t}\n\t\tr.buff = append(r.buff, res.entry)\n\t}\n\n\tsize := cos.Min(int(n), len(r.buff))\n\tif size == 0 {\n\t\tsize = len(r.buff)\n\t}\n\treturn r.buff[:size], err\n}", "func (i *queueIndex) getHead() (int, int) {\n\taid := i.indexArena.ReadUint64(0)\n\tpos := i.indexArena.ReadUint64(8)\n\treturn int(aid), int(pos)\n}", "func TakeFirst[T any](inner TryNextor[T], n uint) TryNextor[T] {\n\treturn &take[T]{\n\t\tn: n,\n\t\tinner: inner,\n\t}\n}", "func SumOfFirstNIntegers(n int) int {\n return (1 + n) * n / 2\n}", "func (bids *Bids) Head() int {\n\treturn bids.head\n}", "func head(n *node) *node {\n\t// base case\n\tif n == nil || n.parent == nil {\n\t\treturn n\n\t}\n\treturn head(n.parent)\n}", "func (t *Map) Head() interface{} {\n\tif t.NotEmpty() {\n\t\treturn t.entries[t.keys.Front().Value].val\n\t}\n\treturn nil\n}", "func minSliceElement(a []float64) float64", "func findKthAll(input []int, k int) []int {\n\t// res := make([]int, 0)\n\t/*First method Parition:\n\tJust let k th elements'left is all less than kth elements\n\t*/\n\n\tbegin := 0\n\tend := len(input) - 1\n\tif end < 1 {\n\t\treturn input\n\t}\n\tindex := parition(&input, begin, end)\n\tfor index != k-1 {\n\t\tif index > k-1 {\n\t\t\tend = index - 1\n\t\t\tindex = parition(&input, begin, end)\n\t\t} else if index < k-1 {\n\t\t\tbegin = index + 1\n\t\t\tindex = parition(&input, begin, end)\n\t\t}\n\t}\n\treturn input[:k]\n}", "func (this *MyCircularQueue) Rear() int {\n if this.IsEmpty() { return -1 }\n return this.vals[(this.tail+this.n-1)%this.n] // Note\n}", "func TestExecuteOperationsWithAdditionFirst(t *testing.T) {\n\tassert := assert.New(t)\n\toperands := []int{1, 2, 3, 4, 6}\n\toperators := []rune{'+', '*', '+', '*'}\n\tassert.Equal(126, ExecuteOperationsWithAdditionFirst(operands, operators))\n\tassert.Equal([]int{3, 7, 6}, PerformAdditionOnly(operands, operators))\n}", "func (g *Group) GetFirstElem(operand Operand) *list.Element {\n\tif operand == OperandAny {\n\t\treturn g.Equivalents.Front()\n\t}\n\treturn g.FirstExpr[operand]\n}", "func (e *TopNExec) Next(ctx context.Context, req *chunk.Chunk) error {\n\treq.Reset()\n\tif !e.fetched {\n\t\te.totalLimit = e.limit.Offset + e.limit.Count\n\t\te.Idx = int(e.limit.Offset)\n\t\terr := e.loadChunksUntilTotalLimit(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = e.executeTopN(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\te.fetched = true\n\t}\n\tif e.Idx >= len(e.rowPtrs) {\n\t\treturn nil\n\t}\n\tif !req.IsFull() {\n\t\tnumToAppend := mathutil.Min(len(e.rowPtrs)-e.Idx, req.RequiredRows()-req.NumRows())\n\t\trows := make([]chunk.Row, numToAppend)\n\t\tfor index := 0; index < numToAppend; index++ {\n\t\t\trows[index] = e.rowChunks.GetRow(e.rowPtrs[e.Idx])\n\t\t\te.Idx++\n\t\t}\n\t\treq.AppendRows(rows)\n\t}\n\treturn nil\n}", "func removeNthFromEnd(head *ListNode, n int) *ListNode {\n\ttemp := head\n\tlist := []*ListNode{}\n\tfor temp != nil {\n\t\tlist = append(list, temp)\n\t\ttemp = temp.Next\n\t}\n\tlength := len(list)\n\tthisOne := list[length - n]\n\tbeforeIdx := length - n - 1;\n\tif beforeIdx < 0 {\n\t\treturn thisOne.Next\n\t}\n\tlist[beforeIdx].Next = thisOne.Next\n\treturn head\n}", "func middleOfLinkedList(head *Node) *Node {\n\tsingleP := head\n\tdoubleP := singleP\n\tif singleP == nil {\n\t\treturn singleP\n\t}\n\tfor doubleP != nil && doubleP.Next != nil {\n\t\tsingleP = singleP.Next\n\t\tdoubleP = doubleP.Next.Next\n\t}\n\treturn singleP\n}", "func removeN(head *node, n int) *node {\n\tret := head.next\n\n\tcur := head\n\tfor i := 0; i < n; i++ {\n\t\tcur = cur.next\n\t}\n\t// cur is now the last node we need to extract\n\thead.next = cur.next\n\tcur.next = ret\n\n\treturn ret\n}", "func dataframeHead(_ *starlark.Thread, b *starlark.Builtin, args starlark.Tuple, kwargs []starlark.Tuple) (starlark.Value, error) {\n\tvar nVal starlark.Value\n\n\tif err := starlark.UnpackArgs(\"head\", args, kwargs,\n\t\t\"n?\", &nVal,\n\t); err != nil {\n\t\treturn nil, err\n\t}\n\tself := b.Receiver().(*DataFrame)\n\n\tnumRows, ok := toIntMaybe(nVal)\n\tif !ok {\n\t\t// n defaults to 5 if not given\n\t\tnumRows = 5\n\t}\n\n\tbuilder := newTableBuilder(self.numCols(), 0)\n\tfor rows := newRowIter(self); !rows.Done() && numRows > 0; rows.Next() {\n\t\tr := rows.GetRow()\n\t\tbuilder.pushRow(r.data)\n\t\tnumRows--\n\t}\n\n\treturn &DataFrame{\n\t\tcolumns: self.columns,\n\t\tbody: builder.body(),\n\t}, nil\n}", "func middleNode(head *ListNode) *ListNode {\r\n\tvar (\r\n\t\tp *ListNode\r\n\t\tl, i, m int\r\n\t)\r\n\tfor l, p = 0, head; p != nil; p = p.Next {\r\n\t\tl += 1\r\n\t}\r\n\t\r\n\tfor m, i, p = l / 2, 0, head; i < m; i++ {\r\n\t\tp = p.Next\r\n\t}\r\n\treturn p\r\n}", "func (trcn *TestRetrievalClientNode) GetChainHead(ctx context.Context) (shared.TipSetToken, abi.ChainEpoch, error) {\n\treturn shared.TipSetToken{}, 0, nil\n}", "func (l *List) Head() *Node {\n\t// when the size is 0 return nil because the next node of the root\n\t// node is itself\n\tif l.Size == 0 {\n\t\treturn nil\n\t}\n\treturn l.root.next\n}", "func GetNthByMinHeapInt64(data []int64, k int) int64 {\n\th := make(minHeapInt64, k + 1)\n\tfor i := 0; i < k + 1; i++ {\n\t\th[i] = data[i]\n\t\tj := i\n\n\t\tfor j != 0 && h[heapParent(j)] > h[j] {\n\t\t\th[heapParent(j)], h[j] = h[j], h[heapParent(j)]\n\t\t\tj = heapParent(j)\n\t\t}\n\t}\n\tfor i := k + 1; i < len(data); i++ {\n\t\tif data[i] > h[0] {\n\t\t\th[0] = data[i]\n\t\t\th.minHeapify(0)\n\t\t}\n\t}\n\treturn h[0]\n}", "func partition(head *ListNode, x int) *ListNode {\n\tif head == nil || head.Next == nil {\n\t\treturn head\n\t}\n\tdummy, dummy2 := &ListNode{}, &ListNode{}\n\tdummy.Next = head\n\tp, cur := dummy, dummy2\n\n\t// find the first item not small than x.\n\tfor p = dummy; p.Next != nil; {\n\t\tif p.Next.Val < x {\n\t\t\tp = p.Next\n\t\t} else {\n\t\t\tcur.Next = p.Next\n\t\t\tcur = cur.Next\n\t\t\tp.Next = p.Next.Next\n\t\t}\n\t}\n\tcur.Next = nil\n\tp.Next = dummy2.Next\n\n\treturn dummy.Next\n}", "func (trpn *TestRetrievalProviderNode) GetChainHead(ctx context.Context) (shared.TipSetToken, abi.ChainEpoch, error) {\n\treturn []byte{42}, 0, trpn.ChainHeadError\n}", "func Min[E Ordered](list []E) E {\n\tmin := list[0]\n\tfor _, x := range list {\n\t\tif x < min {\n\t\t\tmin = x\n\t\t}\n\t}\n\treturn min\n}", "func removeNthFromEnd(head *ListNode, n int) *ListNode {\n\tif head == nil || n == 0 {\n\t\treturn nil\n\t}\n\tvar (\n\t\tend = head\n\t\ti = 1\n\t)\n\tfor ; i <= n && end.Next != nil; i++ {\n\t\tend = end.Next\n\t}\n\tpre := head\n\tfor end.Next != nil {\n\t\tend = end.Next\n\t\tpre = pre.Next\n\t}\n\tif pre.Next != nil && i != n {\n\t\tpre.Next = pre.Next.Next\n\t} else {\n\t\treturn head.Next\n\t}\n\treturn head\n}", "func GetHead(l *List) *List {\n\tfor l.prev != nil {\n\t\tl = l.prev\n\t}\n\treturn l\n}", "func (req *Request) Head() *ResultSet {\n\treturn req.do(http.MethodHead)\n}", "func (m *MockFullNode) ChainHead(arg0 context.Context) (*types0.TipSet, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ChainHead\", arg0)\n\tret0, _ := ret[0].(*types0.TipSet)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (e *edge) Head() Node {\n\treturn e.v\n}", "func nmin(a ...int) int {\r\n\tret := a[0]\r\n\tfor _, e := range a {\r\n\t\tret = min(ret, e)\r\n\t}\r\n\treturn ret\r\n}", "func nmin(a ...int) int {\r\n\tret := a[0]\r\n\tfor _, e := range a {\r\n\t\tret = min(ret, e)\r\n\t}\r\n\treturn ret\r\n}", "func nmin(a ...int) int {\r\n\tret := a[0]\r\n\tfor _, e := range a {\r\n\t\tret = min(ret, e)\r\n\t}\r\n\treturn ret\r\n}", "func nmin(a ...int) int {\r\n\tret := a[0]\r\n\tfor _, e := range a {\r\n\t\tret = min(ret, e)\r\n\t}\r\n\treturn ret\r\n}", "func (_m *Database) IndexHead() [1]uint64 {\n\tret := _m.Called()\n\n\tvar r0 [1]uint64\n\tif rf, ok := ret.Get(0).(func() [1]uint64); ok {\n\t\tr0 = rf()\n\t} else {\n\t\tif ret.Get(0) != nil {\n\t\t\tr0 = ret.Get(0).([1]uint64)\n\t\t}\n\t}\n\n\treturn r0\n}", "func (s *List) PopHead() (interface{}, error) {\n\tif s.IsEmpty() {\n\t\treturn 0, ErrEmpty\n\t}\n\n\tv := s.Head.Data\n\tif s.isLastOne() {\n\t\ts.onEmpty()\n\t} else {\n\t\ts.Head = s.Head.Next\n\t\ts.Head.Prev = nil\n\t}\n\treturn v, nil\n}", "func (list *List) Head() *Node {\n\treturn list.head\n}", "func Head(s string) rune {\n\tif s == \"\" {\n\t\tpanic(\"empty list\")\n\t}\n\tr, _ := utf8.DecodeRuneInString(s)\n\treturn r\n}", "func middleNode(head *ListNode) *ListNode {\n\tslow,fast := head,head\n\n\tfor fast.Next !=nil && fast.Next.Next != nil {\n\t\tslow = slow.Next\n\t\tfast = fast.Next.Next\n\t}\n\n\treturn slow\n}", "func (r *radNode) getFirst() *radNode {\n\tn := r.desc\n\tif n == nil {\n\t\treturn nil\n\t}\n\tif n.sis == nil {\n\t\treturn n\n\t}\n\tkey := string(n.prefix)\n\tfor d := n.sis; d != nil; d = d.sis {\n\t\tk := string(d.prefix)\n\t\tif k < key {\n\t\t\tn = d\n\t\t\tkey = k\n\t\t}\n\t}\n\treturn n\n}", "func minSlice(out, a, b []float64)", "func RemoveNthFromEnd(head *ListNode, n int) *ListNode {\n\thead = &ListNode{ // append a starter to the head\n\t\tVal: -1,\n\t\tNext: head,\n\t}\n\t// find the first n+1 th node\n\tstartNode, node := head, head\n\tfor i := 0; i < n; i++ { // we want find the node before n, the first node is the current node\n\t\tif node == nil {\n\t\t\tbreak\n\t\t}\n\t\t// move to the next head\n\t\tnode = node.Next\n\t}\n\tif node == nil {\n\t\t// not exists\n\t\treturn nil\n\t}\n\t// exists\n\tendNode := node\n\n\t// moving both start and end node, until the end node reach the end of linked list\n\tfor endNode.Next != nil {\n\t\tendNode = endNode.Next\n\t\tstartNode = startNode.Next\n\t}\n\n\t// now, the start node is the node a node before n-th node in the linked list\n\tstartNode.Next = startNode.Next.Next\n\n\treturn head.Next\n}", "func (np *NodePool) First(bucket int32) NodeIndex {\n\treturn np.first[bucket]\n}", "func (s *stringStack) Head() string {\n\tlast := s.list.Back()\n\tif last == nil {\n\t\treturn \"\"\n\t}\n\treturn last.Value.(string)\n}", "func selectKth(l []interface{}, mid int, left int, right int, cmp func(x interface{}, y interface{}) (int, error)) (pivotIdx int) {\n\t// Go does not provide a function for getting a random value\n\t// from a range, so this bounds the random value between\n\t// left and right\n\tidx := rand.RangeInt(left, right)\n\tpivotIdx = partition(l, left, right, idx, cmp)\n\tif (left + mid - 1) == pivotIdx {\n\t\treturn\n\t}\n\n\tif (left + mid - 1) < pivotIdx {\n\t\treturn selectKth(l, mid, left, pivotIdx-1, cmp)\n\t} else {\n\t\treturn selectKth(l, mid-(pivotIdx-left+1), pivotIdx+1, right, cmp)\n\t}\n}", "func (s *List) PushHead(x interface{}) {\n\ti := newItem(x)\n\tif !s.addFirstItem(i) {\n\t\ts.Head.Prev = i\n\t\ti.Next = s.Head\n\t\ts.Head = i\n\t}\n}", "func (s *SeriesServiceOp) First(ctx context.Context, opt *SeriesOptions) (Series, *Response, error) {\n\tss, r, err := s.List(ctx, opt)\n\tif err != nil || len(ss) == 0 {\n\t\treturn Series{}, r, err\n\t}\n\treturn ss[0], r, err\n}", "func (iter *PermutationIterator) First() []interface{} {\n\titer.Reset()\n\treturn iter.Next()\n}", "func getIntersectionNode(headA, headB *ListNode) *ListNode {\n\tif headA == nil || headB == nil {\n\t\treturn nil\n\t}\n\tpa, pb := headA, headB\n\tround := 2\n\tfor pa != pb {\n\t\tpa = pa.Next\n\t\tpb = pb.Next\n\t\tif pa == nil || pb == nil {\n\t\t\tif round == 0 {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tround--\n\t\t}\n\t\tif pa == nil {\n\t\t\tpa = headB\n\t\t}\n\t\tif pb == nil {\n\t\t\tpb = headA\n\t\t}\n\t}\n\treturn pa\n}", "func safeHead(slice []string) string {\n\tif len(slice) > 0 {\n\t\treturn Head(slice)\n\t}\n\n\treturn \"\"\n}" ]
[ "0.66164404", "0.5852609", "0.580193", "0.55647224", "0.54286236", "0.52772564", "0.5217181", "0.51278394", "0.5118447", "0.51058024", "0.51055855", "0.5101355", "0.50714004", "0.505273", "0.4997914", "0.49927148", "0.49915862", "0.49413338", "0.492538", "0.49159002", "0.49055237", "0.49021345", "0.48456916", "0.4834777", "0.4794596", "0.47903907", "0.47738194", "0.47609013", "0.47594273", "0.47530538", "0.4733689", "0.4725338", "0.46997672", "0.46578673", "0.4646249", "0.46399644", "0.46097943", "0.46049306", "0.4585947", "0.45686695", "0.45669532", "0.45597774", "0.45557758", "0.45514232", "0.45198724", "0.4518126", "0.45154542", "0.44950843", "0.44880074", "0.44783366", "0.4442034", "0.44299766", "0.44170195", "0.4413313", "0.44132698", "0.4405384", "0.44021732", "0.4395963", "0.43947968", "0.43704078", "0.43664575", "0.43648815", "0.4361904", "0.4354722", "0.43517664", "0.43442312", "0.43340132", "0.43310678", "0.43233508", "0.431517", "0.4314494", "0.43128115", "0.43120915", "0.42970267", "0.4294594", "0.42945117", "0.42882568", "0.42756718", "0.42749494", "0.4273288", "0.42572704", "0.42572704", "0.42572704", "0.42572704", "0.42569292", "0.42561978", "0.42543745", "0.4253023", "0.4248938", "0.42263567", "0.42232546", "0.42227498", "0.42186856", "0.4210487", "0.42101848", "0.4209973", "0.42087823", "0.42067295", "0.42022783", "0.4193701" ]
0.82317895
0
Intersect returns the intersection of a and operand. Duplicate elements are removed. Ordering is not preserved.
func Intersect(a, operand []string) []string { intersection := make([]string, Max(len(a), len(operand))) uniqA := make(map[string]bool, len(a)) for _, elem := range a { uniqA[elem] = true } i := 0 uniqOp := make(map[string]bool, len(operand)) for _, elem := range operand { if uniqA[elem] && !uniqOp[elem] { intersection[i] = elem i++ } uniqOp[elem] = true } return intersection[:i] }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func intersect(a interface{}, b interface{}) []interface{} {\n\tset := make([]interface{}, 0)\n\tav := reflect.ValueOf(a)\n\n\tfor i := 0; i < av.Len(); i++ {\n\t\tel := av.Index(i).Interface()\n\t\tif contains(b, el) {\n\t\t\tset = append(set, el)\n\t\t}\n\t}\n\n\treturn set\n}", "func Intersection[T comparable](a, b Set[T]) Set[T] {\n\tresult := a.New()\n\tfor _, e := range b.Elements() {\n\t\tif a.Contains(e) {\n\t\t\tresult.Add(e)\n\t\t}\n\t}\n\tfor _, e := range a.Elements() {\n\t\tif b.Contains(e) {\n\t\t\tresult.Add(e)\n\t\t}\n\t}\n\treturn result\n}", "func Intersect(a, b Set) Set {\n\treturn a.Where(func(v Value) bool { return b.Has(v) })\n}", "func Intersection(a, b AnySlice) AnySlice {\n\tmustBeSlice(a)\n\tmustBeSlice(b)\n\n\taVal := reflect.ValueOf(a)\n\tbVal := reflect.ValueOf(b)\n\taCount := aVal.Len()\n\tbCount := bVal.Len()\n\toutput := makeFilterSlice(a, 0, aCount+bCount)\n\tkeys := make(map[interface{}]bool)\n\n\tfor i := 0; i < aCount; i++ {\n\t\tkeys[aVal.Index(i).Interface()] = true\n\t}\n\tfor i := 0; i < bCount; i++ {\n\t\tkey := bVal.Index(i)\n\t\tif _, present := keys[key.Interface()]; present {\n\t\t\toutput = reflect.Append(output, key)\n\t\t}\n\t}\n\treturn output.Interface()\n}", "func (bst *BST) Intersection(other *BST) []interface{} {\n\treturn set_operation(bst.Root, other.Root, setIntersect)\n}", "func (a seriesIDs) intersect(other seriesIDs) seriesIDs {\n\tl := a\n\tr := other\n\n\t// we want to iterate through the shortest one and stop\n\tif len(other) < len(a) {\n\t\tl = other\n\t\tr = a\n\t}\n\n\t// they're in sorted order so advance the counter as needed.\n\t// That is, don't run comparisons against lower values that we've already passed\n\tvar i, j int\n\n\tids := make([]uint64, 0, len(l))\n\tfor i < len(l) && j < len(r) {\n\t\tif l[i] == r[j] {\n\t\t\tids = append(ids, l[i])\n\t\t\ti++\n\t\t\tj++\n\t\t} else if l[i] < r[j] {\n\t\t\ti++\n\t\t} else {\n\t\t\tj++\n\t\t}\n\t}\n\n\treturn seriesIDs(ids)\n}", "func intersectionSelector(a, b Selector) Selector {\n\treturn func(n *Node) bool {\n\t\treturn a(n) && b(n)\n\t}\n}", "func (b *Bag) intersection(c Bag) Bag {\n\tbag := make(Bag)\n\tfor k, vb := range *b {\n\t\tvc, ok := c[k]\n\t\tif ok {\n\t\t\tif vb < vc {\n\t\t\t\tbag[k] = vb\n\t\t\t} else {\n\t\t\t\tbag[k] = vc\n\t\t\t}\n\t\t}\n\t}\n\treturn bag\n}", "func TestIntersect(t *testing.T) {\n\ttype args struct {\n\t\ta interface{}\n\t\tb interface{}\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs args\n\t\twant interface{}\n\t}{\n\t\t{\n\t\t\tname: \"int64-slice-交集\",\n\t\t\targs: args{\n\t\t\t\ta: []int64{1, 2, 3, 4, 4, 5, 6},\n\t\t\t\tb: []int64{1, 3, 5, 7, 9, 11},\n\t\t\t},\n\t\t\twant: []int64{1, 3, 5},\n\t\t},\n\t}\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tvar ret []int64\n\t\t\tif got := Intersect(tt.args.a, tt.args.b, &ret); !reflect.DeepEqual(ret, tt.want) {\n\t\t\t\tt.Errorf(\"Intersection() = %v, want %v\", got, tt.want)\n\t\t\t}\n\t\t})\n\t}\n}", "func intersect(a, b []model.Fingerprint) []model.Fingerprint {\n\tif a == nil {\n\t\treturn b\n\t}\n\tresult := []model.Fingerprint{}\n\tfor i, j := 0, 0; i < len(a) && j < len(b); {\n\t\tif a[i] == b[j] {\n\t\t\tresult = append(result, a[i])\n\t\t}\n\t\tif a[i] < b[j] {\n\t\t\ti++\n\t\t} else {\n\t\t\tj++\n\t\t}\n\t}\n\treturn result\n}", "func (s *IntSet) Intersection(s2 *IntSet) *IntSet {\n\tout := new(IntSet)\n\tout.Clear()\n\n\tfor key := range s.elem {\n\t\tif s2.Contains(key) {\n\t\t\tout.Add(key)\n\t\t}\n\t}\n\n\treturn out\n}", "func PartsIntersect(a, b []Part) []Part {\n\tm := make(map[string]bool, len(a))\n\tfor _, p := range a {\n\t\tk := p.key()\n\t\tm[k] = true\n\t}\n\tvar d []Part\n\tfor _, p := range b {\n\t\tk := p.key()\n\t\tif m[k] {\n\t\t\td = append(d, p)\n\t\t}\n\t}\n\treturn d\n}", "func intersection(nums1 []int, nums2 []int) []int {\n\tvar ret []int\n\tmapping := map[int]bool{}\n\tfor _, e := range nums1 {\n\t\tmapping[e] = true\n\t}\n\tfor _, e := range nums2 {\n\t\tif _, ok := mapping[e]; ok {\n\t\t\tret = append(ret, e)\n\t\t\tdelete(mapping, e)\n\t\t}\n\t}\n\treturn ret\n}", "func Intersect(a []string, b []string) []string {\n\tvar res []string\n\tfor _, v := range b {\n\t\tif Has(a, v) {\n\t\t\tres = append(res, v)\n\t\t}\n\t}\n\treturn res\n}", "func deriveIntersect(this, that []int) []int {\n\tintersect := make([]int, 0, deriveMin(len(this), len(that)))\n\tfor i, v := range this {\n\t\tif deriveContains(that, v) {\n\t\t\tintersect = append(intersect, this[i])\n\t\t}\n\t}\n\treturn intersect\n}", "func (self *Rectangle) Intersection(b *Rectangle, out *Rectangle) *Rectangle{\n return &Rectangle{self.Object.Call(\"intersection\", b, out)}\n}", "func Intersection[T comparable](a, b []T) []T {\n\tinter := []T{}\n\thash := make(map[T]bool, len(a))\n\tfor _, i := range a {\n\t\thash[i] = false\n\t}\n\tfor _, i := range b {\n\t\tif done, exists := hash[i]; exists && !done {\n\t\t\tinter = append(inter, i)\n\t\t\thash[i] = true\n\t\t}\n\t}\n\treturn inter\n}", "func Intersect(a, b IntSet, reuse []int) IntSet {\n\tswitch {\n\tcase a.Empty() || b.Empty():\n\t\treturn IntSet{}\n\tcase a.Inverse:\n\t\tif b.Inverse {\n\t\t\treturn IntSet{Inverse: true, Set: combine(a.Set, b.Set, reuse)}\n\t\t}\n\t\treturn IntSet{Set: subtract(b.Set, a.Set, reuse)}\n\tcase b.Inverse:\n\t\treturn IntSet{Set: subtract(a.Set, b.Set, reuse)}\n\t}\n\treturn IntSet{Set: intersect(a.Set, b.Set, reuse)}\n}", "func sortedIntersect(a []string, b []string) []string {\n\tset := make([]string, 0)\n\n\tfor _, el := range a {\n\t\tidx := sort.SearchStrings(b, el)\n\t\tif idx < len(b) && b[idx] == el {\n\t\t\tset = append(set, el)\n\t\t}\n\t}\n\n\treturn set\n}", "func Intersect(queries ...Query) VariadicQuery {\n\treturn VariadicQuery{\n\t\tOperator: QueryIntersect,\n\t\tQueries: queries,\n\t}\n}", "func (n *numbers) Intersect() []int {\n\tm := map[int]int{}\n\tfor _, el := range n.nums1 {\n\t\tm[el]++\n\t}\n\tout := make([]int, 0)\n\tfor _, el := range n.nums2 {\n\t\tif m[el] > 0 {\n\t\t\tm[el]--\n\t\t\tout = append(out, el)\n\t\t}\n\t}\n\treturn out\n}", "func (root *TreeNode) intersection(b *TreeNode, res *TreeNode, n *int) *TreeNode {\n\tif root == nil || b == nil {\n\t\treturn res\n\t}\n\n\tif root.left != nil {\n\t\tres = root.left.intersection(b, res, n)\n\t}\n\tif root.right != nil {\n\t\tres = root.right.intersection(b, res, n)\n\t}\n\tif b.exists(root.key) {\n\t\tres, _ = res.insert(root.key, root.key, root.issmaller)\n\t\t*n++\n\t}\n\n\treturn res\n}", "func Intersect(a []string, b []string, s ...[]string) []string {\n\tif len(a) == 0 || len(b) == 0 {\n\t\treturn []string{}\n\t}\n\tset := make([]string, 0)\n\thash := make(map[string]bool)\n\tfor _, el := range a {\n\t\thash[el] = true\n\t}\n\tfor _, el := range b {\n\t\tif _, found := hash[el]; found {\n\t\t\tset = append(set, el)\n\t\t}\n\t}\n\tif len(s) == 0 {\n\t\treturn set\n\t}\n\treturn Intersect(set, s[0], s[1:]...)\n}", "func (s *HashSet)Intersection(other *HashSet) *HashSet{\n\ts.mu.RLock()\n\tdefer s.mu.RUnlock()\n\tvar a,b *HashSet\n\tresult := NewSet()\n\tif s.Len()<other.Len(){\n\t\ta = s\n\t\tb = other\n\t}else {\n\t\ta = other\n\t\tb = s\n\t}\n\tfor item,_:=range a.items{\n\t\tif b.Has(item){\n\t\t\tresult.Insert(item)\n\t\t}\n\t}\n\treturn result\n}", "func (s *Set) Intersection(other *Set) *Set {\n\titems := make([]Value, 0)\n\ts.Range(func(item Value) bool {\n\t\tif other.Contains(item) {\n\t\t\titems = append(items, item)\n\t\t}\n\n\t\treturn true\n\t})\n\n\treturn NewSet(items...)\n}", "func (s Set) Intersection(otherSet Set) Set {\n\n\tinterSet := NewSet()\n\n\tfor k := range s.mem {\n\t\tif otherSet.mem[k] == true {\n\t\t\tinterSet.mem[k] = true\n\t\t\t// Adds an element to the intersection set if it is a member of both sets.\n\t\t}\n\t}\n\n\treturn interSet\n}", "func (p *intersect) Intersect(nums1 []int, nums2 []int) []int {\n\tsort.Ints(nums1)\n\tsort.Ints(nums2)\n\ti := 0\n\tj := 0\n\tinter := make([]int, 0)\n\tfor i < len(nums1) && j < len(nums2) {\n\t\tif nums1[i] == nums2[j] {\n\t\t\tinter = append(inter, nums1[i])\n\t\t\ti++\n\t\t\tj++\n\t\t} else if nums1[i] > nums2[j] {\n\t\t\tj++\n\t\t} else {\n\t\t\ti++\n\t\t}\n\t}\n\treturn inter\n}", "func (s *IntSet) Intersection(y *IntSet) *IntSet {\n\ti := NewIntSet([]int{})\n\n\tfor _, m := range s.Members() {\n\t\tif y.Contains(m) {\n\t\t\ti.Add(m)\n\t\t}\n\t}\n\n\treturn i\n}", "func Intersection(s1, s2 Set) Set {\n\ttoReturn := New()\n\tfor k := range s1 {\n\t\tif _, ok := s2[k]; ok {\n\t\t\ttoReturn.Add(k)\n\t\t}\n\t}\n\treturn toReturn\n}", "func (s *StrSet) Intersection(o *StrSet) StrSet {\n\tn := NewStr()\n\tfor el := range s.els {\n\t\tif _, ok := o.els[el]; ok {\n\t\t\tn.Add(el)\n\t\t}\n\t}\n\treturn n\n}", "func (s *Set) Intersect(s1 *Set) *Set {\n\ts2 := NewSet()\n\tfor _, val := range s.elements {\n\t\tfor _, val1 := range s1.elements {\n\t\t\tif val == val1 {\n\t\t\t\ts2.Add(val)\n\t\t\t}\n\t\t}\n\t}\n\treturn s2\n}", "func (r Range) Intersection(b Range) (intersection Range) {\n\tif (r.Pos >= b.Pos && r.Pos < b.End()) || (b.Pos >= r.Pos && b.Pos < r.End()) {\n\t\tintersection.Pos = max(r.Pos, b.Pos)\n\t\tintersection.Size = min(r.End(), b.End()) - intersection.Pos\n\t}\n\treturn\n}", "func intersectB(nums1 []int, nums2 []int) []int {\n\tif len(nums2) > len(nums1) {\n\t\tnums1, nums2 = nums2, nums1\n\t}\n\n\ttaken := make([]bool, len(nums2))\n\tresp := []int{}\n\tfor i := 0; i < len(nums1); i++ {\n\t\tfor j := 0; j < len(nums2); j++ {\n\t\t\tif nums1[i] == nums2[j] && !taken[j] {\n\t\t\t\tresp = append(resp, nums1[i])\n\t\t\t\ttaken[j] = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\treturn resp\n}", "func TestIntersection(t *testing.T) {\n\ttests := []struct {\n\t\tnote string\n\t\trules []string\n\t\texpected interface{}\n\t}{\n\t\t{\"intersection_0_sets\", []string{`p = x { intersection(set(), x) }`}, \"[]\"},\n\t\t{\"intersection_2_sets\", []string{`p = x { intersection({set(), {1, 2}}, x) }`}, \"[]\"},\n\t\t{\"intersection_2_sets\", []string{`p = x { s1 = {1, 2, 3}; s2 = {2}; intersection({s1, s2}, x) }`}, \"[2]\"},\n\t\t{\"intersection_3_sets\", []string{`p = x { s1 = {1, 2, 3}; s2 = {2, 3, 4}; s3 = {4, 5, 6}; intersection({s1, s2, s3}, x) }`}, \"[]\"},\n\t\t{\"intersection_4_sets\", []string{`p = x { s1 = {\"a\", \"b\", \"c\", \"d\"}; s2 = {\"b\", \"c\", \"d\"}; s3 = {\"c\", \"d\"}; s4 = {\"d\"}; intersection({s1, s2, s3, s4}, x) }`}, \"[\\\"d\\\"]\"},\n\t}\n\n\tdata := loadSmallTestData()\n\n\tfor _, tc := range tests {\n\t\trunTopDownTestCase(t, data, tc.note, tc.rules, tc.expected)\n\t}\n}", "func (s *set) Intersect(other Set) Set {\n\to := other.(*set)\n\tn, m := s.Len(), o.Len()\n\tss := s\n\tso := o\n\tif m < n {\n\t\tss = o\n\t\tso = s\n\t\tn = m\n\t}\n\n\tr := newset(n)\n\tss.Foreach(func(x *Term) {\n\t\tif so.Contains(x) {\n\t\t\tr.Add(x)\n\t\t}\n\t})\n\treturn r\n}", "func (c MethodsCollection) Intersect() pIntersect {\n\treturn pIntersect{\n\t\tMethod: c.MustGet(\"Intersect\"),\n\t}\n}", "func ExampleIntSet_Intersect() {\n\ts1 := gset.NewIntSet()\n\ts1.Add([]int{1, 2, 3}...)\n\tvar s2 gset.IntSet\n\ts2.Add([]int{1, 2, 3, 4}...)\n\tfmt.Println(s2.Intersect(s1).Slice())\n\n\t// May Output:\n\t// [1 2 3]\n}", "func (b *BitSet) Intersection(c *BitSet) *BitSet {\n\tif c == nil {\n\t\treturn nil\n\t}\n\n\tres := New(0)\n\n\tfor key, bValue := range b.set {\n\t\tif cValue, found := c.set[key]; found {\n\t\t\tresValue := bValue & cValue\n\t\t\tif resValue != 0 {\n\t\t\t\tres.set[key] = resValue\n\t\t\t}\n\t\t}\n\t}\n\n\treturn res\n\n}", "func (s1 Segment) Intersect(e2 Edge) (Vector, bool) {\n\treturn Intersect(s1, e2)\n}", "func IntervalIntersection(a, b TimeInterval) []TimeInterval {\n\n\t// b and a don't overlap\n\tif !IntervalOverlap(a, b) {\n\t\treturn []TimeInterval{}\n\t}\n\n\tresult := TimeInterval{}\n\n\tif b.Start.Before(a.Start) {\n\t\tresult.Start = a.Start\n\t} else if a.Start.Before(b.Start) {\n\t\tresult.Start = b.Start\n\t} else {\n\t\t//a.Start.Equal(b.Start)\n\t\tresult.Start = b.Start\n\t}\n\n\tif b.End.After(a.End) {\n\t\tresult.End = a.End\n\t} else if a.End.After(b.End) {\n\t\tresult.End = b.End\n\t} else {\n\t\t//a.End.Equal(b.End)\n\t\tresult.End = b.End\n\t}\n\n\treturn []TimeInterval{result}\n}", "func intersection(a, b map[string]bool) map[string]bool {\n\tr := make(map[string]bool)\n\tfor k := range a {\n\t\tif b[k] {\n\t\t\tr[k] = true\n\t\t}\n\t}\n\treturn r\n}", "func Union(a, operand []string) []string {\n\tuniq := make(map[string]bool, len(a)+len(operand))\n\tfor _, elem := range a {\n\t\tuniq[elem] = true\n\t}\n\tfor _, elem := range operand {\n\t\tuniq[elem] = true\n\t}\n\tunion := make([]string, len(uniq))\n\ti := 0\n\tfor k := range uniq {\n\t\tunion[i] = k\n\t\ti++\n\t}\n\treturn union[:i]\n}", "func (ons *orderedNodeSet) intersection(ons2 *orderedNodeSet) *orderedNodeSet {\n\tintersection := newOrderedNodeSet()\n\tif ons2 == nil {\n\t\treturn intersection\n\t}\n\n\tfor _, k := range ons.elements() {\n\t\tif ons2.contains(k) {\n\t\t\tintersection.add(k)\n\t\t}\n\t}\n\n\treturn intersection\n}", "func intersection(x string, y string) string {\n\tvar intersection string\n\tfor i, char := range x {\n\t\tif x[i] == y[i] {\n\t\t\tintersection += string(char)\n\t\t}\n\t}\n\treturn intersection\n}", "func (v Set) Intersect(committeeSet uint64) Set {\n\tif committeeSet == All || committeeSet == v.Whole() {\n\t\treturn v[:]\n\t}\n\n\tc := New()\n\n\tfor i, elem := range v {\n\t\t// looping on all bits to see which one is set to 1\n\t\tif ((committeeSet >> uint(i)) & 1) != 0 {\n\t\t\tc = append(c, elem)\n\t\t}\n\t}\n\n\treturn c\n}", "func NewIntersecter() Intersecter {\n\treturn &intersect{}\n}", "func intersect(nums1 []int, nums2 []int) []int {\n\tm := make(map[int]int)\n\tret := make([]int, 0)\n\tfor _, num := range nums1 {\n\t\tm[num]++\n\t}\n\tfor _, num := range nums2 {\n\t\tif m[num] > 0 {\n\t\t\tret = append(ret, num)\n\t\t\tm[num]--\n\t\t}\n\t}\n\treturn ret\n}", "func TestIntersection(t *T) {\n\t// Degenerate case\n\tempty := NewSet()\n\tassert.Equal(t, 0, Size(empty.Intersection(empty)))\n\n\tints1 := []interface{}{0, 1, 2}\n\tints2 := []interface{}{1, 2, 3}\n\tints3 := []interface{}{4, 5, 6}\n\tintsi := []interface{}{1, 2}\n\ts1 := NewSet(ints1...)\n\ts2 := NewSet(ints2...)\n\ts3 := NewSet(ints3...)\n\n\tassert.Equal(t, 0, Size(s1.Intersection(empty)))\n\tassert.Equal(t, 0, Size(empty.Intersection(s1)))\n\n\tsi := s1.Intersection(s2)\n\tassert.Equal(t, 0, Size(s1.Intersection(s3)))\n\tassertSeqContentsSet(t, ints1, s1)\n\tassertSeqContentsSet(t, ints2, s2)\n\tassertSeqContentsSet(t, ints3, s3)\n\tassertSeqContentsSet(t, intsi, si)\n}", "func (actionSet ActionSet) Intersection(sset ActionSet) ActionSet {\n\tnset := NewActionSet()\n\tfor k := range actionSet {\n\t\tif _, ok := sset[k]; ok {\n\t\t\tnset.Add(k)\n\t\t}\n\t}\n\n\treturn nset\n}", "func (s *IntSet) Intersect(t *IntSet) {\n\tfor i, tword := range t.words {\n\t\ts.words[i] &= tword\n\n\t}\n}", "func (s *Int64) Intersect(other Int64) Int64 {\n\tres := NewInt64WithSize(min(s.Size(), other.Size()))\n\n\t// loop over the smaller set\n\tif len(s.m) <= len(other.m) {\n\t\tfor val := range s.m {\n\t\t\tif other.Contains(val) {\n\t\t\t\tres.Add(val)\n\t\t\t}\n\t\t}\n\t} else {\n\t\tfor val := range other.m {\n\t\t\tif s.Contains(val) {\n\t\t\t\tres.Add(val)\n\t\t\t}\n\t\t}\n\t}\n\treturn res\n}", "func Intersect[T comparable](list1 []T, list2 []T) []T {\n\tresult := []T{}\n\tseen := map[T]struct{}{}\n\n\tfor _, elem := range list1 {\n\t\tseen[elem] = struct{}{}\n\t}\n\n\tfor _, elem := range list2 {\n\t\tif _, ok := seen[elem]; ok {\n\t\t\tresult = append(result, elem)\n\t\t}\n\t}\n\n\treturn result\n}", "func Intersect[T comparable](list1 []T, list2 []T) []T {\n\tresult := []T{}\n\tseen := map[T]struct{}{}\n\n\tfor _, elem := range list1 {\n\t\tseen[elem] = struct{}{}\n\t}\n\n\tfor _, elem := range list2 {\n\t\tif _, ok := seen[elem]; ok {\n\t\t\tresult = append(result, elem)\n\t\t}\n\t}\n\n\treturn result\n}", "func SliceIntersects(a, b interface{}) bool {\n\taValue, bValue := reflect.ValueOf(a), reflect.ValueOf(b)\n\taValueKind, bValueKind := aValue.Kind(), bValue.Kind()\n\n\tif aValueKind != reflect.Slice || bValueKind != reflect.Slice {\n\t\tpanic(fmt.Sprintf(\"one of parameters is not a slice: (%v, %v)\", aValueKind, bValueKind))\n\t}\n\tfor i := 0; i < bValue.Len(); i++ {\n\t\tfor j := 0; j < aValue.Len(); j++ {\n\t\t\tif bValue.Index(i).Interface() == aValue.Index(j).Interface() {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false\n}", "func (resourceSet ResourceSet) Intersection(sset ResourceSet) ResourceSet {\n\tnset := NewResourceSet()\n\tfor k := range resourceSet {\n\t\tif _, ok := sset[k]; ok {\n\t\t\tnset.Add(k)\n\t\t}\n\t}\n\n\treturn nset\n}", "func (i *Result) Intersection() []geom.Coord {\n\treturn i.intersection\n}", "func (self *Rectangle) IntersectionI(args ...interface{}) *Rectangle{\n return &Rectangle{self.Object.Call(\"intersection\", args)}\n}", "func Fintersection(lista, listb []string) []string {\n\tlistx := []string{}\n\tFcompress(&lista)\n\tFcompress(&listb)\n\tfor _, i := range lista {\n\t\tif Fmember(listb, i) {\n\t\t\tlistx = append(listx, i)\n\t\t}\n\t}\n\treturn listx\n}", "func (rs Ranges) Intersection(r Range) (newRs Ranges) {\n\tif len(rs) == 0 {\n\t\treturn rs\n\t}\n\tfor !r.IsEmpty() {\n\t\tvar curr Range\n\t\tvar found bool\n\t\tcurr, r, found = rs.Find(r)\n\t\tif found {\n\t\t\tnewRs.Insert(curr)\n\t\t}\n\t}\n\treturn newRs\n}", "func Intersection(shapes ...Shape) Shape {\n\treturn intersectionGroup{\n\t\tshapes: shapes,\n\t}\n}", "func IntSliceIntersects(a, b []int) (rb bool) {\n\trb = false\n\tfor _, k := range a {\n\t\tfor _, l := range b {\n\t\t\tif k == l {\n\t\t\t\trb = true\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func (s *Set) Intersection(other *Set) *Set {\n\tnewSet := NewSizedSet(len(s.m) + len(other.m))\n\tfor elem := range s.m {\n\t\tif _, found := other.m[elem]; found {\n\t\t\tnewSet.m[elem] = SetNothing{}\n\t\t}\n\t}\n\tfor elem := range other.m {\n\t\tif _, found := s.m[elem]; found {\n\t\t\tnewSet.m[elem] = SetNothing{}\n\t\t}\n\t}\n\treturn newSet\n}", "func intersect(b, c *Block, postnum []int, idom []*Block) *Block {\n\t// TODO: This loop is O(n^2). It used to be used in nilcheck,\n\t// see BenchmarkNilCheckDeep*.\n\tfor b != c {\n\t\tif postnum[b.ID] < postnum[c.ID] {\n\t\t\tb = idom[b.ID]\n\t\t} else {\n\t\t\tc = idom[c.ID]\n\t\t}\n\t}\n\treturn b\n}", "func (obj *object) Intersect(other Object) [][3]*Term {\n\tr := [][3]*Term{}\n\tobj.Foreach(func(k, v *Term) {\n\t\tif v2 := other.Get(k); v2 != nil {\n\t\t\tr = append(r, [3]*Term{k, v, v2})\n\t\t}\n\t})\n\treturn r\n}", "func (s *HashSet) Intersection(other *HashSet) *HashSet {\n\tresult := NewHashSet()\n\tif s == other {\n\t\tresult.AddSet(s)\n\t} else {\n\t\tif other != nil && other.Size() > 0 {\n\t\t\tfor k := range other.GetMap() {\n\t\t\t\tif s.Has(&k) {\n\t\t\t\t\tresult.Add(&k)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn result\n}", "func (g *Group) LocalIntersect(r *algebra.Ray) ([]*Intersection, bool) {\n\txs := make([]*Intersection, 0, 0)\n\tvar hit bool = false\n\n\t// Get the AABB of the group\n\tmin, max := g.GetBounds()\n\n\tif min == nil {\n\t\treturn xs, false\n\t}\n\tif GetBoundsTransform(min, max, g.GetTransform()).Intersect(r.Transform(g.GetTransform())) == false {\n\t\treturn xs, false\n\t}\n\n\tfor _, s := range g.shapes {\n\t\tm := s.GetTransform()\n\t\tri := r.Transform(m.Inverse())\n\t\tshapeXs, shapeHit := s.LocalIntersect(ri)\n\t\thit = hit || shapeHit\n\t\tfor _, t := range shapeXs {\n\t\t\txs = append(xs, t)\n\t\t}\n\t}\n\treturn xs, hit\n}", "func intersection(nums1 []int, nums2 []int) []int {\n\tmaps := make(map[int]int)\n\tfor i := 0; i < len(nums1); i++ {\n\t\tmaps[nums1[i]] = 1\n\t}\n\tres := []int{}\n\tfor j := 0; j < len(nums2); j++ {\n\t\tif v, ok := maps[nums2[j]]; ok && v == 1 {\n\t\t\tmaps[nums2[j]] += 1\n\t\t\tres = append(res, nums2[j])\n\t\t}\n\t}\n\treturn res\n}", "func (r Ray) Intersect(s Ray) (v Vec, ok bool) {\n\tif r.Colinear(s) {\n\t\treturn\n\t}\n\n\tv = r.LineIntersect(s)\n\n\treturn v, r.InAABB(v) && s.InAABB(v)\n}", "func (s stringSet) intersect(o stringSet) stringSet {\n\tns := newStringSet()\n\tfor k := range s {\n\t\tif _, ok := o[k]; ok {\n\t\t\tns[k] = struct{}{}\n\t\t}\n\t}\n\tfor k := range o {\n\t\tif _, ok := s[k]; ok {\n\t\t\tns[k] = struct{}{}\n\t\t}\n\t}\n\treturn ns\n}", "func (r *Rights) Intersect(b *Rights) *Rights {\n\tif r == nil {\n\t\treturn &Rights{}\n\t}\n\tres := make([]Right, 0)\n\trs, bs := makeRightsSet(r), makeRightsSet(b)\n\tfor right := range rs {\n\t\tif _, ok := bs[right]; ok {\n\t\t\tres = append(res, right)\n\t\t}\n\t}\n\treturn &Rights{Rights: res}\n}", "func (set *AppleSet) Intersect(other *AppleSet) *AppleSet {\n\tif set == nil || other == nil {\n\t\treturn nil\n\t}\n\n\tintersection := NewAppleSet()\n\n\tset.s.RLock()\n\tother.s.RLock()\n\tdefer set.s.RUnlock()\n\tdefer other.s.RUnlock()\n\n\t// loop over smaller set\n\tif set.Size() < other.Size() {\n\t\tfor v := range set.m {\n\t\t\tif other.Contains(v) {\n\t\t\t\tintersection.doAdd(v)\n\t\t\t}\n\t\t}\n\t} else {\n\t\tfor v := range other.m {\n\t\t\tif set.Contains(v) {\n\t\t\t\tintersection.doAdd(v)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn intersection\n}", "func (oss *OrderedStringSet) Intersection(setB *OrderedStringSet) *OrderedStringSet {\n\tintersection := NewOrderedStringSet()\n\tif setB == nil {\n\t\treturn intersection\n\t}\n\n\tfor _, m := range oss.Elements() {\n\t\tif setB.Contains(m) {\n\t\t\tintersection.Add(m)\n\t\t}\n\t}\n\n\treturn intersection\n}", "func (mr *MockMutableListMockRecorder) Intersect(other interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Intersect\", reflect.TypeOf((*MockMutableList)(nil).Intersect), other)\n}", "func (t Torus) Intersect(a, b Rectangle) Rectangle {\n\ta, b = t.AlignRects(a, b)\n\treturn a.Intersect(b)\n}", "func Intersection(nums1 []int, nums2 []int) []int {\n\tres := make([]int, 0)\n\tseen := make(map[int]bool)\n\t// Sort the smallest one\n\tif len(nums1) < len(nums2) {\n\t\tsort.Ints(nums1)\n\t\t// Search every element of bigger array in smaller array\n\t\t// and print the element if found\n\t\tfor _, v := range nums2 {\n\t\t\tif Search(nums1, v) != -1 && !seen[v] {\n\t\t\t\tres = append(res, v)\n\t\t\t\tseen[v] = true\n\t\t\t}\n\t\t}\n\t} else {\n\t\tsort.Ints(nums2)\n\t\t// Search every element of bigger array in smaller array\n\t\t// and print the element if found\n\t\tfor _, v := range nums1 {\n\t\t\tif Search(nums2, v) != -1 && !seen[v] {\n\t\t\t\tres = append(res, v)\n\t\t\t\tseen[v] = true\n\t\t\t}\n\t\t}\n\t}\n\treturn res\n}", "func (t *Tree) Intersection(x *Tree) (res *Tree) {\n\tif t == nil || x == nil {\n\t\treturn nil\n\t}\n\tres = New()\n\tt.lock.RLock()\n\tx.lock.RLock()\n\tdefer t.lock.RUnlock()\n\tdefer x.lock.RUnlock()\n\n\tn := 0\n\tnewRoot := t.root.intersection(x.root, res.root, &n)\n\n\treturn &Tree{\n\t\troot: newRoot,\n\t\tCount: n,\n\t}\n}", "func (s StringSet) Intersect(other StringSet) StringSet {\n\tresultSet := make(StringSet, len(s))\n\tfor val := range s {\n\t\tif other[val] {\n\t\t\tresultSet[val] = true\n\t\t}\n\t}\n\n\treturn resultSet\n}", "func (r Ray) LineIntersect(s Ray) (point Vec) {\n\t/*\n\t\tequation is derived from system of equations with\n\t\ttwo unknowns where equations are r.Formula and s.Formula\n\t\tfrom which we can derive x of intersection point\n\n\t\tstarting with:\n\t\t\tr.V.Y*X - r.V.X*Y - r.V.Y*r.O.X + r.V.X*r.O.Y = 0\n\t\tand:\n\t\t\ts.V.Y*X - s.V.X*Y - s.V.Y*s.O.X + s.V.X*s.O.Y = 0\n\n\t\tget y from first one:\n\t\t\tr.V.Y*X - r.V.Y*r.O.X + r.V.X*r.O.Y = r.V.X*Y\n\t\t\t(r.V.Y*X - r.V.Y*r.O.X + r.V.X*r.O.Y)/r.V.X = Y\n\n\t\tthen we substitute and get x:\n\t\t\ts.V.Y*X - s.V.X * (r.V.Y*X - r.V.Y*r.O.X + r.V.X*r.O.Y) / r.V.X - s.V.Y*s.O.X + s.V.X*s.O.Y = 0 // * r.V.X\n\t\t\ts.V.Y*X*r.V.X - s.V.X*r.V.Y*X + s.V.X*r.V.Y*r.O.X - s.V.X*r.V.X*r.O.Y - s.V.Y*s.O.X*r.V.X + s.V.X*s.O.Y*r.V.X = 0 // - s.V.Y*X*r.V.X + s.V.X*r.V.Y*X\n\t\t\ts.V.X*r.V.Y*r.O.X - s.V.X*r.V.X*r.O.Y - s.V.Y*s.O.X*r.V.X + s.V.X*s.O.Y*r.V.X = s.V.X*r.V.Y*X - s.V.Y*X*r.V.X // simplify\n\t\t\ts.V.X * (r.V.Y*r.O.X + r.V.X * (s.O.Y - r.O.Y)) - s.V.Y*s.O.X*r.V.X = X * (s.V.X*r.V.Y - s.V.Y*r.V.X) // / (s.V.X*r.V.Y - s.V.Y*r.V.X)\n\t\t\t(s.V.X * (r.V.Y*r.O.X + r.V.X * (s.O.Y - r.O.Y)) - s.V.Y*s.O.X*r.V.X) / (s.V.X*r.V.Y - s.V.Y*r.V.X) = X\n\t*/\n\n\tpoint.X = (s.V.X*(r.V.Y*r.O.X+r.V.X*(s.O.Y-r.O.Y)) - s.V.Y*s.O.X*r.V.X) / (s.V.X*r.V.Y - s.V.Y*r.V.X)\n\n\tif r.V.X == 0 {\n\t\tpoint.Y = s.ProjectX(point.X)\n\t} else {\n\t\tpoint.Y = r.ProjectX(point.X)\n\t}\n\n\treturn\n}", "func IntersectionPolicy(sub ...GCPolicy) GCPolicy { return intersectionPolicy{sub} }", "func (mr *MockListMockRecorder) Intersect(other interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Intersect\", reflect.TypeOf((*MockList)(nil).Intersect), other)\n}", "func (s Set) Intersect(other Set) {\n\tintersect := New()\n\n\tfor item := range other {\n\t\tif s.Has(item) {\n\t\t\tintersect.Insert(item)\n\t\t}\n\t}\n\n\tvar remove []string\n\tfor item := range s {\n\t\tif !intersect.Has(item) {\n\t\t\tremove = append(remove, item)\n\t\t}\n\t}\n\n\tfor _, r := range remove {\n\t\ts.Remove(r)\n\t}\n}", "func (s Sphere) Intersect(r *Ray) (*Intersection, bool) {\n\n\tl := vector.Subtract(&r.origin, &s.origin)\n\ta := vector.Dot(&r.direction, &r.direction)\n\tb := 2 * vector.Dot(&r.direction, &l)\n\tc := vector.Dot(&l, &l) - s.radius*s.radius\n\n\tr0, r1, err := solveQuadratic(a, b, c)\n\n\tif err {\n\t\treturn nil, false\n\t}\n\n\tt0 := math.Min(r0, r1)\n\tt1 := math.Max(r0, r1)\n\n\tif t0 < 0 && t1 < 0 {\n\t\treturn nil, false\n\t}\n\n\ti := Intersection{thing: s, r: *r, dist: t0}\n\n\treturn &i, true\n}", "func Intersection(s, t Interface) (u Interface) {\n\tu = s.Copy()\n\tfor _, x := range s.Members() {\n\t\tif !t.Contains(x) {\n\t\t\tu.Discard(x)\n\t\t}\n\t}\n\treturn\n}", "func (s *ConcurrentSet) Intersection(others ...Set) Set {\n\tvar n sync.Map\n\n\tsize := uint32(0)\n\ts.hash.Range(func(k, v interface{}) bool {\n\t\texistAll := true\n\t\tfor _, set := range others {\n\t\t\tif !set.Contains(k) {\n\t\t\t\texistAll = false\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif existAll {\n\t\t\tn.Store(k, nothing{})\n\t\t\tsize++\n\t\t}\n\t\treturn true\n\t})\n\n\treturn &ConcurrentSet{n, size}\n}", "func IntersectAll(queries ...Query) VariadicQuery {\n\treturn VariadicQuery{\n\t\tOperator: QueryIntersectAll,\n\t\tQueries: queries,\n\t}\n}", "func intersectPreferences(a []uint8, b []uint8) (intersection []uint8) {\n\tvar j int\n\tfor _, v := range a {\n\t\tfor _, v2 := range b {\n\t\t\tif v == v2 {\n\t\t\t\ta[j] = v\n\t\t\t\tj++\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\treturn a[:j]\n}", "func Intersect(set1 []string, set2 []string) []string {\n\thelperSet := make(map[string]bool)\n\tvar andSet []string\n\n\tfor _, entry := range set1 {\n\t\thelperSet[entry] = false\n\t}\n\n\tfor _, candidate := range set2 {\n\t\tif _, ok := helperSet[candidate]; ok {\n\t\t\tandSet = append(andSet, candidate)\n\t\t}\n\t}\n\n\treturn andSet\n}", "func intervalIntersection(A [][]int, B [][]int) [][]int {\n\tia := 0\n\tib := 0\n\tmerged := [][]int{}\n\tfor {\n\t\tif ia >= len(A) || ib >= len(B) {\n\t\t\tbreak\n\t\t}\n\t\tvar a, b []int\n\t\tif ia != len(A) {\n\t\t\ta = A[ia]\n\t\t}\n\t\tif ib != len(B) {\n\t\t\tb = B[ib]\n\t\t}\n\n\t\tif a[0] < b[0] {\n\t\t\tif a[1] >= b[1] {\n\t\t\t\tmerged = append(merged, []int{b[0], b[1]})\n\t\t\t\tib++\n\t\t\t} else if a[1] >= b[0] {\n\t\t\t\tmerged = append(merged, []int{b[0], a[1]})\n\t\t\t\tia++\n\t\t\t} else {\n\t\t\t\tia++\n\t\t\t}\n\n\t\t} else if a[0] == b[0] {\n\t\t\t// a and b start together\n\t\t\tif a[1] > b[1] {\n\t\t\t\t// b contained in a\n\t\t\t\tmerged = append(merged, []int{a[0], b[1]})\n\n\t\t\t\tib++\n\t\t\t} else if a[1] == b[1] {\n\t\t\t\t// full overlap\n\t\t\t\tmerged = append(merged, []int{a[0], a[1]})\n\t\t\t\tia++\n\t\t\t\tib++\n\t\t\t} else /* a[1] < b[1] */ {\n\t\t\t\t// a contained in b\n\t\t\t\tmerged = append(merged, []int{a[0], b[1]})\n\t\t\t\tia++\n\t\t\t}\n\t\t} else {\n\t\t\t// a > b\n\t\t\tif b[1] >= a[1] {\n\t\t\t\t// containment: a in b\n\t\t\t\tmerged = append(merged, []int{a[0], a[1]})\n\t\t\t\tia++\n\t\t\t\tif b[1] == a[1] {\n\t\t\t\t\tib++\n\t\t\t\t}\n\t\t\t} else if b[1] >= a[0] {\n\t\t\t\tmerged = append(merged, []int{a[0], b[1]})\n\t\t\t\tib++\n\t\t\t} else {\n\t\t\t\t// no overlap\n\t\t\t\tib++\n\t\t\t}\n\t\t}\n\t}\n\treturn merged\n}", "func (s *IPSet) Intersect(b *IPSet) {\n\tb = b.Clone()\n\tb.Complement()\n\ts.RemoveSet(b)\n}", "func (s *Sphere) Intersect(pt, ur Vector3) float64 {\n\t// if the vector is normalized, a is always 1\n\tb := ur.Scale(2).Dot(pt.Sub(s.point))\n\tc := pt.Sub(s.point).Dot(pt.Sub(s.point)) - s.radius*s.radius\n\tdisc := b*b - 4*c\n\tif disc < 0 {\n\t\treturn math.Inf(0)\n\t}\n\tt0 := (-b - math.Sqrt(disc)) / 2\n\tif t0 > 0 {\n\t\treturn t0 - EPSILON\n\t}\n\tt1 := (-b + math.Sqrt(disc)) / 2\n\treturn t1 - EPSILON\n}", "func (s *Uint64) Intersect(other Uint64) Uint64 {\n\tres := NewUint64WithSize(min(s.Size(), other.Size()))\n\n\t// loop over the smaller set\n\tif len(s.m) <= len(other.m) {\n\t\tfor val := range s.m {\n\t\t\tif other.Contains(val) {\n\t\t\t\tres.Add(val)\n\t\t\t}\n\t\t}\n\t} else {\n\t\tfor val := range other.m {\n\t\t\tif s.Contains(val) {\n\t\t\t\tres.Add(val)\n\t\t\t}\n\t\t}\n\t}\n\treturn res\n}", "func intersects(\n\tm_minx, m_miny, m_maxx, m_maxy float64,\n\to_minx, o_miny, o_maxx, o_maxy float64,\n) bool {\n\t//not disjoint\n\treturn !(o_minx > m_maxx || o_maxx < m_minx || o_miny > m_maxy || o_maxy < m_miny)\n}", "func Intersect(firstSet cpuset.CPUSet, secondSet cpuset.CPUSet) []int {\n\tcommonSet := firstSet.Intersection(secondSet)\n\treturn commonSet.List()\n}", "func (a Possibility) Intersect(b Possibility) Possibility {\n\tif a == Impossible || b == Impossible {\n\t\treturn Impossible\n\t}\n\tif a == Maybe {\n\t\tif b == Maybe {\n\t\t\treturn Maybe\n\t\t}\n\t\ta, b = b, a\n\t}\n\t// a is True or False\n\t// b is True, False or Maybe\n\tif b == Maybe || a == b {\n\t\treturn a\n\t}\n\treturn Impossible\n}", "func (nset *nodeSet) intersection(nset2 *nodeSet) *nodeSet {\n\tintersection := newNodeSet()\n\tif nset2 == nil {\n\t\treturn intersection\n\t}\n\n\tfor k := range nset.nodes {\n\t\tif nset2.contains(k) {\n\t\t\tintersection.add(k)\n\t\t}\n\t}\n\treturn intersection\n\n}", "func Intersect(s1 []string, s2 []string) []string {\n\tmb := make(map[string]struct{}, len(s2))\n\tfor _, v := range s2 {\n\t\tmb[v] = struct{}{}\n\t}\n\n\tintersect := make([]string, 0)\n\tfor _, v := range s1 {\n\t\tif _, ok := mb[v]; ok {\n\t\t\tintersect = append(intersect, v)\n\t\t}\n\t}\n\treturn intersect\n}", "func (p *RPrism) Intersect(r Ray) (float64, *Material, Vec3, Vec3, bool) {\n\tqueries := []*rprismIntersectQ{\n\t\t{\n\t\t\t[3]float64{r.V.X, r.V.Y, r.V.Z},\n\t\t\t[3]float64{r.D.X, r.D.Y, r.D.Z},\n\t\t\tp.Pos.X, p.Pos.Y, p.Pos.Y + p.Dim[1], p.Pos.Z, p.Pos.Z + p.Dim[2],\n\t\t\tVec3{-1, 0, 0},\n\t\t},\n\t\t{\n\t\t\t[3]float64{r.V.X, r.V.Y, r.V.Z},\n\t\t\t[3]float64{r.D.X, r.D.Y, r.D.Z},\n\t\t\tp.Pos.X + p.Dim[0], p.Pos.Y, p.Pos.Y + p.Dim[1], p.Pos.Z, p.Pos.Z + p.Dim[2],\n\t\t\tVec3{1, 0, 0},\n\t\t},\n\t\t{\n\t\t\t[3]float64{r.V.Y, r.V.Z, r.V.X},\n\t\t\t[3]float64{r.D.Y, r.D.Z, r.D.X},\n\t\t\tp.Pos.Y, p.Pos.Z, p.Pos.Z + p.Dim[2], p.Pos.X, p.Pos.X + p.Dim[0],\n\t\t\tVec3{0, -1, 0},\n\t\t},\n\t\t{\n\t\t\t[3]float64{r.V.Y, r.V.Z, r.V.X},\n\t\t\t[3]float64{r.D.Y, r.D.Z, r.D.X},\n\t\t\tp.Pos.Y + p.Dim[1], p.Pos.Z, p.Pos.Z + p.Dim[2], p.Pos.X, p.Pos.X + p.Dim[0],\n\t\t\tVec3{0, 1, 0},\n\t\t},\n\t\t{\n\t\t\t[3]float64{r.V.Z, r.V.X, r.V.Y},\n\t\t\t[3]float64{r.D.Z, r.D.X, r.D.Y},\n\t\t\tp.Pos.Z, p.Pos.X, p.Pos.X + p.Dim[0], p.Pos.Y, p.Pos.Y + p.Dim[1],\n\t\t\tVec3{0, 0, -1},\n\t\t},\n\t\t{\n\t\t\t[3]float64{r.V.Z, r.V.X, r.V.Y},\n\t\t\t[3]float64{r.D.Z, r.D.X, r.D.Y},\n\t\t\tp.Pos.Z + p.Dim[2], p.Pos.X, p.Pos.X + p.Dim[0], p.Pos.Y, p.Pos.Y + p.Dim[1],\n\t\t\tVec3{0, 0, 1},\n\t\t},\n\t}\n\tnearest := math.MaxFloat64\n\tfound := false\n\tvar normal Vec3\n\tfor _, q := range queries {\n\t\td, ok := rprismIntersects(q)\n\t\tif ok {\n\t\t\tif d > minDistance && d < nearest {\n\t\t\t\tfound = true\n\t\t\t\tnearest = d\n\t\t\t\tnormal = q.normal\n\t\t\t}\n\t\t}\n\t}\n\n\tif !found {\n\t\treturn 0, nil, Vec3{}, Vec3{}, false\n\t}\n\tpt := r.At(nearest)\n\treturn nearest, p.Mat, pt, normal, found\n}", "func Intersection() {\n\n}", "func IntersectInts(x []int, y []int) []int {\n\ttraceMap := make(map[int]bool)\n\tresult := make([]int, 0)\n\tfor _, v := range x {\n\t\ttraceMap[v] = true\n\t}\n\tfor _, v := range y {\n\t\tif traceMap[v] {\n\t\t\tresult = append(result, v)\n\t\t}\n\t}\n\treturn result\n}", "func (ts List) Intersection() List {\n\treturn ts.IntersectionWithHandler(func(intersectingEvent1, intersectingEvent2, intersectionSpan T) T {\n\t\treturn intersectionSpan\n\t})\n}" ]
[ "0.6938771", "0.6470136", "0.6355641", "0.62724626", "0.6205251", "0.61451906", "0.6133328", "0.6062865", "0.604192", "0.5919244", "0.5915263", "0.59103686", "0.5897044", "0.58723134", "0.5854885", "0.58476067", "0.58396256", "0.5808019", "0.5781952", "0.5766504", "0.57609963", "0.5736789", "0.57056004", "0.57007545", "0.5687982", "0.5676356", "0.5670762", "0.56660235", "0.56534517", "0.5648309", "0.5633577", "0.5622809", "0.56127214", "0.56038785", "0.55945486", "0.5587837", "0.55619997", "0.55602413", "0.5540261", "0.54906875", "0.5474582", "0.5428652", "0.54277235", "0.5400283", "0.53901184", "0.538065", "0.53776085", "0.5373746", "0.5373571", "0.535735", "0.53441924", "0.533112", "0.533112", "0.532211", "0.53056985", "0.5304094", "0.5303666", "0.5286916", "0.5285439", "0.52792674", "0.5272508", "0.5267111", "0.52593863", "0.52579945", "0.5245263", "0.523386", "0.5229889", "0.5218977", "0.5217608", "0.52031714", "0.51900774", "0.5166832", "0.51650506", "0.5153346", "0.51503503", "0.51460236", "0.5134603", "0.5125977", "0.5124047", "0.51194644", "0.5113056", "0.5102253", "0.50609845", "0.50590825", "0.5057275", "0.50415313", "0.50160754", "0.50060606", "0.49976295", "0.49927217", "0.4988369", "0.49826875", "0.49770042", "0.4949581", "0.4948215", "0.49379048", "0.4936619", "0.49327213", "0.49303883", "0.4924527" ]
0.7582888
0
Reverse returns a copy of operand with the elements in reverse order.
func Reverse(operand []string) []string { reversed := make([]string, len(operand)) for i := range operand { reversed[len(operand)-i-1] = operand[i] } return reversed }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *T) Reverse() *T { return &T{lhs: t.rhs, rhs: t.lhs} }", "func (*Functions) Reverse(in interface{}) interface{} {\n\tv := reflect.ValueOf(in)\n\tc := v.Len()\n\tout := reflect.MakeSlice(v.Type(), c, c)\n\tfor i := 0; i < c; i++ {\n\t\tout.Index(i).Set(v.Index(c - i - 1))\n\t}\n\treturn out.Interface()\n}", "func Reverse(scope *Scope, tensor tf.Output, dims tf.Output) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Reverse\",\n\t\tInput: []tf.Input{\n\t\t\ttensor, dims,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (a Slice[T]) Reverse() Slice[T] {\n\tfor i := len(a)/2 - 1; i >= 0; i-- {\n\t\topp := len(a) - 1 - i\n\t\ta[i], a[opp] = a[opp], a[i]\n\t}\n\treturn a\n}", "func Brrev(op Op) Op", "func reverse(args []interface{}) []interface{} {\n\treversed := make([]interface{}, len(args))\n\n\tj := 0\n\tfor i := len(args) - 1; i >= 0; i-- {\n\t\treversed[j] = args[i]\n\t\tj++\n\t}\n\n\treturn reversed\n}", "func Reverse(slice interface{}) {\n\tvalue := reflect.ValueOf(slice)\n\tif value.Kind() == reflect.Ptr {\n\t\tvalue = value.Elem()\n\t}\n\tif value.Kind() != reflect.Slice {\n\t\treturn\n\t}\n\n\ttemp := reflect.New(value.Index(0).Type()).Elem()\n\tfor i, j := 0, value.Len()-1; i < j; i, j = i+1, j-1 {\n\t\ttemp.Set(value.Index(i))\n\t\tvalue.Index(i).Set(value.Index(j))\n\t\tvalue.Index(j).Set(temp)\n\t}\n}", "func Reverse(slice interface{}) {\n\tv := reflect.ValueOf(slice)\n\tl := v.Len()\n\tfor i := 0; i < l/2; i++ {\n\t\ta, b := v.Index(i), v.Index(l-1-i)\n\t\tt := a.Interface()\n\t\ta.Set(b)\n\t\tb.Set(reflect.ValueOf(t))\n\t}\n}", "func (s Runes) Reverse() Runes {\n\tsz := len(s)\n\trs := make(Runes, sz)\n\tif sz > 0 {\n\t\tfor n := 0; n <= sz/2; n++ {\n\t\t\trs[n], rs[sz-n-1] = s[sz-n-1], s[n]\n\t\t}\n\t}\n\treturn rs\n}", "func (elem *Items) Reverse() {\n\teLen := len(*elem)\n\tvar target int\n\tfor i := eLen/2 - 1; i >= 0; i-- {\n\t\ttarget = eLen - 1 - i\n\t\t(*elem)[i], (*elem)[target] = (*elem)[target], (*elem)[i]\n\t}\n}", "func (q *Query) Reverse() *Query {\n\tq.reverse = true\n\treturn q\n}", "func (i Interval) Reverse() Interval {\n\treturn Interval{Start: i.End, End: i.Start}\n}", "func Reverse[T any](collection []T) []T {\n\tlength := len(collection)\n\thalf := length / 2\n\n\tfor i := 0; i < half; i = i + 1 {\n\t\tj := length - 1 - i\n\t\tcollection[i], collection[j] = collection[j], collection[i]\n\t}\n\n\treturn collection\n}", "func Reverse[T any](slice []T) {\n\tfor i, j := 0, len(slice)-1; i < j; i, j = i+1, j-1 {\n\t\tslice[i], slice[j] = slice[j], slice[i]\n\t}\n}", "func reverse(input []int, left, right int) {\n\ttmp := 0\n\tfor left < right {\n\t\ttmp = input[left]\n\t\tinput[left], input[right] = input[right], tmp\n\t\tleft++\n\t\tright--\n\t}\n}", "func Reverse(l List) List {\n\tfoldFunc := func(carry, elem interface{}) interface{} {\n\t\treturn Cons(elem, carry)\n\t}\n\treturn Foldl(foldFunc, Mzero(), l).(List)\n}", "func (a *ArrayObject) reverse() *ArrayObject {\n\tarrLen := len(a.Elements)\n\treversedArrElems := make([]Object, arrLen)\n\n\tfor i, element := range a.Elements {\n\t\treversedArrElems[arrLen-i-1] = element\n\t}\n\n\tnewArr := &ArrayObject{\n\t\tbaseObj: &baseObj{class: a.class},\n\t\tElements: reversedArrElems,\n\t}\n\n\treturn newArr\n}", "func Inverse(arg interface{}) Value {\n\treturn Reverse(arg)\n}", "func ReverseInPlace(arr []int) []int {\n\tfor i, j := 0, len(arr)-1; i < j; i, j = i+1, j-1 {\n\t\tarr[i], arr[j] = arr[j], arr[i]\n\t}\n\treturn arr\n}", "func Reverse(slice AnySlice) {\n\tmustBeSlice(slice)\n\tcount := reflect.ValueOf(slice).Len()\n\tswapper := reflect.Swapper(slice)\n\ti := 0\n\tj := count - 1\n\tfor i < j {\n\t\tswapper(i, j)\n\t\ti++\n\t\tj--\n\t}\n}", "func (l *List) Reverse() *List {\n\tr := List{}\n\tfor data, err := l.PopBack(); err == nil; data, err = l.PopBack() {\n\t\tr.PushBack(data)\n\t}\n\tl.first, l.last = r.first, r.last\n\treturn l\n}", "func Reverse(slice []interface{}) {\n\tfor i, j := 0, len(slice)-1; i < j; i, j = i+1, j-1 {\n\t\tslice[i], slice[j] = slice[j], slice[i]\n\t}\n}", "func Reverse(seq Seq) Seq {\n\treturn RevAppend(seq, nil)\n}", "func execReverse(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := sort.Reverse(args[0].(sort.Interface))\n\tp.Ret(1, ret)\n}", "func ExampleReverse() {\n\tsli := []int{1, 2, 3, 4, 5, 6, 7, 8, 9, 10}\n\tfor i := 0; i < len(sli)/2; i++ {\n\t\tsli[i], sli[len(sli)-i-1] = sli[len(sli)-i-1], sli[i]\n\t}\n\tfmt.Println(sli)\n\n\t// Output:\n\t// [10 9 8 7 6 5 4 3 2 1]\n}", "func Reverse(seq Sequence) Sequence {\n\tvar ff FeatureSlice\n\tfor _, f := range seq.Features() {\n\t\tff = ff.Insert(Feature{f.Key, f.Loc.Reverse(Len(seq)), f.Props.Clone()})\n\t}\n\tseq = WithFeatures(seq, ff)\n\n\tp := make([]byte, Len(seq))\n\tcopy(p, seq.Bytes())\n\tflip.Bytes(p)\n\tseq = WithBytes(seq, p)\n\n\treturn seq\n}", "func Reverse(arg interface{}) Value {\n\tif val, ok := arg.(Value); ok {\n\t\treturn val.Reverse()\n\t}\n\treturn value{value: arg, color: ReverseFm}\n}", "func (q Query) Reverse() Query {\n\treturn Query{\n\t\tIterate: func() Iterator {\n\t\t\tnext := q.Iterate()\n\t\t\tvar readDone = false\n\t\t\tvar readError error\n\t\t\tvar index int\n\n\t\t\tvar items = make([]Record, 0, 16)\n\n\t\t\treturn func(ctx Context) (item Record, err error) {\n\t\t\t\tif !readDone {\n\t\t\t\t\tif readError != nil {\n\t\t\t\t\t\terr = readError\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\tfor {\n\t\t\t\t\t\tcurrent, err := next(ctx)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\tif !IsNoRows(err) {\n\t\t\t\t\t\t\t\treadError = err\n\t\t\t\t\t\t\t\treturn Record{}, err\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\titems = append(items, current)\n\t\t\t\t\t}\n\n\t\t\t\t\tindex = len(items) - 1\n\t\t\t\t\treadDone = true\n\t\t\t\t}\n\n\t\t\t\tif index < 0 {\n\t\t\t\t\terr = ErrNoRows\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\titem = items[index]\n\t\t\t\tindex--\n\t\t\t\treturn\n\t\t\t}\n\t\t},\n\t}\n}", "func (v Data) Reverse() {\n\tfor i, j := 0, len(v)-1; i < j; i, j = i+1, j-1 {\n\t\tv.Swap(i, j)\n\t}\n}", "func SliceReverse(a interface{}) {\n\trv := reflect.ValueOf(a)\n\n\tif !rv.IsValid() {\n\t\treturn\n\t}\n\n\tif rv.Kind() != reflect.Slice && rv.Kind() != reflect.Array {\n\t\treturn\n\t}\n\n\tfor i := 0; i < rv.Len()/2; i++ {\n\t\ttemp := rv.Index(rv.Len() - i - 1).Interface()\n\t\trv.Index(rv.Len() - i - 1).Set(rv.Index(i))\n\t\trv.Index(i).Set(reflect.ValueOf(temp))\n\t}\n}", "func Reverse(src []byte) []byte {\n\tdst := make([]byte, len(src))\n\tfor i := len(src); i > 0; i-- {\n\t\tdst[len(src)-i] = src[i-1]\n\t}\n\treturn dst\n}", "func Reverse(src []byte) []byte {\n\tdst := make([]byte, len(src))\n\tfor i := len(src); i > 0; i-- {\n\t\tdst[len(src)-i] = src[i-1]\n\t}\n\treturn dst\n}", "func (s *SliceInt) Reverse() *SliceInt {\n\tif s == nil {\n\t\treturn nil\n\t}\n\n\tfor i := len(s.data)/2 - 1; i >= 0; i-- {\n\t\topp := len(s.data) - 1 - i\n\t\ts.data[i], s.data[opp] = s.data[opp], s.data[i]\n\t}\n\n\treturn s\n}", "func (this *Tuple) Reverse() {\n\tfor i, j := 0, this.Len()-1; i < j; i, j = i+1, j-1 {\n\t\tthis.data[i], this.data[j] = this.data[j], this.data[i]\n\t}\n}", "func (g *Graph) ReverseSub(x1 Node, x2 Node) Node {\n\treturn g.NewOperator(fn.NewReverseSubScalar(x1, x2), x1, x2)\n}", "func reverse(dst []string) []string {\n\tlength := len(dst)\n\tfor i := 0; i < length/2; i++ {\n\t\tdst[i], dst[length-i-1] = dst[length-i-1], dst[i]\n\t}\n\treturn dst\n}", "func (v IntVec) Reverse() {\n\tfor i, j := 0, len(v)-1; i < j; i, j = i+1, j-1 {\n\t\tv.Swap(i, j)\n\t}\n}", "func (c *StoreCandidates) Reverse() *StoreCandidates {\n\tfor i := len(c.Stores)/2 - 1; i >= 0; i-- {\n\t\topp := len(c.Stores) - 1 - i\n\t\tc.Stores[i], c.Stores[opp] = c.Stores[opp], c.Stores[i]\n\t}\n\treturn c\n}", "func (p *IntArray) Reverse() {\n\ttmp := *p\n\tvar new_array IntArray\n\tlength := len(tmp)\n\n\tfor i := length - 1; i >= 0; i-- {\n\t\tnew_array = append(new_array, tmp[i])\n\t}\n\t*p = new_array\n}", "func Reverse[T any](ss []T) []T {\n\t// Avoid the allocation. If there is one element or less it is already\n\t// reversed.\n\tif len(ss) < 2 {\n\t\treturn ss\n\t}\n\n\tsorted := make([]T, len(ss))\n\tfor i := 0; i < len(ss); i++ {\n\t\tsorted[i] = ss[len(ss)-i-1]\n\t}\n\n\treturn sorted\n}", "func reverse(dst, src []byte) []byte {\n\tl := len(dst)\n\tfor i, j := 0, l-1; i < (l+1)/2; {\n\t\tdst[i], dst[j] = src[j], src[i]\n\t\ti++\n\t\tj--\n\t}\n\treturn dst\n}", "func (l List) Reverse() List {\n\ts := make(List, len(l))\n\tfor i, j := 0, len(l)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = l[j], l[i]\n\t}\n\treturn s\n}", "func ReverseSlice(s interface{}) {\n\tsize := reflect.ValueOf(s).Len()\n\tswap := reflect.Swapper(s)\n\tfor i, j := 0, size-1; i < j; i, j = i+1, j-1 {\n\t\tswap(i, j)\n\t}\n}", "func reverse(rs []*histRecord) {\n\tfor i := 0; i < len(rs)/2; i++ {\n\t\trs[i], rs[len(rs)-i-1] = rs[len(rs)-i-1], rs[i]\n\t}\n}", "func Reverse(m Mapping) Mapping {\n\tsrc, dst := m.Mapping()\n\treturn Map(dst, src)\n}", "func (d *Deque) Reverse() Sequence {\n\treturn &Deque{\n\t\thead: d.tail,\n\t\ttail: d.head,\n\t}\n}", "func (this *Sorter) Reverse() *Sorter {\n\tthis.reversed = !this.reversed\n\treturn this\n}", "func (h *Hash) Reverse() Hash {\n\tfor i, b := range h[:HashSize/2] {\n\t\th[i], h[HashSize-1-i] = h[HashSize-1-i], b\n\t}\n\treturn *h\n}", "func reverse(a *[arraySize]int) {\n\tfor i, j := 0, len(a)-1; i < j; i, j = i+1, j-1 {\n\t\ta[i], a[j] = a[j], a[i]\n\t}\n}", "func Reverse(data interface{}) (interface{}, error) {\n\tvar err error\n\n\tresult := func(err *error) interface{} {\n\t\tdefer catch(err)\n\n\t\tif !isNonNilData(err, \"data\", data) {\n\t\t\treturn nil\n\t\t}\n\n\t\tdataValue, dataType, _, dataValueLen := inspectData(data)\n\n\t\tif !isSlice(err, \"data\", dataValue) {\n\t\t\treturn nil\n\t\t}\n\n\t\tresult := makeSlice(dataType)\n\n\t\tif dataValueLen == 0 {\n\t\t\treturn result.Interface()\n\t\t}\n\n\t\tfor i := 0; i < dataValueLen; i++ {\n\t\t\tresult = reflect.Append(result, dataValue.Index(dataValueLen-1-i))\n\t\t}\n\n\t\treturn result.Interface()\n\t}(&err)\n\n\treturn result, err\n}", "func Reverse(ctx *sapphire.CommandContext) {\n\n\tvar chars []string = strings.Split(ctx.JoinedArgs(), \"\")\n\n\tfor i, j := 0, len(chars)-1; i < j; i, j = i+1, j-1 {\n\t\tchars[i], chars[j] = chars[j], chars[i]\n\t}\n\n\tctx.Reply(sapphire.Escape(strings.Join(chars, \"\")))\n}", "func Reverse(nodes []graph.Node) {\n\tfor i, j := 0, len(nodes)-1; i < j; i, j = i+1, j-1 {\n\t\tnodes[i], nodes[j] = nodes[j], nodes[i]\n\t}\n}", "func (l *List) Reverse() *List {\n\tcurrent := l.head\n\tl.head, l.tail = l.tail, l.head\n\n\tfor current != nil {\n\t\tcurrent.prev, current.next = current.next, current.prev\n\t\tcurrent = current.prev\n\t}\n\n\treturn l\n}", "func reverse(arr []int64) []int64 {\n\tarrLen := len(arr)\n\tnewArr := make([]int64, arrLen)\n\tfor i := arrLen - 1; i > -1; i-- {\n\t\tnewArr[arrLen-(i+1)] = arr[i]\n\t}\n\treturn newArr\n}", "func ReverseSlice(s interface{}) {\n\treflectValue := reflect.ValueOf(s)\n\tif reflectValue.Kind() != reflect.Slice {\n\t\treturn\n\t}\n\tsize := reflectValue.Len()\n\tswap := reflect.Swapper(s)\n\tfor i, j := 0, size-1; i < j; i, j = i+1, j-1 {\n\t\tswap(i, j)\n\t}\n}", "func (s *stack) Reverse() *stack {\n\tstack := NewStack()\n\tfor node := s.top; node != nil; {\n\t\tstack.Push(node.value)\n\t\tnode = node.prev\n\t}\n\treturn stack\n}", "func reverse(runes []rune, start, end int) []rune {\n\tfor i, j := start, end; i < j; i, j = i+1, j-1 {\n\t\trunes[i], runes[j] = runes[j], runes[i]\n\t}\n\treturn runes\n}", "func BitReverse(a []fr.Element) {\n\tn := uint64(len(a))\n\tnn := uint64(64 - bits.TrailingZeros64(n))\n\n\tfor i := uint64(0); i < n; i++ {\n\t\tirev := bits.Reverse64(i) >> nn\n\t\tif irev > i {\n\t\t\ta[i], a[irev] = a[irev], a[i]\n\t\t}\n\t}\n}", "func (l *List) Reverse() *List {\n\tl.head, l.tail = l.tail, l.head\n\tcurr := l.head\n\tfor curr != nil {\n\t\tcurr.prev, curr.next = curr.next, curr.prev\n\t\tcurr = curr.next\n\t}\n\treturn l\n}", "func (ll *Doubly[T]) Reverse() {\n\tvar Prev, Next *Node[T]\n\tcur := ll.Head\n\n\tfor cur != nil {\n\t\tNext = cur.Next\n\t\tcur.Next = Prev\n\t\tcur.Prev = Next\n\t\tPrev = cur\n\t\tcur = Next\n\t}\n\n\tll.Head = Prev\n}", "func Reverse(s []int) {\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = s[j], s[i]\n\n\t}\n}", "func (s Players) Reverse() (out Players) {\n\tfor i := len(s) - 1; i >= 0; i-- {\n\t\tout = append(out, s[i])\n\t}\n\treturn\n}", "func reverse(s []int64) {\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = s[j], s[i]\n\t}\n}", "func reverse(items []string) []string {\n\tl := len(items)\n\tvar reversed = make([]string, l)\n\tfor i, v := range items {\n\t\treversed[l-i-1] = v\n\t}\n\treturn reversed\n}", "func (v Int32Vec) Reverse() {\n\tfor i, j := 0, len(v)-1; i < j; i, j = i+1, j-1 {\n\t\tv.Swap(i, j)\n\t}\n}", "func reverse(s []int) {\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = s[j], s[i]\n\t} \n}", "func (l *List) Reverse() {\n\tnode := l.head\n\tnewHead, newLast := l.last, l.head\n\n\tfor node != nil {\n\t\tnode.previous, node.next = node.Next(), node.Prev()\n\t\tnode = node.Prev()\n\t}\n\n\tl.head, l.last = newHead, newLast\n}", "func (array Array) Reverse() Array {\n\treversedArray := New()\n\tfor i := len(array) - 1; i >= 0; i-- {\n\t\treversedArray = reversedArray.Push(array[i])\n\t}\n\treturn reversedArray\n}", "func (p *Particle) Reverse() {\n\tp.ReverseX()\n\tp.ReverseY()\n}", "func Reverse(sort []string) {\n\tns := len(sort) - 1\n\tfor i := 0; i < (ns+1)/2; i++ {\n\t\tsort[i], sort[ns-i] = sort[ns-i], sort[i]\n\t}\n}", "func Reverse(arr []int) []int {\n\tstack := doubly_linked_list.NewStack()\n\tfor _, item := range arr {\n\t\tstack.Push(item)\n\t}\n\ti := 0\n\tfor !stack.IsEmpty() {\n\t\tif item, ok := stack.Pop(); !ok {\n\t\t\tarr[i] = item.(int)\n\t\t\ti++\n\t\t}\n\t}\n\tfmt.Println(stack.Size())\n\treturn arr\n}", "func reverse(s []int) []int {\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = s[j], s[i]\n\t}\n\treturn s\n}", "func reverse(s []int) {\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = s[j], s[i]\n\t}\n}", "func reverse(s []int) {\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = s[j], s[i]\n\t}\n}", "func reverse(s []int) {\n\tfor i, j := 0, len(s) - 1; i < j; i, j = i + 1, j - 1 {\n\t\ts[i], s[j] = s[j], s[i]\n\t}\n}", "func ExampleReverse(){\n\tfmt.Println(Reverse(\"Hello, World\"))\n\t// Output: dlroW ,olleH\n}", "func reverse(s []int) { // we expect a slice , hence arrays will not work\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\t(s)[i], (s)[j] = (s)[j], (s)[i]\n\t}\n}", "func reverse(commits []*vcsinfo.LongCommit) {\n\ttotal := len(commits)\n\tfor i := 0; i < total/2; i++ {\n\t\tcommits[i], commits[total-i-1] = commits[total-i-1], commits[i]\n\t}\n}", "func reverse(s []int){\n\tfor i,j := 0,len(s)-1;i < j;i,j = i + 1,j-1{\n\t\ts[i],s[j] = s[j],s[i]\n\t}\n}", "func reverse(number []byte) []byte {\n\tres := make([]byte, len(number))\n\th := 0\n\tt := len(number) - 1\n\tfor t >= 0 {\n\t\tres[h] = number[t]\n\t\th++\n\t\tt--\n\t}\n\treturn res\n}", "func reverse(start, end int, arr []int) {\n\tfor start < end {\n\n\t\t// swapping both indexes\n\t\t// using left and right pointers\n\t\tarr[start], arr[end] = arr[end], arr[start]\n\n\t\t// moving forward\n\t\tstart++\n\t\t// moving backwards\n\t\tend--\n\t}\n}", "func reverse(s []int) {\r\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\r\n\t\ts[i], s[j] = s[j], s[i]\r\n\t}\r\n}", "func Right(arr []int, n int) {\n\tmisc.Reverse(arr[n:])\n\tmisc.Reverse(arr[:n])\n\tmisc.Reverse(arr[:])\n}", "func reverse(s []string) []string {\n\tfor head, tail := 0, len(s)-1; head < tail; head, tail = head+1, tail-1 {\n\t\ts[head], s[tail] = s[tail], s[head]\n\t}\n\treturn s\n}", "func (p PagesGroup) Reverse() PagesGroup {\n\tfor i, j := 0, len(p)-1; i < j; i, j = i+1, j-1 {\n\t\tp[i], p[j] = p[j], p[i]\n\t}\n\n\treturn p\n}", "func reverse(init int, end int, arr []byte) {\n\tsize := len(arr)\n\tfor {\n\t\tif init >= end {\n\t\t\tbreak\n\t\t}\n\t\taux := arr[init%size]\n\t\tarr[init%size] = arr[end%size]\n\t\tarr[end%size] = aux\n\t\tinit++\n\t\tend--\n\t}\n}", "func reverse(in []string) []string {\n\tif len(in) == 0 {\n\t\treturn in\n\t}\n\treturn append(reverse(in[1:]), in[0])\n}", "func rev(s []int) {\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = s[j], s[i]\n\t}\n}", "func (mdl *Model) Reverse() {\n\treturn\n}", "func reverseSlice(orderedSlice []Message) []Message {\n\tlast := len(orderedSlice) - 1\n\tfor i := 0; i < len(orderedSlice)/2; i++ {\n\t\torderedSlice[i], orderedSlice[last-i] = orderedSlice[last-i], orderedSlice[i]\n\t}\n\n\treturn orderedSlice\n}", "func reverseComplement(in *bytes.Buffer) []byte {\n\tvar result []byte = make([]byte, in.Len(), in.Len())\n\tfor pos := in.Len() - 1; pos >= 0; pos-- {\n\t\tcurrent, ok := in.ReadByte()\n\t\tcheckResult(ok)\n\t\tresult[pos] = reverse(current)\n\t}\n\treturn result\n}", "func rev(in []byte) {\n\ts := len(in)\n\tfor i := 0; i < len(in)/2; i++ {\n\t\tin[i], in[s-1-i] = in[s-1-i], in[i]\n\t}\n}", "func (hc Code) Reversed() Code {\n\treturn MakeReversedCode(hc.Size, hc.Bits)\n}", "func reverse(s []int) {\n\t// start: 5 4 3 2 1 0\n\t// step: i=0, j=5\n\t// \t\t 0 4 3 2 1 5\n\t// step: i=1, j=4\n\t// \t\t 0 1 3 2 4 5\n\t// step: i=2, j=3\n\t// end: 0 1 2 3 4 5\n\tfor i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 {\n\t\ts[i], s[j] = s[j], s[i]\n\t}\n}", "func BSWAPQ(r operand.Op) { ctx.BSWAPQ(r) }", "func Reverse(input string) string {\n\tvar chars []byte\n\n\tl := len(input)\n\n\tfor i := l - 1; i >= 0; i-- {\n\t\tchars = append(chars, input[i])\n\t}\n\n\treturn string(chars)\n}", "func reverse(numbers []int) []int {\n\tfor i, j := 0, len(numbers)-1; i < j; i, j = i+1, j-1 {\n\t\tnumbers[i], numbers[j] = numbers[j], numbers[i]\n\t}\n\n\treturn numbers\n}", "func (i *Iter) Reversal() *stripe.Reversal {\n\treturn i.Current().(*stripe.Reversal)\n}", "func (l IntList) Reverse() IntList {\n\tnewList := make(IntList, 0)\n\tfor i := (l.Length() - 1); i >= 0; i-- {\n\t\tnewList = newList.Append(IntList([]int{l[i]}))\n\t}\n\treturn newList\n}", "func Reverse(input string) string {\n\tvar reverse string\n\trunes := []rune(input)\n\tlength := len(runes)\n\tfor i := 0; i < length; i++ {\n\t\treverse += string(runes[length-1-i])\n\t}\n\treturn reverse\n}" ]
[ "0.6923477", "0.6782499", "0.6638562", "0.6633635", "0.66281265", "0.65916026", "0.6445335", "0.64214706", "0.6296603", "0.6294488", "0.62126815", "0.6190993", "0.61606294", "0.6137002", "0.6126894", "0.61138034", "0.61090255", "0.60554713", "0.60536623", "0.6051185", "0.6050039", "0.6034056", "0.6000249", "0.59990776", "0.59598947", "0.59446484", "0.5944222", "0.59337616", "0.5925186", "0.5923292", "0.5885604", "0.5885604", "0.58783805", "0.5877229", "0.58755857", "0.58713937", "0.58336884", "0.5832294", "0.58238405", "0.5817141", "0.5805176", "0.5769159", "0.57587856", "0.5744355", "0.57412744", "0.57276756", "0.5717599", "0.5715222", "0.5712221", "0.5703926", "0.56957513", "0.5687662", "0.56752986", "0.5671374", "0.56538147", "0.5653301", "0.56137747", "0.56014794", "0.5592604", "0.5578593", "0.5528289", "0.55152655", "0.55049014", "0.549953", "0.5477589", "0.5418811", "0.5403496", "0.5403232", "0.54032266", "0.5402352", "0.54010135", "0.5395224", "0.53946984", "0.53946984", "0.5370692", "0.5369741", "0.53589875", "0.53542984", "0.53385687", "0.53315413", "0.53289664", "0.532262", "0.5320558", "0.5312604", "0.5308028", "0.5300127", "0.52995104", "0.52966094", "0.529355", "0.52744925", "0.5272786", "0.52564347", "0.52513236", "0.52391183", "0.52224547", "0.521883", "0.5213185", "0.5191742", "0.5189683", "0.5189245" ]
0.7513851
0
Seq generates a sequence of ints from first to last. If incr is specified (an optional third argument), then the sequence will increment by incr. Otherwise, incr defaults to 1. Incr may be negative to generate a sequence of descending ints.
func Seq(first, last int, incr ...int) []int { j := 1 if len(incr) > 1 { panic("Seq incr cannot be more than one value in length") } if len(incr) == 1 { if (incr[0]) == 0 { panic("Seq incr value cannot be zero") } j = incr[0] } var values []int current := first if j > 0 { for current <= last { values = append(values, current) current += j } } else { for current >= last { values = append(values, current) current += j } } return values }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func intSeq() func() int {\n\ti := 0\n\treturn func() int {\n\t\ti++\n\t\treturn i\n\t}\n}", "func intSeq() func() int {\n\ti := 0\n\treturn func() int {\n\t\ti++\n\t\treturn i\n\t}\n}", "func (d *DataPacket) SequenceIncr() {\n\td.data[111]++\n}", "func Sequence(start int, includedEnd int, interval int) []int {\n\tif interval <= 0 {\n\t\tlogger.Error(\"Cannot generate sequence with interval smaller or equal to zero\")\n\t\treturn make([]int, 0)\n\t}\n\tn := 1 + ((includedEnd - start) / interval)\n\tseq := make([]int, n)\n\tfor i := 0; i < n; i++ {\n\t\tseq[i] = start + (i-1)*interval\n\t}\n\treturn seq\n}", "func seq(id string, vals ...int64) chan Ordered {\n\tch := make(chan Ordered)\n\n\tgo func() {\n\t\tfor _, v := range vals {\n\t\t\tch <- item{id, v}\n\t\t}\n\t\tclose(ch)\n\t}()\n\n\treturn ch\n}", "func incSeq(seq *[24]byte) {\n\tn := uint32(1)\n\tfor i := 0; i < 8; i++ {\n\t\tn += uint32(seq[i])\n\t\tseq[i] = byte(n)\n\t\tn >>= 8\n\t}\n}", "func intSeq() func() int {\n i := 0\n return func() int {\n i += 1\n return i\n }\n}", "func NewInt(start int, incr uint, end int, width uint) Sequence {\n\tstep := int(incr)\n\tif end < start {\n\t\tstep = -step\n\t}\n\n\tseq := &intSequence{data: make(chan string), step: step, end: end, width: width, widthExceededMutex: sync.RWMutex{}}\n\n\tgo seq.push(start)\n\n\treturn seq\n}", "func (b *Bar) Incr(n int) {\n\tif n < 1 || isClosed(b.done) {\n\t\treturn\n\t}\n\tb.incrCh <- int64(n)\n}", "func (e *Etcd) Sequence(ctx context.Context, key string) (int64, error) {\n\tfor {\n\t\tval, err := e.Get(ctx, key)\n\t\tif cli.IsKeyNotFound(err) {\n\t\t\terr = e.Set(ctx, key, \"20000\")\n\t\t\treturn 20000, err\n\t\t} else if err != nil {\n\t\t\treturn -1, err\n\t\t}\n\n\t\tival, err := strconv.ParseInt(val, 10, 64)\n\t\tif err != nil {\n\t\t\treturn -1, err\n\t\t}\n\n\t\terr = e.Cas(ctx, key, val, fmt.Sprintf(\"%d\", ival+1))\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\treturn ival + 1, nil\n\t}\n}", "func intSeq() func() int {\n\ti := 0\n\treturn func() int { // anonymous function\n\t\ti += 1\n\t\treturn i\n\t}\n}", "func (c *TRandomLong) Sequence(min int64, max int64) []int64 {\n\tif min < 0 {\n\t\tmin = 0\n\t}\n\tif max < min {\n\t\tmax = min\n\t}\n\n\tcount := c.NextLong(min, max)\n\n\tresult := make([]int64, count, count)\n\tfor i := range result {\n\t\tresult[i] = int64(i)\n\t}\n\n\treturn result\n}", "func (buf *ListBuffer) Incr(idx BufferIndex) BufferIndex {\n\tif idx == NilIndex {\n\t\treturn NilIndex\n\t} \n\treturn buf.Buffer[idx].Next\n}", "func Append(seqs ...Seq) Seq {\n\tnextseq := make([]Seq, 0)\n\tif len(seqs) == 0 {\n\t\treturn nil\n\t} else if len(seqs) == 1 {\n\t\tif seqs[0] != nil {\n\t\t\tif seqs[0].Rest() != nil {\n\t\t\t\tnextseq = append(nextseq, seqs[0].Rest())\n\t\t\t\treturn cons(seqs[0].First(), Append(nextseq...))\n\t\t\t}\n\t\t\treturn cons(seqs[0].First(), nil)\n\t\t}\n\t\treturn nil\n\t}\n\tif seqs[0] == nil {\n\t\treturn Append(seqs[1:]...)\n\t}\n\tif seqs[0].Rest() != nil {\n\t\tnextseq = append(nextseq, Append(seqs[0].Rest()))\n\t\tnextseq = append(nextseq, seqs[1:]...)\n\t\treturn cons(seqs[0].First(), Append(nextseq...))\n\t}\n\tnextseq = append(nextseq, seqs[1:]...)\n\treturn cons(seqs[0].First(), Append(nextseq...))\n}", "func (session *Session) Incr(column string, arg ...interface{}) *Session {\n\tsession.Statement.Incr(column, arg...)\n\treturn session\n}", "func Seq(rows int, cols int) (matrix Matrix) {\n\tmatrix = NewMatrix(rows, cols)\n\n\t// looks awkward, but it's the most performant way\n\tseq := 1.0\n\tfor _, row := range matrix {\n\t\tfor i := range row {\n\t\t\trow[i] = seq\n\t\t\tseq += 1.0\n\t\t}\n\t}\n\n\treturn matrix\n}", "func Seq(l List) {\n\tfor !IsEmpty(l) {\n\t\tHead(l)\n\t\tl = Tail(l)\n\t}\n}", "func increaseSeq(u *uint32) {\n\tif *u == math.MaxUint32 {\n\t\t*u = 1\n\t} else {\n\t\t*u++\n\t}\n}", "func (statement *Statement) Incr(column string, arg ...interface{}) *Statement {\n\tif len(arg) > 0 {\n\t\tstatement.IncrColumns.Add(column, arg[0])\n\t} else {\n\t\tstatement.IncrColumns.Add(column, 1)\n\t}\n\treturn statement\n}", "func (d *DB) addSeq(delta uint64) {\n\tatomic.AddUint64(&d.seq, delta)\n}", "func (k *Kvs) Incr(key string) (string, error) {\n\tval := k.kv[key]\n\tvalInt, err := strconv.Atoi(val)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tk.kv[key] = convStr(valInt + 1)\n\treturn k.kv[key], nil\n}", "func (u *Int32) Incr() int32 {\n\treturn u.Add(1)\n}", "func (r *RateCounter) Incr(val int64) {\n\tr.counter.Incr(val)\n\tgo r.scheduleDecrement(val)\n}", "func Seq(args ...interface{}) *Pattern {\n\treturn Seq2(args)\n}", "func (c *Client) Incr(stat string) error {\n\treturn c.Increment(stat, 1, 1)\n}", "func (session *Session) Incr(column string, arg ...interface{}) *Session {\n\tsession.Session = session.Session.Incr(column, arg...)\n\treturn session\n}", "func (ss SliceType) Sequence(params ...int) SliceType {\n\tvar creator = func(i int) ElementType {\n\t\treturn ElementType(i)\n\t}\n\n\treturn ss.SequenceUsing(creator, params...)\n}", "func (t *NFSv42Client) SequenceArgs() NfsArgop4 {\n\treturn Sequence(t.Sid, t.Seq, 0, 0, false)\n}", "func (store *sqlStore) IncrNextTargetMsgSeqNum() error {\n\tstore.cache.IncrNextTargetMsgSeqNum()\n\treturn store.SetNextTargetMsgSeqNum(store.cache.NextTargetMsgSeqNum())\n}", "func (s *SliceOfInt) Concat(items []int) *SliceOfInt {\n\ts.items = append(s.items, items...)\n\treturn s\n}", "func (c *Client) Incr(name string) error {\n\treturn c.Increment(name, 1, 1)\n}", "func inc(n int) int {\n\treturn n + 1\n}", "func inc(n int) int {\n\tn++\n\tif Repeat > 0 {\n\t\tn %= Repeat\n\t}\n\treturn n\n}", "func nextSequence() func() int {\n\tcount := 0\n\treturn func() int {\n\t\tcount = count + 1\n\t\treturn count\n\t}\n}", "func (p *PromCounter) Incr(count int64) error {\n\tp.ctr.Add(float64(count))\n\treturn nil\n}", "func NewSequence(n uint64) *Sequence {\n\treturn &Sequence{n: n}\n}", "func (id ID) Seq() uint64 {\n\t_, seq := id.genseq()\n\treturn seq\n}", "func NewRptSeq(val int) RptSeqField {\n\treturn RptSeqField{quickfix.FIXInt(val)}\n}", "func (c *CountString) Incr() *CountString {\n\tc.str = \"\"\n\tvar ok bool\n\tfor i, j := range c.indexes {\n\t\tj++\n\t\tif j < len(countStringSet) {\n\t\t\tc.indexes[i] = j\n\t\t\tok = true\n\t\t\tbreak\n\t\t}\n\t\tc.indexes[i] = 0\n\t}\n\tif !ok {\n\t\tif len(c.indexes) < cap(c.indexes) {\n\t\t\tc.indexes = append(c.indexes, 0)\n\t\t} else {\n\t\t\t// to zero\n\t\t\tc.indexes = make([]int, 1, cap(c.indexes))\n\t\t}\n\t}\n\treturn c\n}", "func TestPrependSeq(t *T) {\n\t//Normal case\n\tintl1 := []interface{}{3, 4}\n\tintl2 := []interface{}{0, 1, 2}\n\tl1 := NewList(intl1...)\n\tl2 := NewList(intl2...)\n\tnl := l1.PrependSeq(l2)\n\tassertSaneList(l1, t)\n\tassertSaneList(l2, t)\n\tassertSaneList(nl, t)\n\tassertSeqContents(l1, intl1, t)\n\tassertSeqContents(l2, intl2, t)\n\tassertSeqContents(nl, []interface{}{0, 1, 2, 3, 4}, t)\n\n\t// Degenerate cases\n\tblank1 := NewList()\n\tblank2 := NewList()\n\tnl = blank1.PrependSeq(blank2)\n\tassertEmpty(blank1, t)\n\tassertEmpty(blank2, t)\n\tassertEmpty(nl, t)\n\n\tnl = blank1.PrependSeq(l1)\n\tassertEmpty(blank1, t)\n\tassertSaneList(nl, t)\n\tassertSeqContents(nl, intl1, t)\n\n\tnl = l1.PrependSeq(blank1)\n\tassertEmpty(blank1, t)\n\tassertSaneList(nl, t)\n\tassertSeqContents(nl, intl1, t)\n}", "func SeqSum(x *big.Int) (res *big.Int) {\n\tif x.Cmp(B0) == 0 {\n\t\treturn B0\n\t}\n\tlx := big.NewInt(int64(math.Log2(float64(x.Int64()))))\n\tif lx.Mod(lx, B2).Cmp(B0) == 0 {\n\t\tres = seqSumEven(x)\n\t} else {\n\t\tres = seqSumOdd(x)\n\t}\n\n\treturn res\n}", "func (g *Generator) Next() ([]int, error) {\n\t// Check more outcomes can be generated.\n\tif g.current >= g.total {\n\t\treturn nil, errors.New(\"out of range\")\n\t}\n\n\t// Build current outcome. Since we update sequence numbers from left, outcome\n\t// generated from rifht, so build result starting from end of slice.\n\tres := []int{}\n\tfor i := len(g.position) - 1; i >= 0; i-- {\n\t\tres = append(res, g.elements[g.position[i]])\n\t}\n\n\t// Update position numbers and increase current counter.\n\tg.updateCounter()\n\n\treturn res, nil\n}", "func Sequence(seq int) option {\n\treturn func(w *Writer) {\n\t\tw.seq = seq\n\t}\n}", "func (s *Streams) Seq() *immute.Sequence {\n\treturn s.Buffer.Seq()\n}", "func (s *dbStore) NextCmdSeq() (int, error) {\n\tvar seq uint64\n\terr := s.db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(bucketCmd))\n\t\tseq = b.Sequence() + 1\n\t\treturn nil\n\t})\n\treturn int(seq), err\n}", "func TestPrependSeq(t *T) {\n\t//Normal case\n\tintl1 := []interface{}{3, 4}\n\tintl2 := []interface{}{0, 1, 2}\n\tl1 := NewList(intl1...)\n\tl2 := NewList(intl2...)\n\tnl := l1.PrependSeq(l2)\n\tassertSaneList(l1, t)\n\tassertSaneList(l2, t)\n\tassertSaneList(nl, t)\n\tassert.Equal(t, intl1, ToSlice(l1))\n\tassert.Equal(t, intl2, ToSlice(l2))\n\tassert.Equal(t, []interface{}{0, 1, 2, 3, 4}, ToSlice(nl))\n\n\t// Degenerate cases\n\tblank1 := NewList()\n\tblank2 := NewList()\n\tnl = blank1.PrependSeq(blank2)\n\tassert.Equal(t, 0, Size(blank1))\n\tassert.Equal(t, 0, Size(blank2))\n\tassert.Equal(t, 0, Size(nl))\n\n\tnl = blank1.PrependSeq(l1)\n\tassert.Equal(t, 0, Size(blank1))\n\tassertSaneList(nl, t)\n\tassert.Equal(t, intl1, ToSlice(nl))\n\n\tnl = l1.PrependSeq(blank1)\n\tassert.Equal(t, 0, Size(blank1))\n\tassertSaneList(nl, t)\n\tassert.Equal(t, intl1, ToSlice(nl))\n}", "func (a *AtomicInt64) Incr() int64 {\n\treturn atomic.AddInt64((*int64)(a), 1)\n}", "func (p *PromGauge) Incr(count int64) error {\n\tp.ctr.Add(float64(count))\n\treturn nil\n}", "func AllocateSeq(prgrm *CXProgram, size int) (offset int) {\n\tresult := prgrm.Heap.HeapPointer\n\tnewFree := result + size\n\n\tif newFree > INIT_HEAP_SIZE {\n\t\t// call GC\n\t\tMarkAndCompact(prgrm)\n\t\tresult = prgrm.Heap.HeapPointer\n\t\tnewFree = prgrm.Heap.HeapPointer + size\n\n\t\tif newFree > INIT_HEAP_SIZE {\n\t\t\t// heap exhausted\n\t\t\tpanic(\"heap exhausted\")\n\t\t}\n\t}\n\n\tprgrm.Heap.HeapPointer = newFree\n\n\treturn result\n}", "func (s *dbStore) CmdsWithSeq(from, upto int) ([]Cmd, error) {\n\tvar cmds []Cmd\n\terr := s.IterateCmds(from, upto, func(cmd Cmd) {\n\t\tcmds = append(cmds, cmd)\n\t})\n\treturn cmds, err\n}", "func (l *Lifecycle) NextSequence() error {\n\tcommand := []string{\n\t\t\"peer lifecycle chaincode querycommitted\",\n\t\tfmt.Sprintf(\"--channelID %v\", l.Channel),\n\t\tfmt.Sprintf(\"--name %v\", l.Chaincode),\n\t\tfmt.Sprintf(\"-o %v\", os.Getenv(\"ORDERER_ADDRESS\")),\n\t\tfmt.Sprintf(\"--tls %v\", \"true\"),\n\t\tfmt.Sprintf(\"--cafile %v\", os.Getenv(\"ORDERER_CA\")),\n\t\tfmt.Sprintf(\"--peerAddresses %v\", os.Getenv(\"CORE_PEER_ADDRESS\")),\n\t\tfmt.Sprintf(\"--tlsRootCertFiles %v\", os.Getenv(\"CORE_PEER_TLS_ROOTCERT_FILE\")),\n\t\t\"-O json\",\n\t}\n\n\tresponse, err := l.execute(strings.Join(command, \" \"))\n\tif err == nil {\n\t\tvar committed QueryCommitted\n\t\terr = json.Unmarshal(response.Output.Bytes(), &committed)\n\t\tl.Sequence = committed.Sequence + 1\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Client) IncrBy(stat string, n int) error {\n\treturn c.Increment(stat, n, 1)\n}", "func (s Broker) Incr(name string) {\n\ts.Count(name, 1)\n}", "func (s *Stats) Incr(stat string, count int64, tags ...Tag) {\n\tif 0 != count {\n\t\ts.trans.bufLock.Lock()\n\t\tlastLen := len(s.trans.buf)\n\n\t\ts.trans.buf = append(s.trans.buf, []byte(s.metricPrefix)...)\n\t\ts.trans.buf = append(s.trans.buf, []byte(stat)...)\n\t\tif s.trans.tagFormat.Placement == TagPlacementName {\n\t\t\ts.trans.buf = s.formatTags(s.trans.buf, tags)\n\t\t}\n\t\ts.trans.buf = append(s.trans.buf, ':')\n\t\ts.trans.buf = strconv.AppendInt(s.trans.buf, count, 10)\n\t\ts.trans.buf = append(s.trans.buf, []byte(\"|c\")...)\n\t\tif s.trans.tagFormat.Placement == TagPlacementSuffix {\n\t\t\ts.trans.buf = s.formatTags(s.trans.buf, tags)\n\t\t}\n\t\ts.trans.buf = append(s.trans.buf, '\\n')\n\n\t\ts.trans.checkBuf(lastLen)\n\t\ts.trans.bufLock.Unlock()\n\t}\n}", "func (t *Table) Increments(colNm string) *Table {\n\tt.columns = append(t.columns, &column{Name: colNm, ColumnType: TypeSerial, IsPrimaryKey: true})\n\treturn t\n}", "func Incr(stat string, count int64, tagsInput map[string]string) {\n\tif on {\n\t\tstdClient.Incr(stat, count, convertTags(tagsInput)...)\n\t}\n}", "func (i *Meta) Sequence() int {\n\tupper := (i.Step / 1000)\n\tlower := (i.Step % 100)\n\treturn upper + lower\n}", "func (l ArgLocation) Incr(c rune, tabstop int) Location {\n\t// Begin by advancing the beginning\n\tl.B = l.E\n\n\t// Advance the ending column\n\tif c != EOF {\n\t\tl.E.C++\n\t}\n\n\treturn l\n}", "func PrependItems(slice []int, values ...int) []int {\n\treturn append(values, slice...)\n}", "func (io *Io) NextInts(n int) []int {\n\tres := make([]int, n)\n\tfor i := 0; i < n; i++ {\n\t\tres[i] = io.NextInt()\n\t}\n\treturn res\n}", "func (k *Kvs) IncrBy(key string, val int) (string, error) {\n\tv := k.kv[key]\n\tvalInt, err := strconv.Atoi(v)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tk.kv[key] = convStr(valInt + val)\n\treturn k.kv[key], err\n}", "func NewSequenced(left, right Expr) *Sequenced {\n\treturn &Sequenced{\n\t\tLeft: left,\n\t\tRight: right,\n\t}\n}", "func NewSeqImpl(conf Conf, initializer core.Initializer, data []core.Elemt, distrib Distrib) Impl {\n\treturn Impl{\n\t\tbuffer: core.NewDataBuffer(data, conf.FrameSize),\n\t\tinitializer: initializer,\n\t\tuniform: distuv.Uniform{Max: 1, Min: 0, Src: conf.RGen},\n\t\tstore: NewCenterStore(conf.RGen),\n\t\tstrategy: &SeqStrategy{},\n\t\tdistrib: distrib,\n\t}\n}", "func Incr(key string) (int64, error) {\n\treturn getClient().Incr(key).Result()\n}", "func (io *Io) NextInts(n int) []int {\n\tret := make([]int, n)\n\tfor i := 0; i < n; i++ {\n\t\tret[i] = io.NextInt()\n\t}\n\treturn ret\n}", "func Seq(x interface{}, args ...interface{}) Node {\n\treturn Node{&ast.Set{\n\t\tStartOp: \"[\",\n\t\tEndOp: \"]\",\n\t\tX: ToNode(x).Node,\n\t\tY: ArgsList(args...).Node,\n\t}}\n}", "func (client *MemcachedClient4T) Incr(key string, value uint64) (uint64, error) {\n\treturn client.parse.IncrOrDecr(\"incr\", key, value)\n}", "func IntRange(args ...int) <-chan int {\n\tvar start, stop, step int\n\tswitch len(args) {\n\tcase 1: // 1 argument: stop\n\t\tstart = 0\n\t\tstop = args[0]\n\t\tstep = 1\n\tcase 2: // 2 arguments: start, stop\n\t\tstart = args[0]\n\t\tstop = args[1]\n\t\tstep = 1\n\tcase 3: // 3 arguments: start, stop, step\n\t\tstart = args[0]\n\t\tstop = args[1]\n\t\tstep = args[2]\n\tdefault: // invalid argument count\n\t\tpanic(\"IntRange takes 1 to 3 arguments.\")\n\t}\n\tch := make(chan int)\n\tif step >= 0 {\n\t\t// increment case\n\t\tgo func() {\n\t\t\tfor i := start; i < stop; i += step {\n\t\t\t\tch <- i\n\t\t\t}\n\t\t\tclose(ch)\n\t\t}()\n\t} else {\n\t\t// decrement case\n\t\tgo func() {\n\t\t\tfor i := start; i > stop; i += step {\n\t\t\t\tch <- i\n\t\t\t}\n\t\t\tclose(ch)\n\t\t}()\n\t}\n\treturn ch\n}", "func incFn(args []reflect.Value) []reflect.Value {\n\treturn []reflect.Value{reflect.ValueOf(args[0].Interface().(int) + 1)}\n}", "func (s *String) Incr(delta int64) (int64, error) {\n\tvalue := s.Meta.Value\n\tif value != nil {\n\t\tv, err := strconv.ParseInt(string(value), 10, 64)\n\t\tif err != nil {\n\t\t\treturn 0, ErrInteger\n\t\t}\n\t\tdelta = v + delta\n\t}\n\n\tvs := strconv.FormatInt(delta, 10)\n\tif err := s.Set([]byte(vs)); err != nil {\n\t\treturn 0, err\n\t}\n\treturn delta, nil\n\n}", "func NewSequence(gid int, elements ...Element) *Sequence {\n\treturn &Sequence{\n\t\telement: element{gid},\n\t\telements: elements,\n\t}\n}", "func (r *RBucket) NextSequence() (uint64, error) {\n\tvar resp uint64\n\treq := &NextSequenceRequest{}\n\treq.BucketID = r.id\n\treq.ContextID = r.parent\n\n\terr := r.r.call(\"srv.NextSequence\", req, &resp)\n\treturn resp, err\n}", "func GenerateCSeq() int {\n\treturn rand.Int() % 50000\n}", "func InSequence(ch chan Result, work []int) []Result {\n\tresults := make([]Result, len(work))\n\tgo buildInSeq(ch, work, results)\n\treturn results\n}", "func (s *SliceInt) Prepend(elems ...int) *SliceInt {\n\tif s == nil {\n\t\treturn nil\n\t}\n\ts.data = append(elems, s.data...)\n\treturn s\n}", "func (ctr *Ctr) Incr() uint32 { return atomic.AddUint32((*uint32)(unsafe.Pointer(ctr)), 1) }", "func IncrDecr(s storage.Storage, req request.Request) response.Response {\n\tgetBody := ExtractCalcBody(req)\n\n\tvalue, err := s.Get([]byte(string(getBody.Body)))\n\n\tif err == storage.ErrKeyNotFound {\n\t\tinitialValue := make([]byte, 8)\n\t\tbinary.BigEndian.PutUint64(initialValue, getBody.InitialValue)\n\n\t\t_, err := s.Set(\n\t\t\tstorage.Record{\n\t\t\t\tKey: string(getBody.Body),\n\t\t\t\tValue: initialValue,\n\t\t\t\tExp: getBody.Expiration,\n\t\t\t\tCAS: req.Cas,\n\t\t\t},\n\t\t)\n\n\t\tif err != nil {\n\t\t\treturn *response.BuildResponse(req, req.Opcode, packet.StatusInternalError, []byte{}, []byte{})\n\t\t}\n\n\t\treturn *response.BuildResponse(req, req.Opcode, packet.StatusNoError, []byte{}, initialValue)\n\t}\n\n\tif err != nil {\n\t\treturn *response.BuildResponse(req, req.Opcode, packet.StatusInternalError, []byte{}, []byte{})\n\t}\n\n\texistNum, err := strconv.Atoi(string(value.Value))\n\tif err != nil {\n\t\treturn *response.BuildResponse(req, req.Opcode, packet.StatusIncrDecrOnNumericValue, []byte{}, []byte(\"Value: Non-numeric server-side value for incr or decr\"))\n\n\t}\n\n\tif ((req.Opcode == packet.CmdDecrement || req.Opcode == packet.CmdDecrementQ) && (existNum-1 < 0)) || ((req.Opcode == packet.CmdIncrement || req.Opcode == packet.CmdIncrementQ) && (4294967295 < existNum+1)) {\n\t\treturn *response.BuildResponse(req, req.Opcode, packet.StatusIncrDecrOnNumericValue, []byte{}, []byte(\"Value: Non-numeric server-side value for incr or decr\"))\n\t}\n\n\tif util.Debug > 1 {\n\t\tfmt.Fprintf(os.Stderr, \"incr/decr: Value Size: %d, Prev: %d, Incr=%d, Decr=%d\\n\", len(value.Value), existNum, existNum+1, existNum-1)\n\t}\n\n\tresp := make([]byte, 8)\n\tvar v []byte\n\tif req.Opcode == packet.CmdIncrement || req.Opcode == packet.CmdIncrementQ {\n\t\tv = []byte(fmt.Sprint(uint32(existNum + 1)))\n\t\tbinary.BigEndian.PutUint64(resp, uint64(existNum+1))\n\t}\n\n\tif req.Opcode == packet.CmdDecrement || req.Opcode == packet.CmdDecrementQ {\n\t\tv = []byte(fmt.Sprint(uint32(existNum - 1)))\n\t\tbinary.BigEndian.PutUint64(resp, uint64(existNum-1))\n\t}\n\n\tvalue.Value = v\n\n\t_, err = s.Set(*value)\n\n\tif err != nil {\n\t\treturn *response.BuildResponse(req, req.Opcode, packet.StatusInternalError, []byte{}, []byte{})\n\t}\n\n\treturn *response.BuildResponse(req, req.Opcode, packet.StatusNoError, []byte{}, resp)\n\n}", "func newSequenceCursor(parent *sequenceCursor, seq sequence, idx int) *sequenceCursor {\n\td.PanicIfTrue(seq == nil)\n\tseqLen := seq.seqLen()\n\tif idx < 0 {\n\t\tidx += seqLen\n\t\td.PanicIfFalse(idx >= 0)\n\t}\n\n\treturn &sequenceCursor{parent, seq, idx, seqLen}\n}", "func (s *SliceOfInt64) Concat(items []int64) *SliceOfInt64 {\n\ts.items = append(s.items, items...)\n\treturn s\n}", "func NewSeq(s []byte) Seq {\n\treturn Seq{\n\t\tLength: len(s),\n\t\tSeq: contract(s),\n\t}\n}", "func (c *Client) IncrBy(name string, n int) error {\n\treturn c.Increment(name, n, 1)\n}", "func (be InmemBackend) Incr(key []byte, value uint) (int, error) {\n\treturn be.Increment(key, int(value), false)\n}", "func DecadePrimes(start int64) IntStream {\n\tif start < 1 {\n\t\tstart = 1\n\t}\n\tif start > math.MaxInt64/10 {\n\t\tstart = math.MaxInt64 / 10\n\t}\n\treturn &decadeStream{primes: Primes(10*start + 1)}\n}", "func (b *Bucket) NextSequence() (uint64, error) {\n seq, err := b.bucket.NextSequence()\n return seq, err\n}", "func (v *VersionVector) Incr(actor string) LamportTime {\n\tv.l.Lock()\n\n\tt, exists := v.dots[actor]\n\tif !exists {\n\t\tt = LamportTime(1)\n\t\tv.dots[actor] = t\n\t} else {\n\t\tt++\n\t\tv.dots[actor] = t\n\t}\n\n\tv.l.Unlock()\n\treturn t\n}", "func CycleInt(s []int) <-chan int {\n\tch := make(chan int, 1)\n\tgo func() {\n\t\tfor i := 0;; i++ {\n\t\t\tch <- s[i%len(s)]\n\t\t}\n\t}()\n\treturn ch\n}", "func (ub *UpdateBuilder) Incr(field string) string {\n\tf := Escape(field)\n\treturn fmt.Sprintf(\"%s = %s + 1\", f, f)\n}", "func rsI(n int) []int {\n\tt := make([]int, n)\n\tfor i := 0; i < n; i++ {\n\t\tt[i] = rnI()\n\t}\n\treturn t\n}", "func WrapWithSeq(v interface{}, seq Seq) SeqWrapper {\n\treturn &seqWrapper{\n\t\tseq: seq,\n\t\tv: v,\n\t}\n}", "func NewSequence(width int, numPeriods int) Sequence {\n\treturn make(Sequence, Width64bits+numPeriods*width)\n}", "func (c *Counter) Incr() {\n\tc.mu.Lock()\n\tc.count++\n\tc.mu.Unlock()\n\tfmt.Println(\"Write counter:\", c.count)\n}", "func makeIncremented(x []float64, inc int, extra int) []float64 {\n\tif inc == 0 {\n\t\tpanic(\"zero inc\")\n\t}\n\tabsinc := inc\n\tif absinc < 0 {\n\t\tabsinc = -inc\n\t}\n\txcopy := make([]float64, len(x))\n\tif inc > 0 {\n\t\tcopy(xcopy, x)\n\t} else {\n\t\tfor i := 0; i < len(x); i++ {\n\t\t\txcopy[i] = x[len(x)-i-1]\n\t\t}\n\t}\n\n\t// don't use NaN because it makes comparison hard\n\t// Do use a weird unique value for easier debugging\n\tcounter := 100.0\n\tvar xnew []float64\n\tfor i, v := range xcopy {\n\t\txnew = append(xnew, v)\n\t\tif i != len(x)-1 {\n\t\t\tfor j := 0; j < absinc-1; j++ {\n\t\t\t\txnew = append(xnew, counter)\n\t\t\t\tcounter++\n\t\t\t}\n\t\t}\n\t}\n\tfor i := 0; i < extra; i++ {\n\t\txnew = append(xnew, counter)\n\t\tcounter++\n\t}\n\treturn xnew\n}", "func Benchmark_SliceIntIncr4(b *testing.B) {\n\tb.ResetTimer()\n\tfor i := 0; i <= b.N; i++ {\n\t\tn := [10]int{}\n\t\tfor i := range n {\n\t\t\tv := &n[i]\n\t\t\t*v++\n\t\t}\n\t\t// 寫入\n\t}\n}", "func DoSequence(step func(it int) error, count int) error {\n\tfor i := 0; i < count; i++ {\n\t\terr := step(i)\n\t\t// break the loop as soon as we get an error\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (q *queue) AppendedSeq() int64 {\n\tq.rwMutex.RLock()\n\tdefer q.rwMutex.RUnlock()\n\n\treturn q.appendedSeq.Load()\n}", "func (c *Counter) IncrBy(nlines, nbytes int, active float64) {\n\tc.cur += nlines\n\tc.curBytes += nbytes\n\tc.curActiveSecs += active\n\tc.pRecords.Add(float64(nlines))\n\tc.pBytes.Add(float64(nbytes))\n\tc.pActiveSecs.Add(active)\n}", "func Reverse(seq Seq) Seq {\n\treturn RevAppend(seq, nil)\n}", "func (c *LoggerClient) Incr(name string) {\n\tc.Count(name, 1)\n}", "func resetSeq(tasks *[]db.Task) error {\n\ttx := db.Conn.Begin()\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\ttx.Rollback()\n\t\t}\n\t}()\n\tif err := tx.Error; err != nil {\n\t\treturn err\n\t}\n\n\tif err := tx.Model(&db.Task{}).Where(\"1 = 1\").Update(\"seq\", nil).Error; err != nil {\n\t\treturn err\n\t}\n\n\tfor i := 0; i < len(*tasks); i++ {\n\t\tthisTask := &(*tasks)[i]\n\t\tthisTask.Seq = i + 1\n\t\tif err := tx.Save(&thisTask).Error; err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn tx.Commit().Error\n}", "func Benchmark_SliceIntIncr3(b *testing.B) {\n\tb.ResetTimer()\n\tfor i := 0; i <= b.N; i++ {\n\t\tn := [10]int{}\n\t\tfor i := range n {\n\t\t\tv := n[i]\n\t\t\tv++\n\t\t}\n\t}\n}" ]
[ "0.58424926", "0.58424926", "0.5653298", "0.56301075", "0.5609903", "0.5451971", "0.5306408", "0.52724165", "0.52414966", "0.52021176", "0.5157603", "0.5133184", "0.5119144", "0.51074183", "0.50727046", "0.50644374", "0.50640833", "0.5040232", "0.49489027", "0.49045926", "0.4893106", "0.48904562", "0.4883187", "0.48734292", "0.48604834", "0.48420942", "0.48009074", "0.47958338", "0.4791672", "0.47399107", "0.47354066", "0.47119978", "0.46802902", "0.46800792", "0.46692938", "0.465096", "0.4649049", "0.46426177", "0.46368688", "0.46224687", "0.46183178", "0.46107668", "0.45855054", "0.45799214", "0.45712802", "0.45591778", "0.45539254", "0.4540617", "0.4533506", "0.4529399", "0.45243502", "0.45140415", "0.45113015", "0.45098212", "0.4494518", "0.44889835", "0.44868043", "0.44846386", "0.44777033", "0.44776022", "0.4472501", "0.44624674", "0.4457572", "0.4440359", "0.4434848", "0.44241965", "0.442248", "0.44144708", "0.44118828", "0.4407354", "0.44070864", "0.440494", "0.44017866", "0.43962306", "0.43748868", "0.43704233", "0.43694124", "0.43623507", "0.4354872", "0.43335074", "0.43266505", "0.4326292", "0.4317777", "0.43014383", "0.4295906", "0.42838338", "0.42823625", "0.42822227", "0.42714906", "0.42685944", "0.4264087", "0.42526013", "0.4250733", "0.42374638", "0.42277095", "0.42064694", "0.42053923", "0.4200148", "0.41991898", "0.41980553" ]
0.7631241
0
Shuffle returns a copy of operand with the elements shuffled pseudorandomly.
func Shuffle(operand []string) []string { shuffled := make([]string, len(operand)) for i, p := range pseudo.Perm(len(operand)) { shuffled[i] = operand[p] } return shuffled }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (a Slice[T]) Shuffle() Slice[T] {\n\tfor i := len(a) - 1; i > 0; i-- {\n\t\tj := rand.Intn(i + 1)\n\t\ta[i], a[j] = a[j], a[i]\n\t}\n\treturn a\n}", "func (indis Individuals) shuffle(generator *rand.Rand) Individuals {\n\tvar shuffled = make(Individuals, len(indis))\n\tfor i, v := range generator.Perm(len(indis)) {\n\t\tshuffled[v] = indis[i]\n\t}\n\treturn shuffled\n}", "func (this *Solution) Shuffle() []int {\n\t// 随机交换元素的位置\n\ttmp := make([]int, len(this.a))\n\tcopy(tmp, this.a)\n\tfor i := 0; i < len(tmp); i++ {\n\t\tr := rand.Intn(len(tmp))\n\t\ttmp[i], tmp[r] = tmp[r], tmp[i]\n\t}\n\treturn tmp\n}", "func Shuffle(slice interface{}) {\n\trv := reflect.ValueOf(slice)\n\tswap := reflect.Swapper(slice)\n\tlength := rv.Len()\n\tfor i := length - 1; i > 0; i-- {\n\t\tj := rand.Intn(i + 1)\n\t\tswap(i, j)\n\t}\n}", "func (d *Dataset) Shuffle() (output Dataset) {\n\tdata := *(*d).Data\n\tfor i := range data {\n\t\tj := rand.Intn(i + 1)\n\t\tdata[i], data[j] = data[j], data[i]\n\t}\n\toutput.Data = &data\n\toutput.Size = (*d).Size\n\treturn\n}", "func (this *Solution) Shuffle() []int {\r\n n := len(this.raw)\r\n words := make([]int, n)\r\n copy(words, this.raw)\r\n for i:=n; i>1; i-- {\r\n seed := rand.Intn(i)\r\n words[i-1], words[seed] = words[seed], words[i-1]\r\n }\r\n return words\r\n}", "func Shuffle[T any](collection []T) []T {\n\trand.Shuffle(len(collection), func(i, j int) {\n\t\tcollection[i], collection[j] = collection[j], collection[i]\n\t})\n\n\treturn collection\n}", "func (s *Shuffle) Shuffle() []int {\n\tori := make([]int, len(s.ori))\n\tcopy(ori, s.ori)\n\trd := make([]int, len(ori))\n\tfor i := len(rd) - 1; i >= 0; i-- {\n\t\tpos := rand.Intn(i + 1)\n\t\trd[i] = ori[pos]\n\t\tori[pos], ori[i] = ori[i], ori[pos]\n\t}\n\treturn rd\n}", "func (this *Solution) Shuffle() []int {\n\tresult := make([]int, this.size)\n\tcopy(result, this.nums)\n\tfor i := this.size - 1; i >= 0; i-- {\n\t\tj := rand.Intn(this.size)\n\t\tresult[i], result[j] = result[j], result[i]\n\t}\n\treturn result\n}", "func (d Deck) Shuffle() {\n\tsource := rand.NewSource(time.Now().UnixNano())\n\tr := rand.New(source)\n\n\tfor i := range d {\n\t\tnewPos := r.Intn(len(d) - 1)\n\n\t\td[i], d[newPos] = d[newPos], d[i]\n\t}\n}", "func Shuffle(n int, swap func(i, j int)) { globalRand.Shuffle(n, swap) }", "func Shuffle(n int, swap func(i, j int)) { globalRand.Shuffle(n, swap) }", "func (d *Dense) Shuffle(r *rand.Rand) {\n\tr.Shuffle(d.len, d.swap)\n}", "func (this *Solution) Shuffle() []int {\n\tvar randIndex int\n\tlength := len(this.nums)\n\n\tfor i := 0; i < length; i++ {\n\t\trandIndex = rand.Intn(length)\n\t\tthis.nums[i], this.nums[randIndex] = this.nums[randIndex], this.nums[i]\n\t}\n\n\treturn this.nums\n}", "func (this *Solution) Shuffle() []int {\n\n\tl := len(this.current)\n\ttmp := make([]int, l)\n\tcopy(tmp, this.current)\n\tfor i:=range this.current{\n\t\tindex := i+rand.Int()%(l-i)\n\t\ttmp[i], tmp[index] = tmp[index], tmp[i]\n\t}\n\treturn tmp\n}", "func (this *Solution) Shuffle() []int {\ntmp:=make([]int,len(this.arr))\ncopy(tmp,this.arr)\nfor i:=0;i<len(tmp);i++{\n\tr:=rand.Intn(len(tmp))\n\ttmp[i],tmp[r] = tmp[r],tmp[i]\n}\nreturn tmp\n}", "func Shuffle(values []float64) {\n\tvar tmp float64\n\tvar j int\n\tfor i := len(values) - 1; i > 0; i-- {\n\t\tj = rand.Int() % i\n\t\ttmp = values[j]\n\t\tvalues[j] = values[i]\n\t\tvalues[i] = tmp\n\t}\n}", "func (this *Solution) Shuffle() []int {\n\tlength := len(this.cp)\n\tfor i := int32(length - 1); i >= 0; i-- {\n\t\tidx := rand.Int31n(i + 1)\n\t\tthis.cp[i], this.cp[idx] = this.cp[idx], this.cp[i]\n\t}\n\treturn this.cp\n}", "func RandomShuffle(scope *Scope, value tf.Output, optional ...RandomShuffleAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"RandomShuffle\",\n\t\tInput: []tf.Input{\n\t\t\tvalue,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (this *Solution) Shuffle() []int {\n\tnums := make([]int, len(this.nums))\n\tcopy(nums, this.nums)\n\trand.Shuffle(len(nums), func(i int, j int) {\n\t\tnums[i], nums[j] = nums[j], nums[i]\n\t})\n\treturn nums\n}", "func (this *SolutionShuffle) Shuffle() []int {\n\tres := make([]int, len(this.origin))\n\ttemp := append([]int{}, this.origin...)\n\tfor i := len(this.origin); i > 0; i-- {\n\t\t// gen random index\n\t\tidex := rand.Intn(i)\n\t\t// swap that index number with last index number\n\t\ttemp[len(temp)-1], temp[idex] = temp[idex], temp[len(temp)-1]\n\t\t// pop that number and add to array\n\t\tres[i-1] = temp[len(temp)-1]\n\t\ttemp = temp[:len(temp)-1]\n\t}\n\treturn res\n}", "func Shuffle(n int, swap func(i, j int)) {\n\tglobalSource.Shuffle(n, swap)\n}", "func StatelessShuffle(scope *Scope, value tf.Output, key tf.Output, counter tf.Output, alg tf.Output) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"StatelessShuffle\",\n\t\tInput: []tf.Input{\n\t\t\tvalue, key, counter, alg,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (list *List) Shuffle() {\n\tfor i := range list.data {\n\t\tj := rand.Intn(i + 1)\n\t\tlist.data[i], list.data[j] = list.data[j], list.data[i]\n\t}\n}", "func (f *Facts) Shuffle() []string {\n\tt := time.Now()\n\trand.Seed(int64(t.Nanosecond()))\n\n\tarr := f.behaviorsAccess\n\tfor i := len(arr) - 1; i > 0; i-- {\n\t\tj := rand.Intn(i)\n\t\tarr[i], arr[j] = arr[j], arr[i]\n\t}\n\treturn arr\n}", "func (d deck) shuffle() {\n\tsource := rand.NewSource(time.Now().UnixNano())\n\tr := rand.New(source)\n\t// now generate rnd num and swap positions\n\tfor i := range d {\n\t\tnewPosition := r.Intn(len(d) - 1)\n\t\td[i], d[newPosition] = d[newPosition], d[i] // swap postions in slice\n\t}\n}", "func (this *Solution) Shuffle() []int {\n\tarrLen := len(this.Arr)\n\tna := make([]int, arrLen, arrLen)\n\tcopy(na, this.Arr)\n\trand.Shuffle(arrLen, func(i, j int) {\n\t\tna[i], na[j] = na[j], na[i]\n\t})\n\treturn na\n}", "func (s *Solution) Shuffle() []int {\n\ttmp := make([]int, len(s.arr))\n\tcopy(tmp, s.arr)\n\tfor i := 0; i < len(s.arr); i++ {\n\t\tr := rand.Intn(len(s.arr))\n\t\ttmp[i], tmp[r] = tmp[r], tmp[i]\n\t}\n\n\treturn tmp\n}", "func (d deck) shuffle() {\n\trand.Seed(time.Now().UnixNano())\n\t// initial implementation\n\t// maxNdxPlus1 := len(d)\n\t// for i := range d {\n\t// \tnewPosition := rand.Intn(maxNdxPlus1)\n\t// \td[i], d[newPosition] = d[newPosition], d[i]\n\t// }\n\n\t// alternate implementation (using builtin Shuffle() function)\n\trand.Shuffle(len(d), func(i, j int) {\n\t\td[i], d[j] = d[j], d[i]\n\t})\n}", "func Shuffle(a []int) {\n\trand.Seed(time.Now().UnixNano())\n\n\tfor i := range a {\n\t\tj := rand.Intn(i + 1)\n\t\ta[i], a[j] = a[j], a[i]\n\t}\n}", "func (d deck) shuffle() {\n\t// To generate new random number, generating a new seed every time\n\tsource := rand.NewSource(time.Now().UnixNano()) // time.Now().UnixNano() generates new int64 every time\n\tr := rand.New(source)\n\n\tfor i := range d {\n\t\tnewPosition := r.Intn(len(d) - 1)\n\t\td[i], d[newPosition] = d[newPosition], d[i]\n\t}\n}", "func Shuffle(n int, swap func(i, j int)) {\n\tglobalRand.Shuffle(n, swap)\n}", "func (ss SliceType) Shuffle(source rand.Source) SliceType {\n\tn := len(ss)\n\n\t// Avoid the extra allocation.\n\tif n < 2 {\n\t\treturn ss\n\t}\n\n\t// go 1.10+ provides rnd.Shuffle. However, to support older versions we copy\n\t// the algorithm directly from the go source: src/math/rand/rand.go below,\n\t// with some adjustments:\n\tshuffled := make([]ElementType, n)\n\tcopy(shuffled, ss)\n\n\trnd := rand.New(source)\n\n\tutil.Shuffle(rnd, n, func(i, j int) {\n\t\tshuffled[i], shuffled[j] = shuffled[j], shuffled[i]\n\t})\n\n\treturn shuffled\n}", "func (d deck) shuffle() {\n\t//seed value = source\n\t//pass some value to NewSource > use time package\n\t//func Now & func (Time)UnixNano >every single time we run program we will use a new time\n\t//to generate number of type int64, we use that as seed as source object\n\tsource := rand.NewSource(time.Now().UnixNano())\n\n\t//type Rand\n\tr := rand.New(source)\n\n\t//for index or simply i (per convention) -- we need to get the element that we need to iterate over\n\tfor i := range d {\n\t\t//use rand function in math package -- pseudo random generator (by default go uses the exact same seed value)\n\t\t//len is the length of slice\n\t\t//newPosition := rand.Intn(len(d) - 1)\n\n\t\t//with fix to the same seed value (src) -> func New(src Source) *Rand\n\t\tnewPosition := r.Intn(len(d) - 1)\n\n\t\t//swap elements, take whatever is at newPosition and assign to i, take whatever is in i and assign to newPosition\n\t\td[i], d[newPosition] = d[newPosition], d[i]\n\t}\n}", "func Shuffle(cards []Card) []Card {\n\tfor i := range cards {\n\t\tj := rand.Intn(len(cards))\n\t\tcards[i], cards[j] = cards[j], cards[i]\n\t}\n\n\treturn cards\n}", "func (d Deck) Shuffle() {\n\tvar numCards = len(d.cards)\n\tvar newCardsArray = make([]Card, numCards)\n\tvar perm = rand.Perm(numCards)\n\tfor i, j := range perm {\n\t\tnewCardsArray[j] = d.cards[i]\n\t}\n\td.cards = newCardsArray\n}", "func (d deck) pseduoshuffle() {\n\tfor i := range d {\n\t\tnewPosition := rand.Intn(len(d) - 1) //pseudo random generator where we cannot specify a source\n\t\td[i], d[newPosition] = d[newPosition], d[i]\n\n\t}\n}", "func (d deck) shuffle() {\n\tsource := rand.NewSource(time.Now().UnixNano())\n\tr := rand.New(source)\n\n\tfor i := range d {\n\t\tnewPosition := r.Intn(len(d) - 1)\n\t\td[i], d[newPosition] = d[newPosition], d[i]\n\t}\n}", "func (list List) Shuffle() {\n\trand.Seed(time.Now().Unix())\n\trand.Shuffle(len(list), func(i, j int) {\n\t\tlist[i], list[j] = list[j], list[i]\n\t})\n}", "func (d *deck) shuffle() {\n\trand.Seed(time.Now().UnixNano())\n\tfor i := len(d) - 1; i > 0; i-- {\n\t\tj := rand.Intn(i + 1)\n\t\td[i], d[j] = d[j], d[i]\n\t}\n}", "func (d deck) shuffle() {\n\tsource := rand.NewSource(time.Now().UnixNano())\n\tr := rand.New(source)\n\n\tfor i := range d { // i = index\n\t\tnewPosition := r.Intn(len(d) - 1)\n\n\t\td[i], d[newPosition] = d[newPosition], d[i] //swap them\n\t}\n}", "func (d deck) shuffle() {\n\t//create new random source, seed value\n\tsource := rand.NewSource(time.Now().UnixNano())\n\tr := rand.New(source)\n\n\t//calling upon r of type rand we can replace rand.Intn() with r.Intn() to generate a random source seed for the Intn()\n\tfor i := range d {\n\t\tnewPosition := r.Intn(len(d) - 1)\n\n\t\td[i], d[newPosition] = d[newPosition], d[i]\n\t}\n}", "func ShufflePermutation(\n\tpermutation [][2]int,\n\tseed int64,\n) {\n\trand.Seed(seed)\n\trand.Shuffle(\n\t\tlen(permutation),\n\t\tfunc(i, j int) {\n\t\t\tpermutation[i], permutation[j] = permutation[j], permutation[i]\n\t\t},\n\t)\n}", "func (d deck) shuffle() {\n\tdeckLength := len(d)\n\n\tsource := rand.NewSource(time.Now().UnixNano())\n\tr := rand.New(source)\n\n\tfor i := range d {\n\t\tr := r.Intn(deckLength)\n\t\td[i], d[r] = d[r], d[i]\n\t}\n}", "func (d deck) shuffle(){\n\n\n\t// method - 1\n\t//source := rand.NewSource(time.Now().UnixNano())\n\t//r := rand.New(source)\n\n\t// method - 2\n\trand.Seed(time.Now().UnixNano())\n\tmax := len(d)\n\tfor i := range d{\n\t\t//newPosition := r.Intn(max-i)+i\n\t\tnewPosition := rand.Intn(max-i)+i\t\t// generate random number between i and len(deck)\n\t\td[i], d[newPosition] = d[newPosition], d[i]\n\t}\n}", "func (d deck) shuffle() {\n\trand.Seed(time.Now().UnixNano())\n\n\tfor i := range d {\n\t\tswapPos := rand.Intn(len(d) - 1)\n\t\td[i], d[swapPos] = d[swapPos], d[i]\n\t}\n}", "func (c *StoreCandidates) Shuffle() *StoreCandidates {\n\trand.Shuffle(len(c.Stores), func(i, j int) { c.Stores[i], c.Stores[j] = c.Stores[j], c.Stores[i] })\n\treturn c\n}", "func (shoe *Shoe) Shuffle() {\n\tc := shoe.Cards\n\trand.Seed(time.Now().UnixNano())\n\trand.Shuffle(\n\t\tlen(c),\n\t\tfunc(i, j int) {\n\t\t\tc[i], c[j] = c[j], c[i]\n\t\t})\n}", "func (d deck) shuffle() {\n\tsource := rand.NewSource(time.Now().UnixNano()) //generate a new seed for random number generator\n\tr := rand.New(source)\n\n\tfor i := range d {\n\t\tnewPosition := r.Intn(len(d) - 1)\n\n\t\td[i], d[newPosition] = d[newPosition], d[i]\n\t}\n}", "func (this *Solution) Shuffle() []int {\n\tvar temp []int\n\tresult := []int{}\n\tif len(this.changed) == 0 {\n\t\ttemp = this.initial\n\t} else {\n\t\ttemp = this.changed\n\t}\n\tfor i, d := range temp {\n\t\tif i == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tresult = append(result, d)\n\t}\n\tresult = append(result, temp[0])\n\tthis.changed = result\n\treturn result\n}", "func Shuffle(deck Deck) (shuffledDeck Deck) {\n\tshuffledDeck = deck\n\tlength := len(deck)\n\tfor i := 1; i < length; i++ {\n\t\trandom := rand.Intn(i + 1)\n\t\tif i != random {\n\t\t\tshuffledDeck[random], shuffledDeck[i] = shuffledDeck[i], shuffledDeck[random]\n\t\t}\n\t}\n\treturn\n}", "func (d deck) shuffle() {\n\t// type Rand is a source of random numbers. If we create a value of type Rand, we have the ability to specify the seed, or source of randomness for our number generator.\n\t// To make a value of type Rand, and to make a source, we have to call newSource() and passing in a random int64 number, and pass source to our random number generator\n\t// In summary: we're using current time in nanoseconds to generate a new int64 number every time we start the program. We use that as the \"seed\" to generate a new source object, and we use that source object as the basis for our new random number generator.\n\tsource := rand.NewSource(time.Now().UnixNano()) // https://golang.org/pkg/math/rand/#Source\n\tr := rand.New(source) // r is random number generator of type Rand\n\n\tfor idx := range d {\n\t\tnewPosition := r.Intn(len(d) - 1) // if you have a value of type Rand, you can call Intn https://golang.org/pkg/math/rand/#Rand.Intn\n\n\t\td[idx], d[newPosition] = d[newPosition], d[idx] // swap elements at both idx and\n\t}\n\n}", "func (this *Solution) Shuffle() []int {\n\tshuffled := make([]int, len(this.NumOrigin))\n\tperm := this.random.Perm(len(this.NumOrigin))\n\tfor k, v := range perm {\n\t\tshuffled[k] = this.NumOrigin[v]\n\t}\n\n\treturn shuffled\n}", "func (d *Deck) shuffle() {\n\tdeck := d.cards\n\tfor i := 1; i < len(deck); i++ {\n\t\tr := rand.Intn(i + 1)\n\t\tif i != r {\n\t\t\tdeck[r], deck[i] = deck[i], deck[r]\n\t\t}\n\t}\n}", "func shuffle(a []int, n int) []int {\n\tfor i := 0; i < n; i++ {\n\t\tv := a[n+i]\n\t\tfor x := n + i; x > 2*i+1; x-- {\n\t\t\ta[x] = a[x-1]\n\t\t}\n\t\ta[2*i+1] = v\n\t}\n\treturn a\n}", "func (s *summary) shuffle(rng RNG) {\n\tfor i := len(s.means) - 1; i > 1; i-- {\n\t\ts.Swap(i, rng.Intn(i+1))\n\t}\n}", "func (d deck) shuffle() {\n\t// Get a new Source\n\tsrc := rand.NewSource(time.Now().UTC().UnixNano())\n\n\t// Get a new Rand\n\trandomizer := rand.New(src)\n\n\t// shuffle the cards in the deck\n\tfor i := range d {\n\t\tnewIndx := randomizer.Intn(len(d))\n\t\td[i], d[newIndx] = d[newIndx], d[i]\n\t}\n\n}", "func (d deck) shuffle() {\n\t/* \t1. create a trully random generator, starts whit a source object (is like a seed)\n\t \tevery sigle time the code runs use the Time object to create a unique value as a source (or seed)\n\t \tfor random number generator, NewSource needs and int64 as param, and UnixNano returns a int64 from actual time */\n\tsource := rand.NewSource(time.Now().UnixNano())\n\t//using the source object creates a random generator object *Rand type\n\tr := rand.New(source)\n\t//2. iterate the deck of cards, we will use only the index of the slice\n\tfor i := range d {\n\t\t//*Rand Object has Intn() method to creates a trully a random int beetwen 0 and lenght of slice - 1\n\t\tnewPosition := r.Intn(len(d) - 1)\n\t\t//4. in every i slice position, swap the values of the indexes, newPosition is a random number\n\t\t//so, the position of cards always will be random\n\t\td[i], d[newPosition] = d[newPosition], d[i]\n\n\t}\n\n}", "func Shuffle(deck []Card) []Card {\n\tr := rand.New(rand.NewSource(time.Now().Unix()))\n\tshuffled := make([]Card, len(deck))\n\tj := 0\n\tfor _, i := range r.Perm(len(deck)) {\n\t\tshuffled[j] = deck[i]\n\t\tj++\n\t}\n\treturn shuffled\n}", "func (s *SliceInt) Shuffle() *SliceInt {\n\tif s == nil {\n\t\treturn nil\n\t}\n\n\tfor i := len(s.data) - 1; i > 0; i-- {\n\t\tj := rand.Intn(i + 1)\n\t\ts.data[i], s.data[j] = s.data[j], s.data[i]\n\t}\n\treturn s\n}", "func Shuffle(cards []Card) []Card {\n\tr := rand.New(rand.NewSource(time.Now().Unix()))\n\tr.Shuffle(len(cards), func(i, j int) {\n\t\tcards[i], cards[j] = cards[j], cards[i]\n\t})\n\treturn cards\n}", "func Shuffle(availableNumbers []string) []string {\n\trand.Seed(time.Now().UnixNano())\n\tfor i := len(availableNumbers) - 1; i > 0; i-- {\n\t\tj := rand.Intn(i + 1)\n\t\tavailableNumbers[i], availableNumbers[j] = availableNumbers[j], availableNumbers[i]\n\t}\n\treturn availableNumbers\n}", "func ShuffleJumps(s []jump) {\n\tfor i := range s {\n\t\tj := rand.Intn(i + 1)\n\t\ts[i], s[j] = s[j], s[i]\n\t}\n}", "func (d *Deck) Shuffle() error {\n\tfor i := 0; i < d.size; i++ {\n\t\tr := rand.Intn(i + 1)\n\t\tif i != r {\n\t\t\td.Cards[r], d.Cards[i] = d.Cards[i], d.Cards[r]\n\t\t}\n\t}\n\treturn nil\n}", "func shuffleAds() {\n\tdest := make([]AdObject, len(Ads))\n\tperm := rand.Perm(len(Ads))\n\tfor i, v := range perm {\n\t\tdest[v] = Ads[i]\n\t}\n}", "func (d *Deck) Shuffle(seed int64) {\n\tif seed == UniqueShuffle {\n\t\tseed = time.Now().UnixNano()\n\t}\n\trand.Seed(seed)\n\tfor i := 0; i < 52; i++ {\n\t\tr := i + rand.Intn(52-i)\n\t\td.Cards[r], d.Cards[i] = d.Cards[i], d.Cards[r]\n\t}\n}", "func (deck *Deck) Shuffle() {\n\tdeck.cards = make([]Card, len(fullDeck.cards))\n\tcopy(deck.cards, fullDeck.cards)\n\trand.Shuffle(len(deck.cards), func(i, j int) {\n\t\tdeck.cards[i], deck.cards[j] = deck.cards[j], deck.cards[i]\n\t})\n}", "func (pts testDatapoints) Shuffle() {\n\tfor i := len(pts) - 1; i > 0; i-- {\n\t\tif j := rand.Intn(i + 1); i != j {\n\t\t\tpts[i], pts[j] = pts[j], pts[i]\n\t\t}\n\t}\n}", "func (p *PCG64) Shuffle(n int, swap func(i, j int)) {\n\tif n < 0 {\n\t\tpanic(\"fastrand: invalid argument to Shuffle\")\n\t}\n\n\t// Fisher-Yates shuffle: https://en.wikipedia.org/wiki/Fisher%E2%80%93Yates_shuffle\n\t// Shuffle really ought not be called with n that doesn't fit in 32 bits.\n\t// Not only will it take a very long time, but with 2³¹! possible permutations,\n\t// there's no way that any PRNG can have a big enough internal state to\n\t// generate even a minuscule percentage of the possible permutations.\n\t// Nevertheless, the right API signature accepts an int n, so handle it as best we can.\n\ti := n - 1\n\tfor ; i > 1<<31-1-1; i-- {\n\t\tj := int(p.Int63n(int64(i + 1)))\n\t\tswap(i, j)\n\t}\n\tfor ; i > 0; i-- {\n\t\tj := int(p.Int31n(int32(i + 1)))\n\t\tswap(i, j)\n\t}\n}", "func shufflePoints(ps []Point) {\n\tfor i := 0; i < len(ps)-1; i++ {\n\t\tj := rand.Intn(len(ps)-i) + i\n\t\tps[i], ps[j] = ps[j], ps[i]\n\t}\n}", "func TestShuffle(t *testing.T) {\n\tdeck := New()\n\tShuffle(deck)\n}", "func (d *Deck) Shuffle() *Deck {\n\tvar shuffled []card.Card\n\tfor len(d.Cards) > 0 {\n\t\tindex := d.PickIndex()\n\t\tshuffled = append(shuffled, d.Cards[index])\n\t\td.remove(index)\n\t}\n\td.Cards = shuffled\n\treturn d\n}", "func NewShuffle(seed int64) Processor {\n\treturn shuffle{\n\t\trng: rand.New(rand.NewSource(seed)),\n\t}\n}", "func NewShuffler() *Shuffler { return &Shuffler{} }", "func ShuffleRandom(servers []types.DatabaseServer) []types.DatabaseServer {\n\trand.New(rand.NewSource(time.Now().UnixNano())).Shuffle(\n\t\tlen(servers), func(i, j int) {\n\t\t\tservers[i], servers[j] = servers[j], servers[i]\n\t\t})\n\treturn servers\n}", "func randomize(it graph.Nodes, src rand.Source) graph.Nodes {\n\tnodes := graph.NodesOf(it)\n\tvar shuffle func(int, func(i, j int))\n\tif src == nil {\n\t\tshuffle = rand.Shuffle\n\t} else {\n\t\tshuffle = rand.New(src).Shuffle\n\t}\n\tshuffle(len(nodes), func(i, j int) {\n\t\tnodes[i], nodes[j] = nodes[j], nodes[i]\n\t})\n\treturn iterator.NewOrderedNodes(nodes)\n}", "func (d *Deck) Shuffle() {\n\td.cards = rand.Perm(d.numCards)\n}", "func shuffleInPlace(text []byte, contract uint32) []byte {\n\tif contract == 0 {\n\t\treturn text\n\t}\n\tsalt := make([]byte, 4)\n\tbinary.BigEndian.PutUint32(salt[0:4], contract)\n\n\tresult := duplicateSlice(text)\n\tfor i, v, p := len(result)-1, 0, 0; i > 0; i-- {\n\t\tp += int(salt[v])\n\t\tj := (int(salt[v]) + v + p) % i\n\t\tresult[i], result[j] = result[j], result[i]\n\t\tv = (v + 1) % len(salt)\n\t}\n\treturn result\n}", "func shuffle_array(arr []int) {\n rand.Shuffle(len(arr),\n func(i, j int) {\n arr[i], arr[j] = arr[j], arr[i] })\n}", "func Shuffle(sz []string) []string {\n\tif len(sz) <= 0 {\n\t\treturn nil\n\t}\n\n\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\tfor i := len(sz) - 1; i > 0; i-- {\n\t\tj := r.Intn(i + 1)\n\t\ttmp := sz[i]\n\t\tsz[i] = sz[j]\n\t\tsz[j] = tmp\n\t}\n\treturn sz\n}", "func Shuffle(random io.Reader, n int, swap func(i, j int)) *ShuffleMap {\n\tif n < 0 || n > (1<<31-1-1) {\n\t\tpanic(\"invalid argument to Shuffle\")\n\t}\n\n\tidx := make([]int, n)\n\tfor i := range idx {\n\t\tidx[i] = i\n\t}\n\tperm := make([]int, n)\n\n\t// Fisher-Yates shuffle: https://en.wikipedia.org/wiki/Fisher%E2%80%93Yates_shuffle\n\tfor i := n - 1; i > 0; i-- {\n\t\tj := int(uniformRandom31(random, int32(i+1)))\n\t\tswap(i, int(j))\n\t\tidx[i], idx[j] = idx[j], idx[i]\n\t\tperm[idx[i]] = i\n\t}\n\treturn &ShuffleMap{perm}\n}", "func Shuffle235(well []string, count int64) []string {\n\tvar choice int64\n\tvar old string\n\n\tfor i := count - 1; i > 1; i-- {\n\t\tchoice = rand.Int63n(i)\n\t\told = well[i]\n\t\twell[i] = well[choice]\n\t\twell[choice] = old\n\t}\n\treturn well\n}", "func (q *Question) ShuffleAnswers() []string {\n\tanswers := append(q.IncorrectAnswers, q.CorrectAnswer)\n\n\tshuffle.Strings(answers)\n\n\treturn answers\n}", "func ShuffleArray(array []int) {\n\tfor i := 0; i < len(array); i++ {\n\t\tk := rand.Int() % (i + 1)\n\t\ttemp := array[k]\n\t\tarray[k] = array[i]\n\t\tarray[i] = temp\n\t}\n}", "func ArrayInt64Shuffle(arr []int64) []int64 {\n\tn := len(arr)\n\tfor i := n - 1; i >= 0; i-- {\n\t\tj := IntRand(0, i+1)\n\t\tarr[i], arr[j] = arr[j], arr[i]\n\t}\n\treturn arr\n}", "func ShuffleInts(a []int) {\n\tfor i := range a {\n\t\tj := rand.Intn(i + 1)\n\t\ta[i], a[j] = a[j], a[i]\n\t}\n}", "func ShuffleDeck() Rule {\n\treturn func(deck []Card) []Card {\n\t\trand.Shuffle(len(deck), func(i, j int) {\n\t\t\tdeck[i], deck[j] = deck[j], deck[i]\n\t\t})\n\t\treturn deck\n\t}\n}", "func (d *Deck) Shuffle() {\n\n\tShuffleCards(d.Cards)\n}", "func randomOperation(r *rand.Rand) Operation {\n\toperations := []Operation{\n\t\tOpBondOrUnbond,\n\t\tOpAddTokens,\n\t\tOpRemoveShares,\n\t}\n\tr.Shuffle(len(operations), func(i, j int) {\n\t\toperations[i], operations[j] = operations[j], operations[i]\n\t})\n\treturn operations[0]\n}", "func randomOp() *opDesc {\n\tr := rand.Float64()\n\tfor _, op := range ops {\n\t\tif r < op.normFreq {\n\t\t\treturn op\n\t\t}\n\t}\n\treturn ops[len(ops)-1]\n}", "func (c StringArrayCollection) Shuffle() Collection {\n\tvar d = make([]string, len(c.value))\n\tcopy(d, c.value)\n\trand.Seed(time.Now().UnixNano())\n\trand.Shuffle(len(c.value), func(i, j int) { d[i], d[j] = d[j], d[i] })\n\treturn StringArrayCollection{\n\t\tvalue: d,\n\t}\n}", "func IntShuffle(values []int) {\n\tvar j, tmp int\n\tfor i := len(values) - 1; i > 0; i-- {\n\t\tj = rand.Int() % i\n\t\ttmp = values[j]\n\t\tvalues[j] = values[i]\n\t\tvalues[i] = tmp\n\t}\n}", "func RandomIndexShuffle(scope *Scope, index tf.Output, seed tf.Output, max_index tf.Output, optional ...RandomIndexShuffleAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"RandomIndexShuffle\",\n\t\tInput: []tf.Input{\n\t\t\tindex, seed, max_index,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (br *BlackRock) Shuffle(m uint64) uint64 {\n\tvar c uint64\n\tc = br.encrypt(br.rounds, br.a, br.b, m, br.seed)\n\tfor {\n\t\tif c >= br.inputSize {\n\t\t\tc = br.encrypt(br.rounds, br.a, br.b, c, br.seed)\n\t\t} else {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn c\n}", "func ShuffleInts(x []int) []int {\n\tfor i := len(x) - 1; i > 0; i-- {\n\t\tj := rand.Intn(i + 1)\n\t\tx[i], x[j] = x[j], x[i]\n\t}\n\treturn x\n}", "func Slice(slice interface{}) {\n\trv := reflect.ValueOf(slice)\n\tswap := reflect.Swapper(slice)\n\trand.Shuffle(rv.Len(), swap)\n}", "func popRandom(txs []Tx, rng *rand.Rand) []Tx {\n\tnewlen := len(txs) - 1\n\ti := rng.Intn(len(txs))\n\ttxs[i] = txs[newlen]\n\treturn txs[:newlen]\n}", "func (br *BlackRock) Unshuffle(m uint64) uint64 {\n\tvar c uint64\n\n\tc = br.unencrypt(br.rounds, br.a, br.b, m, br.seed)\n\tfor {\n\t\tif c >= br.inputSize {\n\t\t\tc = br.unencrypt(br.rounds, br.a, br.b, c, br.seed)\n\t\t} else {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn c\n}", "func shuffleCards(cards []*Card) []*Card {\n\tshuffled := make([]*Card, len(cards))\n\tcopy(shuffled, cards)\n\n\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\tr.Shuffle(len(shuffled), func(i, j int) {\n\t\tshuffled[i], shuffled[j] = shuffled[j], shuffled[i]\n\t})\n\n\treturn shuffled\n}", "func copyRandomList(head *Node) *Node {\n\tvar result *Node\n\n\t// loop 1, duplicate each node\n\tloop := head\n\tfor loop != nil {\n\t\tnext := loop.Next\n\t\tloop.Next = &Node{\n\t\t\tVal: loop.Val,\n\t\t\tNext: next,\n\t\t}\n\t\tloop = loop.Next.Next\n\t}\n\n\t// add random for duplicate node\n\tloop = head\n\tfor loop != nil {\n\t\tif loop.Random != nil {\n\t\t\tloop.Next.Random = loop.Random.Next\n\t\t}\n\t\tloop = loop.Next.Next\n\t}\n\n\t// pick up duplicate node\n\tloop = head\n\tif head != nil {\n\t\tresult = head.Next\n\t}\n\tfor loop != nil {\n\t\tif loop.Next.Next == nil {\n\t\t\tloop.Next = nil\n\t\t\tbreak\n\t\t}\n\t\ttmp := loop.Next\n\t\tloop.Next = loop.Next.Next\n\t\tloop = tmp\n\t}\n\treturn result\n}" ]
[ "0.65571165", "0.63301414", "0.63115364", "0.6239861", "0.62270623", "0.62163925", "0.62152797", "0.62081397", "0.6158927", "0.61263853", "0.60523444", "0.60523444", "0.6043424", "0.6010341", "0.60028875", "0.59959465", "0.5987183", "0.59851146", "0.5976741", "0.59650725", "0.5954744", "0.5875736", "0.587019", "0.5866285", "0.58549786", "0.58408076", "0.5829281", "0.579915", "0.5798456", "0.57936615", "0.5770798", "0.57606626", "0.5758242", "0.5752122", "0.57476145", "0.57259756", "0.5721523", "0.5709148", "0.5707955", "0.5697503", "0.56974626", "0.5690554", "0.5689447", "0.5689027", "0.56786543", "0.56766665", "0.56598717", "0.56496245", "0.5648165", "0.5640758", "0.5626251", "0.5595028", "0.5586982", "0.5573644", "0.5571168", "0.5558916", "0.55385464", "0.5535225", "0.55315214", "0.5522931", "0.55205995", "0.5492053", "0.5491092", "0.5447351", "0.54305714", "0.5411751", "0.5406113", "0.5402594", "0.5401509", "0.5372574", "0.5364841", "0.53517747", "0.53462654", "0.53424597", "0.5322661", "0.5273441", "0.52681375", "0.52572954", "0.52365506", "0.5206043", "0.5193271", "0.51805645", "0.5170899", "0.51419455", "0.5126597", "0.51177", "0.50978196", "0.5080086", "0.50658303", "0.5029642", "0.5026963", "0.5009636", "0.4995135", "0.49717066", "0.49674028", "0.4950909", "0.49488854", "0.49412262", "0.4913778", "0.48937914" ]
0.7436856
0
Sort returns a copy of operand sorted by sort.Strings.
func Sort(operand []string) []string { sorted := make([]string, len(operand)) for i := range operand { sorted[i] = operand[i] } sort.Strings(sorted) return sorted }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p StringSlice) Sort() { Sort(p) }", "func Strings(a []string) { Sort(StringSlice(a)) }", "func (s StringSlice) Sort() {\n\tsort.Sort(s)\n}", "func TestSort() {\n\ts := StringSlice{\"name1\", \"name3\", \"name2\"}\n\tfmt.Println(\"Input: \", s)\n\tsort.Sort(s)\n\t// Std lib: sort.SortStrings(s) :3\n\tfmt.Println(\"Result: \", s)\n}", "func (q *BasicQuery) Sort(sstr string) Query {\n\tq.Sorts.Add(sstr)\n\treturn q\n}", "func (g Group) Sort() Group {\n\tsort.Strings(g.str)\n\treturn g\n}", "func copyAndSortStringSlice(s []string) []string {\n\tsc := make([]string, 0, len(s))\n\tsc = append(sc, s...)\n\n\tsort.Strings(sc)\n\treturn sc\n}", "func Sort(input []string) {\n\tsort.Sort(slice(input))\n}", "func Sort(ss []string) []string {\n\tif ss == nil {\n\t\treturn nil\n\t}\n\tss2 := make([]string, len(ss))\n\tcopy(ss2, ss)\n\tsort.Strings(ss2)\n\treturn ss2\n}", "func TestMyStringComparator(t *testing.T) {\n\ttables := []struct {\n\t\toperand string\n\t\toperator string\n\t\tmyVal string\n\t\texpected bool\n\t\terr error\n\t}{\n\t\t{\"random\", \">\", \"myName\", \"random\" > \"myName\", nil},\n\t\t{\"12\", \"!=\", \"myName\", \"12\" != \"myName\", nil},\n\t\t{\"12\", \"=\", \"myName\", \"12\" == \"myName\", nil},\n\t\t{\"12\", \"<=\", \"myName\", \"12\" <= \"myName\", nil},\n\t\t{\"12\", \">=\", \"myName\", \"12\" >= \"myName\", nil},\n\t\t{\"12\", \"<\", \"myName\", \"12\" < \"myName\", nil},\n\t\t{\"name\", \"like\", \"_x%\", false, nil},\n\t\t{\"12\", \"randomoperator\", \"myName\", false, ErrUnsupportedSyntax},\n\t}\n\tfor _, table := range tables {\n\t\tmyVal, err := stringEval(table.operand, table.operator, table.myVal)\n\t\tif err != table.err {\n\t\t\tt.Error()\n\t\t}\n\t\tif myVal != table.expected {\n\t\t\tt.Error()\n\t\t}\n\t}\n}", "func sortStringLists(f *File, w *Rewriter) {\n\tsortStringList := func(x *Expr) {\n\t\tSortStringList(*x)\n\t}\n\n\tWalk(f, func(e Expr, stk []Expr) {\n\t\tswitch v := e.(type) {\n\t\tcase *CallExpr:\n\t\t\tif f.Type == TypeDefault || f.Type == TypeBzl {\n\t\t\t\t// Rule parameters, not applicable to .bzl or default file types\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif leaveAlone(stk, v) {\n\t\t\t\treturn\n\t\t\t}\n\t\t\trule := callName(v)\n\t\t\tfor _, arg := range v.List {\n\t\t\t\tif leaveAlone1(arg) {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tas, ok := arg.(*AssignExpr)\n\t\t\t\tif !ok || leaveAlone1(as) {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tkey, ok := as.LHS.(*Ident)\n\t\t\t\tif !ok {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tcontext := rule + \".\" + key.Name\n\t\t\t\tif w.SortableDenylist[context] {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif w.IsSortableListArg[key.Name] ||\n\t\t\t\t\t\tw.SortableAllowlist[context] ||\n\t\t\t\t\t\t(!disabled(\"unsafesort\") && allowedSort(context)) {\n\t\t\t\t\tif doNotSort(as) {\n\t\t\t\t\t\tdeduplicateStringList(as.RHS)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfindAndModifyStrings(&as.RHS, sortStringList)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\tcase *AssignExpr:\n\t\t\tif disabled(\"unsafesort\") {\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// \"keep sorted\" comment on x = list forces sorting of list.\n\t\t\tif keepSorted(v) {\n\t\t\t\tfindAndModifyStrings(&v.RHS, sortStringList)\n\t\t\t}\n\t\tcase *KeyValueExpr:\n\t\t\tif disabled(\"unsafesort\") {\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// \"keep sorted\" before key: list also forces sorting of list.\n\t\t\tif keepSorted(v) {\n\t\t\t\tfindAndModifyStrings(&v.Value, sortStringList)\n\t\t\t}\n\t\tcase *ListExpr:\n\t\t\tif disabled(\"unsafesort\") {\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// \"keep sorted\" comment above first list element also forces sorting of list.\n\t\t\tif len(v.List) > 0 && (keepSorted(v) || keepSorted(v.List[0])) {\n\t\t\t\tfindAndModifyStrings(&e, sortStringList)\n\t\t\t}\n\t\t}\n\t})\n}", "func Strings(s []string, caseInsensitive bool) {\n\tif caseInsensitive {\n\t\tsort.Sort(stringSlice(s))\n\t} else {\n\t\tsort.Strings(s)\n\t}\n}", "func StringComparator(a, b interface{}) int {\n\ts1 := a.(string)\n\ts2 := b.(string)\n\n\tmin := len(s2)\n\n\tif len(s1) < len(s2) {\n\t\tmin = len(s1)\n\t}\n\n\tdiff := 0\n\n\tfor i := 0; i < min && diff == 0; i++ {\n\t\tdiff = int(s1[i]) - int(s2[i])\n\t}\n\n\tif diff == 0 {\n\t\tdiff = len(s1) - len(s2)\n\t}\n\n\tif diff < 0 {\n\t\treturn -1\n\t}\n\n\tif diff > 0 {\n\t\treturn 1\n\t}\n\n\treturn 0\n}", "func Strings(s []string) int {\n\treturn Sort(sort.StringSlice(s))\n}", "func (s *Sort) Execute() error {\n\tmatches := make([]string, 0)\n\n\tfields := strings.Split(s.args.fields, \",\")\n\n\tfor _, file := range s.files {\n\t\tfileMatches, err := searchFile(s.args, file)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tmatches = append(matches, fileMatches...)\n\n\t\tif s.args.duplicate {\n\t\t\tnewMatches := make([]string, 0)\n\n\t\t\tset := map[string]struct{}{}\n\n\t\t\tfor _, match := range matches {\n\t\t\t\tif _, ok := set[match]; !ok {\n\t\t\t\t\tnewMatches = append(newMatches, match)\n\t\t\t\t\tset[match] = struct{}{}\n\t\t\t\t}\n\t\t\t}\n\t\t\tmatches = newMatches\n\t\t}\n\t}\n\n\tif len(s.args.fields) > 0 {\n\t\tfor _, v := range fields {\n\t\t\tfield, err := strconv.Atoi(v)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.New(\"invalid field value\")\n\t\t\t}\n\t\t\tif field <= 0 {\n\t\t\t\treturn errors.New(\"fields are numbered from 1\")\n\t\t\t}\n\n\t\t\tsort.Slice(matches, func(i, j int) bool {\n\t\t\t\tlhs := strings.Split(matches[i], \" \")\n\t\t\t\trhs := strings.Split(matches[j], \" \")\n\n\t\t\t\tif len(lhs) <= field-1 || len(rhs) <= field-1 {\n\t\t\t\t\treturn lhs[0] < rhs[0]\n\t\t\t\t}\n\n\t\t\t\tif s.args.reverse {\n\t\t\t\t\treturn strings.Split(matches[i], \" \")[field-1] >\n\t\t\t\t\t\tstrings.Split(matches[j], \" \")[field-1]\n\t\t\t\t} else {\n\t\t\t\t\treturn strings.Split(matches[i], \" \")[field-1] <\n\t\t\t\t\t\tstrings.Split(matches[j], \" \")[field-1]\n\t\t\t\t}\n\t\t\t})\n\t\t}\n\t} else {\n\t\tif s.args.reverse {\n\t\t\tsort.Sort(sort.Reverse(sort.StringSlice(matches)))\n\t\t} else {\n\t\t\tsort.Strings(matches)\n\t\t}\n\t}\n\n\ts.result = matches\n\n\treturn nil\n}", "func SortString(w string) string {\n\ts := strings.Split(w, \"\")\n\tsort.Strings(s)\n\treturn strings.Join(s, \"\")\n}", "func customSortString(S string, T string) string {\n var buffer bytes.Buffer\n \n TCount := map[byte]int{} \n for _, letterRune := range(T) {\n letter := byte(letterRune)\n TCount[letter]++\n } \n \n for _, letterRune := range(S) {\n letter := byte(letterRune)\n if count, ok := TCount[letter]; ok {\n for i := 0; i < count; i++ {\n buffer.WriteByte(letter)\n }\n delete(TCount, letter)\n }\n }\n \n for letterRune, count := range(TCount) {\n letter := byte(letterRune)\n for i := 0; i < count; i++ {\n buffer.WriteByte(letter)\n }\n }\n \n return buffer.String()\n}", "func StringComparator(a, b string) int {\n\tif a > b {\n\t\treturn +1\n\t}\n\tif a < b {\n\t\treturn -1\n\t}\n\treturn 0\n}", "func BuildByCmpSort(str string) []int {\n\tstrLen := len(str)\n\n\t// Store suffixes and their sufarr in an array of structures.\n\t// The structure is needed to sort the suffixes alphabatically\n\t// and maintain their old sufarr while sorting\n\tfor i := 0; i < strLen; i++ {\n\t\tsufarr[i].i = i\n\t\tsufarr[i].rank[0] = int(str[i])\n\t\tsufnrr[i] = i\n\t}\n\n\t// At his point, all suffixes are sorted according to first\n\t// 2 characters. Let us sort suffixes according to first 4\n\t// characters, then first 8 and so on\n\tfor i, k, l := 0, 1, 0; k < strLen; k <<= 1 {\n\t\tl = k\n\n\t\t// Assign next rank to every suffix\n\t\tfor i = 0; i < strLen; i++ {\n\t\t\tadjSuf := sufarr[i].i + l\n\t\t\tif adjSuf < strLen {\n\t\t\t\tsufarr[i].rank[1] = sufarr[sufnrr[adjSuf]].rank[0]\n\t\t\t} else {\n\t\t\t\tsufarr[i].rank[1] = -1\n\t\t\t}\n\t\t}\n\n\t\t// Sort the suffixes according to first k characters\n\t\tsort.Sort(suffixes(sufarr[:strLen]))\n\n\t\t// Assigning rank and index values to first suffix\n\t\t// Assigning rank to suffixes\n\t\tr := 0\n\t\tfor i = 1; i < strLen; i++ {\n\t\t\tsufnrr[sufarr[i-1].i] = i - 1\n\t\t\tif Equal(&sufarr, i, i-1) {\n\t\t\t\tsufarr[i-1].rank[0] = r\n\t\t\t} else {\n\t\t\t\tsufarr[i-1].rank[0] = r\n\t\t\t\tr++\n\t\t\t}\n\t\t}\n\t\tsufnrr[sufarr[i-1].i] = i - 1\n\t\tsufarr[i-1].rank[0] = r\n\t}\n\n\tsuffixArray := make([]int, strLen)\n\t// Store sufarr of all sorted suffixes in the suffix array\n\tfor i := 0; i < strLen; i++ {\n\t\tsuffixArray[i] = sufarr[i].i\n\t}\n\n\treturn suffixArray\n}", "func (c *Combination) Sort() {\n\tsort.Strings(c.Numbers)\n}", "func StringsSortByIndex(d *[]string, sortedIds []int) {\n\tnewd := make([]string, len(*d))\n\n\tfor i := range sortedIds {\n\t\tnewd[i] = (*d)[sortedIds[i]]\n\t}\n\n\t(*d) = newd\n}", "func sortString(s string) string {\n\tchars := strings.Split(s, \"\")\n\tsort.Strings(chars)\n\treturn strings.Join(chars, \"\")\n}", "func Sort(word string) (sorted string) {\n\t// Convert the string to []rune.\n\tvar r []rune\n\tfor _, char := range word {\n\t\tr = append(r, char)\n\t}\n\n\tsort.Slice(r, func(i, j int) bool {\n\t\treturn r[i] < r[j]\n\t})\n\n\tsorted = string(r)\n\treturn\n}", "func (sm StringMap) Sort() StringMap {\n\tsort.SliceStable(*sm.orig, func(i, j int) bool { return (*sm.orig)[i].Key < (*sm.orig)[j].Key })\n\treturn sm\n}", "func (s Sort) String() string {\n\treturn string(s)\n}", "func s(s string) string {\n\tz := strings.Split(s, \"|\")\n\tsort.StringSlice(z).Sort()\n\treturn strings.Join(z, \"|\")\n}", "func (ss sortbase) Sort(s interface{}, fieldName string) {}", "func Sort_String(str string) string {\n\t\tnewStr := strings.Split(str, \"\")\n\t\tsort.Strings(newStr)\n\t\treturn strings.Join(newStr,\"\")\n\t}", "func sortSliceString(dataSlice sliceStringInterface, keyOrder, sortDirection string) sliceStringInterface {\n\tindex := make(PairList, len(dataSlice))\n\ti := 0\n\tfor k, v := range dataSlice {\n\t\tindex[i] = Pair{v[keyOrder].(string), k}\n\t\ti++\n\t}\n\tsort.Sort(index)\n\n\torderedDataSlice := make(sliceStringInterface, len(dataSlice))\n\tif sortDirection == \"asc\" {\n\t\tfor k, v := range index {\n\t\t\torderedDataSlice[k] = dataSlice[v.Value]\n\t\t}\n\t} else {\n\t\tfor k, v := range index {\n\t\t\torderedDataSlice[len(dataSlice)-k-1] = dataSlice[v.Value]\n\t\t}\n\t}\n\treturn orderedDataSlice\n}", "func (s *GoSort) GoSort() {\n sort.Sort(s)\n}", "func (sm StringMap) Sort() StringMap {\n\tsort.SliceStable(*sm.orig, func(i, j int) bool {\n\t\t// Intention is to move the nil values at the end.\n\t\treturn (*sm.orig)[i].Key < (*sm.orig)[j].Key\n\t})\n\treturn sm\n}", "func Sort[S ~[]E, E constraints.Ordered](x S) { expslices.Sort(x) }", "func (s InputSecureFileArray) Sort(less func(a, b InputSecureFile) bool) InputSecureFileArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func customSortString(order string, str string) string {\n\tvar res string\n\tcnts := [26]int{}\n\tfor i := 0; i < len(str); i++ {\n\t\tcnts[str[i]-'a']++\n\t}\n\n\tfor i := 0; i < len(order); i++ {\n\t\ts := order[i]\n\t\tif cnts[s-'a'] > 0 {\n\t\t\tres += strings.Repeat(string(s), cnts[s-'a'])\n\t\t}\n\t\tcnts[s-'a'] = 0\n\t}\n\n\tfor i := 0; i < 26; i++ {\n\t\tif cnts[i] > 0 {\n\t\t\tres += strings.Repeat(string(rune(i+'a')), cnts[i])\n\t\t}\n\t}\n\treturn res\n}", "func (s Sort) String() string {\n\treturn awsutil.Prettify(s)\n}", "func ByString(data StringInterface) {\n\tl := data.Len()\n\tparallelSort(data, radixSortString, task{end: l})\n\n\t// check results!\n\tfor i := 1; i < l; i++ {\n\t\tif data.Less(i, i-1) {\n\t\t\tif data.Key(i) > data.Key(i-1) {\n\t\t\t\tpanic(keyPanicMessage)\n\t\t\t}\n\t\t\tpanic(panicMessage)\n\t\t}\n\t}\n}", "func (s EncryptedChatDiscardedArray) Sort(less func(a, b EncryptedChatDiscarded) bool) EncryptedChatDiscardedArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func Strings(in []string) {\n\tvar mi int\n\n\tfor i := 0; i < len(in); i++ {\n\t\tmi = i\n\n\t\tfor j := i; j < len(in); j++ {\n\t\t\tif in[j] < in[mi] {\n\t\t\t\tmi = j\n\t\t\t}\n\t\t}\n\n\t\tif i != mi {\n\t\t\tin[i], in[mi] = in[mi], in[i]\n\t\t}\n\t}\n}", "func (s SecurePlainPhoneArray) Sort(less func(a, b SecurePlainPhone) bool) SecurePlainPhoneArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (s URLAuthResultAcceptedArray) Sort(less func(a, b URLAuthResultAccepted) bool) URLAuthResultAcceptedArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (s FileLocationArray) Sort(less func(a, b FileLocation) bool) FileLocationArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func execStringsAreSorted(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := sort.StringsAreSorted(args[0].([]string))\n\tp.Ret(1, ret)\n}", "func TestString(t *testing.T, sortFn func([]string)) {\n\tseed := time.Now().UnixNano()\n\tt.Log(\"Seed for random cases:\", seed)\n\trand.Seed(seed)\n\n\tfor name, list := range map[string][]string{\n\t\t\"sorted\": []string{\"apple\", \"banana\", \"cat\", \"dog\"},\n\t\t\"reverse\": []string{\"dog\", \"cat\", \"ball\", \"alphabet\"},\n\t\t\"duplicates\": []string{\"alphabet\", \"ball\", \"cat\", \"alphabet\", \"ball\", \"cat\"},\n\t\t\"similar\": []string{\"apple\", \"app\", \"alligator\", \"all\", \"all-in\", \"alphabet\", \"apoplexy\", \"apology\", \"apologize\"},\n\t\t\"random-1000-20\": func() []string {\n\t\t\talphabet := \"abcdefghijklmnopqrstuvwxyz\"\n\t\t\tvar list []string\n\t\t\tfor i := 0; i < 1000; i++ {\n\t\t\t\tvar sb strings.Builder\n\t\t\t\tfor j := 0; j < 20; j++ {\n\t\t\t\t\tsb.WriteByte(alphabet[rand.Intn(len(alphabet))])\n\t\t\t\t\tif rand.Intn(3) == 0 {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tlist = append(list, sb.String())\n\t\t\t}\n\t\t\treturn list\n\t\t}(),\n\t} {\n\t\tt.Run(name, func(t *testing.T) {\n\t\t\twant := make([]string, len(list))\n\t\t\tfor i, val := range list {\n\t\t\t\twant[i] = val\n\t\t\t}\n\t\t\tsort.Strings(want)\n\t\t\tsortFn(list)\n\t\t\terrorCount := 0\n\t\t\tif len(list) != len(want) {\n\t\t\t\tt.Fatalf(\"got len %d; want %d\", len(list), len(want))\n\t\t\t}\n\t\t\tfor i := 0; i < len(want); i++ {\n\t\t\t\tif errorCount >= 5 {\n\t\t\t\t\tt.Fatalf(\"additional errors omitted for brevity\")\n\t\t\t\t}\n\t\t\t\tif list[i] != want[i] {\n\t\t\t\t\terrorCount++\n\t\t\t\t\t// t.Errorf(\"list[%d] = %d; want %d\", i, list[i], want[i])\n\t\t\t\t}\n\t\t\t}\n\t\t})\n\t}\n}", "func frequencySort(s string) string {\n\ta := make([]FreqItem, 128)\n\tfor i := range a {\n\t\ta[i].byt = (byte)(i)\n\t}\n\tfor _, c := range s {\n\t\ta[c].freq++\n\t}\n\tsort.Sort(FreqItemSlice(a))\n\n\tvar buf bytes.Buffer\n\tfor _, x := range a {\n\t\tfor x.freq > 0 {\n\t\t\tbuf.WriteByte(x.byt)\n\t\t\tx.freq--\n\t\t}\n\t}\n\treturn buf.String()\n}", "func MySort(content []string, options sortOpts) []string {\n\twordSelector := getWordSelector(options.col)\n\treverser := getReverser(options.reverse)\n\n\tsort.SliceStable(content, func (i, j int) bool {\n\t\tlhs, rhs := wordSelector(content[i]), wordSelector(content[j])\n\n\t\tif options.numeric {\n\t\t\tlhsInt, err := strconv.ParseInt(lhs, 10, 32)\n\n\t\t\tif err == nil {\n\t\t\t\trhsInt, err := strconv.ParseInt(rhs, 10, 32)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn reverser(lhsInt < rhsInt)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tif options.reg {\n\t\t\tlhs, rhs = strings.ToLower(lhs), strings.ToLower(rhs)\n\t\t}\n\n\t\treturn reverser(lhs < rhs)\n\t})\n\n\tif options.first {\n\t\tcontent = filterUnique(content, options.reg)\n\t}\n\n\treturn content\n}", "func (s UserAuthPasswordArray) Sort(less func(a, b UserAuthPassword) bool) UserAuthPasswordArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func SortStringList(x Expr) {\n\tlist, ok := x.(*ListExpr)\n\tif !ok || len(list.List) < 2 {\n\t\treturn\n\t}\n\n\tif doNotSort(list.List[0]) {\n\t\tlist.List = deduplicateStringExprs(list.List)\n\t\treturn\n\t}\n\n\tforceSort := keepSorted(list) || keepSorted(list.List[0])\n\n\t// TODO(bazel-team): Decide how to recognize lists that cannot\n\t// be sorted. Avoiding all lists with comments avoids sorting\n\t// lists that say explicitly, in some form or another, why they\n\t// cannot be sorted. For example, many cc_test rules require\n\t// certain order in their deps attributes.\n\tif !forceSort {\n\t\tif line, _ := hasComments(list); line {\n\t\t\tdeduplicateStringList(list)\n\t\t\treturn\n\t\t}\n\t}\n\n\tlist.List = sortStringExprs(list.List)\n}", "func (l *Loader) SortSub(s Sym) Sym {\n\n\tif s == 0 || l.sub[s] == 0 {\n\t\treturn s\n\t}\n\n\t// Sort symbols using a slice first. Use a stable sort on the off\n\t// chance that there's more than once symbol with the same value,\n\t// so as to preserve reproducible builds.\n\tsl := []symWithVal{}\n\tfor ss := l.sub[s]; ss != 0; ss = l.sub[ss] {\n\t\tsl = append(sl, symWithVal{s: ss, v: l.SymValue(ss)})\n\t}\n\tsort.Stable(bySymValue(sl))\n\n\t// Then apply any changes needed to the sub map.\n\tns := Sym(0)\n\tfor i := len(sl) - 1; i >= 0; i-- {\n\t\ts := sl[i].s\n\t\tl.sub[s] = ns\n\t\tns = s\n\t}\n\n\t// Update sub for outer symbol, then return\n\tl.sub[s] = sl[0].s\n\treturn sl[0].s\n}", "func (s EmojiListArray) Sort(less func(a, b EmojiList) bool) EmojiListArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func SortString(str string) string {\n\tr := []rune(str)\n\tsort.Sort(runesStr(r))\n\treturn string(r)\n}", "func SelectionSortString(list []string) {\n\tfor i := 0; i < len(list); i++ {\n\t\tlowestIndex := i\n\n\t\tfor j := i + 1; j < len(list); j++ {\n\t\t\tif list[j] < list[lowestIndex] {\n\t\t\t\tlowestIndex = j\n\t\t\t}\n\t\t}\n\t\tlist[i], list[lowestIndex] = list[lowestIndex], list[i]\n\t}\n}", "func (x ComparableSlice[T]) Sort() { sort.Sort(x) }", "func SortString(s string) string {\n\tsr := SortableString(s)\n\tsort.Sort(sr)\n\n\treturn string(sr)\n}", "func (s IPPortSecretArray) Sort(less func(a, b IPPortSecret) bool) IPPortSecretArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (s InputSecureFileUploadedArray) Sort(less func(a, b InputSecureFileUploaded) bool) InputSecureFileUploadedArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func sortNames(slice []string) []string {\n\tsort.Sort(StringSlice(slice))\n\treturn slice\n}", "func execSearchStrings(_ int, p *gop.Context) {\n\targs := p.GetArgs(2)\n\tret := sort.SearchStrings(args[0].([]string), args[1].(string))\n\tp.Ret(2, ret)\n}", "func SortByDistance(base string, paths []string, baseDist Distance) {\n\ttype elem struct {\n\t\tdist Distance\n\t\tpath string\n\t}\n\n\telems := make([]elem, len(paths))\n\tfor i, p := range paths {\n\t\telems[i] = elem{\n\t\t\tDistanceBetween(base, p, baseDist),\n\t\t\tp,\n\t\t}\n\t}\n\n\tsort.Slice(elems, func(i, j int) bool {\n\t\tei, ej := elems[i], elems[j]\n\n\t\tif ei.dist == ej.dist {\n\t\t\treturn strings.Compare(ei.path, ej.path) < 0\n\t\t}\n\n\t\treturn ei.dist < ej.dist\n\t})\n\n\tfor idx, elem := range elems {\n\t\tpaths[idx] = elem.path\n\t}\n}", "func (s Set) Sort() Set {\r\n\tif len(s) < 2 {\r\n\t\treturn s\r\n\t}\r\n\r\n\tleft, right := 0, len(s)-1\r\n\tpivotIndex := len(s) / 2\r\n\r\n\ts[pivotIndex], s[right] = s[right], s[pivotIndex]\r\n\r\n\tfor i := range s {\r\n\t\tif s[i] < s[right] {\r\n\t\t\ts[i], s[left] = s[left], s[i]\r\n\t\t\tleft++\r\n\t\t}\r\n\t}\r\n\r\n\ts[left], s[right] = s[right], s[left]\r\n\r\n\t// Recursively call sort\r\n\tSet(s[:left]).Sort()\r\n\tSet(s[left+1:]).Sort()\r\n\r\n\treturn s\r\n}", "func (s Select) Sort(sorts ...string) Select {\n\ts.BaseQuery.sort = buildSort(sorts...)\n\ts.BaseQuery.hasSort = true\n\n\treturn s\n}", "func SortFunc[S ~[]E, E any](x S, cmp func(a, b E) int) { expslices.SortFunc(x, cmp) }", "func sortWord(word string) string {\n\tw := []rune(word)\n\trs := RuneSlice(w)\n\tsort.Sort(rs)\n\treturn string(rs)\n}", "func StringsDo(op Op, s []string, t ...string) []string {\n\tdata := sort.StringSlice(append(s, t...))\n\tn := op(data, len(s))\n\treturn data[:n]\n}", "func (s URLAuthResultRequestArray) Sort(less func(a, b URLAuthResultRequest) bool) URLAuthResultRequestArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (r IPRules) Sort() IPRules {\n\tsort.SliceStable(r, func(i, j int) bool {\n\t\tif r[i].Re.Pattern != nil && r[j].Re.Pattern != nil {\n\t\t\treturn *r[i].Re.Pattern < *r[j].Re.Pattern\n\t\t}\n\t\tif r[i].Re.Pattern != nil {\n\t\t\treturn true\n\t\t}\n\t\treturn false\n\t})\n\n\treturn r\n}", "func execmStringSliceLess(_ int, p *gop.Context) {\n\targs := p.GetArgs(3)\n\tret := args[0].(sort.StringSlice).Less(args[1].(int), args[2].(int))\n\tp.Ret(3, ret)\n}", "func (s ByteSlice) Sort() {\n\tsort.Sort(s)\n}", "func main() {\n\n\tex1 := []string{\"hello\", \"world\", \"after\", \"all\"}//after\n\tsort.Sort(sort.Reverse(ByLength(ex1)))\n\tfmt.Println(ex1[2])\n\tex2 := []string{\"hello\",\"world\",\"before\",\"all\"}//world\n\tsort.Sort(sort.Reverse(ByLength(ex2)))\n\tfmt.Println(ex2[2])\n}", "func (s URLAuthResultClassArray) Sort(less func(a, b URLAuthResultClass) bool) URLAuthResultClassArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func main() {\n fruits := []string{\"peach\", \"banana\", \"kiwi\"}\n sort.Sort(byLength(fruits))\n fmt.Println(fruits)\n}", "func safeSort(in []string) []string {\n\tif sort.StringsAreSorted(in) {\n\t\treturn in\n\t}\n\tout := make([]string, len(in))\n\tcopy(out, in)\n\tsort.Strings(out)\n\treturn out\n}", "func (res *Result) Sort() *Result {\n\tslices.SortFunc(res.rows, func(a, b []string) int {\n\t\treturn slices.Compare(a, b)\n\t})\n\treturn res\n}", "func (this *Sorter) StoreStrings(dest List) <-chan int {\n\tthis.storeIn(dest.key)\n\treturn IntCommand(this.key, this.key.args(\"sort\", this.sortargs()...)...)\n}", "func ComparatorString(a, b interface{}) int {\n\treturn strings.Compare(qn_conv.String(a), qn_conv.String(b))\n}", "func (s SecretVersions) Sort() {\n\tsort.Sort(s)\n}", "func (s RuneSlice) Sort() {\n\tsort.Sort(s)\n}", "func (p Int64Slice) Sort() { sort.Sort(p) }", "func (s EncryptedChatArray) Sort(less func(a, b EncryptedChat) bool) EncryptedChatArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (s SecurePlainEmailArray) Sort(less func(a, b SecurePlainEmail) bool) SecurePlainEmailArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func SortString(s string) string {\n\tr := []rune(s)\n\tsort.Sort(sortRunes(r))\n\treturn string(r)\n}", "func SortAndConcat(param map[string]string) string {\n\tvar keys []string\n\tfor k := range param {\n\t\tkeys = append(keys, k)\n\t}\n\n\tvar sortedParam []string\n\tsort.Strings(keys)\n\tfor _, k := range keys {\n\t\t// fmt.Println(k, \"=\", param[k])\n\t\tsortedParam = append(sortedParam, k+\"=\"+param[k])\n\t}\n\n\treturn strings.Join(sortedParam, \"&\")\n}", "func ExampleOrd() {\n\tvar (\n\t\tstr = `goframe`\n\t\tresult = gstr.Ord(str)\n\t)\n\n\tfmt.Println(result)\n\n\t// Output:\n\t// 103\n}", "func (c *Commands) Sort() {\n\tsort.Slice(c.list, func(i, j int) bool {\n\t\treturn c.list[i].Name < c.list[j].Name\n\t})\n}", "func (s InputSecureFileClassArray) Sort(less func(a, b InputSecureFileClass) bool) InputSecureFileClassArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (s HelpAppUpdateArray) Sort(less func(a, b HelpAppUpdate) bool) HelpAppUpdateArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (s FileLocationUnavailableArray) Sort(less func(a, b FileLocationUnavailable) bool) FileLocationUnavailableArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (s BigIntSlice) Sort() {\n\tsort.Sort(s)\n}", "func (s FileLocationClassArray) Sort(less func(a, b FileLocationClass) bool) FileLocationClassArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func main() {\n\tfruits := []string{\"peach\", \"banana\", \"kiwi\"}\n\t// Convert original fruits slice to byLength and then use sort.Sort on the typed slice\n\tsort.Sort(byLength(fruits))\n\tfmt.Println(fruits)\n}", "func (s String) Sorted() []string {\n\tslice := s.ToSlice()\n\tsort.Strings(slice)\n\treturn slice\n}", "func frequencySort(s string) string {\r\n\thT := make(map[string]int)\r\n\tvar res = \"\"\r\n\tfor i, _ := range s {\r\n\t\tss := s[i:i+1]\r\n\t\tif _, ok := hT[ss]; ok {\r\n\t\t\thT[ss] += 1\r\n\t\t} else {\r\n\t\t\thT[ss] = 1\r\n\t\t}\r\n\t}\r\n\tl := len(hT)\r\n\tfor i := 0; i < l; i++ {\r\n\t\tmax, key := ^int(^uint(0) >> 1), \"\"\r\n\t\tfor k, v := range hT {\r\n\t\t\tif max < v {\r\n\t\t\t\tmax, key = v, k\r\n\t\t\t}\r\n\t\t}\r\n\t\tfor j := 0; j < max; j++ {\r\n\t\t\tres += key\r\n\t\t}\r\n\t\tdelete(hT, key)\r\n\t}\r\n\treturn res\r\n}", "func (s SecurePlainDataClassArray) Sort(less func(a, b SecurePlainDataClass) bool) SecurePlainDataClassArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func NewWithStringComparator(values ...interface{}) *Set {\n\tset := &Set{ValueType: utils.TypeString, tree: rbt.NewWithStringComparator()}\n\tif len(values) > 0 {\n\t\tset.Add(values...)\n\t}\n\treturn set\n}", "func DedupeSortSlice(s []string, modifier func(string) string) []string {\n\to := DedupeSlice(s, modifier)\n\tsort.Strings(o)\n\treturn o\n}", "func (s EncryptedChatRequestedArray) Sort(less func(a, b EncryptedChatRequested) bool) EncryptedChatRequestedArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func SortString(s string) string {\n\tr := []rune(s)\n\tsort.Sort(sortChars(r))\n\treturn string(r)\n}", "func (s BigRatSlice) Sort() {\n\tsort.Sort(s)\n}", "func sortStringByChar(s string) string {\n\tsChars := []rune(s)\n\tsort.Sort(sortChars(sChars))\n\treturn string(sChars)\n}", "func (s UserAuthArray) Sort(less func(a, b UserAuth) bool) UserAuthArray {\n\tsort.Slice(s, func(i, j int) bool {\n\t\treturn less(s[i], s[j])\n\t})\n\treturn s\n}", "func (p SliceUI) Sort() { sort.Sort(p) }" ]
[ "0.7126395", "0.671989", "0.62521285", "0.61218864", "0.6060801", "0.5976071", "0.59110546", "0.57168454", "0.5669356", "0.5645189", "0.55945086", "0.5568044", "0.5536487", "0.54952943", "0.5461731", "0.5431923", "0.53972733", "0.53659576", "0.53563726", "0.5344095", "0.5343517", "0.53407377", "0.5322107", "0.5319862", "0.5305165", "0.530322", "0.5297083", "0.5294941", "0.52900124", "0.52805626", "0.52669984", "0.52541983", "0.5236875", "0.5226516", "0.52053714", "0.51785654", "0.5171951", "0.5165217", "0.5161143", "0.51609796", "0.5140017", "0.51343197", "0.51313365", "0.5127159", "0.51137525", "0.51082635", "0.5108246", "0.5096971", "0.5079956", "0.5078578", "0.5074999", "0.50722045", "0.5068833", "0.5058161", "0.50362533", "0.5017927", "0.50123066", "0.4998622", "0.49977073", "0.49791163", "0.4974581", "0.49711964", "0.4953405", "0.49515152", "0.49481302", "0.49369904", "0.49293572", "0.49249822", "0.4915513", "0.48983854", "0.48968667", "0.48922902", "0.48889813", "0.48660752", "0.48557225", "0.4846353", "0.4843504", "0.48417336", "0.4831341", "0.4825043", "0.48168245", "0.48139098", "0.48135695", "0.48109597", "0.48024952", "0.48002094", "0.4799485", "0.47902763", "0.47871104", "0.4785578", "0.47788942", "0.4777676", "0.4766489", "0.4753341", "0.47526246", "0.47455758", "0.47297746", "0.47296008", "0.4725", "0.472319" ]
0.7650956
0
Tail returns the last n elements of operand. If less than n elements are in operand, it returns all of operand.
func Tail(n int, operand []string) []string { if len(operand) < n { return operand } return operand[len(operand)-n:] }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (ts TrickSlice) Last(n int) TrickSlice {\n\tv := reflect.Value(ts)\n\tif n > v.Len() {\n\t\tn = v.Len()\n\t}\n\treturn TrickSlice(v.Slice3(v.Len()-n, v.Len(), v.Len()))\n}", "func Head(n int, operand []string) []string {\n\tif len(operand) < n {\n\t\treturn operand\n\t}\n\treturn operand[:n]\n}", "func (s *exprStack) popN(n int) []Expression {\n\ttrace_util_0.Count(_util_00000, 166)\n\tif n > s.len() || n < 0 {\n\t\ttrace_util_0.Count(_util_00000, 168)\n\t\tn = s.len()\n\t}\n\ttrace_util_0.Count(_util_00000, 167)\n\tidx := s.len() - n\n\texprs := s.stack[idx:]\n\ts.stack = s.stack[:idx]\n\treturn exprs\n}", "func Tail(l List) List {\n\tif l == nil {\n\t\tl = Mzero()\n\t}\n\tif _, ok := l.(uint); ok {\n\t\treturn unit{}\n\t}\n\tll := l.([2]interface{})\n\tf := ll[1].(func() List)\n\treturn f()\n}", "func (*Functions) Tail(start int, array interface{}) interface{} {\n\tv := reflect.ValueOf(array)\n\tc := v.Len() - start\n\tout := reflect.MakeSlice(v.Type(), c, c)\n\tfor i := 0; i < c; i++ {\n\t\tout.Index(i).Set(v.Index(start + i))\n\t}\n\treturn out.Interface()\n}", "func Tail(arr interface{}) interface{} {\n\tvalue := redirectValue(reflect.ValueOf(arr))\n\tvalueType := value.Type()\n\n\tkind := value.Kind()\n\n\tif kind == reflect.Array || kind == reflect.Slice {\n\t\tlength := value.Len()\n\n\t\tif length <= 1 {\n\t\t\treturn arr\n\t\t}\n\n\t\treturn value.Slice(1, length).Interface()\n\t}\n\n\tpanic(fmt.Sprintf(\"Type %s is not supported by Initial\", valueType.String()))\n}", "func (a Slice[T]) Lasts(count int) Slice[T] {\n\treturn a[len(a)-count:]\n}", "func Tail(data interface{}) (interface{}, error) {\n\tvar err error\n\n\tresult := func(err *error) interface{} {\n\t\tdefer catch(err)\n\n\t\tif !isNonNilData(err, \"data\", data) {\n\t\t\treturn nil\n\t\t}\n\n\t\tdataValue, dataType, _, dataValueLen := inspectData(data)\n\n\t\tif !isSlice(err, \"data\", dataValue) {\n\t\t\treturn nil\n\t\t}\n\n\t\tif dataValueLen == 0 {\n\t\t\treturn makeSlice(dataType).Interface()\n\t\t}\n\n\t\tresult := dataValue.Slice(1, dataValueLen)\n\t\treturn result.Interface()\n\t}(&err)\n\n\treturn result, err\n}", "func (a Arguments) Tail() Arguments {\n\tif len(a.args) <= 1 {\n\t\treturn Arguments{bin: a.bin, args: []string{}}\n\t}\n\treturn Arguments{bin: a.bin, args: a.args[1:]}\n}", "func (q *OperationQueue) Top(n int) []*SignedOperation {\n\tanswer := []*SignedOperation{}\n\tfor _, item := range q.set.Values() {\n\t\tanswer = append(answer, item.(*SignedOperation))\n\t\tif len(answer) == n {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn answer\n}", "func tailRecursive(number int, product int) int {\n\tproduct = product * number\n\n\tif number == 1 {\n\t\treturn product\n\t}\n\treturn tailRecursive(number-1, product)\n}", "func (this *MyCircularQueue) Rear() int {\n if this.IsEmpty() { return -1 }\n return this.vals[(this.tail+this.n-1)%this.n] // Note\n}", "func (list *TList) Tail() *list.Element {\n\treturn list.list.Back()\n}", "func (s *LogStore) Tail(n int) string {\n\treturn s.tailHelper(n, s.spans, true)\n}", "func Right(arr []int, n int) {\n\tmisc.Reverse(arr[n:])\n\tmisc.Reverse(arr[:n])\n\tmisc.Reverse(arr[:])\n}", "func removeNthFromEndSlice(head *ListNode, n int) *ListNode {\n\tif head == nil || head.Next == nil {\n\t\treturn nil\n\t}\n\ts := []*ListNode{}\n\tfor {\n\t\ts = append(s, head)\n\t\thead = head.Next\n\t\tif head == nil {\n\t\t\tbreak\n\t\t}\n\t}\n\tx := len(s) - n\n\tif x > 0 {\n\t\ts[x-1].Next = s[x].Next\n\t\treturn s[0]\n\t}\n\treturn s[1]\n}", "func (p *Permutator) NextN(count int) interface{} { \n\tif count <= 0 || p.left() == 0 {\n\t\treturn reflect.MakeSlice(reflect.SliceOf(p.value.Type()), 0, 0).Interface()\n\t}\n \n cap := p.left()\n\tif cap > count {\n\t\tcap = count\n\t}\n\n result := reflect.MakeSlice(reflect.SliceOf(p.value.Type()), cap, cap)\n\n length := 0 \n for index := 0; index < cap; index++ { \n if _, ok := p.Next(); ok {\n length++\n list := p.copySliceValue()\n result.Index(index).Set(list)\n }\n }\n\n list := reflect.MakeSlice(result.Type(), length, length)\n reflect.Copy(list, result)\n \n return list.Interface()\n}", "func (l *List) Tail() interface{} {\n\tl.RLock()\n\tdefer l.RUnlock()\n\n\tif l.length == 0 {\n\t\treturn nil\n\t}\n\treturn l.tail.value\n}", "func (q *Queue) Tail() uint64 { return q.tail }", "func Last(arr interface{}) interface{} {\n\tvalue := redirectValue(reflect.ValueOf(arr))\n\tvalueType := value.Type()\n\n\tkind := value.Kind()\n\n\tif kind == reflect.Array || kind == reflect.Slice {\n\t\tif value.Len() == 0 {\n\t\t\treturn nil\n\t\t}\n\n\t\treturn value.Index(value.Len() - 1).Interface()\n\t}\n\n\tpanic(fmt.Sprintf(\"Type %s is not supported by Last\", valueType.String()))\n}", "func TakeRight(data interface{}, size int) (interface{}, error) {\n\tvar err error\n\n\tresult := func(err *error) interface{} {\n\t\tdefer catch(err)\n\n\t\tif !isNonNilData(err, \"data\", data) {\n\t\t\treturn nil\n\t\t}\n\n\t\tdataValue, dataType, _, dataValueLen := inspectData(data)\n\n\t\tif !isSlice(err, \"data\", dataValue) {\n\t\t\treturn nil\n\t\t}\n\n\t\tif !isZeroOrPositiveNumber(err, \"size\", size) {\n\t\t\treturn data\n\t\t}\n\n\t\tresult := makeSlice(dataType)\n\n\t\tif dataValueLen == 0 {\n\t\t\treturn result.Interface()\n\t\t}\n\n\t\tforEachSlice(dataValue, dataValueLen, func(each reflect.Value, i int) {\n\t\t\tif i >= (dataValueLen - size) {\n\t\t\t\tresult = reflect.Append(result, each)\n\t\t\t}\n\t\t})\n\n\t\treturn result.Interface()\n\t}(&err)\n\n\treturn result, err\n}", "func tailCall(i Instruction, ls *LuaState) {\n\ta, b, _ := i.ABC()\n\ta += 1\n\n\t// todo: optimize tail call!\n\tc := 0\n\tnArgs := _pushFuncAndArgs(a, b, ls)\n\tls.Call(nArgs, c-1)\n\t_popResults(a, c, ls)\n}", "func Last(obj interface{}) interface{} {\n\ttypeOfObj := reflect.TypeOf(obj)\n\tvalueOfObj := reflect.ValueOf(obj)\n\tif typeOfObj.Kind() != reflect.Array && typeOfObj.Kind() != reflect.Slice {\n\t\tpanic(\"make sure obj is array or slice\")\n\t}\n\tif valueOfObj.Len() == 0 {\n\t\tpanic(\"make sure obj is array not empty\")\n\t}\n\n\treturn valueOfObj.Index(valueOfObj.Len() - 1).Interface()\n}", "func Last(data interface{}) (interface{}, error) {\n\tvar err error\n\n\tresult := func(err *error) interface{} {\n\t\tdefer catch(err)\n\n\t\tif !isNonNilData(err, \"data\", data) {\n\t\t\treturn nil\n\t\t}\n\n\t\tdataValue, _, _, dataValueLen := inspectData(data)\n\n\t\tif !isSlice(err, \"data\", dataValue) {\n\t\t\treturn nil\n\t\t}\n\n\t\tif dataValueLen == 0 {\n\t\t\treturn nil\n\t\t}\n\n\t\treturn dataValue.Index(dataValueLen - 1).Interface()\n\t}(&err)\n\n\treturn result, err\n}", "func (a Slice[T]) Last() *T {\n\treturn a.At(len(a) - 1)\n}", "func NthRest(seq Seq, n uint) Seq {\n\treturn Drop(seq, n)\n}", "func sliceForAppend(in []byte, n int) (head, tail []byte) {\n\tif total := len(in) + n; cap(in) >= total {\n\t\thead = in[:total]\n\t} else {\n\t\thead = make([]byte, total)\n\t\tcopy(head, in)\n\t}\n\ttail = head[len(in):]\n\treturn\n}", "func removeNthFromEnd(head *ListNode, n int) *ListNode {\n\tdummyHead := ListNode{Next: head}\n\n\tvar length int\n\tfor cursor := &dummyHead; cursor.Next != nil; cursor = cursor.Next {\n\t\tlength++\n\t}\n\n\tif n <= 0 || n > length {\n\t\treturn dummyHead.Next\n\t}\n\n\tpreIdx := length - n\n\tpreNode := &dummyHead\n\tfor i := 0; i < preIdx; i++ {\n\t\tpreNode = preNode.Next\n\t}\n\n\tdelNode := preNode.Next\n\tpreNode.Next = delNode.Next\n\tdelNode.Next = nil // avoid memory leaks\n\n\treturn dummyHead.Next\n}", "func TestRemoveNthFromEnd(t *testing.T) {\n\thead := GenListNode([]int{1, 2, 3, 4, 5})\n\tresult := removeNthFromEnd(head, 2)\n\tLogListNode(result)\n}", "func removeNthFromEnd(head *ListNode, n int) *ListNode {\n\ttemp := head\n\tlist := []*ListNode{}\n\tfor temp != nil {\n\t\tlist = append(list, temp)\n\t\ttemp = temp.Next\n\t}\n\tlength := len(list)\n\tthisOne := list[length - n]\n\tbeforeIdx := length - n - 1;\n\tif beforeIdx < 0 {\n\t\treturn thisOne.Next\n\t}\n\tlist[beforeIdx].Next = thisOne.Next\n\treturn head\n}", "func (this *Tuple) Right(n int) *Tuple {\n\tlength := this.Len()\n\tn = max(0, length-n)\n\treturn this.Slice(n, length)\n}", "func (p *Stack) PopNArgs(arity int) []interface{} {\n\n\tpstk := p.data\n\tn := len(pstk)\n\tif n >= arity {\n\t\targs := make([]interface{}, arity)\n\t\tn -= arity\n\t\tfor i := 0; i < arity; i++ {\n\t\t\targs[i] = pstk[n+i]\n\t\t}\n\t\tp.data = pstk[:n]\n\t\treturn args\n\t}\n\tpanic(\"unexpected argument count\")\n}", "func (l List) Last() interface{} {\n\tn := l.Head\n\tfor n.Next != nil {\n\t\tn = n.Next\n\t}\n\treturn n.Data\n}", "func DropRight[T any](collection []T, n int) []T {\n\tif len(collection) <= n {\n\t\treturn make([]T, 0)\n\t}\n\n\tresult := make([]T, len(collection)-n)\n\tfor i := len(collection) - 1 - n; i >= 0; i-- {\n\t\tresult[i] = collection[i]\n\t}\n\n\treturn result\n}", "func lastRemaining(n int, m int) int {\n\ts := make([]int, n)\n\tfor i := 0; i < n; i++ {\n\t\ts[i] = i\n\t}\n\tresult := 0\n\tindex := 1\n\tpoint := 0 //指向数组\n\tfor {\n\t\tif len(s) == 1 {\n\t\t\tresult = s[0]\n\t\t\tbreak\n\t\t}\n\t\tif index == m {\n\t\t\ttemp := make([]int, point)\n\t\t\tcopy(temp, s[:point])\n\t\t\ts = append(temp, s[point+1:]...)\n\t\t\tindex = 1\n\t\t\tpoint--\n\t\t} else {\n\t\t\tindex++\n\t\t}\n\t\tif point + 1 >= len(s) {\n\t\t\tpoint = 0\n\t\t} else {\n\t\t\tpoint++\n\t\t}\n\t}\n\treturn result\n}", "func (l *List) Last() interface{} {\n\tif l.len != 0 {\n\t\treturn l.tail.Value()\n\t}\n\treturn nil\n}", "func (p *SliceOfMap) LastN(n int) ISlice {\n\tif n == 0 {\n\t\treturn NewSliceOfMapV()\n\t}\n\treturn p.Slice(absNeg(n), -1)\n}", "func (r EventStorageAdapter) PopN(n int64) ([]api.RedisStoredEventDTO, error) {\n\n\ttoReturn := make([]api.RedisStoredEventDTO, 0)\n\n\telMutex.Lock()\n\tlrange := r.client.LRange(r.eventsListNamespace(), 0, n-1)\n\tif lrange.Err() != nil {\n\t\tlog.Error.Println(\"Fetching events\", lrange.Err().Error())\n\t\telMutex.Unlock()\n\t\treturn nil, lrange.Err()\n\t}\n\ttotalFetchedEvents := int64(len(lrange.Val()))\n\n\tidxFrom := n\n\tif totalFetchedEvents < n {\n\t\tidxFrom = totalFetchedEvents\n\t}\n\n\tres := r.client.LTrim(r.eventsListNamespace(), idxFrom, -1)\n\tif res.Err() != nil {\n\t\tlog.Error.Println(\"Trim events\", res.Err().Error())\n\t\telMutex.Unlock()\n\t\treturn nil, res.Err()\n\t}\n\telMutex.Unlock()\n\n\t//JSON unmarshal\n\tlistOfEvents := lrange.Val()\n\tfor _, se := range listOfEvents {\n\t\tstoredEventDTO := api.RedisStoredEventDTO{}\n\t\terr := json.Unmarshal([]byte(se), &storedEventDTO)\n\t\tif err != nil {\n\t\t\tlog.Error.Println(\"Error decoding event JSON\", err.Error())\n\t\t\tcontinue\n\t\t}\n\t\ttoReturn = append(toReturn, storedEventDTO)\n\t}\n\n\treturn toReturn, nil\n}", "func TakeLast(values ts.Datapoints) float64 {\n\tfor i := len(values) - 1; i >= 0; i-- {\n\t\tif !math.IsNaN(values[i].Value) {\n\t\t\treturn values[i].Value\n\t\t}\n\t}\n\n\treturn math.NaN()\n}", "func (t *Token) Last() *Token {\n\ttmp := t\n\tfor tmp.Next != nil {\n\t\ttmp = tmp.Next\n\t}\n\treturn tmp\n}", "func (ns Nodes) Last() Nodes {\n\tl := len(ns)\n\tif l < 2 {\n\t\treturn ns\n\t}\n\treturn Nodes{ns[l-1]}\n}", "func (tp *Tiles) TakeLast() *TilePlacement {\n\tlength := len(*tp)\n\tif length == 0 {\n\t\treturn nil\n\t}\n\n\tp := &(*tp)[length-1]\n\t*tp = (*tp)[:length-1]\n\treturn p\n}", "func Last(input interface{}, data map[string]interface{}) interface{} {\n\tvalue := reflect.ValueOf(input)\n\tkind := value.Kind()\n\n\tif kind != reflect.Array && kind != reflect.Slice {\n\t\treturn input\n\t}\n\tlen := value.Len()\n\tif len == 0 {\n\t\treturn input\n\t}\n\treturn value.Index(len - 1).Interface()\n}", "func removeNthFromEnd(head *ListNode, n int) *ListNode {\n\tif head == nil || n == 0 {\n\t\treturn nil\n\t}\n\tvar (\n\t\tend = head\n\t\ti = 1\n\t)\n\tfor ; i <= n && end.Next != nil; i++ {\n\t\tend = end.Next\n\t}\n\tpre := head\n\tfor end.Next != nil {\n\t\tend = end.Next\n\t\tpre = pre.Next\n\t}\n\tif pre.Next != nil && i != n {\n\t\tpre.Next = pre.Next.Next\n\t} else {\n\t\treturn head.Next\n\t}\n\treturn head\n}", "func (me Tokens) Last(matches func(*Token) bool) *Token {\n\tif len(me) == 0 {\n\t\treturn nil\n\t}\n\tif matches != nil {\n\t\tfor i := len(me) - 1; i >= 0; i-- {\n\t\t\tif t := &me[i]; matches(t) {\n\t\t\t\treturn t\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\treturn &me[len(me)-1]\n}", "func RemoveNthFromEnd(head *ListNode, n int) *ListNode {\n\treturn removeNthFromEnd(head, n)\n}", "func (s *Stack[T]) Last() (T, bool) {\n\tif s.IsEmpty() {\n\t\treturn *new(T), false\n\t}\n\treturn (*s)[(len(*s) - 1)], true\n}", "func (x *Lazy) Take(n int, ar AnyValue) {\n\toar := reflect.ValueOf(ar)\n\tfor i := 0; i < n; i++ {\n\t\tif v, ok := x.omap[i]; ok {\n\t\t\toar.Index(i).Set(v)\n\t\t\tcontinue\n\t\t}\n\t\tvar v = []reflect.Value{x.iar.Index(i)}\n\t\tfor j := 0; j < len(x.fns); j++ {\n\t\t\tv = x.fns[j].Call(v)\n\t\t}\n\t\toar.Index(i).Set(v[0])\n\t\tx.omap[i] = oar.Index(i)\n\t}\n}", "func tail(log []string, n int) string {\n\tw := out.NewBufferWriter()\n\tn = int(math.Min(float64(n), 1000))\n\ttailN := int(math.Max(0, float64(len(log)-n)))\n\tfor _, line := range log[tailN:] {\n\t\tif strings.Contains(line, \"ERROR\") || strings.Contains(line, \"Error:\") {\n\t\t\tout.Reply(w, out.Red)\n\t\t} else if strings.Contains(line, \"WARN\") {\n\t\t\tout.Reply(w, out.Yellow)\n\t\t} else {\n\t\t\tout.Reply(w, out.Grey)\n\t\t}\n\t\tout.ReplyNL(w, line)\n\t}\n\tout.ReplyNL(w, out.Yellow, fmt.Sprintf(\"[time now %s]\", time.Now().String()))\n\treturn w.String()\n}", "func (iter *radixIterator) Last() {\n\tresIter, err := iter.miTxn.GetReverse(nil)\n\tif err != nil {\n\t\titer.err = err\n\t\treturn\n\t}\n\titer.resIter = resIter\n\titer.isReverser = true\n\titer.Prev()\n}", "func rep(p []float64) []float64 {\n\tif lastOp == nil {\n\t\tfatalf(\"No operator to rep\")\n\t}\n\tswitch lastOp.f.(type) {\n\tcase func(float64, float64) float64:\n\tcase func(int64, int64) int64:\n\tdefault:\n\t\tfatalf(\"Invalid operator for rep\")\n\t}\n\tfor len(stack) > 1 {\n\t\tlastOp.exec()\n\t}\n\treturn stack\n}", "func (l *List) Tail() *Node {\n\t// when the size is 0 return nil because the next node of the root\n\t// code is itself\n\tif l.Size == 0 {\n\t\treturn nil\n\t}\n\treturn l.root.prev\n}", "func last(input interface{}) interface{} {\n\tif input == nil {\n\t\treturn nil\n\t}\n\tarr := reflect.ValueOf(input)\n\tif arr.Len() == 0 {\n\t\treturn nil\n\t}\n\treturn arr.Index(arr.Len() - 1).Interface()\n}", "func (l *List) tail() *List {\n\tnewL := new(List)\n\tnode := l.Head.Next\n\tfor node != nil {\n\t\tnewL.Insert(node.Data)\n\t\tnode = node.Next\n\t}\n\tll := newL.reverse()\n\treturn ll\n}", "func Last10(storage Storage) ([]Block, error) {\n\treturn storage.Last10()\n}", "func SplitAfterN(sep string, n int, operand string) []string {\n\treturn strings.SplitAfterN(operand, sep, n)\n}", "func (root *TreeNode) topN(n int) (res []interface{}) {\n\tif root == nil {\n\t\treturn res\n\t}\n\n\tif root.right != nil {\n\t\ttmp := root.right.topN(n)\n\t\tfor _, k := range tmp {\n\t\t\tif len(res) == n {\n\t\t\t\treturn res\n\t\t\t}\n\n\t\t\tres = append(res, k)\n\t\t}\n\t}\n\n\tif len(res) < n {\n\t\tres = append(res, root.Values...)\n\t}\n\n\tif len(res) == n {\n\t\treturn res\n\t}\n\n\tif root.left != nil {\n\t\ttmp := root.left.topN(n - len(res))\n\t\tfor _, k := range tmp {\n\t\t\tif len(res) == n {\n\t\t\t\treturn res\n\t\t\t}\n\t\t\tres = append(res, k)\n\t\t}\n\t}\n\treturn res\n}", "func Last(collection interface{}) interface{} {\n\ttypeOfCollection := reflect.TypeOf(collection)\n\tif typeOfCollection.Kind() != reflect.Array && typeOfCollection.Kind() != reflect.Slice {\n\t\tpanic(\"collection should be array or slice\")\n\t}\n\tvalueOfCollection := reflect.ValueOf(collection)\n\n\treturn valueOfCollection.Index(valueOfCollection.Len() - 1).Interface()\n}", "func (c *Termination) Get(n int) Value {\n\tif n >= c.pushIndex {\n\t\treturn NilValue\n\t}\n\treturn c.args[n]\n}", "func NthToLast(head *LinkedListNode, k int) *LinkedListNode {\n\tif k <= 0 {\n\t\treturn nil\n\t}\n\n\tp1, p2 := head, head\n\n\tfor i := 0; i < k-1; i++ {\n\t\tif p2 == nil {\n\t\t\treturn nil\n\t\t}\n\t\tp2 = p2.next\n\t}\n\tif p2 == nil {\n\t\treturn nil\n\t}\n\n\tfor p2.next != nil {\n\t\tp1 = p1.next\n\t\tp2 = p2.next\n\t}\n\n\treturn p1\n}", "func HdTail(l List) (interface{}, List) {\n\treturn Head(l), Tail(l)\n}", "func (o *Option) Last() *Option { \n if o == nil { return nil }\n for o.isNotLast { o = o.next }\n return o\n}", "func (l *ListNode) Tail() *ListNode {\n\tn := l\n\tfor {\n\t\tif n.Next == nil {\n\t\t\tbreak\n\t\t}\n\t\tn = n.Next\n\t}\n\treturn n\n}", "func (p *SliceOfMap) DropLastN(n int) ISlice {\n\tif n == 0 {\n\t\treturn p\n\t}\n\treturn p.Drop(absNeg(n), -1)\n}", "func KthToLast(l *list.List, k int) []interface{} {\n\tc := l.Count() - k\n\tif c <= 0 {\n\t\treturn []interface{}{}\n\t}\n\n\tat := l.First()\n\tfor i := 0; i < k && at != nil; i++ {\n\t\tat = at.Next()\n\t}\n\n\tresult := make([]interface{}, 0, c)\n\tfor ; at != nil; at = at.Next() {\n\t\tresult = append(result, at.Value)\n\t}\n\treturn result\n}", "func (l *List) Last() interface{} {\n\tif l.len == 0 {\n\t\treturn 0\n\t}\n\treturn l.lastElement.Value\n}", "func main() {\n\tfmt.Println(getKthFromEnd(&ListNode{\n\t\tVal: 1,\n\t\tNext: &ListNode{\n\t\t\tVal: 2,\n\t\t\tNext: &ListNode{\n\t\t\t\tVal: 3,\n\t\t\t\tNext: &ListNode{\n\t\t\t\t\tVal: 4,\n\t\t\t\t\tNext: &ListNode{\n\t\t\t\t\t\tVal: 5,\n\t\t\t\t\t\tNext: nil,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}, 2))\n}", "func (mes *MarkerEncodingScheme) Tail(b byte, pos int) checked.Bytes { return mes.tails[int(b)][pos-1] }", "func (t *Map) Tail() interface{} {\n\tif t.NotEmpty() {\n\t\treturn t.entries[t.keys.Back().Value].val\n\t}\n\treturn nil\n}", "func (d *Deque) Rest() Sequence {\n\t_, r, _ := d.Split()\n\treturn r\n}", "func (b *Buffer) Get(n int) []interface{} {\n\t// reset any invalid values\n\tif n > b.size || n < 0 {\n\t\tn = b.size\n\t}\n\n\tb.RLock()\n\tdefer b.RUnlock()\n\n\t// create a delta\n\tdelta := b.size - n\n\n\t// if all the values are less than delta\n\tif len(b.vals) < delta {\n\t\treturn b.vals\n\t}\n\n\t// return the delta set\n\treturn b.vals[delta:]\n}", "func (n Nodes) Slice() []*Node", "func (v IntVec) Last() int {\n\treturn v[len(v)-1]\n}", "func LastRemaining2(n, m int) (int, bool) {\n\tif n < 1 || m < 1 || n < m {\n\t\treturn 0, false\n\t}\n\n\tnumRing := ring.New(n)\n\tfor i := 0; i < n; i++ {\n\t\tnumRing.Value = i\n\t\tnumRing = numRing.Next()\n\t}\n\n\tfor numRing.Len() > 1 {\n\t\tfor i := 1; i < m-1; i++ {\n\t\t\tnumRing = numRing.Next()\n\t\t}\n\n\t\tnumRing.Unlink(1)\n\t\tnumRing = numRing.Next()\n\t}\n\treturn numRing.Value.(int), true\n}", "func (i *queueIndex) getTail() (int, int) {\n\taid := i.indexArena.ReadUint64(16)\n\tpos := i.indexArena.ReadUint64(24)\n\treturn int(aid), int(pos)\n}", "func (s *sliding) takeTill(end LogPosition) []Update {\n\treturn s.log[:s.mutable-s.start][:end-s.start]\n}", "func lastEntry(ents []Entry) Entry {\n return ents[len(ents)-1]\n}", "func (s SecurePlainPhoneArray) Last() (v SecurePlainPhone, ok bool) {\n\tif len(s) < 1 {\n\t\treturn\n\t}\n\treturn s[len(s)-1], true\n}", "func SliceNTokens(b []byte, tok byte, n int) (s []byte, nLeft int) {\n\tfor i := range b {\n\t\tif b[i] == tok {\n\t\t\tn--\n\t\t\tif n == 0 { return b[:i+1], 0 }\n\t\t}\n\t}\n\n\treturn b, n\n}", "func (this *MyCircularQueue) Rear() int {\n if this.IsEmpty() {\n return -1\n }\n return this.Items[this.TailIndex]\n}", "func SplitN(sep string, n int, operand string) []string { return strings.SplitN(operand, sep, n) }", "func Bottom(items []Value, count int) []Value {\n\tif len(items) <= count || count == 0 {\n\t\treturn items\n\t}\n\tvalues := valueSlice(items)\n\tsort.Sort(sort.Reverse(values))\n\treturn values[0:count]\n}", "func (c Chain) End() *big.Int {\n\treturn c[len(c)-1]\n}", "func RemoveLastElemFromTop(c *gin.Context) { dequeueTop([]qMessage{}) }", "func (s *MQImpressionsStorage) PopN(n int64) ([]dtos.Impression, error) {\n\tvar toReturn []dtos.Impression\n\tvar totalItems int\n\n\t// Mutexing queue\n\ts.mutexQueue.Lock()\n\tdefer s.mutexQueue.Unlock()\n\n\tif int64(s.queue.Len()) >= n {\n\t\ttotalItems = int(n)\n\t} else {\n\t\ttotalItems = s.queue.Len()\n\t}\n\n\ttoReturn = make([]dtos.Impression, totalItems)\n\tfor i := 0; i < totalItems; i++ {\n\t\ttoReturn[i] = s.queue.Remove(s.queue.Front()).(dtos.Impression)\n\t}\n\n\treturn toReturn, nil\n}", "func (da *DoubleArray) TailLen() int {\n\treturn len(da.tail)\n}", "func (s *Series) End() int64 {\n\tif s.Len() == 0 {\n\t\treturn -1\n\t}\n\treturn s.Start() + int64(s.Step()*(s.Len()-1))\n}", "func (s *SinglyLinkedList) RemoveNthFromEnd(n int) *Node {\n\tdummy := &Node{0, nil}\n\tdummy = s.Head\n\tfirst := dummy\n\tsecond := dummy\n\tfor i := 1; i <= n+1; i++ {\n\t\tfirst = first.Next\n\t}\n\n\tfor first != nil {\n\t\tfirst = first.Next\n\t\tsecond = second.Next\n\t}\n\tsecond.Next = second.Next.Next\n\treturn dummy.Next\n}", "func (v Int32Vec) Last() int32 {\n\treturn v[len(v)-1]\n}", "func (l *List) Last() *Node {\n\treturn l.tail\n}", "func (l *List) Last() *Node {\n\treturn l.tail\n}", "func Nth(data interface{}, i int) (interface{}, error) {\n\tvar err error\n\n\tresult := func(err *error) interface{} {\n\t\tdefer catch(err)\n\n\t\tif !isNonNilData(err, \"data\", data) {\n\t\t\treturn nil\n\t\t}\n\n\t\tdataValue, _, _, dataValueLen := inspectData(data)\n\n\t\tif !isSlice(err, \"data\", dataValue) {\n\t\t\treturn nil\n\t\t}\n\n\t\tif dataValueLen == 0 {\n\t\t\treturn nil\n\t\t}\n\n\t\tif i < 0 {\n\t\t\ti = dataValueLen + i\n\t\t}\n\n\t\tif i < dataValueLen {\n\t\t\treturn dataValue.Index(i).Interface()\n\t\t}\n\n\t\treturn nil\n\t}(&err)\n\n\treturn result, err\n}", "func (d *Deque[T]) Last() *DequeIterator[T] {\n\treturn d.IterAt(d.Size() - 1)\n}", "func (e Expr) LastArg() Expr {\n\treturn e.Args[len(e.Args)-1]\n}", "func (l *List) GetLast(v interface{} /* val */) *El {\n\tcur := l.search(v, false, false)\n\n\tif cur == &l.zero || l.less(cur.val, v) {\n\t\treturn nil\n\t}\n\n\treturn cur\n}", "func (b *QueueBuffer) Next(n int) []byte {\n\tif x := len(*b) - n; x >= 0 {\n\t\tp := make([]byte, n)\n\t\tcopy(p, (*b)[x:])\n\t\t*b = (*b)[:x]\n\t\treturn p\n\t}\n\tp := *b\n\t*b = nil\n\treturn p\n}", "func right(x uint, n uint) uint {\n\tif level(x) == 0 {\n\t\treturn x\n\t}\n\n\tr := x ^ (0x03 << (level(x) - 1))\n\tfor r > 2*(n-1) {\n\t\tr = left(r)\n\t}\n\treturn r\n}", "func (self *Dll) getTail() *node {\n\treturn self.tail\n}", "func lastIP(subnet types.IPNet) net.IP {\n\tvar end net.IP\n\tfor i := 0; i < len(subnet.IP); i++ {\n\t\tend = append(end, subnet.IP[i]|^subnet.Mask[i])\n\t}\n\tif subnet.IP.To4() != nil {\n\t\tend[3]--\n\t}\n\n\treturn end\n}", "func (h headTailIndex) tail() uint32 {\n\treturn uint32(h)\n}" ]
[ "0.5990556", "0.58987325", "0.56341475", "0.5606524", "0.55992526", "0.55949426", "0.5534373", "0.5496476", "0.5344239", "0.530208", "0.5287378", "0.5271894", "0.52395886", "0.5224553", "0.5187527", "0.51836413", "0.5181041", "0.51473564", "0.513726", "0.5094199", "0.5035168", "0.50177073", "0.50142944", "0.4977229", "0.49629915", "0.49519277", "0.4945917", "0.49437147", "0.49324244", "0.4927155", "0.4918332", "0.48905048", "0.48654345", "0.4854677", "0.48433223", "0.48236182", "0.4790729", "0.47623235", "0.47392362", "0.4733173", "0.47242224", "0.4717119", "0.4715268", "0.47127137", "0.4710322", "0.4689629", "0.46870703", "0.46710297", "0.4669212", "0.46627167", "0.46571887", "0.46552265", "0.46530133", "0.46524253", "0.4637771", "0.46285608", "0.4622424", "0.46157634", "0.46105728", "0.46028057", "0.45907447", "0.4587597", "0.45849922", "0.45801625", "0.4579888", "0.45756453", "0.45731026", "0.45652494", "0.45604405", "0.4547976", "0.4538386", "0.4538215", "0.45280516", "0.45234534", "0.45145842", "0.45132187", "0.45111415", "0.44966325", "0.4486776", "0.44854882", "0.44745648", "0.4465821", "0.44653395", "0.44542065", "0.44519722", "0.44519603", "0.44451866", "0.44414082", "0.4431972", "0.44314906", "0.44314906", "0.44268718", "0.4425141", "0.44181234", "0.4417185", "0.44100145", "0.4409049", "0.44007564", "0.43994525", "0.43989855" ]
0.8165838
0
Union returns the union of a and operand. Duplicate elements are removed. Ordering is not preserved.
func Union(a, operand []string) []string { uniq := make(map[string]bool, len(a)+len(operand)) for _, elem := range a { uniq[elem] = true } for _, elem := range operand { uniq[elem] = true } union := make([]string, len(uniq)) i := 0 for k := range uniq { union[i] = k i++ } return union[:i] }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Union[T comparable](a, b Set[T]) Set[T] {\n\tresult := a.New()\n\tresult.Add(a.Elements()...)\n\tresult.Add(b.Elements()...)\n\treturn result\n}", "func (s *QuickUnionSet) Union(a, b int) {\n\tfrom := s.Root(a)\n\tto := s.Root(b)\n\ts.SetID(from, to)\n}", "func (r *Rights) Union(b ...*Rights) *Rights {\n\treturn makeRightsSet(append(b, r)...).rights()\n}", "func (a Possibility) Union(b Possibility) Possibility {\n\tif a == Impossible || b == Impossible {\n\t\treturn Impossible\n\t}\n\tif a.Equals(b) == True {\n\t\treturn a\n\t}\n\treturn Maybe\n}", "func Union(a, b Set) Set {\n\tfor e := b.Enumerator(); e.MoveNext(); {\n\t\ta = a.With(e.Current())\n\t}\n\treturn a\n}", "func Union(elements ...Shape) Shape {\n\treturn Group{space.NoTransformation(), elements}\n}", "func (a ValueAggregation) Union(other ValueAggregation) ValueAggregation {\r\n\treturn ValueAggregation{\r\n\t\tFirstValue: a.FirstValue,\r\n\t\tLastValue: other.LastValue,\r\n\t\tSampleCount: a.SampleCount + other.SampleCount,\r\n\t\tMaximum: maxFloat(a.Maximum, other.Maximum),\r\n\t\tMinimum: minFloat(a.Minimum, other.Minimum),\r\n\t\tSum: a.Sum + other.Sum,\r\n\t\tSumSquare: a.SumSquare + other.SumSquare,\r\n\t\tBuckets: bucketMerge(a.Buckets, other.Buckets),\r\n\t}\r\n}", "func Union(acculator map[int]bool, toAdd map[int]bool) map[int]bool {\n\tif acculator == nil {\n\t\tpanic(\"Cannot perform union on nil accumulator\")\n\t}\n\tif toAdd != nil {\n\t\tfor e := range toAdd {\n\t\t\tacculator[e] = true\n\t\t}\n\t}\n\treturn acculator\n}", "func unionSelector(a, b Selector) Selector {\n\treturn func(n *Node) bool {\n\t\treturn a(n) || b(n)\n\t}\n}", "func (s *Set) Union(s1 *Set) *Set {\n\ts2 := NewSet()\n\tfor _, val := range s.elements {\n\t\ts2.Add(val)\n\t}\n\tfor _, val := range s1.elements {\n\t\ts2.Add(val)\n\t}\n\treturn s2\n}", "func Union(queries ...Query) VariadicQuery {\n\treturn VariadicQuery{\n\t\tOperator: QueryUnion,\n\t\tQueries: queries,\n\t}\n}", "func SliceUnion(a, b []interface{}) []interface{} {\n\tm := make(map[interface{}]bool)\n\n\t// iterate through slice a, adding values as\n\t// keys in m\n\tfor _, v := range a {\n\t\tm[v] = true\n\t}\n\n\t// iterate through slice b, adding values not\n\t// in map m to slice a\n\tfor _, v := range b {\n\t\tif _, ok := m[v]; !ok {\n\t\t\ta = append(a, v)\n\t\t}\n\t}\n\n\t// return union of slices a and b\n\treturn a\n}", "func (wqu *wQuickUnion) Union(p, q int) {\n\ti := wqu.Find(p)\n\tj := wqu.Find(q)\n\tif i != j {\n\t\tif wqu.sz[i] < wqu.sz[j] {\n\t\t\twqu.id[i] = j\n\t\t\twqu.sz[j] += wqu.sz[i]\n\t\t} else {\n\t\t\twqu.id[j] = i\n\t\t\twqu.sz[i] += wqu.sz[j]\n\t\t}\n\t}\n}", "func (self *Rectangle) Union(b *Rectangle) *Rectangle{\n return &Rectangle{self.Object.Call(\"union\", b)}\n}", "func (a seriesIDs) union(other seriesIDs) seriesIDs {\n\tl := a\n\tr := other\n\tids := make([]uint64, 0, len(l)+len(r))\n\tvar i, j int\n\tfor i < len(l) && j < len(r) {\n\t\tif l[i] == r[j] {\n\t\t\tids = append(ids, l[i])\n\t\t\ti++\n\t\t\tj++\n\t\t} else if l[i] < r[j] {\n\t\t\tids = append(ids, l[i])\n\t\t\ti++\n\t\t} else {\n\t\t\tids = append(ids, r[j])\n\t\t\tj++\n\t\t}\n\t}\n\n\t// now append the remainder\n\tif i < len(l) {\n\t\tids = append(ids, l[i:]...)\n\t} else if j < len(r) {\n\t\tids = append(ids, r[j:]...)\n\t}\n\n\treturn ids\n}", "func (b ButtonState) Union(other ButtonState) ButtonState {\n\tvar result ButtonState\n\tresult[0] = b[0] | other[0]\n\tresult[1] = b[1] | other[1]\n\tresult[2] = b[2] | other[2]\n\treturn result\n}", "func (bst *BST) Union(other *BST) []interface{} {\n\treturn set_operation(bst.Root, other.Root, setUnion)\n}", "func (b *BitSet) Union(c *BitSet) *BitSet {\n\tif c == nil {\n\t\treturn nil\n\t}\n\n\tres := new(BitSet)\n\n\tfor key, bValue := range b.set {\n\t\tres.set[key] |= bValue\n\t}\n\tfor key, cValue := range c.set {\n\t\tres.set[key] |= cValue\n\t}\n\treturn res\n}", "func Unionize(lhs, rhs SelectStatement, unionType string, by OrderBy, limit *Limit, lock string) *Union {\n\n\tunion, isUnion := lhs.(*Union)\n\tif isUnion {\n\t\tunion.UnionSelects = append(union.UnionSelects, &UnionSelect{UnionType: unionType, Statement: rhs})\n\t\tunion.OrderBy = by\n\t\tunion.Limit = limit\n\t\tunion.Lock = lock\n\t\treturn union\n\t}\n\n\treturn &Union{FirstStatement: lhs, UnionSelects: []*UnionSelect{{UnionType: unionType, Statement: rhs}}, OrderBy: by, Limit: limit, Lock: lock}\n}", "func (s *Set) Union(other *Set) *Set {\n\tresult := s\n\n\t// Simplest possible solution right now. Would probable be more efficient\n\t// to concatenate two slices of elements from the two sets and create a\n\t// new set from that slice for many cases.\n\tother.Range(func(item Value) bool {\n\t\tresult = result.Add(item)\n\t\treturn true\n\t})\n\n\treturn result\n}", "func Union(s, t Interface) (u Interface) {\n\tu = s.Copy()\n\tfor _, x := range t.Members() {\n\t\tu.Add(x)\n\t}\n\treturn\n}", "func (s *Uint64) Union(other Uint64) Uint64 {\n\tres := NewUint64WithSize(s.Size() + other.Size())\n\n\tfor val := range s.m {\n\t\tres.Add(val)\n\t}\n\tfor val := range other.m {\n\t\tres.Add(val)\n\t}\n\treturn res\n}", "func (s *StrSet) Union(o *StrSet) StrSet {\n\tus := NewStr()\n\tfor el := range s.els {\n\t\tus.Add(el)\n\t}\n\tfor el := range o.els {\n\t\tus.Add(el)\n\t}\n\treturn us\n}", "func union(left, right []string) []string {\n\tu := make([]string, len(left))\n\tcopy(u, left)\noutter:\n\tfor _, r := range right {\n\t\tfor _, l := range left {\n\t\t\tif l == r {\n\t\t\t\tcontinue outter\n\t\t\t}\n\t\t}\n\t\tu = append(u, r)\n\t}\n\treturn u\n}", "func (s *Int64) Union(other Int64) Int64 {\n\tres := NewInt64WithSize(s.Size() + other.Size())\n\n\tfor val := range s.m {\n\t\tres.Add(val)\n\t}\n\tfor val := range other.m {\n\t\tres.Add(val)\n\t}\n\treturn res\n}", "func (s Set) Union(otherSet Set) Set {\n\n\tunionSet := NewSet()\n\n\tfor k := range s.mem {\n\t\tunionSet.mem[k] = true\n\t}\n\t// Adds the elements of the set to the union set.\n\n\tfor k := range otherSet.mem {\n\t\tunionSet.mem[k] = true\n\t}\n\t// Adds the elements of the set in the argument to the union set.\n\n\treturn unionSet\n}", "func (s *set) Union(other Set) Set {\n\tr := NewSet()\n\ts.Foreach(func(x *Term) {\n\t\tr.Add(x)\n\t})\n\tother.Foreach(func(x *Term) {\n\t\tr.Add(x)\n\t})\n\treturn r\n}", "func (oss *OrderedStringSet) Union(setB *OrderedStringSet) *OrderedStringSet {\n\tunion := NewOrderedStringSet(oss.Elements()...)\n\tif setB == nil {\n\t\treturn union\n\t}\n\n\tfor _, m := range setB.Elements() {\n\t\tunion.Add(m)\n\t}\n\n\treturn union\n}", "func (stmt *statement) Union(all bool, query Statement) Statement {\n\tp := posUnion\n\tif len(stmt.parts) > 0 {\n\t\tlast := (&stmt.parts[len(stmt.parts)-1]).position\n\t\tif last >= p {\n\t\t\tp = last + 1\n\t\t}\n\t}\n\tvar index int\n\tif all {\n\t\tindex = stmt.addPart(p, \"UNION ALL \", \"\", query.Args(), \"\")\n\t} else {\n\t\tindex = stmt.addPart(p, \"UNION \", \"\", query.Args(), \"\")\n\t}\n\tpart := &stmt.parts[index]\n\t// Make sure subquery is not dialect-specific.\n\tif query.GetDialect() != DefaultDialect {\n\t\tquery.SetDialect(DefaultDialect)\n\t\tquery.Invalidate()\n\t}\n\tstmt.buffer.WriteString(query.String())\n\tpart.bufHigh = stmt.buffer.Len()\n\t// Close the subquery\n\tquery.Close()\n\n\treturn stmt\n}", "func (s PodSet) Union(o PodSet) PodSet {\n\tresult := PodSet{}\n\tfor key := range s {\n\t\tresult.Insert(key)\n\t}\n\tfor key := range o {\n\t\tresult.Insert(key)\n\t}\n\treturn result\n}", "func Union(subQuery ...SQLProvider) *Query {\n\tq := newQuery()\n\tq.unions = append(q.unions, subQuery...)\n\tq.action = action_union\n\treturn q\n}", "func Union(s1, s2 Set) Set {\n\ttoReturn := New()\n\tfor k := range s1 {\n\t\ttoReturn.Add(k)\n\t}\n\tfor k := range s2 {\n\t\ttoReturn.Add(k)\n\t}\n\treturn toReturn\n}", "func (b *Bag) union(c Bag) Bag {\n\tbag := make(Bag)\n\tfor k, v := range *b {\n\t\tbag[k] += v\n\t}\n\tfor k, v := range c {\n\t\tbag[k] += v\n\t}\n\treturn bag\n}", "func (s Set) Union(others ...Set) Set {\n\tr := make(setUnion, 1, len(others)+1)\n\tr[0] = s.setI\n\tfor _, ss := range others {\n\t\tif ss == None {\n\t\t\tcontinue\n\t\t}\n\t\tif su, ok := ss.setI.(setUnion); ok {\n\t\t\tr = append(r, su...)\n\t\t} else {\n\t\t\tr = append(r, ss.setI)\n\t\t}\n\t}\n\tif len(r) == 1 {\n\t\treturn Set{setI: r[0]}\n\t}\n\treturn Set{setI: r}\n}", "func (s *Set) Union(s2 *Set) *Set {\n\tnew := NewSet() // Pointer to new set\n\n\t// Add all vals in s to new set\n\tfor k := range s.set {\n\t\tnew.AddVal(k)\n\t}\n\n\t// Add vals from s2 not already in new set\n\tfor k := range s2.set {\n\t\t// Will reassign existing vals instead of adding duplicates\n\t\tnew.AddVal(k)\n\t}\n\n\treturn new\n}", "func (s *HashSet) Union(other *HashSet) *HashSet {\n\tresult := s.Clone()\n\tif s != other {\n\t\tresult.AddSet(other)\n\t}\n\treturn result\n}", "func union(a, b []string) [][]rune {\n\tm := make(map[string]bool)\n\tfor _, item := range a {\n\t\tm[item] = true\n\t}\n\tfor _, item := range b {\n\t\tif _, ok := m[item]; !ok {\n\t\t\ta = append(a, item)\n\t\t}\n\t}\n\n\t// Convert a to rune matrix (with x -> words and y -> characters)\n\tout := make([][]rune, len(a))\n\tfor i, word := range a {\n\t\tout[i] = []rune(word)\n\t}\n\treturn out\n}", "func Funion(lista, listb []string) []string {\n\tlistx := append(lista, listb...)\n\tFcompress(&listx)\n\treturn listx\n}", "func (self *Rectangle) Union1O(b *Rectangle, out *Rectangle) *Rectangle{\n return &Rectangle{self.Object.Call(\"union\", b, out)}\n}", "func (set *AppleSet) Union(other *AppleSet) *AppleSet {\n\tif set == nil {\n\t\treturn other\n\t}\n\n\tif other == nil {\n\t\treturn set\n\t}\n\n\tunionedSet := set.Clone()\n\n\tother.s.RLock()\n\tdefer other.s.RUnlock()\n\n\tfor v := range other.m {\n\t\tunionedSet.doAdd(v)\n\t}\n\n\treturn unionedSet\n}", "func (geom Geometry) Union(other Geometry) Geometry {\n\tnewGeom := C.OGR_G_Union(geom.cval, other.cval)\n\treturn Geometry{newGeom}\n}", "func (d *DSU) Union(x, y interface{}) bool {\n\tif !d.Contains(x) || !d.Contains(y) {\n\t\treturn false\n\t}\n\n\tif d.Find(x) == d.Find(y) {\n\t\treturn false\n\t}\n\n\tnodex := d.nodes[d.Find(x)]\n\tnodey := d.nodes[d.Find(y)]\n\n\tif nodex.size > nodey.size {\n\t\tnodey.parent = nodex\n\t\tnodex.size += nodey.size\n\t} else {\n\t\tnodex.parent = nodey\n\t\tnodey.size += nodex.size\n\t}\n\n\treturn true\n}", "func (s *Set) Union(other *Set) *Set {\n\tnewSet := NewSizedSet(len(s.m) + len(other.m))\n\tfor elem := range s.m {\n\t\tnewSet.m[elem] = SetNothing{}\n\t}\n\tfor elem := range other.m {\n\t\tnewSet.m[elem] = SetNothing{}\n\t}\n\treturn newSet\n}", "func (list List_str) Union(lists ...List_str) List_str {\n\n vsf := make(List_str, 0)\n \n //append all lists into list\n for _, lst := range lists {\n vsf = list.Append(lst)\n }\n \n //return uniq (non-duplicate)\n return vsf.Uniq()\n}", "func (q *pathCompression) Union(p, r int) {\n\tqr := q.Root(r)\n\tqp := q.Root(p)\n\tif qr == qp {\n\t\treturn\n\t}\n\n\tif q.Weights[qr] > q.Weights[qp] {\n\t\tq.IDs[qp] = qr\n\t\tq.Weights[qr] += q.Weights[qp]\n\t} else {\n\t\tq.IDs[qr] = qp\n\t\tq.Weights[qp] += q.Weights[qr]\n\t}\n\n}", "func (s *IntSet) Union(s2 *IntSet) *IntSet {\n\tout := new(IntSet)\n\tout.Clear()\n\n\tfor key := range s.elem {\n\t\tout.Add(key)\n\t}\n\n\tfor key := range s2.elem {\n\t\tout.Add(key)\n\t}\n\n\treturn out\n}", "func Union[T comparable](list1 []T, list2 []T) []T {\n\tresult := []T{}\n\n\tseen := map[T]struct{}{}\n\thasAdd := map[T]struct{}{}\n\n\tfor _, e := range list1 {\n\t\tseen[e] = struct{}{}\n\t}\n\n\tfor _, e := range list2 {\n\t\tseen[e] = struct{}{}\n\t}\n\n\tfor _, e := range list1 {\n\t\tif _, ok := seen[e]; ok {\n\t\t\tresult = append(result, e)\n\t\t\thasAdd[e] = struct{}{}\n\t\t}\n\t}\n\n\tfor _, e := range list2 {\n\t\tif _, ok := hasAdd[e]; ok {\n\t\t\tcontinue\n\t\t}\n\t\tif _, ok := seen[e]; ok {\n\t\t\tresult = append(result, e)\n\t\t}\n\t}\n\n\treturn result\n}", "func Union[T comparable](list1 []T, list2 []T) []T {\n\tresult := []T{}\n\n\tseen := map[T]struct{}{}\n\thasAdd := map[T]struct{}{}\n\n\tfor _, e := range list1 {\n\t\tseen[e] = struct{}{}\n\t}\n\n\tfor _, e := range list2 {\n\t\tseen[e] = struct{}{}\n\t}\n\n\tfor _, e := range list1 {\n\t\tif _, ok := seen[e]; ok {\n\t\t\tresult = append(result, e)\n\t\t\thasAdd[e] = struct{}{}\n\t\t}\n\t}\n\n\tfor _, e := range list2 {\n\t\tif _, ok := hasAdd[e]; ok {\n\t\t\tcontinue\n\t\t}\n\t\tif _, ok := seen[e]; ok {\n\t\t\tresult = append(result, e)\n\t\t}\n\t}\n\n\treturn result\n}", "func (s stringSet) union(o stringSet) stringSet {\n\tns := newStringSet()\n\tfor k := range s {\n\t\tns[k] = struct{}{}\n\t}\n\tfor k := range o {\n\t\tns[k] = struct{}{}\n\t}\n\treturn ns\n}", "func (a TimeWindowAggregation) Union(other TimeWindowAggregation) TimeWindowAggregation {\r\n\treturn TimeWindowAggregation{\r\n\t\tVa: a.Va.Union(other.Va),\r\n\t\tTw: a.Tw.Union(other.Tw),\r\n\t}\r\n}", "func (p Polygon) Union(p2 Polygonal) Polygon {\n\treturn p.op(p2, polyclip.UNION)\n}", "func (s Set) Union(os Set) bool {\n\tb := false\n\tfor i := range os.m {\n\t\tif s.Contains(i) {\n\t\t\tcontinue\n\t\t}\n\t\ts.add(i)\n\t\tb = true\n\t}\n\treturn b\n}", "func AsUnion(dt DataType) *Union {\n\tswitch t := dt.(type) {\n\tcase *UserTypeExpr:\n\t\treturn AsUnion(t.Type)\n\tcase *ResultTypeExpr:\n\t\treturn AsUnion(t.Type)\n\tcase *Union:\n\t\treturn t\n\tdefault:\n\t\treturn nil\n\t}\n}", "func Union(list1 []Node, list2 []Node) []Node {\n\n\tidSet := map[string]Node{}\n\n\tfor _, n := range list1 {\n\t\tidSet[n.String()] = n\n\t}\n\tfor _, n := range list2 {\n\t\tif _, ok := idSet[n.String()]; !ok {\n\t\t\tidSet[n.String()] = n\n\t\t}\n\t}\n\n\tres := make([]Node, len(idSet))\n\ti := 0\n\tfor _, n := range idSet {\n\t\tres[i] = n\n\t\ti++\n\t}\n\n\treturn res\n}", "func (s StringSet) Union(other StringSet) StringSet {\n\tresultSet := make(StringSet, len(s))\n\tfor val := range s {\n\t\tresultSet[val] = true\n\t}\n\n\tfor val := range other {\n\t\tresultSet[val] = true\n\t}\n\n\treturn resultSet\n}", "func Union(a, b RegInvImage) RegInvImage {\n\tfor imageName, digestTags := range b {\n\t\t// If a does not have this image at all, then it's a simple\n\t\t// injection.\n\t\tif a[imageName] == nil {\n\t\t\ta[imageName] = digestTags\n\t\t\tcontinue\n\t\t}\n\t\tfor digest, tags := range digestTags {\n\t\t\t// If a has the image but not this digest, inject just this digest\n\t\t\t// and all associated tags.\n\t\t\tif a[imageName][digest] == nil {\n\t\t\t\ta[imageName][digest] = tags\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// If c has the digest already, try to inject those tags in b that\n\t\t\t// are not already in a.\n\t\t\ttagSlice := TagSlice{}\n\t\t\tfor tag := range tags.Union(a[imageName][digest]) {\n\t\t\t\tif tag == \"latest\" {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\ttagSlice = append(tagSlice, tag)\n\t\t\t}\n\t\t\ta[imageName][digest] = tagSlice\n\t\t}\n\t}\n\n\treturn a\n}", "func NewUnion(l, r Regex) Regex {\n\tswitch l.(type) {\n\tcase *empty:\n\t\treturn r\n\tdefault:\n\t\treturn &union{\n\t\t\tl: l,\n\t\t\tr: r,\n\t\t}\n\t}\n}", "func Union(sets ...Set) Set {\n\tif len(sets) == 0 {\n\t\treturn None\n\t}\n\n\tr := make(setUnion, 0, len(sets))\n\tfor _, set := range sets {\n\t\tif set == None {\n\t\t\tcontinue\n\t\t}\n\t\tif su, ok := set.setI.(setUnion); ok {\n\t\t\tr = append(r, su...)\n\t\t} else {\n\t\t\tr = append(r, set.setI)\n\t\t}\n\t}\n\tif len(r) == 1 {\n\t\treturn Set{setI: r[0]}\n\t}\n\treturn Set{setI: r}\n}", "func (qf *QuickFind) Union(p, q int) {\n\tpID, pErr := qf.Find(p)\n\tif pErr != nil {\n\t\treturn\n\t}\n\tqID, qErr := qf.Find(q)\n\tif qErr != nil {\n\t\treturn\n\t}\n\tif pID == qID {\n\t\treturn\n\t}\n\tfor i := 0; i < len(qf.id); i++ {\n\t\tif qf.id[i] == pID {\n\t\t\tqf.id[i] = qID\n\t\t}\n\t}\n\tqf.count--\n}", "func (h *HyperLogLog32) Union(a, b *HyperLogLog32) error {\n\tif a.p != b.p {\n\t\treturn errors.New(\"card: mismatched precision\")\n\t}\n\tta := reflect.TypeOf(b.hash)\n\tif reflect.TypeOf(b.hash) != ta {\n\t\treturn errors.New(\"card: mismatched hash function\")\n\t}\n\tif h.hash != nil && reflect.TypeOf(h.hash) != ta {\n\t\treturn errors.New(\"card: mismatched hash function\")\n\t}\n\n\tif h != a && h != b {\n\t\t*h = HyperLogLog32{p: a.p, m: a.m, hash: h.hash, register: make([]uint8, a.m)}\n\t}\n\tfor i, r := range a.register {\n\t\th.register[i] = max(r, b.register[i])\n\t}\n\treturn nil\n}", "func (u *QuickFind) Union(p, q int) {\n\tpid := u.ids[p]\n\tqid := u.ids[q]\n\n\tfor i := range u.ids {\n\t\tif u.ids[i] == pid {\n\t\t\tu.ids[i] = qid\n\t\t}\n\t}\n}", "func (b *Bound) Union(other *Bound) *Bound {\n\tb.Extend(other.SouthWest())\n\tb.Extend(other.NorthWest())\n\tb.Extend(other.SouthEast())\n\tb.Extend(other.NorthEast())\n\n\treturn b\n}", "func (r Rect) Union(s Rect) Rect {\n\treturn R(\n\t\tmath.Min(r.Min.X, s.Min.X),\n\t\tmath.Min(r.Min.Y, s.Min.Y),\n\t\tmath.Max(r.Max.X, s.Max.X),\n\t\tmath.Max(r.Max.Y, s.Max.Y),\n\t)\n}", "func union(first, second *mapNode, less func(a, b interface{}) bool, overwrite bool) *mapNode {\n\tif first == nil {\n\t\treturn second.incref()\n\t}\n\tif second == nil {\n\t\treturn first.incref()\n\t}\n\n\tif first.weight < second.weight {\n\t\tsecond, first, overwrite = first, second, !overwrite\n\t}\n\n\tleft, mid, right := split(second, first.key, less, false)\n\tvar result *mapNode\n\tif overwrite && mid != nil {\n\t\tresult = mid.shallowCloneWithRef()\n\t} else {\n\t\tresult = first.shallowCloneWithRef()\n\t}\n\tresult.weight = first.weight\n\tresult.left = union(first.left, left, less, overwrite)\n\tresult.right = union(first.right, right, less, overwrite)\n\tleft.decref()\n\tmid.decref()\n\tright.decref()\n\treturn result\n}", "func TestUnion(t *T) {\n\t// Degenerate case\n\tempty := NewSet()\n\tassert.Equal(t, 0, Size(empty.Union(empty)))\n\n\tints1 := []interface{}{0, 1, 2}\n\tints2 := []interface{}{3, 4, 5}\n\tintsu := append(ints1, ints2...)\n\ts1 := NewSet(ints1...)\n\ts2 := NewSet(ints2...)\n\n\tassertSeqContentsSet(t, ints1, s1.Union(empty))\n\tassertSeqContentsSet(t, ints1, empty.Union(s1))\n\n\tsu := s1.Union(s2)\n\tassertSeqContentsSet(t, ints1, s1)\n\tassertSeqContentsSet(t, ints2, s2)\n\tassertSeqContentsSet(t, intsu, su)\n}", "func (rg Range) Union(r Range) Range {\n\tif rg.Max.X < r.Max.X {\n\t\trg.Max.X = r.Max.X\n\t}\n\tif rg.Max.Y < r.Max.Y {\n\t\trg.Max.Y = r.Max.Y\n\t}\n\tif rg.Min.X > r.Min.X {\n\t\trg.Min.X = r.Min.X\n\t}\n\tif rg.Min.Y > r.Min.Y {\n\t\trg.Min.Y = r.Min.Y\n\t}\n\treturn rg\n}", "func TestUnion(t *testing.T) {\n\ttests := []struct {\n\t\tnote string\n\t\trules []string\n\t\texpected interface{}\n\t}{\n\t\t{\"union_0_sets\", []string{`p = x { union(set(), x) }`}, \"[]\"},\n\t\t{\"union_2_sets\", []string{`p = x { union({set(), {1, 2}}, x) }`}, \"[1, 2]\"},\n\t\t{\"union_2_sets\", []string{`p = x { s1 = {1, 2, 3}; s2 = {2}; union({s1, s2}, x) }`}, \"[1, 2, 3]\"},\n\t\t{\"union_3_sets\", []string{`p = x { s1 = {1, 2, 3}; s2 = {2, 3, 4}; s3 = {4, 5, 6}; union({s1, s2, s3}, x) }`}, \"[1, 2, 3, 4, 5, 6]\"},\n\t\t{\"union_4_sets\", []string{`p = x { s1 = {\"a\", \"b\", \"c\", \"d\"}; s2 = {\"b\", \"c\", \"d\"}; s3 = {\"c\", \"d\"}; s4 = {\"d\"}; union({s1, s2, s3, s4}, x) }`}, \"[\\\"a\\\", \\\"b\\\", \\\"c\\\", \\\"d\\\"]\"},\n\t}\n\n\tdata := loadSmallTestData()\n\n\tfor _, tc := range tests {\n\t\trunTopDownTestCase(t, data, tc.note, tc.rules, tc.expected)\n\t}\n}", "func Union(one Set, other Set) Set {\n\tif other == nil && one == nil {\n\t\treturn nil\n\t}\n\tif other == nil {\n\t\treturn one.Copy()\n\t}\n\tif one == nil {\n\t\treturn other.Copy()\n\t}\n\toneLen := one.Len()\n\totherLen := other.Len()\n\tif oneLen == 0{\n\t\treturn other.Copy()\n\t}\n\tif otherLen == 0{\n\t\treturn one.Copy()\n\t}\n\tcopyset := other.Copy()\n\tfor _, key := range one.Elements() {\n\t\tcopyset.Add(key)\n\t}\n\treturn copyset\n}", "func (uf *UnionFind) Union(x, y int) {\n\tuf.checkRange(x, y)\n\tpx := uf.parent(x)\n\tpy := uf.parent(y)\n\tif px != py {\n\t\tuf.volume--\n\t\tif uf.size[px] < uf.size[py] {\n\t\t\tuf.id[px] = py\n\t\t\tuf.size[py] += uf.size[px]\n\t\t} else {\n\t\t\tuf.id[py] = px\n\t\t\tuf.size[px] += uf.size[py]\n\t\t}\n\t}\n}", "func (c MethodsCollection) Union() pUnion {\n\treturn pUnion{\n\t\tMethod: c.MustGet(\"Union\"),\n\t}\n}", "func (n *NFA) Union(ns ...*NFA) *NFA {\n\tstart, final := State(0), State(1)\n\tunion := NewNFA(start, States{final})\n\tfactory := newStateFactory(final)\n\n\tnfas := append([]*NFA{n}, ns...)\n\tfor id, nfa := range nfas {\n\t\tfor _, kv := range nfa.trans.KeyValues() {\n\t\t\ts := factory.StateFor(id, kv.Key)\n\t\t\tfor _, kv := range kv.Val.KeyValues() {\n\t\t\t\ta := kv.Key\n\n\t\t\t\tnext := make(States, len(kv.Val))\n\t\t\t\tfor i, t := range kv.Val {\n\t\t\t\t\tnext[i] = factory.StateFor(id, t)\n\t\t\t\t}\n\n\t\t\t\tunion.Add(s, a, next)\n\t\t\t}\n\t\t}\n\n\t\tss := factory.StateFor(id, nfa.Start)\n\t\tunion.Add(start, E, States{ss})\n\n\t\tfor _, f := range nfa.Final {\n\t\t\tff := factory.StateFor(id, f)\n\t\t\tunion.Add(ff, E, States{final})\n\t\t}\n\t}\n\n\treturn union\n}", "func (g *unionGraph) Union(a, b unionNode) {\n\tinA := g.find(a)\n\tinB := g.find(b)\n\tif inA != inB {\n\t\tif inA.rank > inB.rank {\n\t\t\tinB.parent = inA\n\t\t} else if inA.rank < inB.rank {\n\t\t\tinA.parent = inB\n\t\t} else { // inA.rank == inB.rank\n\t\t\tinA.parent = inB\n\t\t\tinB.rank++\n\t\t}\n\t}\n}", "func (u OpUnion) Clone() OpUnion {\n\tclone := OpUnion{\n\t\tType: u.Type,\n\t\tTransformation: u.Transformation,\n\t}\n\tif u.Type == pipeline.RollupOpType {\n\t\tclone.Rollup = u.Rollup.Clone()\n\t}\n\treturn clone\n}", "func (set Int64Set) Union(other Int64Set) Int64Set {\n\tunionedSet := set.Clone()\n\tfor v := range other {\n\t\tunionedSet.doAdd(v)\n\t}\n\treturn unionedSet\n}", "func (r Rectangle) Union(s Rectangle) Rectangle {\n\tif r.Empty() {\n\t\treturn s\n\t}\n\tif s.Empty() {\n\t\treturn r\n\t}\n\tif r.Min.X > s.Min.X {\n\t\tr.Min.X = s.Min.X\n\t}\n\tif r.Min.Y > s.Min.Y {\n\t\tr.Min.Y = s.Min.Y\n\t}\n\tif r.Max.X < s.Max.X {\n\t\tr.Max.X = s.Max.X\n\t}\n\tif r.Max.Y < s.Max.Y {\n\t\tr.Max.Y = s.Max.Y\n\t}\n\treturn r\n}", "func (qf *QuickFind) Union(p int, q int) error {\n\tif p < 0 || q < 0 {\n\t\treturn errors.New(\"p and q must be greater or equal to 0\")\n\t}\n\tif p > len(qf.ids) || q > len(qf.ids) {\n\t\treturn fmt.Errorf(\"p and q must be smaller or equal to %d\", len(qf.ids))\n\t}\n\tpid := qf.ids[p]\n\tqid := qf.ids[q]\n\n\t// p and q are already connected\n\tif pid == qid {\n\t\treturn nil\n\t}\n\t// p and q are already connected\n\tfor i, v := range qf.ids {\n\t\tif v == pid {\n\t\t\tqf.ids[i] = qid\n\t\t}\n\t}\n\treturn nil\n}", "func (s *Set[T]) Union(set Set[T]) {\n\ts.resize(2 * set.Len())\n\tfor elt := range set {\n\t\t(*s)[elt] = struct{}{}\n\t}\n}", "func (u *OpUnion) Reset() { *u = OpUnion{} }", "func (ts List) Union() List {\n\treturn ts.UnionWithHandler(func(mergeInto, mergeFrom, mergeSpan T) T {\n\t\treturn mergeSpan\n\t})\n}", "func NewUnion(exs ...Extractor) *Union {\n\tvar union = &Union{Extractors: make([]Extractor, len(exs))}\n\tfor i, ex := range exs {\n\t\tunion.Extractors[i] = ex\n\t}\n\treturn union\n}", "func (s Set) Union(other Set) {\n\tfor k := range other {\n\t\ts.Insert(k)\n\t}\n}", "func (s UserSet) Union(other m.UserSet) m.UserSet {\n\tres := s.Collection().Call(\"Union\", other)\n\tresTyped := res.(models.RecordSet).Collection().Wrap(\"User\").(m.UserSet)\n\treturn resTyped\n}", "func (s StringSet) Union(other StringSet) StringSet {\n\tresult := make(StringSet)\n\tfor v := range s {\n\t\tresult[v] = struct{}{}\n\t}\n\tfor v := range other {\n\t\tresult[v] = struct{}{}\n\t}\n\treturn result\n}", "func (as1 Attributes) Union(as2 Attributes) (Attributes, error) {\n\tfor k, v := range as2 {\n\t\tif _, ok := as1[k]; ok {\n\t\t\treturn nil, fmt.Errorf(\"attribute %v already exists\", k)\n\t\t}\n\t\tas1[k] = v\n\t}\n\treturn as1, nil\n}", "func (mp MultiPolygon) Union(p2 Polygonal) Polygon {\n\treturn mp.op(p2, polyclip.UNION)\n}", "func (set Set) Union(ctx context.Context, keys ...string) ([]string, error) {\n\treq := newRequestSize(2+len(keys), \"\\r\\n$6\\r\\nSUNION\\r\\n$\")\n\treq.addStringAndStrings(set.name, keys)\n\treturn set.c.cmdStrings(ctx, req)\n}", "func (s *IntSet) Union(y *IntSet) *IntSet {\n\tu := NewIntSet([]int{})\n\n\tfor _, m := range s.Members() {\n\t\tu.Add(m)\n\t}\n\n\tfor _, m := range y.Members() {\n\t\tu.Add(m)\n\t}\n\n\treturn u\n}", "func (ev Vars) Union(other Vars) Vars {\n\tnewSet := NewVars()\n\tfor key, value := range ev {\n\t\tnewSet[key] = value\n\t}\n\tfor key, value := range other {\n\t\tnewSet[key] = value\n\t}\n\treturn newSet\n}", "func (rs Reasons) Union(other Reasons) Reasons {\n\tmerged := make(Reasons)\n\tfor r := range rs {\n\t\tmerged.Add(r)\n\t}\n\tfor r := range other {\n\t\tmerged.Add(r)\n\t}\n\treturn merged\n}", "func (self *Rectangle) UnionI(args ...interface{}) *Rectangle{\n return &Rectangle{self.Object.Call(\"union\", args)}\n}", "func (sd *SelectDataset) Union(other *SelectDataset) *SelectDataset {\n\treturn sd.withCompound(exp.UnionCompoundType, other.CompoundFromSelf())\n}", "func union(a, b map[string]bool) map[string]bool {\n\tr := make(map[string]bool)\n\tfor k := range a {\n\t\tr[k] = true\n\t}\n\tfor k := range b {\n\t\tr[k] = true\n\t}\n\treturn r\n}", "func Or(a, b Dense) Dense {\n\tshort, long := a, b\n\tif b.len < a.len {\n\t\tshort, long = b, a\n\t}\n\trLen := long.len\n\tif short.negated {\n\t\trLen = short.len\n\t}\n\tr := Dense{\n\t\tbits: make([]byte, 0, BytesFor(rLen)),\n\t\tlen: rLen,\n\t\tnegated: a.negated || b.negated,\n\t}\n\tfor i := range short.bits {\n\t\tr.bits = append(r.bits, a.bits[i]|b.bits[i])\n\t}\n\tif !short.negated {\n\t\tfor i := len(short.bits); i < len(long.bits); i++ {\n\t\t\tr.bits = append(r.bits, long.bits[i])\n\t\t}\n\t}\n\treturn r\n}", "func oruint8s(a, b uint8) uint8", "func union(a, b map[string]bool) {\n\tfor k, v := range b {\n\t\tif v {\n\t\t\ta[k] = v\n\t\t}\n\t}\n}", "func (arr *filterTableArr) UnionEqual(other filterTableArr) *filterTableArr {\n\tfor i, el := range other {\n\t\tif el {\n\t\t\tarr[i] = true\n\t\t}\n\t}\n\n\treturn arr\n}", "func Uniq(data interface{}) (interface{}, error) {\n\treturn Union(data)\n}", "func Union(s1, s2 string) string {\n\tvar intersect strings.Builder\n\tset := make(map[rune]bool)\n\tfor _, char := range s1 {\n\t\tif _, ok := set[char]; !ok {\n\t\t\tset[char] = true\n\t\t\tintersect.WriteRune(char)\n\t\t}\n\t}\n\tfor _, char := range s2 {\n\t\tif _, ok := set[char]; !ok {\n\t\t\tset[char] = true\n\t\t\tintersect.WriteRune(char)\n\t\t}\n\t}\n\treturn intersect.String()\n}", "func (p *planner) UnionClause(n *parser.UnionClause, desiredTypes []parser.Datum, autoCommit bool) (planNode, error) {\n\tvar emitAll = false\n\tvar emit unionNodeEmit\n\tswitch n.Type {\n\tcase parser.UnionOp:\n\t\tif n.All {\n\t\t\temitAll = true\n\t\t} else {\n\t\t\temit = make(unionNodeEmitDistinct)\n\t\t}\n\tcase parser.IntersectOp:\n\t\tif n.All {\n\t\t\temit = make(intersectNodeEmitAll)\n\t\t} else {\n\t\t\temit = make(intersectNodeEmitDistinct)\n\t\t}\n\tcase parser.ExceptOp:\n\t\tif n.All {\n\t\t\temit = make(exceptNodeEmitAll)\n\t\t} else {\n\t\t\temit = make(exceptNodeEmitDistinct)\n\t\t}\n\tdefault:\n\t\treturn nil, errors.Errorf(\"%v is not supported\", n.Type)\n\t}\n\n\tleft, err := p.newPlan(n.Left, desiredTypes, autoCommit)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tright, err := p.newPlan(n.Right, desiredTypes, autoCommit)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tleftColumns := left.Columns()\n\trightColumns := right.Columns()\n\tif len(leftColumns) != len(rightColumns) {\n\t\treturn nil, fmt.Errorf(\"each %v query must have the same number of columns: %d vs %d\", n.Type, len(left.Columns()), len(right.Columns()))\n\t}\n\tfor i := 0; i < len(leftColumns); i++ {\n\t\tl := leftColumns[i]\n\t\tr := rightColumns[i]\n\t\t// TODO(dan): This currently checks whether the types are exactly the same,\n\t\t// but Postgres is more lenient:\n\t\t// http://www.postgresql.org/docs/9.5/static/typeconv-union-case.html.\n\t\tif !l.Typ.TypeEqual(r.Typ) {\n\t\t\treturn nil, fmt.Errorf(\"%v types %s and %s cannot be matched\", n.Type, l.Typ.Type(), r.Typ.Type())\n\t\t}\n\t\tif l.hidden != r.hidden {\n\t\t\treturn nil, fmt.Errorf(\"%v types cannot be matched\", n.Type)\n\t\t}\n\t}\n\n\tnode := &unionNode{\n\t\tright: right,\n\t\tleft: left,\n\t\trightDone: false,\n\t\tleftDone: false,\n\t\temitAll: emitAll,\n\t\temit: emit,\n\t\tscratch: make([]byte, 0),\n\t}\n\treturn node, nil\n}", "func (box *LLBox) Union(box2 *LLBox) *LLBox {\n\treturn &LLBox{N: math.Max(box.N, box2.N), W: math.Min(box.W, box2.W), S: math.Min(box.S, box2.S), E: math.Max(box.E, box2.E)}\n}" ]
[ "0.7149548", "0.6900931", "0.6745773", "0.66751015", "0.6497951", "0.6349197", "0.6344535", "0.63178736", "0.631593", "0.6304623", "0.6292116", "0.6256469", "0.6236854", "0.6230655", "0.617909", "0.61442053", "0.61380035", "0.6113462", "0.6104106", "0.6093799", "0.6038361", "0.6030144", "0.6025214", "0.59990627", "0.59879726", "0.59846526", "0.5941574", "0.59181035", "0.58945495", "0.5866677", "0.5862", "0.58577603", "0.58466804", "0.58299726", "0.5825114", "0.58054936", "0.5795908", "0.5786629", "0.5777685", "0.57508594", "0.5742535", "0.5694971", "0.5672435", "0.5647581", "0.5644116", "0.56439334", "0.5629449", "0.5629449", "0.56272626", "0.5617794", "0.559934", "0.5592824", "0.55857635", "0.5582924", "0.55822706", "0.55747646", "0.556873", "0.55647", "0.5553579", "0.5545002", "0.55390674", "0.5535647", "0.5530556", "0.5528946", "0.55200756", "0.54834396", "0.5482239", "0.54804474", "0.54760015", "0.5459316", "0.54531777", "0.54288703", "0.5419452", "0.5409077", "0.54054576", "0.53873944", "0.5360933", "0.5342097", "0.5329692", "0.53103614", "0.5301763", "0.5298599", "0.5285792", "0.5281903", "0.5269373", "0.52589494", "0.5244746", "0.52281827", "0.52131885", "0.52106315", "0.5203057", "0.5168523", "0.5146339", "0.51272756", "0.50805575", "0.5074453", "0.5065293", "0.5056388", "0.50400186", "0.503059" ]
0.7705332
0
/ Time Now uses time.Now to return the current time as a time.Time instance.
func Now() time.Time { return time.Now() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *Time) Now() time.Time {\n\treturn t.current\n}", "func Now() Time {\n\treturn NewTime(time.Now())\n}", "func TimeNow() Time {\n\treturn Time{\n\t\ttime.Now(),\n\t}\n}", "func Now() Time {\n\treturn Time{format(time.Now())}\n}", "func Now() Time {\n\tvar t Time\n\tt.FromNSec(uint64(gotime.Now().UnixNano()))\n\treturn t\n}", "func Now() Time {\n\treturn Time(time.Now().UnixNano() / 1000)\n}", "func Now() time.Time {\n\treturn now()\n}", "func Now() time.Time { return now() }", "func Now() time.Time { return time.Now().UTC() }", "func Now() time.Time { return time.Now().UTC() }", "func Now() Time {\n\treturn DefaultScheduler.Now()\n}", "func Now() time.Time {\n\treturn CurrentClock().Now()\n}", "func TimeNow() time.Time {\n\treturn time.Now().UTC()\n}", "func NowTime() time.Time {\n\treturn ExtractTimeFromDatetime(time.Now())\n}", "func nowTime() time.Time {\n\treturn time.Now().UTC()\n}", "func Now() time.Time {\n\treturn time.Date(int(Year.Status().(uint16)), time.Month(Month.Status().(uint8)), int(Day.Status().(uint16)), int(Hour.Status().(uint8)), int(Minute.Status().(uint8)), int(Second.Status().(uint8)), 0, time.FixedZone(\"custom\", int(Tz.Status().(int8)) * 3600))\n}", "func Now() time.Time {\n\treturn Clock.Now()\n}", "func Now() *time.Time {\n\tt := time.Now()\n\treturn &t\n}", "func (t Time) GetNow() time.Time {\n\treturn now()\n}", "func Now() time.Time {\n\treturn c.Now()\n}", "func now() time.Time {\n\treturn time.Now().UTC()\n}", "func Now() time.Time {\n\treturn time.Unix(0, time.Now().UnixNano()/1e6*1e6)\n}", "func (c *Clock) Now() time.Time {\n\treturn time.Now().UTC().Truncate(time.Second)\n}", "func (Clock) Now() time.Time {\n\treturn time.Now()\n}", "func Now() time.Time {\n\treturn Work.Now()\n}", "func GetNow() time.Time {\n\treturn time.Now().UTC()\n}", "func GetNow() time.Time {\n\treturn time.Now().UTC()\n}", "func (d *dummyClock) Now() time.Time {\n\treturn time.Time{}\n}", "func (defaultClock) Now() time.Time {\n\treturn time.Now()\n}", "func (t *DefaultClock) Now() time.Time {\n\treturn time.Now()\n}", "func Now() time.Time {\n\tif IsTest() {\n\t\treturn now\n\t}\n\n\treturn time.Now()\n}", "func (s *Scheduler) Now() Time {\n\ts.lock.RLock()\n\tdefer s.lock.RUnlock()\n\treturn s.now()\n}", "func Now() time.Time {\n\treturn time.Now().In(_defaultLocation)\n}", "func (realClocker) Now() time.Time { return time.Now() }", "func (c *Clock) Now() time.Time { return time.Now() }", "func Now() Time {\n\t/*\n\t\tft := windows.Filetime{}\n\t\twindows.GetSystemTimeAsFileTime(&ft)\n\t\treturn Time(ft.Nanoseconds() / nanosecondsScale)\n\t*/\n\t// On Windows, it turns out to be much faster to simply call time.Now().Unix(), so do it\n\treturn Time(time.Now().Unix())\n}", "func (fc *fakeClock) Now() time.Time {\n\treturn fc.time\n}", "func (c *RunningClock) Now() time.Time {\n\treturn time.Now()\n}", "func (c stubClocker) Now() time.Time { return c.t }", "func (u *Util) Now() time.Time {\n\tif controlDuration != 0 {\n\t\treturn time.Now().Add(controlDuration).UTC()\n\t}\n\treturn time.Now().UTC()\n}", "func (m *timeSource) Now() mstime.Time {\n\treturn mstime.Now()\n}", "func (RealClock) Now() time.Time {\n\treturn time.Now()\n}", "func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func (s systemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func CurrentTime() time.Time {\n\treturn time.Now()\n}", "func (c *MockClock) Now() time.Time {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\treturn c.now\n}", "func (ns *Namespace) Now() _time.Time {\n\treturn _time.Now()\n}", "func (clock *Clock) Now() time.Time {\n\tvar now time.Time\n\tclock.do(func(s *state) {\n\t\tnow = s.t\n\t\ts.t = s.t.Add(clock.step)\n\t})\n\treturn now\n}", "func Now() DateTime {\n\treturn DateTimeFromTime(time.Now())\n}", "func (tc *TestClock) Now() time.Time {\n\ttc.l.RLock()\n\tdefer tc.l.RUnlock()\n\treturn tc.t\n}", "func Now() time.Time {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tvar now time.Time\n\tif testMode {\n\t\tnow = testNow()\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.In(localtz.Get())\n}", "func (c *StoppedClock) Now() time.Time {\n\treturn c.time\n}", "func TimeUTCNow() Time {\n\treturn Time{\n\t\ttime.Now().Truncate(time.Second).UTC(),\n\t}\n}", "func (wc WallClock) Now() time.Time {\n\treturn time.Now()\n}", "func (c *StoppedClock) Now() time.Time {\n\treturn c.t\n}", "func (f FakeTimeSource) Now() time.Time {\n\treturn f.FakeTime\n}", "func (c *CumulativeClock) Now() time.Time {\n\treturn c.current\n}", "func (c *FakeClock) Now() time.Time {\n\tc.steps++\n\treturn c.Time(c.steps)\n}", "func (t *TimeService) Now(request *NowRequest) (*NowResponse, error) {\n\trsp := &NowResponse{}\n\treturn rsp, t.client.Call(\"time\", \"Now\", request, rsp)\n}", "func (c *Context) Now() time.Time {\n\treturn c.currentTime\n}", "func Now() Date {\n\tn := time.Now()\n\treturn Of(n)\n}", "func (f *FixedTimeSource) Now() time.Time {\n\treturn f.fakeTime\n}", "func (m *Mock) Now() time.Time {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\treturn m.now\n}", "func Now() time.Time {\n\treturn time.Now().In(LOCATION)\n}", "func (t *Timer) Now() time.Duration {\n\treturn nowFunc().Sub(t.Zero)\n}", "func CurrentTime() Time {\n\tif atomic.LoadUint32(&isRunning) != 0 {\n\t\treturn currentTime\n\t}\n\n\treturn Time(time.Now().UnixNano() / 1e6)\n}", "func (f *FakeTimeSource) Now() time.Time {\n\tf.mu.RLock()\n\tdefer f.mu.RUnlock()\n\treturn f.now\n}", "func NowDate() time.Time {\n\treturn ExtractDateFromDatetime(time.Now())\n}", "func (m *Mock) Now() time.Time {\n\tm.Lock()\n\tdefer m.Unlock()\n\treturn m.now\n}", "func Now() Timespec {\n\treturn NewTimespec(time.Now())\n}", "func New(t time.Time) *Now {\n\treturn &Now{t}\n}", "func (c *webCtx) Now() time.Time {\n\treturn c.now\n}", "func (rs *requestContext) Now() time.Time {\n\treturn rs.now\n}", "func (s *Scheduler) now() Time {\n\treturn s.provider.Now().Add(s.timeOffset)\n}", "func (c *Clock) Now() time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tstep := c.step\n\tif c.skipStep {\n\t\tstep = 0\n\t\tc.skipStep = false\n\t}\n\tc.advanceLocked(rt, step)\n\n\treturn c.present\n}", "func (a *IncrementingFakeTimeSource) Now() time.Time {\n\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\n\ta.NextIncrement++\n\n\treturn adjustedTime\n}", "func UTCNow() time.Time {\n\treturn time.Now().UTC()\n}", "func GetCurrentDateTimeInTime() time.Time {\n\treturn time.Now()\n}", "func (p *PredefinedFake) Now() time.Time {\n\tadjustedTime := p.Base.Add(p.Delays[p.Next])\n\tp.Next++\n\treturn adjustedTime\n}", "func (em *eventManager) Now() time.Time {\n\tem.mu.Lock()\n\tdefer em.mu.Unlock()\n\treturn em.now\n}", "func GetCurrentTime() uint64 {\n\treturn ComposeTSByTime(time.Now(), 0)\n}", "func (p PT) Now() int64 {\n\tif p.Seconds {\n\t\treturn time.Now().Unix()\n\t}\n\treturn time.Now().UnixNano()\n}", "func ExampleNow() {\n\tt := gtime.Now()\n\tfmt.Println(t)\n\n\t// May Output:\n\t// 2021-11-06 13:41:08\n}", "func (tr *TextRegion) TimeNow() {\n\ttr.Time.Now()\n}", "func GetCurrentTime() typeutil.Timestamp {\n\treturn ComposeTSByTime(time.Now(), 0)\n}", "func defaultNow() time.Time {\n\treturn time.Now().UTC()\n}", "func (c *ClockVal) Now() {\n\tc.CurrentTime = NowTime()\n}", "func TimeNowUTC() time.Time {\n\treturn time.Now().UTC()\n}", "func dbNow() time.Time {\n\treturn roundTime(time.Now())\n}", "func Now(upToSecond ...bool) *TimeStamp {\n\treturn TimeFrom(time.Now(), upToSecond...)\n}", "func defaultNow() time.Time {\n\treturn time.Now()\n}", "func (c *chrono) Now() time.Time {\n\tvar now time.Time\n\tif c.paused {\n\t\tnow = c.pausedAt\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.Add(c.skew)\n}", "func (c CurrentTime) GetCurrentDayAndTime() time.Time {\n\treturn time.Now()\n}", "func (s *Scheduler) now() time.Time {\n\tif s.nowTimeFn == nil {\n\t\treturn time.Now()\n\t} else {\n\t\treturn s.nowTimeFn()\n\t}\n}", "func PtrTimeNow() *time.Time {\n\tn := time.Now()\n\treturn &n\n}", "func Now(ctx Context) time.Time {\n\treturn getWorkflowEnvironment(ctx).Now()\n}", "func getToday() time.Time {\n\treturn time.Now().UTC()\n}", "func Time() time.Time {\n\treturn time.Time{}\n}", "func (t Time) Time() time.Time {\n\treturn time.Time(t)\n}" ]
[ "0.8948937", "0.89447755", "0.8840452", "0.882197", "0.8772744", "0.8706519", "0.87028384", "0.8692103", "0.8672435", "0.8672435", "0.8658959", "0.86543334", "0.86413795", "0.85889804", "0.8583817", "0.8577014", "0.85453176", "0.8479161", "0.8455715", "0.844247", "0.84101653", "0.8381427", "0.82933307", "0.8272688", "0.82672775", "0.82309926", "0.82309926", "0.8228011", "0.81911665", "0.81514883", "0.8141234", "0.8125734", "0.8115866", "0.81057614", "0.8075424", "0.8074386", "0.806354", "0.80599576", "0.8041903", "0.80171525", "0.80130243", "0.79913205", "0.79593724", "0.79593724", "0.7951844", "0.795115", "0.79380804", "0.7928444", "0.79276365", "0.78876066", "0.7876858", "0.78654414", "0.78651994", "0.7860265", "0.7855687", "0.7853525", "0.7840071", "0.7826191", "0.77921885", "0.77802074", "0.777839", "0.77571964", "0.7741011", "0.7706042", "0.7698949", "0.7694271", "0.7686414", "0.76771766", "0.76488626", "0.7646484", "0.76143146", "0.76069975", "0.7605061", "0.7566364", "0.7560675", "0.7554028", "0.75211185", "0.7520461", "0.7498385", "0.7479606", "0.7453865", "0.744098", "0.7430931", "0.7427847", "0.7401144", "0.7380361", "0.7371275", "0.73705125", "0.7368227", "0.73207104", "0.73110956", "0.730729", "0.72388613", "0.72136366", "0.7200867", "0.72000057", "0.7197437", "0.71959805", "0.7173729", "0.71685666" ]
0.8504319
17
ParseTime uses time.Parse to return a time.Time instance of operand parsed according to format.
func ParseTime(format, operand string) (time.Time, error) { return time.Parse(format, operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func parseTime(format string, timeStr string) time.Time {\n\tt, err := time.Parse(format, timeStr)\n\tcheck(err)\n\treturn t\n}", "func ParseTime(time string) (string, error) {\n\tmatches := TimeRegex.FindStringSubmatch(time)\n\tif len(matches) != 3 {\n\t\treturn \"\", errors.New(\"wrong time format - only minutes or HH:mm format allowed\")\n\t}\n\thoursString := matches[1]\n\tminutesString := matches[2]\n\tif len(hoursString) == 0 {\n\t\treturn strconv.Itoa(getMinutes(minutesString)), nil\n\t}\n\treturn strconv.Itoa(getHours(hoursString)*60 + getMinutes(minutesString)), nil\n}", "func ParseTime(text string) (t time.Time, err error) {\r\n\tfor _, layout := range timeFormats {\r\n\t\tt, err = time.Parse(layout, text)\r\n\t\tif err == nil {\r\n\t\t\treturn\r\n\t\t}\r\n\t}\r\n\treturn\r\n}", "func timeParse(ts string) (time.Time, error) {\n\tformat := timeFormat[:len(ts)]\n\tt, err := time.ParseInLocation(format, ts, time.Local)\n\treturn t, err\n}", "func ParseTime(s string) Time {\n\ttp := timeParser{p: s}\n\n\t// TODO: limit length of elements so that\n\t// 20060102T000000 is not parsed as year 20060102\n\n\tyear := tp.val(\":-\")\n\tmonth := tp.xval(\":-\")\n\tday := tp.xval(\"tT\")\n\n\tif tp.prec == 0 {\n\t\treturn Time{}\n\t}\n\n\thour := tp.val(\":\")\n\tmin := tp.val(\":\")\n\tsec := tp.val(\".\")\n\n\tnsec, ndenom, ok := tp.rat(\"\")\n\tif ok {\n\t\tfor ndenom < 1e9 {\n\t\t\tnsec, ndenom = nsec*10, ndenom*10\n\t\t}\n\t}\n\n\tloc := tp.loc()\n\n\treturn Time{\n\t\tTime: time.Date(year, time.Month(month), day, hour, min, sec, nsec, loc),\n\t\tPrec: tp.prec,\n\t\tHasLoc: tp.hasLoc,\n\t}\n}", "func ParseTime(s string) (Time, error) {\n\tif s == \"\" {\n\t\treturn Time{}, nil\n\t}\n\tif !timeRe.MatchString(s) {\n\t\treturn Time{}, fmt.Errorf(\"parse time: expected hhmmss.ss format, got '%s'\", s)\n\t}\n\thour, _ := strconv.Atoi(s[:2])\n\tminute, _ := strconv.Atoi(s[2:4])\n\tsecond, _ := strconv.ParseFloat(s[4:], 64)\n\twhole, frac := math.Modf(second)\n\treturn Time{true, hour, minute, int(whole), int(round(frac * 1000))}, nil\n}", "func ParseTime(t string) (time.Time, error) {\n\tlayout := \"2006-01-02 15:04:05 ZO700\"\n\treturn time.Parse(layout, t)\n}", "func ParseTime(s string, timeLayout string) (tm time.Time, err error) {\n\treturn time.Parse(timeLayout, s)\n}", "func Parse(format, s string) time.Time {\n\tt, _ := time.Parse(format, s)\n\treturn t\n}", "func TimeParse(t TimeStr) (time.Time, error) {\n\treturn time.Parse(time.RFC3339, string(t))\n}", "func (t Time) Parse() (ts time.Time, err error) {\n\ts := strings.TrimSpace(string(t))\n\t// Support for Bitpipe timestamps using 24:00:00\n\ts = strings.Replace(s, \" 24:\", \" 00:\", 1)\n\tfor _, layout := range formats {\n\t\tif ts, err = time.Parse(layout, s); err == nil {\n\t\t\treturn\n\t\t}\n\t}\n\t// As in time.Parse(), return UTC for the first arg, which will come out\n\t// of the previous calls to time.Parse()\n\treturn ts, errors.New(\"Could not parse \" + s)\n}", "func parseTime(ts string) (time.Time, error) {\n\treturn time.ParseInLocation(\"02-Jan-2006 03:04:05 PM\", ts, tz)\n}", "func ParseTime(s string) (time.Time, error) {\n\tvar t time.Time\n\tvar err error\n\n\t// ISO\n\tt, err = time.Parse(time.RFC3339, s)\n\tif err == nil {\n\t\treturn t, nil\n\t}\n\n\t// ISO Date\n\tt, err = time.Parse(ISODate, s)\n\tif err == nil {\n\t\treturn t, nil\n\t}\n\n\t// PostgreSQL\n\tt, err = time.Parse(PostgresTime, s)\n\tif err == nil {\n\t\treturn t, nil\n\t}\n\n\t// epoch seconds\n\tt64, err := strconv.ParseFloat(s, 64)\n\tif err == nil {\n\t\ts, ns := math.Modf(t64)\n\t\treturn time.Unix(int64(s), int64(ns*1e9)), nil\n\t}\n\n\treturn time.Time{}, PTOErrorf(\"%s not parseable as time\", s).StatusIs(http.StatusBadRequest)\n}", "func ParseTime(timespec string) (Time, error) {\n\tt, err := time.Parse(TsLayout, timespec)\n\treturn Time(t), err\n}", "func Parse(str string) (time.Time, error) {\n\tdatefmt, err := DateParser(str)\n\tif err != nil {\n\t\treturn time.Time{}, errors.Wrap(err, NewParseError(str, \"Parsing date format\"))\n\t}\n\ttimefmt, err := TimeParser(str)\n\tif err != nil {\n\t\treturn time.Time{}, errors.Wrap(err, NewParseError(str, \"Parsing time format\"))\n\t}\n\n\treturn time.Parse(datefmt+timefmt, str)\n}", "func parseTime(timeString string) (time.Time, error) {\n\tif timeString == \"\" {\n\t\treturn time.Time{}, errors.New(\"need time string\")\n\t}\n\n\tt, err := time.Parse(dateFormat, timeString)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\n\t// time.Parse() is \"clever\" but also doesn't check anything more\n\t// granular than a second, so let's be completely paranoid and check\n\t// via regular expression too.\n\tmatched := dateFormatRE.FindAllStringSubmatch(timeString, -1)\n\tif matched == nil {\n\t\treturn time.Time{},\n\t\t\tfmt.Errorf(\"expected time in format %q, got %q\", dateFormatPattern, timeString)\n\t}\n\n\treturn t, nil\n}", "func parseTime(timeString string) (time.Time, error) {\n\tfor _, form := range timeFormat {\n\t\ttimes, err := time.Parse(form, timeString)\n\t\tif err == nil {\n\t\t\t//Parse time value successful\n\t\t\treturn times, nil\n\t\t}\n\t}\n\t//Parse time value unsuccessful\n\treturn time.Now(), fmt.Errorf(\"parsing time %q error\", timeString)\n}", "func ParseTime(s string) (time.Time, error) {\n\treturn time.Parse(time.RFC3339, s)\n}", "func ParseTime(s string) (time.Time, error) {\n\tt, err := time.Parse(logutil.SlowLogTimeFormat, s)\n\tif err != nil {\n\t\t// This is for compatibility.\n\t\tt, err = time.Parse(logutil.OldSlowLogTimeFormat, s)\n\t\tif err != nil {\n\t\t\terr = errors.Errorf(\"string \\\"%v\\\" doesn't has a prefix that matches format \\\"%v\\\", err: %v\", s, logutil.SlowLogTimeFormat, err)\n\t\t}\n\t}\n\treturn t, err\n}", "func parseTime(rawtime string) time.Time {\n\trawtime = strings.Replace(rawtime, \"下午\", \"PM \", -1)\n\trawtime = strings.Replace(rawtime, \"上午\", \"AM \", -1)\n\ttm, err := time.Parse(\"2006-1-2 PM 3:04\", rawtime)\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn time.Now().AddDate(-10, 0, 0)\n\t}\n\treturn tm\n\n}", "func TimeParse() time.Time {\n\tnow := time.Now()\n\tthen := now.AddDate(0, 0, -1).Format(\"2006-01-02T15:04:05:0700\")\n\tsince, _ := time.Parse(\"2006-01-02\", then)\n\treturn since\n}", "func ParseTime(s string) (int64, error) {\n\tif t, err := strconv.ParseFloat(s, 64); err == nil {\n\t\ts, ns := math.Modf(t)\n\t\tns = math.Round(ns*1000) / 1000\n\t\ttm := time.Unix(int64(s), int64(ns*float64(time.Second)))\n\t\treturn TimeToMillis(tm), nil\n\t}\n\tif t, err := time.Parse(time.RFC3339Nano, s); err == nil {\n\t\treturn TimeToMillis(t), nil\n\t}\n\treturn 0, httpgrpc.Errorf(http.StatusBadRequest, \"cannot parse %q to a valid timestamp\", s)\n}", "func parseTime(str string) (*time.Time, error) {\n\tif str == \"\" {\n\t\treturn nil, nil\n\t}\n\tt, err := time.Parse(twilioTimeLayout, str)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &t, err\n}", "func parseTime(s string) (hour, minute int64, err error) {\n\ttime := strings.Split(s, \":\")\n\thour, err = strconv.ParseInt(time[0], 10, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\tminute, err = strconv.ParseInt(time[1], 10, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func ParseDate(date string) (time.Time, error)", "func Parse(layout, value string) (*time.Time, error) {\n\tt, err := time.Parse(layout, value)\n\treturn &t, err\n}", "func ParseTimeFormat(timeIn string) (time.Time, bool) {\n\tvar errorFlag bool = false\n\tresponse, err := time.Parse(TimeFormat, timeIn)\n\tif err != nil {\n\t\tlog.Println(\"Error: Could not parse time. <stats/calculations.go->GetTimeDifferenceParseTimeFormat. \")\n\t\terrorFlag = true\n\t}\n\t//log.Printf(\"Response is: %v\", response)\n\treturn response, errorFlag\n}", "func ParseTime(s string) (time.Duration, error) {\n\tbaseRegex := \"^(\\\\d+)\"\n\tregInputToHour := map[string]int{\n\t\tbaseRegex + \"min\": 1,\n\t\tbaseRegex + \"h\": 60 * 1,\n\t\tbaseRegex + \"day\": 60 * 24,\n\t\tbaseRegex + \"week\": 60 * 24 * 7,\n\t\tbaseRegex + \"month\": 60 * 24 * 30,\n\t\tbaseRegex + \"year\": 60 * 24 * 365,\n\t}\n\thours, err := func() (float64, error) {\n\t\tfor k, v := range regInputToHour {\n\t\t\tre := regexp.MustCompile(k)\n\t\t\tmatch := re.FindStringSubmatch(s)\n\t\t\tif len(match) == 2 {\n\t\t\t\tnumStr := match[1]\n\t\t\t\tnum, err := strconv.ParseFloat(numStr, 64)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn num * float64(v), nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn 0, fmt.Errorf(\"cannot parse %s into a valid duration\", s)\n\t}()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn time.Duration(time.Duration(hours) * time.Minute), nil\n}", "func Parse(str string, format ...string) (time.Time, error) {\n\treturn newZtime.Parse(str, format...)\n}", "func parseTime(from string, offset *time.Time) (time.Time, error) {\n\tvar start time.Time\n\tif offset != nil {\n\t\tstart = *offset\n\t} else {\n\t\tstart = time.Now()\n\t}\n\ty, m, d, dur, r, err := parseISO8601Duration(from)\n\tif err == nil {\n\t\tif r != -1 {\n\t\t\treturn time.Time{}, errors.Errorf(\"repetitions are not allowed\")\n\t\t}\n\t\treturn start.AddDate(y, m, d).Add(dur), nil\n\t}\n\tif dur, err = time.ParseDuration(from); err == nil {\n\t\treturn start.Add(dur), nil\n\t}\n\tif t, err := time.Parse(time.RFC3339, from); err == nil {\n\t\treturn t, nil\n\t}\n\treturn time.Time{}, errors.Errorf(\"unsupported time/duration format %q\", from)\n}", "func parseTime(timeString string) (time.Time, error) {\n\tif timeString == \"now\" {\n\t\treturn time.Now(), nil\n\t}\n\n\t// handle fuzzy time strings\n\tre := regexp.MustCompile(\"([0-9]+)\\\\s+(second|minute|hour|day)s?\\\\s+ago\")\n\tmatches := re.FindStringSubmatch(timeString)\n\tif len(matches) == 3 {\n\t\tvalue, err := strconv.ParseInt(matches[1], 10, 64)\n\t\tif err != nil {\n\t\t\treturn time.Now(), err\n\t\t}\n\n\t\tvar duration time.Duration\n\t\tswitch matches[2] {\n\t\tcase \"second\":\n\t\t\tduration = time.Second\n\t\tcase \"minute\":\n\t\t\tduration = time.Minute\n\t\tcase \"hour\":\n\t\t\tduration = time.Hour\n\t\tcase \"day\":\n\t\t\tduration = time.Hour * 24\n\t\tdefault:\n\t\t\treturn time.Now(), fmt.Errorf(\"Unknown time unit, %s\", matches[2])\n\t\t}\n\n\t\treturn time.Now().Add(-1 * time.Duration(value) * duration), nil\n\t}\n\n\trfcTime, err := time.Parse(time.RFC3339, timeString)\n\tif err != nil {\n\t\treturn time.Now(), err\n\t}\n\treturn rfcTime, nil\n}", "func parseTime(s string) (time.Time, error) {\n\t// attempt to parse time as RFC3339 string\n\tt, err := time.Parse(time.RFC3339Nano, s)\n\tif err == nil {\n\t\treturn t, nil\n\t}\n\n\t// attempt to parse time as float number of unix seconds\n\tif f, err := strconv.ParseFloat(s, 64); err == nil {\n\t\tsec, dec := math.Modf(f)\n\t\treturn time.Unix(int64(sec), int64(dec*(1e9))), nil\n\t}\n\n\t// attempt to parse time as json marshaled value\n\tif err := json.Unmarshal([]byte(s), &t); err == nil {\n\t\treturn t, nil\n\t}\n\n\treturn time.Time{}, err\n}", "func ParseAsFormat(input *string, format string) (*time.Time, error) {\n\tif input == nil {\n\t\treturn nil, nil\n\t}\n\n\tval, err := time.Parse(format, *input)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"parsing %q: %+v\", *input, err)\n\t}\n\n\treturn &val, nil\n}", "func parseTime(t string) (time.Time, error) {\n\tepoch, err := strconv.ParseInt(t, 10, 64)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\treturn time.Unix(epoch, 0), nil\n}", "func parseTime(s string) (time.Time, error) {\n\treturn time.Parse(time.RFC1123Z, s)\n}", "func ParseTime(s string, fmt string) (ti time.Time, err error) {\n\tvar i int64\n\tswitch fmt {\n\tcase \"ns\", \"nanosecond\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i), nil\n\t\t}\n\tcase \"us\", \"microsecond\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Microsecond)), nil\n\t\t}\n\tcase \"ms\", \"millisecond\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Millisecond)), nil\n\t\t}\n\tcase \"s\", \"second\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Second)), nil\n\t\t}\n\tcase \"mi\", \"minute\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Minute)), nil\n\t\t}\n\tcase \"h\", \"hour\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Hour)), nil\n\t\t}\n\tcase \"d\", \"day\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Hour)*24), nil\n\t\t}\n\tdefault:\n\t\tif fmt == \"\" {\n\t\t\tfmt = time.RFC3339Nano\n\t\t}\n\t\tif ti, e := time.ParseInLocation(fmt, s, time.Local); e != nil {\n\t\t\treturn ti, errs.New(e)\n\t\t} else {\n\t\t\treturn ti, nil\n\t\t}\n\t}\n}", "func parseTime(accum string) (time.Time, error) {\n\tformats := []string{time.RFC822, time.RFC822Z, time.RFC1123, time.RFC1123Z,\n\t\ttime.RFC3339, \"2006-1-2\"}\n\n\tfor _, format := range formats {\n\t\ttm, err := time.Parse(format, accum)\n\t\tif err == nil {\n\t\t\treturn tm, nil\n\t\t}\n\t}\n\n\treturn time.Time{}, MkError(\"Could not parse time: %s\", accum)\n}", "func (t DateString) Parse() (time.Time, error) {\n\treturn time.ParseInLocation(\"200601021504\", string(t), Timezone)\n}", "func parseTime(s string) (time.Time, error) {\n\tt, err := time.ParseInLocation(time.RFC1123, s, time.UTC)\n\tif err != nil {\n\t\treturn time.Time{}, fmt.Errorf(\"parsing time: %w\", err)\n\t}\n\n\treturn t, nil\n}", "func (t *Time) ParseFrom(b []byte) error {\n\n\tif t == nil {\n\t\treturn _ERR_NIL_TIME_RECEIVER\n\t}\n\n\tvar i = 0\n\tfor n := len(b); i < n && b[i] <= ' '; i++ { }\n\n\t// Minimum required len: 4 (hhmm - w/o separators, w/o seconds).\n\tif len(b[i:]) < 4 {\n\t\treturn _ERR_NOT_ISO8601_TIME\n\t}\n\n\tx, valid := batoi(b[i], b[i+1])\n\tif !valid || x < 0 || x > 23 {\n\t\treturn _ERR_BAD_HOUR\n\t}\n\n\ti += 2\n\thh := Hour(x)\n\twasSeparator := false\n\n\t// Skip separator\n\tif !(b[i] >= '0' && b[i] <= '9') {\n\t\ti++\n\t\twasSeparator = true\n\t}\n\n\t// At this code point, len(b) may == 1. Check it.\n\tif len(b[i:]) == 1 {\n\t\treturn _ERR_NOT_ISO8601_TIME\n\t}\n\n\tx, valid = batoi(b[i], b[i+1])\n\tif !valid || x < 0 || x > 59 {\n\t\treturn _ERR_BAD_MINUTE\n\t}\n\n\ti += 2\n\tmm := Minute(x)\n\tss := Second(0)\n\n\t// At this code point user may provide \"hhmm\" w/o seconds.\n\t// Check whether seconds are provided.\n\tif l := len(b[i:]); l > 0 {\n\t\t// We need 2 symbols if there was no separator, or 3 symbols otherwise.\n\t\tif (l == 1 && !wasSeparator) || (l == 2 && wasSeparator) {\n\t\t\treturn _ERR_NOT_ISO8601_TIME\n\t\t}\n\t\tif wasSeparator {\n\t\t\ti++\n\t\t}\n\t\tx, valid = batoi(b[i], b[i+1])\n\t\tif !valid || x < 0 || x > 59 {\n\t\t\treturn _ERR_BAD_SECOND\n\t\t}\n\t\tss = Second(x)\n\t}\n\n\n\tif !IsValidTime(hh, mm, ss) {\n\t\treturn _ERR_BAD_CORRESP_TIME\n\t}\n\n\t*t = NewTime(hh, mm, ss)\n\treturn nil\n}", "func ParseTime(s string) (time.Time, error) {\n\tconst msg = \"failed to parse Gerrit timestamp %q\"\n\tif len(s) < 2 || s[0] != '\"' || s[len(s)-1] != '\"' {\n\t\treturn time.Time{}, errors.Reason(msg, s).Err()\n\t}\n\tt, err := time.Parse(GerritTimestampLayout, s[1:len(s)-1])\n\tif err != nil {\n\t\treturn time.Time{}, errors.Annotate(err, msg, s).Err()\n\t}\n\treturn t, nil\n}", "func ParseTimeString(t string) (time.Time, error) {\n\treturn dateparse.ParseLocal(t)\n}", "func ParseHeader(timeStr string) (time.Time, error) {\n\tfor _, dateFormat := range httpTimeFormats {\n\t\tt, err := time.Parse(dateFormat, timeStr)\n\t\tif err == nil {\n\t\t\treturn t, nil\n\t\t}\n\t}\n\treturn time.Time{}, ErrMalformedDate\n}", "func parseTime(cmdArgs []string) (*goment.Goment, error) {\n\tg, _ := goment.New()\n\n\ttime := cmdArgs[1]\n\n\t// Whether it's necessary to add 12 hours to the time (goment expects a 24 hour format).\n\thasPM := regexp.MustCompile(`(?i)pm`).MatchString(time)\n\t// Cleanup the time input.\n\ttime = regexp.MustCompile(`(?i)(pm|am)`).ReplaceAllString(time, \"\")\n\n\t// 13:37, 13.37, 4:20am, 4.20am\n\tvar timeParts []string\n\tif strings.Contains(time, \":\") {\n\t\ttimeParts = strings.Split(time, \":\")\n\t} else if strings.Contains(time, \".\") {\n\t\ttimeParts = strings.Split(time, \".\")\n\t} else {\n\t\ttimeParts = []string{time}\n\t}\n\n\thour, err := strconv.ParseInt(timeParts[0], 10, 32)\n\tif err != nil {\n\t\treturn nil, errors.New(\"cannot parse hour\")\n\t}\n\tif hour < 0 || hour > 23 {\n\t\treturn nil, errors.New(\"invalid hour format\")\n\t}\n\tif hasPM && hour <= 12 {\n\t\thour += 12\n\t}\n\tg.SetHour(int(hour))\n\n\tvar minute int64\n\tif len(timeParts) > 1 {\n\t\tminute, err = strconv.ParseInt(timeParts[1], 10, 32)\n\t\tif err != nil {\n\t\t\treturn nil, errors.New(\"cannot parse minute\")\n\t\t}\n\t\tif minute < 0 || minute > 60 {\n\t\t\treturn nil, errors.New(\"invalid minute format\")\n\t\t}\n\t}\n\tg.SetMinute(int(minute))\n\n\tvar second int64\n\tif len(timeParts) > 2 {\n\t\tsecond, err = strconv.ParseInt(timeParts[2], 10, 32)\n\t\tif err != nil {\n\t\t\treturn nil, errors.New(\"cannot parse second\")\n\t\t}\n\t\tif second < 0 || second > 60 {\n\t\t\treturn nil, errors.New(\"invalid second format\")\n\t\t}\n\t}\n\tg.SetSecond(int(second))\n\n\treturn g, nil\n}", "func parseTime(value string) (int64, error) {\n\tif len(value) != 0 {\n\t\tt, err := time.Parse(time.RFC3339, value)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t\tif !t.IsZero() {\n\t\t\treturn t.Unix(), nil\n\t\t}\n\t}\n\treturn 0, nil\n}", "func (c *Client) parseTime(lines []string, err error) (time.Time, error) {\n\tvar t time.Time\n\tif err != nil {\n\t\treturn t, err\n\t}\n\n\tif len(lines) != 1 {\n\t\treturn t, NewInvalidResponseError(\"parseTime: unexpected lines\", lines...)\n\t}\n\n\ti, err := strconv.ParseInt(lines[0], 10, 64)\n\tif err != nil {\n\t\treturn t, NewInvalidResponseError(\"parseTime: parse int\", lines[0])\n\t}\n\n\treturn time.Unix(i, 0), nil\n}", "func ParseTime(input string) (time.Time, error) {\n\t// validate input with regex\n\tmatch, err := regexp.MatchString(`^[0-9]{1,2}\\/[0-9]{1,2}\\/[0-9]{1,4}$`, input)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\tif !match {\n\t\treturn time.Time{}, errors.New(\"invalid time format\")\n\t\t// return error\n\t}\n\tinputArray := strings.Split(input, \"/\")\n\tfor location, dateStamp := range inputArray {\n\t\t// lazy validation of string\n\t\t// if less 2 strings per position format from 1 to 01\n\t\tif len(dateStamp) < 2 {\n\t\t\ttempFmt := fmt.Sprintf(\"0%s\", dateStamp)\n\t\t\t// modify inputArray with the new format\n\t\t\tinputArray[location] = tempFmt\n\n\t\t}\n\n\t}\n\n\tnewFormattedInput := strings.Join(inputArray, \"/\")\n\tparsedDate, err := time.Parse(\"02/01/2006\", newFormattedInput)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\t// validate date range\n\tif !parsedDate.After(MinDate) && parsedDate.Before(MaxDate) {\n\n\t\treturn time.Time{}, errors.New(\"error dates outside of acceptable range\")\n\t}\n\treturn parsedDate, nil\n}", "func StrictTimeParse(value string) (time.Time, error) {\n\treturn time.Parse(DATETIME, value)\n}", "func Parse(value string) (time.Time, error) {\n\tvar t time.Time\n\tvar err error\n\tfor _, layout := range layouts {\n\t\tt, err = time.Parse(layout, value)\n\t\tif err == nil {\n\t\t\treturn t, err\n\t\t}\n\t}\n\treturn t, err\n}", "func Parse(layout, value string) (Time, error) {}", "func (hms *HHMMSS) Parse(s string) (err error) {\n\tll := []string{\n\t\t\"15:04:05.999999999-07:00\",\n\t\t\"15:04:05.999999999Z07:00\",\n\t\t\"15:04:05.999999999\",\n\t\t\"15:04:05.999999-07:00\",\n\t\t\"15:04:05.999999Z07:00\",\n\t\t\"15:04:05.999999\",\n\t\t\"15:04:05.999-07:00\",\n\t\t\"15:04:05.999Z07:00\",\n\t\t\"15:04:05.999\",\n\t\t\"15:04:05-07:00\",\n\t\t\"15:04:05Z07:00\",\n\t\t\"15:04:05\",\n\t\t\"15:04-07:00\",\n\t\t\"15:04Z07:00\",\n\t\t\"15:04\",\n\t\t\"15-07:00\",\n\t\t\"15Z07:00\",\n\t\t\"15\",\n\t}\n\tconst year = \"2006 \"\n\tys := year + s\n\tfor _, l := range ll {\n\t\ttm, e := time.ParseInLocation(year+l, ys, time.Local)\n\t\tif e == nil {\n\t\t\t*(*time.Duration)(hms) = tm.Sub(tm.Truncate(24 * time.Hour))\n\t\t\treturn nil\n\t\t}\n\t\terr = e\n\t}\n\terr = &time.ParseError{\n\t\tValue: s,\n\t\tMessage: \": cannot parse into HHMMSS\",\n\t}\n\treturn\n}", "func Parse(value string) (*Time, error) {\n\tt, err := time.Parse(RFC3339NanoFixed, value)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn New(t), nil\n}", "func FullTimeParse(value string) (time.Time, error) {\n\tlayouts := []string{\n\t\t\"2006-01-02T15:04:05.999999999Z0700\",\n\t\t\"2006-01-02T15:04:05.999999999Z07:00\",\n\t\t\"2006-01-02T15:04:05.999999999Z07\",\n\t}\n\tvar (\n\t\terr error\n\t\tt time.Time\n\t)\n\tfor _, layout := range layouts {\n\t\tif t, err = time.Parse(layout, value); err == nil {\n\t\t\treturn t, nil\n\t\t}\n\t}\n\treturn t, err\n}", "func ParseTime(ts string) (time.Time, error) {\n\traw := []byte(ts)\n\tif len(raw) != 6 {\n\t\treturn time.Time{}, errors.New(\"ts illegal\")\n\t}\n\n\tfor _, c := range raw {\n\t\tif int(c) > len(reversedAlphabet) || reversedAlphabet[c] == -1 {\n\t\t\treturn time.Time{}, errors.New(\"ts illegal\")\n\t\t}\n\t}\n\n\toffset := uint8(reversedAlphabet[raw[5]])\n\n\tym := uint8((offset>>3)*alphabets) + uint8(reversedAlphabet[raw[0]])\n\tyear := int(ym>>4) + epochYear\n\tmonth := time.Month(ym << 4 >> 4)\n\n\tday := reversedAlphabet[raw[1]]\n\n\thour := int((offset<<5>>7)*alphabets) + reversedAlphabet[raw[2]]\n\tminute := int((offset<<6>>7)*alphabets) + reversedAlphabet[raw[3]]\n\tsecond := int((offset<<7>>7)*alphabets) + reversedAlphabet[raw[4]]\n\n\treturn time.Date(year, month, day, hour, minute, second, 0, time.UTC), nil\n}", "func parseTimeString(srctime, format string) (time.Time, error) {\n\n\tts, err := time.Parse(format, srctime)\n\n\treturn ts, err\n}", "func parsePAXTime(t string) (time.Time, error) {\n\tbuf := []byte(t)\n\tpos := bytes.IndexByte(buf, '.')\n\tvar seconds, nanoseconds int64\n\tvar err error\n\tif pos == -1 {\n\t\tseconds, err = strconv.ParseInt(t, 10, 0)\n\t\tif err != nil {\n\t\t\treturn time.Time{}, err\n\t\t}\n\t} else {\n\t\tseconds, err = strconv.ParseInt(string(buf[:pos]), 10, 0)\n\t\tif err != nil {\n\t\t\treturn time.Time{}, err\n\t\t}\n\t\tnano_buf := string(buf[pos+1:])\n\t\t// Pad as needed before converting to a decimal.\n\t\t// For example .030 -> .030000000 -> 30000000 nanoseconds\n\t\tif len(nano_buf) < maxNanoSecondIntSize {\n\t\t\t// Right pad\n\t\t\tnano_buf += strings.Repeat(\"0\", maxNanoSecondIntSize-len(nano_buf))\n\t\t} else if len(nano_buf) > maxNanoSecondIntSize {\n\t\t\t// Right truncate\n\t\t\tnano_buf = nano_buf[:maxNanoSecondIntSize]\n\t\t}\n\t\tnanoseconds, err = strconv.ParseInt(string(nano_buf), 10, 0)\n\t\tif err != nil {\n\t\t\treturn time.Time{}, err\n\t\t}\n\t}\n\tts := time.Unix(seconds, nanoseconds)\n\treturn ts, nil\n}", "func ParseFormatted(input string) (time.Time, string, error) {\n\t// \"Mon, 02 Jan 2006 15:04:05 MST\"\n\tif t, err := time.Parse(time.RFC1123, input); err == nil {\n\t\treturn t, time.RFC1123, nil\n\t}\n\n\t// \"Mon, 02 Jan 2006 15:04:05 -0700\"\n\tif t, err := time.Parse(time.RFC1123Z, input); err == nil {\n\t\treturn t, time.RFC1123Z, nil\n\t}\n\n\t// \"2006-01-02T15:04:05Z07:00\"\n\tif t, err := time.Parse(time.RFC3339, input); err == nil {\n\t\treturn t, time.RFC3339, nil\n\t}\n\n\t// \"2006-01-02T15:04:05.999999999Z07:00\"\n\tif t, err := time.Parse(time.RFC3339Nano, input); err == nil {\n\t\treturn t, time.RFC3339Nano, nil\n\t}\n\n\t// \"02 Jan 06 15:04 MST\"\n\tif t, err := time.Parse(time.RFC822, input); err == nil {\n\t\treturn t, time.RFC822, nil\n\t}\n\n\t// \"02 Jan 06 15:04 -0700\"\n\tif t, err := time.Parse(time.RFC822Z, input); err == nil {\n\t\treturn t, time.RFC822Z, nil\n\t}\n\n\t// \"Monday, 02-Jan-06 15:04:05 MST\"\n\tif t, err := time.Parse(time.RFC850, input); err == nil {\n\t\treturn t, time.RFC850, nil\n\t}\n\n\t// \"Mon Jan _2 15:04:05 2006\"\n\tif t, err := time.Parse(time.ANSIC, input); err == nil {\n\t\treturn t, time.ANSIC, nil\n\t}\n\n\t// \"Mon Jan _2 15:04:05 MST 2006\"\n\tif t, err := time.Parse(time.UnixDate, input); err == nil {\n\t\treturn t, time.UnixDate, nil\n\t}\n\n\t// \"Mon Jan 02 15:04:05 -0700 2006\"\n\tif t, err := time.Parse(time.RubyDate, input); err == nil {\n\t\treturn t, time.RubyDate, nil\n\t}\n\n\t// \"3:04PM\"\n\tif t, err := time.Parse(time.Kitchen, input); err == nil {\n\t\treturn t, time.Kitchen, nil\n\t}\n\n\t// \"Jan _2 15:04:05\"\n\tif t, err := time.Parse(time.Stamp, input); err == nil {\n\t\treturn t, time.Stamp, nil\n\t}\n\n\t// \"Jan _2 15:04:05.000\"\n\tif t, err := time.Parse(time.StampMilli, input); err == nil {\n\t\treturn t, time.StampMilli, nil\n\t}\n\n\t// \"Jan _2 15:04:05.000000\"\n\tif t, err := time.Parse(time.StampMicro, input); err == nil {\n\t\treturn t, time.StampMicro, nil\n\t}\n\n\t// \"Jan _2 15:04:05.000000000\"\n\tif t, err := time.Parse(time.StampNano, input); err == nil {\n\t\treturn t, time.StampNano, nil\n\t}\n\n\t// \"Mon, 02 Jan 2006 15:04:05 GMT\"\n\tif t, err := time.Parse(FormatHTTP, input); err == nil {\n\t\treturn t, FormatHTTP, nil\n\t}\n\n\tif t, err := time.Parse(FormatGo, strings.Split(input, \" m=\")[0]); err == nil {\n\t\treturn t, FormatGo, nil\n\t}\n\n\t// \"2019-01-25 21:51:38\"\n\tif t, err := time.Parse(FormatSimple, input); err == nil {\n\t\treturn t, FormatSimple, nil\n\t}\n\n\treturn time.Time{}, \"\", ErrParseFormatted\n}", "func Parse(amzDateStr string) (time.Time, error) {\n\tfor _, dateFormat := range amzDateFormats {\n\t\tamzDate, err := time.Parse(dateFormat, amzDateStr)\n\t\tif err == nil {\n\t\t\treturn amzDate, nil\n\t\t}\n\t}\n\treturn time.Time{}, ErrMalformedDate\n}", "func ParseAny(datestr string) (time.Time, error) {\n\treturn parseTime(datestr, nil)\n}", "func parseTime(msg []byte) *time.Time {\n\t// convert input to integer\n\ti := binary.BigEndian.Uint32(msg)\n\n\t// convert time from 1900 to Unix time\n\tt := int64(i) - timeGap1900\n\n\tres := time.Unix(t, 0)\n\treturn &res\n}", "func ParseTimeResponse(response string) (*time.Time, error) {\n\telements := strings.SplitN(response, \" \", 4)\n\tif len(elements) != 4 {\n\t\treturn nil, fmt.Errorf(\"invalid timestamp response: %v\", response)\n\t}\n\tunixTimestampLiteral := elements[0][1:]\n\tunixTimestamp, err := strconv.ParseInt(unixTimestampLiteral, 10, 64)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"invalid timestamp: %v\", unixTimestampLiteral)\n\t}\n\tmsecLiteral := elements[1]\n\tmsec, _ := strconv.ParseInt(msecLiteral, 10, 64)\n\tts := time.Unix(unixTimestamp, msec*1000)\n\treturn &ts, nil\n}", "func parsePdTime(pdTime int64) *metav1.Time {\n\tt := metav1.Unix(pdTime, 0)\n\treturn &t\n}", "func mustParseTime(value string) time.Time {\n\tv, err := time.Parse(defaultTimeLayout, value)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn v\n}", "func CheckParse(s string) (time.Time, error) {\n\tvar value time.Time\n\tvar err error\n\tvalue, err = time.Parse(time.RFC3339, s)\n\tif err != nil {\n\t\tvalue, err = time.Parse(\"2006-01-02\", s)\n\t\tif err != nil {\n\t\t\treturn time.Time{}, fmt.Errorf(\"invalid date format (%s) provided\", s)\n\t\t}\n\t}\n\treturn value, nil\n}", "func Parse(layout, d string) (Date, error) {\n\tt, err := time.Parse(layout, d)\n\tif err != nil {\n\t\treturn Date{}, err\n\t}\n\n\treturn Of(t), nil\n}", "func ParseDate(in string) time.Time {\n\tt, err := time.Parse(DateFormat, in)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn t\n}", "func ParseTimeArg(arg string) (time.Time, error) {\n\tdateFmts := []string{\n\t\t\"2006-01-02 3:04:05 PM MST\",\n\t\t\"01-02 3:04:05 PM MST\",\n\t\t\"2006-01-02 3:04 PM MST\",\n\t\t\"01-02 3:04 PM MST\",\n\t\ttime.Kitchen,\n\t\t\"2006-01-02\",\n\t}\n\tvar res time.Time\n\tfor _, dateFmt := range dateFmts {\n\t\t// Special update for kitchen time format to include year, month,\n\t\t// and day.\n\t\tif dateFmt == time.Kitchen {\n\t\t\tloc, _ := time.LoadLocation(\"Local\")\n\t\t\tt, err := time.ParseInLocation(dateFmt, arg, loc)\n\t\t\tif err != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tres = t\n\t\t\tn := time.Now()\n\t\t\t// Month and day default values are already 1!\n\t\t\tres = res.AddDate(n.Year(), int(n.Month())-1, n.Day()-1)\n\t\t\tbreak\n\t\t} else {\n\t\t\tt, err := time.Parse(dateFmt, arg)\n\t\t\tif err == nil {\n\t\t\t\tres = t\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\tif res.IsZero() {\n\t\tmsg := fmt.Sprintf(\"Unable to parse %s using formats %v\", arg, dateFmts)\n\t\treturn res, errors.New(msg)\n\t}\n\tif res.Year() == 0 {\n\t\tres = res.AddDate(time.Now().In(res.Location()).Year(), 0, 0)\n\t}\n\treturn res.UTC(), nil\n}", "func ParseTimeFromString(in string, loc ...*time.Location) time.Time {\n\tvar timestamp time.Time\n\tvar err error\n\tfor _, tf := range TTFormats {\n\t\tif len(loc) > 0 {\n\t\t\ttimestamp, err = time.ParseInLocation(tf, in, loc[0])\n\t\t} else {\n\t\t\ttimestamp, err = time.Parse(tf, in)\n\t\t}\n\t\tif err == nil {\n\t\t\tbreak\n\t\t}\n\t\tlog.Warnf(\"Failed to parse as text: %v in: %s using: %s\", err, in, tf)\n\t\ttimestamp = time.Unix(0, 0)\n\t}\n\treturn timestamp\n}", "func NewTime(t string) (Time, error) {\n\ttime, err := time.Parse(TimeFormat, t)\n\tif err != nil {\n\t\treturn Time{}, err\n\t}\n\n\treturn Time{&time}, nil\n}", "func MustParseTime(timespec string) Time {\n\tts, err := ParseTime(timespec)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn ts\n}", "func mustParseTime(s string) time.Time {\n\tt, err := time.Parse(time.RFC3339, s)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\treturn t\n}", "func TimeFromString(s string) dgo.Time {\n\tts, err := time.Parse(time.RFC3339Nano, s)\n\tif err != nil {\n\t\tpanic(catch.Error(err))\n\t}\n\treturn &timeVal{ts}\n}", "func Parse(s string) (time.Duration, error) {\n\tif contains(invalidStrings, s) || strings.HasSuffix(s, \"T\") {\n\t\treturn 0, ErrInvalidString\n\t}\n\n\tvar (\n\t\tmatch []string\n\t\tprefix string\n\t)\n\n\tif pattern.MatchString(s) {\n\t\tmatch = pattern.FindStringSubmatch(s)\n\t} else {\n\t\treturn 0, ErrUnsupportedFormat\n\t}\n\n\tif strings.HasPrefix(s, \"-\") {\n\t\tprefix = \"-\"\n\t}\n\n\treturn durationFromMatchAndPrefix(match, prefix)\n}", "func timeParse(microsecs string) time.Time {\n\tt := time.Date(1601, time.January, 1, 0, 0, 0, 0, time.UTC)\n\tm, err := strconv.ParseInt(microsecs, 10, 64)\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"%s\\n\", err.Error())\n\t\tos.Exit(1)\n\t}\n\tvar u int64 = 100000000000000\n\tdu := time.Duration(u) * time.Microsecond\n\tf := float64(m)\n\tx := float64(u)\n\tn := f / x\n\tr := int64(n)\n\tremainder := math.Mod(f, x)\n\tiRem := int64(remainder)\n\tvar i int64\n\tfor i = 0; i < r; i++ {\n\t\tt = t.Add(du)\n\t}\n\n\tt = t.Add(time.Duration(iRem) * time.Microsecond)\n\n\t// RFC1123 = \"Mon, 02 Jan 2006 15:04:05 MST\"\n\t// t.Format(time.RFC1123)\n\treturn t\n}", "func Parse(value string, format string) (DateTime, error) {\n\tparsedTime, err := parseToTime(value, format, UTC)\n\treturn DateTimeFromTime(parsedTime), err\n}", "func TimeParse(datetime string) (timestamp time.Time, err error) {\n\tdatetime = strings.Replace(datetime, \"am\", \"AM\", -1)\n\tdatetime = strings.Replace(datetime, \"pm\", \"PM\", -1)\n\ttimeFormats := []string{\n\t\ttime.ANSIC,\n\t\ttime.UnixDate,\n\t\ttime.RubyDate,\n\t\ttime.RFC822,\n\t\ttime.RFC822Z,\n\t\ttime.RFC850,\n\t\ttime.RFC1123,\n\t\ttime.RFC1123Z,\n\t\ttime.RFC3339,\n\t\ttime.RFC3339Nano,\n\t\ttime.Kitchen,\n\t\ttime.Stamp,\n\t\ttime.StampMilli,\n\t\ttime.StampMicro,\n\t\ttime.StampNano,\n\t}\n\n\tcustomDate := []string{\n\t\t\"January 2 2006\",\n\t\t\"Jan 2 2006\",\n\t\t\"1 2 2006\",\n\t\t\"01 2 2006\",\n\t\t\"1 02 2006\",\n\t\t\"01 02 2006\",\n\n\t\t\"2 January 2006\",\n\t\t\"2 Jan 2006\",\n\t\t\"2 1 2006\",\n\t\t\"2 01 2006\",\n\t\t\"02 1 2006\",\n\t\t\"02 01 2006\",\n\n\t\t\"2006 January 2\",\n\t\t\"2006 Jan 2\",\n\t\t\"2006 1 2\",\n\t\t\"2006 01 2\",\n\t\t\"2006 1 02\",\n\t\t\"2006 01 02\",\n\n\t\t\"January 2\",\n\t\t\"Jan 2\",\n\t\t\"1 2\",\n\t\t\"01 2\",\n\t\t\"1 02\",\n\t\t\"01 02\",\n\n\t\t\"2 January\",\n\t\t\"2 Jan\",\n\t\t\"2 1\",\n\t\t\"2 01\",\n\t\t\"02 1\",\n\t\t\"02 01\",\n\t}\n\n\tcustomTime := []string{\n\t\t\"15:04:05\",\n\t\t\"15:04\",\n\n\t\t\"3:04 PM\",\n\t\t\"03:04 PM\",\n\t\t\"3 PM\",\n\t\t\"03 PM\",\n\n\t\t\"3:04PM\",\n\t\t\"03:04PM\",\n\t\t\"3PM\",\n\t\t\"03PM\",\n\t}\n\n\tcustomZone := []string{\n\t\t\"MST\",\n\n\t\t\"GMT-0700\",\n\t\t\"GMT-7\",\n\t\t\"GMT-07\",\n\t\t\"GMT-07:00\",\n\t\t\"GMT-7:00\",\n\n\t\t\"UTC-0700\",\n\t\t\"UTC-7\",\n\t\t\"UTC-07\",\n\t\t\"UTC-07:00\",\n\t\t\"UTC-7:00\",\n\t}\n\n\tfor _, timeFormat := range timeFormats {\n\t\ttimestamp, err = time.Parse(timeFormat, datetime)\n\t\tif err == nil {\n\t\t\treturn timestamp, nil\n\t\t}\n\t}\n\n\t// Run custom formats only if none of the default formats work\n\tfor _, date := range customDate {\n\t\ttimestamp, err = time.Parse(date, datetime)\n\t\tif err == nil {\n\t\t\treturn timestamp, nil\n\t\t}\n\n\t\tfor _, timer := range customTime {\n\t\t\ttimestamp, err = time.Parse(timer, datetime)\n\t\t\tif err == nil {\n\t\t\t\ttimestamp = timestamp.AddDate(time.Now().Year(), int(time.Now().Month())-1, time.Now().Day())\n\t\t\t\treturn timestamp, nil\n\t\t\t}\n\n\t\t\ttimestamp, err = time.Parse(date+\" \"+timer, datetime)\n\t\t\tif err == nil {\n\t\t\t\treturn timestamp, nil\n\t\t\t}\n\n\t\t\ttimestamp, err = time.Parse(timer+\" \"+date, datetime)\n\t\t\tif err == nil {\n\t\t\t\treturn timestamp, nil\n\t\t\t}\n\n\t\t\tfor _, zone := range customZone {\n\t\t\t\ttimestamp, err = time.Parse(date+\" \"+zone, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(timer+\" \"+zone, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\ttimestamp = timestamp.AddDate(time.Now().Year(), int(time.Now().Month())-1, time.Now().Day())\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(zone+\" \"+date, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(zone+\" \"+timer, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\ttimestamp = timestamp.AddDate(time.Now().Year(), int(time.Now().Month())-1, time.Now().Day())\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(date+\" \"+timer+\" \"+zone, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(date+\" \"+zone+\" \"+timer, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(timer+\" \"+date+\" \"+zone, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(timer+\" \"+zone+\" \"+date, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(zone+\" \"+timer+\" \"+date, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(zone+\" \"+date+\" \"+timer, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// Run with dashed date now if none of the non-dashed work\n\tfor _, date := range customDate {\n\t\tdate = dashed(date)\n\n\t\ttimestamp, err = time.Parse(date, datetime)\n\t\tif err == nil {\n\t\t\treturn timestamp, nil\n\t\t}\n\n\t\tfor _, timer := range customTime {\n\t\t\ttimestamp, err = time.Parse(timer, datetime)\n\t\t\tif err == nil {\n\t\t\t\treturn timestamp, nil\n\t\t\t}\n\n\t\t\ttimestamp, err = time.Parse(date+\" \"+timer, datetime)\n\t\t\tif err == nil {\n\t\t\t\treturn timestamp, nil\n\t\t\t}\n\n\t\t\ttimestamp, err = time.Parse(timer+\" \"+date, datetime)\n\t\t\tif err == nil {\n\t\t\t\treturn timestamp, nil\n\t\t\t}\n\n\t\t\tfor _, zone := range customZone {\n\t\t\t\ttimestamp, err = time.Parse(date+\" \"+zone, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(timer+\" \"+zone, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(zone+\" \"+date, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(zone+\" \"+timer, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(date+\" \"+timer+\" \"+zone, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(date+\" \"+zone+\" \"+timer, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(timer+\" \"+date+\" \"+zone, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(timer+\" \"+zone+\" \"+date, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(zone+\" \"+timer+\" \"+date, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\n\t\t\t\ttimestamp, err = time.Parse(zone+\" \"+date+\" \"+timer, datetime)\n\t\t\t\tif err == nil {\n\t\t\t\t\treturn timestamp, nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn timestamp, err\n}", "func (s *SysbenchMutex) parseExecutionTime(data string) (float64, error) {\n\tmatch := executionTimeRE.FindStringSubmatch(data)\n\tif len(match) < 2 {\n\t\treturn 0.0, fmt.Errorf(\"could not find execution time average: %s\", data)\n\t}\n\treturn strconv.ParseFloat(match[1], 64)\n}", "func Time(val interface{}) time.Time {\r\n\tswitch t := val.(type) {\r\n\t// We could use this later.\r\n\tdefault:\r\n\t\ts := String(t)\r\n\t\tfor _, format := range strToTimeFormats {\r\n\t\t\tr, err := time.ParseInLocation(format, s, time.Local)\r\n\t\t\tif err == nil {\r\n\t\t\t\treturn r\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n\treturn time.Time{}\r\n}", "func TimeStr2Time(timeStr string) (time.Time, error) {\n\tlocal := time.Local\n\treturn time.ParseInLocation(\"2006-01-02 15:04:05\", timeStr, local)\n}", "func (p *Parser) Time(i int, context string) Time {\n\ts := p.String(i, context)\n\tif p.err != nil {\n\t\treturn Time{}\n\t}\n\tv, err := ParseTime(s)\n\tif err != nil {\n\t\tp.SetErr(context, s)\n\t}\n\treturn v\n}", "func ParseTimeOrDuration(input string, now time.Time) (time.Time, error) {\n\tif input == \"\" {\n\t\treturn time.Time{}, nil\n\t}\n\n\ttimestamp, err := time.Parse(time.RFC3339, input)\n\tif err == nil {\n\t\treturn timestamp, nil\n\t}\n\n\tduration, err := time.ParseDuration(input)\n\tif err != nil {\n\t\treturn time.Time{}, fmt.Errorf(\n\t\t\t\"Could not parse %s as either timestamp or duration\",\n\t\t\tinput,\n\t\t)\n\t}\n\treturn now.Add(duration), nil\n}", "func ParseTimeString(datestring string) (parsed time.Time) {\n\tdatestring = strings.TrimSpace(datestring)\n\tfmt.Printf(\"Parsing time string \\\"%s\\\"\", datestring)\n\tt := time.Now()\n\tif len(datestring) != 0 {\n\t\tparsedTime, err := dateparse.ParseAny(datestring)\n\t\tif err != nil {\n\t\t\tt = time.Now()\n\t\t\ttoLog(\"ParseTimeString\", \"Failed to parse given date string: \"+datestring)\n\t\t\t// fmt.Println(err)\n\t\t} else {\n\t\t\tt = parsedTime\n\t\t}\n\t}\n\treturn t\n}", "func ToTime(strTime string) *time.Time {\n\tt, err := time.Parse(\"2006-01-02T15:04:05\", strTime)\n\tif err != nil {\n\t\tt, err = time.Parse(\"2006-01-02T15:04:05.000\", strTime)\n\t\tif err != nil {\n\t\t\tlog.Errorf(err, \"Failed to parse datetime: %v\", strTime)\n\t\t}\n\t}\n\treturn &t\n}", "func ParseTimeOfDay(s string) (TimeOfDay, error) {\n\tif len(s) == 5 && s[2] == ':' {\n\t\thour, hErr := strconv.Atoi(s[0:2])\n\t\tmin, mErr := strconv.Atoi(s[3:5])\n\t\tif hErr == nil && 0 <= hour && hour <= 23 && mErr == nil && 0 <= min && min <= 59 {\n\t\t\td := time.Duration(hour)*time.Hour + time.Duration(min)*time.Minute\n\t\t\treturn Duration(d).TimeOfDay(), nil\n\t\t}\n\t}\n\treturn 0, fmt.Errorf(\"ParseTimeOfDay: %q must be of the form HH:MM\", s)\n}", "func handleTime(portions []string) (time.Time, error) {\n\tret := time.Now()\n\tidx := len(portions) - 1\n\tportions[idx] = strings.Replace(portions[idx], \")\", \"\", -1)\n\tisNowToken := regexp.MustCompile(\"now\")\n\t// check if parsing relative timestamps\n\tif isNowToken.MatchString(portions[0]) {\n\t\tfor _, val := range portions[1:] {\n\t\t\t// parse the relative duration\n\t\t\tdur, err := parseIntoDuration(val)\n\t\t\tif err != nil {\n\t\t\t\treturn ret, err\n\t\t\t}\n\t\t\t// adjust the time by the duration amount\n\t\t\tret = ret.Add(dur)\n\t\t}\n\t} else {\n\t\tportions[0] = strings.Replace(portions[idx], \"(\", \"\", -1)\n\t\tportions[0] = strings.Replace(portions[idx], \",\", \"\", -1)\n\t\ttimestring := strings.Join(portions, \" \")\n\t\tlog.Debug(\"parsing\", timestring)\n\t\tfor _, format := range supported_formats {\n\t\t\tt, err := time.Parse(format, timestring)\n\t\t\tif err != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tret = t\n\t\t\tbreak\n\t\t}\n\t}\n\treturn ret, nil\n}", "func ParseHuman(timestr string) (time.Duration, error) {\n\tvar t, counter time.Duration\n\tfor i, l := 0, len(timestr); i < l; i++ {\n\t\tc := timestr[i]\n\t\tif c >= '0' && c <= '9' {\n\t\t\tcounter = counter*10 + time.Duration(c-'0')\n\t\t\tcontinue\n\t\t}\n\n\t\tswitch c {\n\t\tcase 'H':\n\t\t\tt += counter * time.Hour\n\t\tcase 'M':\n\t\t\tt += counter * time.Minute\n\t\tcase 'S':\n\t\t\tt += counter * time.Second\n\t\tcase 'm':\n\t\t\tt += counter * time.Millisecond\n\t\tcase 'u':\n\t\t\tt += counter * time.Microsecond\n\t\tcase 'n':\n\t\t\tt += counter * time.Nanosecond\n\t\tdefault:\n\t\t\treturn 0, ErrWrongHumanTimeFormat\n\t\t}\n\t\tcounter = 0\n\t}\n\n\treturn t, nil\n}", "func TimeParseAny(value string) (*time.Time, error) {\n\tif t, err := time.Parse(\"2006-01-02 15:04:05\", value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.RFC3339, value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.RFC3339Nano, value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.RFC822, value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.RFC822Z, value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.RFC850, value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.RFC1123, value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.RFC1123Z, value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.UnixDate, value); err == nil {\n\t\treturn &t, nil\n\t} else if t, err = time.Parse(time.RubyDate, value); err == nil {\n\t\treturn &t, nil\n\t}\n\treturn nil, fmt.Errorf(\"parse datetime %s error\", value)\n}", "func ParseFileTime(filename string) (time.Time, error) {\n\tbase := strings.TrimSuffix(filename, filepath.Ext(filename))\n\treturn time.Parse(defaults.AuditLogTimeFormat, base)\n}", "func convertTimeFormat(s string) time.Time {\n\tif s != \"\" {\n\t\tinitT, err := time.Parse(\"01/02/2006\", s)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tsT := initT.Format(time.RFC3339)\n\t\tt, err := time.Parse(time.RFC3339, sT)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\treturn t\n\t}\n\tt := new(time.Time)\n\treturn *t\n}", "func ISO8601Parse(iso8601 string) (t time.Time, err error) {\n\tfor _, layout := range []string{\n\t\tiso8601TimeFormat,\n\t\tiso8601TimeFormatLong,\n\t\ttime.RFC3339,\n\t} {\n\t\tt, err = time.Parse(layout, iso8601)\n\t\tif err == nil {\n\t\t\treturn t, nil\n\t\t}\n\t}\n\n\treturn t, err\n}", "func (c *TIMESTAMP) Parse() string {\n\tEmpty := TIMESTAMP{}\n\tif *c != Empty {\n\t\tt, err := time.Parse(\"2006-01-02 03:04:05.0000\", c.Time)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\t\treturn fmt.Sprintf(`\"timestamp\": \"%v\"`, t.Format(\"2006-01-02 03:04:05.0000\"))\n\t}\n\treturn \"\"\n}", "func MustParse(datestr string) time.Time {\n\tt, err := parseTime(datestr, nil)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\treturn t\n}", "func parseDateTime(dateString, timeString string) (time.Time, error) {\n\tdateTime, err := time.Parse(time.RFC3339, dateString + \"T\" + timeString + \"-05:00\")\n\tif err != nil {\n\t\treturn dateTime, errors.New(\"Invalid Date or Time given.\")\n\t}\n\treturn dateTime, err\n}", "func parseTimeQuantum(v string) (TimeQuantum, error) {\n\tq := TimeQuantum(strings.ToUpper(v))\n\tif !q.Valid() {\n\t\treturn \"\", ErrInvalidTimeQuantum\n\t}\n\treturn q, nil\n}", "func ParseDateTime(s string) (time.Time, error) {\n\treturn time.Parse(TimeFormatISO8601DateTime, s)\n}", "func ParseIn(datestr string, loc *time.Location) (time.Time, error) {\n\treturn parseTime(datestr, loc)\n}", "func parseLocalTime(tm string) (time.Time, error) {\n\tloc, _ := time.LoadLocation(\"America/New_York\")\n\treturn time.ParseInLocation(\"2006-1-2T15:04\", tm, loc)\n}", "func Parse(year, month, day, hourMinute string, loc *time.Location) (time.Time, error) {\n\tnow := time.Now().In(loc)\n\n\ty64, err := strconv.ParseInt(year, 10, 0)\n\ty := int(y64)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\tif y < now.Year()-1000 {\n\t\treturn time.Time{}, fmt.Errorf(\"bad year; %d is too far in the past\", y)\n\t}\n\tm, err := strconv.ParseInt(month, 10, 0)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\tif m < 0 || m > 11 {\n\t\treturn time.Time{}, fmt.Errorf(\"bad month: %d is not within [0, 11]\", m)\n\t}\n\t// Month +1 since time.Month is [1, 12].\n\tm = m + 1\n\td64, err := strconv.ParseInt(day, 10, 0)\n\td := int(d64)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\tif d < 1 {\n\t\treturn time.Time{}, fmt.Errorf(\"bad day: %d; can't be negative\", d)\n\t} else if d > daysIn(time.Month(m), y) {\n\t\treturn time.Time{}, fmt.Errorf(\"bad day: %d; only %d days in %v, %d\", d, daysIn(time.Month(m), y), time.Month(m), y)\n\t}\n\tparts := strings.Split(hourMinute, \":\")\n\tif len(parts) != 2 {\n\t\treturn time.Time{}, fmt.Errorf(\"bad hour/minute: %s\", hourMinute)\n\t}\n\th, err := strconv.ParseInt(parts[0], 10, 0)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\tif h < 0 || h > 60 {\n\t\treturn time.Time{}, fmt.Errorf(\"bad hour: %d\", h)\n\t}\n\tmin, err := strconv.ParseInt(parts[1], 10, 0)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\tif min < 0 || min > 60 {\n\t\treturn time.Time{}, fmt.Errorf(\"bad minute: %d\", min)\n\t}\n\n\tt := time.Time(time.Date(int(y), time.Month(m), int(d), int(h), int(min), 0, 0, loc))\n\tif t.After(now) {\n\t\treturn time.Time{}, fmt.Errorf(\"bad time; %v is in the future\", time.Time(t))\n\t}\n\treturn t, nil\n}", "func ParseInLocation(layout, value string, loc *time.Location) (*time.Time, error) {\n\tt, err := time.ParseInLocation(layout, value, loc)\n\treturn &t, err\n}", "func ParseDate(date string, dateFormat string) time.Time {\n\tvar (\n\t\terr error\n\t\tgoDate time.Time\n\t)\n\n\tif len(dateFormat) < 1 {\n\t\treturn time.Time{}\n\t}\n\n\tif goDate, err = time.ParseInLocation(convertDateFormatToGoFormat(dateFormat), date, time.Local); err != nil {\n\t\treturn time.Time{}\n\t}\n\n\treturn goDate\n}" ]
[ "0.7695394", "0.7498485", "0.739932", "0.73629034", "0.72838485", "0.7269359", "0.72644264", "0.72294974", "0.7206954", "0.72009856", "0.7193034", "0.71692765", "0.71193546", "0.71139544", "0.7111959", "0.7111016", "0.70746815", "0.7039669", "0.70367527", "0.70133245", "0.69696605", "0.69518906", "0.68858546", "0.6884347", "0.6882774", "0.6861639", "0.6855178", "0.6837714", "0.68290395", "0.67795247", "0.6755868", "0.67225087", "0.6714587", "0.6712771", "0.6695461", "0.6681565", "0.6674278", "0.66548085", "0.6578731", "0.6578442", "0.6493845", "0.64639497", "0.6434205", "0.64256084", "0.64163494", "0.63344884", "0.6287604", "0.62787116", "0.62649393", "0.624962", "0.6202766", "0.6198085", "0.6176968", "0.6148519", "0.61278254", "0.6095338", "0.6063504", "0.60558397", "0.6048803", "0.59985495", "0.5960909", "0.59539974", "0.59525937", "0.5947057", "0.59180266", "0.5915016", "0.5912741", "0.5901253", "0.5895092", "0.58779216", "0.5876816", "0.58714104", "0.58576214", "0.585197", "0.5813822", "0.58060145", "0.5772402", "0.57519907", "0.57168424", "0.5709924", "0.5704291", "0.570194", "0.5693374", "0.56781626", "0.5671949", "0.56531304", "0.5652516", "0.5634366", "0.5626422", "0.56225777", "0.56159914", "0.5611096", "0.5609152", "0.5601947", "0.55799574", "0.5576954", "0.55755675", "0.5574519", "0.5542494", "0.5521337" ]
0.8885299
0
/ Regular Expressions Matches use regexp.MatchString to check if operand matches pattern.
func Matches(pattern string, operand string) (bool, error) { return regexp.MatchString(pattern, operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Match(regx string, arg string) bool {\n\tmatched, err := regexp.MatchString(\"^(\"+regx+\")$\", arg)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn matched\n}", "func main() {\n\tfmt.Println(isMatch(\"mississippi\", \"mis*is*p*.\"))\n\tfmt.Println(isMatch(\"aab\", \"c*a*b\"))\n}", "func (String) Matches(pattern string) bool { return boolResult }", "func RegexpEqual(x *syntax.Regexp, y *syntax.Regexp,) bool", "func matchFunc(a, b string) bool {\n\tmatched, _ := regexp.MatchString(b, a)\n\treturn matched\n}", "func isMatch(s string, p string) bool {\n\n}", "func Contains(substr, operand string) bool { return strings.Contains(operand, substr) }", "func Match(r Regex, s string) bool {\n\tfor _, c := range s {\n\t\tr = r.Derivative(c)\n\t}\n\treturn r.Accepting()\n}", "func (v *Value) Match(expr string) bool {\n\t// Compile the regular expression.\n\tre, err := v.script.compileRegexp(expr)\n\tif err != nil {\n\t\treturn false // Fail silently\n\t}\n\n\t// Return true if the expression matches the value, interpreted as a\n\t// string.\n\tloc := re.FindStringIndex(v.String())\n\tif loc == nil {\n\t\tv.script.RStart = 0\n\t\tv.script.RLength = -1\n\t\treturn false\n\t}\n\tv.script.RStart = loc[0] + 1\n\tv.script.RLength = loc[1] - loc[0]\n\treturn true\n}", "func TestMatch(t *testing.T) {\n\tpattern := \"^(B|b)rian$\"\n\tif match, err := regexp.Match(pattern, []byte(\"Brian\")); match != true {\n\t\tt.Errorf(\"Brian did not match %q %v\", pattern, err)\n\t}\n\n\tif match, err := regexp.Match(pattern, []byte(\"brian\")); match != true {\n\t\tt.Errorf(\"brian did not match %q %v\", pattern, err)\n\t}\n}", "func QuoteRegex(operand string) string { return regexp.QuoteMeta(operand) }", "func (a *Assertions) Regexp(rx interface{}, str interface{}, msgAndArgs ...interface{}) bool {\n\treturn Regexp(a.t, rx, str, msgAndArgs...)\n}", "func Match(t Testing, reg, str interface{}, formatAndArgs ...interface{}) bool {\n\tif !tryMatch(reg, str) {\n\t\treturn Fail(t,\n\t\t\tfmt.Sprintf(\"Expect string(%s) to match regexp(%s)\", fmt.Sprint(str), fmt.Sprint(reg)),\n\t\t\tformatAndArgs...)\n\t}\n\n\treturn true\n}", "func TestPattern2(t *testing.T) {\n\tre := MustCompile(\"a$\")\n\tif !re.MatchString(\"a\") {\n\t\tt.Errorf(\"expect to match\\n\")\n\t}\n\tif re.MatchString(\"ab\") {\n\t\tt.Errorf(\"expect to mismatch\\n\")\n\t}\n}", "func MatchString(infix, matchString string) bool {\n\tn := Compile(infix)\n\treturn n.Matches(matchString)\n}", "func (s *sqlStrConcat) Match(n ast.Node, c *gosec.Context) (*gosec.Issue, error) {\n\tif node, ok := n.(*ast.BinaryExpr); ok {\n\t\tif start, ok := node.X.(*ast.BasicLit); ok {\n\t\t\tif str, e := gosec.GetString(start); e == nil {\n\t\t\t\tif !s.MatchPatterns(str) {\n\t\t\t\t\treturn nil, nil\n\t\t\t\t}\n\t\t\t\tif _, ok := node.Y.(*ast.BasicLit); ok {\n\t\t\t\t\treturn nil, nil // string cat OK\n\t\t\t\t}\n\t\t\t\tif second, ok := node.Y.(*ast.Ident); ok && s.checkObject(second, c) {\n\t\t\t\t\treturn nil, nil\n\t\t\t\t}\n\t\t\t\treturn gosec.NewIssue(c, n, s.ID(), s.What, s.Severity, s.Confidence), nil\n\t\t\t}\n\t\t}\n\t}\n\treturn nil, nil\n}", "func Match(t TestingT, r, v interface{}, extras ...interface{}) bool {\n\treg, ok := tryMatch(r, v)\n\tif !ok {\n\t\t_, acts := toString(nil, v)\n\n\t\tErrorf(t, \"Expect to match regexp\", []labeledOutput{\n\t\t\t{\n\t\t\t\tlabel: labelMessages,\n\t\t\t\tcontent: formatExtras(extras...),\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"-regexp\",\n\t\t\t\tcontent: fmt.Sprintf(\"%#v\", reg.String()),\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"+value\",\n\t\t\t\tcontent: fmt.Sprintf(\"%#v\", acts),\n\t\t\t},\n\t\t})\n\t}\n\n\treturn ok\n}", "func matchExps(value string, expressions []*regexp.Regexp) bool {\n\tfor _, exp := range expressions {\n\t\tif exp.MatchString(value) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func matchExps(value string, expressions []*regexp.Regexp) bool {\n\tfor _, exp := range expressions {\n\t\tif exp.MatchString(value) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func regexpMatch(r *regexp.Regexp, s string) bool {\n\treturn r != nil && r.MatchString(s)\n}", "func Match(regex string, text string) bool {\n runerx := compile(regex)\n runetxt := []rune(text)\n \n if len(runerx) > 0 && runerx[0] == '^' {\n return matchhere(runerx[1:], runetxt)\n }\n \n for {\n if matchhere(runerx, runetxt) {\n return true\n }\n if len(runetxt) == 0 {\n return false\n }\n runetxt = runetxt[1:]\n }\n }", "func (t *dataType) Match(r *regexp.Regexp) *dataType {\n\tt.str.Match(r)\n\treturn t\n}", "func ContainsAny(chars, operand string) bool { return strings.ContainsAny(operand, chars) }", "func matchesRegex(pattern, str interface{}) bool {\n\tmatch, err := regexp.MatchString(fmt.Sprint(pattern), fmt.Sprint(str))\n\tif err != nil {\n\t\tlog.Errorf(\"bad regex expression %s\", fmt.Sprint(pattern))\n\t\treturn false\n\t}\n\tscope.Debugf(\"%v regex %v? %v\\n\", pattern, str, match)\n\treturn match\n}", "func Matches(re string) (desc string, f predicate.PredicateFunc) {\n\tdesc = fmt.Sprintf(\"{} =~ /%v/\", re)\n\tf = func(v interface{}) (r bool, ctx []predicate.ContextValue, err error) {\n\t\ts, ok := v.(string)\n\t\tif !ok {\n\t\t\treturn false, nil, fmt.Errorf(\n\t\t\t\t\"value of type '%T' cannot be matched against a regexp\", v)\n\t\t}\n\t\tm, err := regexp.MatchString(re, s)\n\t\tif err != nil {\n\t\t\treturn false, nil, fmt.Errorf(\"failed to compile regexp: %w\", err)\n\t\t}\n\t\treturn m, nil, nil\n\t}\n\treturn\n}", "func Regexp(str string, pattern string) bool {\n\tok, _ := regexp.MatchString(pattern, str)\n\treturn ok\n}", "func Regexp(str string, pattern string) bool {\n\tok, _ := regexp.MatchString(pattern, str)\n\treturn ok\n}", "func (t *StringDataType) Match(r *regexp.Regexp) *StringDataType {\n\treturn t.Validate(func(s string) error {\n\t\tif !r.MatchString(s) {\n\t\t\treturn fmt.Errorf(\"value does not match passed in regex %s\", r.String())\n\t\t}\n\t\treturn nil\n\t})\n}", "func (MatchedText) Matches(pattern string) bool { return boolResult }", "func m(t *testing.T, s, re string) {\n\tt.Helper()\n\tmatched, err := regexp.MatchString(re, s)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif !matched {\n\t\tt.Errorf(\"string does not match pattern %q:\\n%s\", re, s)\n\t}\n}", "func Grep(pattern string, operand []string) ([]string, error) {\n\trex, err := regexp.Compile(pattern)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar matching []string\n\tfor _, elem := range operand {\n\t\tif rex.MatchString(elem) {\n\t\t\tmatching = append(matching, elem)\n\t\t}\n\t}\n\treturn matching, nil\n}", "func Matches(str, pattern string) bool {\n\tmatch, _ := regexp.MatchString(pattern, str)\n\treturn match\n}", "func matches(regex string, input string) bool {\n\tif len(regex) == 0 && len(input) == 0 {\n\t\treturn true\n\t}\n\tvar fi, fr, sr, tr byte\n\tif len(regex) > 0 {\n\t\tfr = regex[0]\n\t}\n\tif len(regex) > 1 {\n\t\tsr = regex[1]\n\t}\n\tif len(regex) > 2 {\n\t\ttr = regex[2]\n\t}\n\tif len(input) > 0 {\n\t\tfi = input[0]\n\t}\n\tif sr == '*' {\n\t\tif fr == fi {\n\t\t\treturn matches(regex, input[1:])\n\t\t} else {\n\t\t\tfor i := 0; i < len(input); i++ {\n\t\t\t\tif input[i] == tr {\n\t\t\t\t\treturn matches(regex[2:], input[i:])\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn false\n\t\t}\n\t} else {\n\t\tif fr == fi {\n\t\t\treturn matches(regex[1:], input[1:])\n\t\t} else {\n\t\t\treturn false\n\t\t}\n\t}\n}", "func Match(regexStr string, textStr string) (result bool, err error) {\n\tregex := []rune(regexStr)\n\ttext := []rune(textStr)\n\tif len(regex) > 0 && regex[0] == '^' {\n\t\treturn matchHere(regex[1:], text)\n\t}\n\tif len(text) == 0 {\n\t\treturn matchHere(regex, text)\n\t}\n\tfor i, _ := range text {\n\t\tr, e := matchHere(regex, text[i:])\n\t\tif r || e != nil {\n\t\t\treturn r, e\n\t\t}\n\t}\n\treturn result, err\n}", "func TestMatchString(t *testing.T) {\n\tpattern, upper, lower := \"^(B|b)rian$\", \"Brian\", \"brian\"\n\n\tif match, err := regexp.MatchString(pattern, upper); match != true {\n\t\tt.Errorf(\"MatchString did not match %q %v\", upper, err)\n\t}\n\n\tif match, err := regexp.MatchString(pattern, lower); match != true {\n\t\tt.Errorf(\"MatchString did not match %q %v\", lower, err)\n\t}\n}", "func Regex(expr ConstString, input String) (Bool, error) {\n\tif expr.HasVariable() {\n\t\treturn nil, fmt.Errorf(\"regex requires a constant expression as its first parameter, but it has a variable parameter\")\n\t}\n\te, err := expr.Eval()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr, err := regexp.Compile(e)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn TrimBool(&regex{\n\t\tr: r,\n\t\tS: input,\n\t\texpr: e,\n\t\thash: Hash(\"regex\", expr, input),\n\t\thasVariable: input.HasVariable(),\n\t}), nil\n}", "func TestPattern1(t *testing.T) {\n\tre := MustCompile(`b\\$a`)\n\tif !re.MatchString(\"b$a\") {\n\t\tt.Errorf(\"expect to match\\n\")\n\t}\n\tre = MustCompile(\"b\\\\$a\")\n\tif !re.MatchString(\"b$a\") {\n\t\tt.Errorf(\"expect to match 2\\n\")\n\t}\n}", "func (re *Regexp) Match(b []byte) bool\t{ return len(re.doExecute(\"\", b, 0)) > 0 }", "func regexEqualMatch(mrb *oruby.MrbState, self oruby.Value) oruby.MrbValue {\n\tvar s string\n\targs := mrb.GetArgs()\n\tdest := args.Item(0)\n\tpos := oruby.MrbFixnum(args.ItemDef(1, oruby.MrbFixnumValue(0)))\n\n\tswitch dest.Type() {\n\tcase oruby.MrbTTSymbol:\n\t\ts = mrb.SymString(oruby.MrbSymbol(dest))\n\tcase oruby.MrbTTString:\n\t\ts = mrb.StrToCstr(dest)\n\tdefault:\n\t\treturn oruby.False\n\t}\n\n\tregx := mrb.Data(self).(*regexp.Regexp)\n\treturn oruby.Bool(regx.MatchString(s[pos:]))\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func (re *Regexp) MatchString(s string) bool\t{ return len(re.doExecute(s, nil, 0)) > 0 }", "func GMatch(value, mask string) bool {\n\tmask = regexp.QuoteMeta(mask)\n\tmask = strings.Replace(mask, `\\*`, \".*\", -1)\n\tmask = strings.Replace(mask, `\\?`, \".\", -1)\n\tmask = strings.Replace(mask, `\\[`, \"[\", -1)\n\tmask = strings.Replace(mask, `\\]`, \"]\", -1)\n\tif r, _ := regexp.Compile(mask); r != nil {\n\t\treturn r.MatchString(value)\n\t}\n\treturn false\n}", "func Test_regex(t *testing.T) {\n\n\tfor _, val := range shouldMatch {\n\t\tif !secretKarmaReg.MatchString(val) {\n\t\t\tt.Errorf(\"String %s should have matched but didn't.\", val)\n\t\t}\n\t}\n\n\t// for _, val := range shouldNotMatch {\n\t// \tif secretKarmaReg.MatchString(val) {\n\t// \t\tt.Errorf(\"String %s should not have have matched but did.\", val)\n\t// \t}\n\t// }\n}", "func MatchRegExp(e RegExp, x Value) bool {\n\t// TODO: add a (global) cache for r.\n\tr, err := regexp.Compile(string(e))\n\tif err != nil {\n\t\treturn false\n\t}\n\tif r.MatchString(string(x)) {\n\t\treturn true\n\t}\n\treturn false\n}", "func matchStr(rgxp string, compare string) bool {\n\tr, err := regexp.Compile(rgxp)\n\tif err != nil {\n\t\tlog.Fatalf(\"invalid regexp: %s\", rgxp)\n\t}\n\treturn r.MatchString(strings.ToLower(compare))\n}", "func NormalMatch(regx string, arg string) bool {\n\tmatched, err := regexp.MatchString(regx, arg)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn matched\n}", "func RegexpMatch(content, pattern string) bool {\n\tre, err := pool.Compile(pattern)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn re.MatchString(content)\n}", "func RegexpAssert(str string,regex string) bool{\n\tvar Regex_x = regexp.MustCompile(regex)\n\tif Regex_x.MatchString(str) == true {\n\t\treturn true\n\t} else {\n\t\tif show == true {\n\t\t\tfmt.Printf(\"Failed! %v and %v are not equal : \\n\")\n\t\t}\n\t\treturn false\n\t}\n}", "func matchhere(regex []rune, text []rune) bool {\n if len(regex) == 0 {\n return true\n }\n if len(regex) > 1 && regex[1] == '*' {\n return matchstar(regex[0], regex[2:], text)\n }\n if regex[0] == '$' && len(regex) == 1 {\n return len(text) == 0\n }\n if len(text) > 0 && (regex[0] == '.' || regex[0] == text[0]) {\n return matchhere(regex[1:], text[1:])\n }\n return false\n }", "func (m RegexpMatcher) Match(s string) bool {\n\treturn m.re.MatchString(s)\n}", "func RegexMatch(key1 string, key2 string) bool {\n\tres, err := regexp.MatchString(key2, key1)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn res\n}", "func RegexMatch(key1 string, key2 string) bool {\n\tres, err := regexp.MatchString(key2, key1)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn res\n}", "func isOperator(x string) bool {\n\tif x == \"+\" || x == \"-\" || x == \"/\" || x == \"*\" {\n\t\treturn true\n\t}\n\treturn false\n}", "func matchRegex(params models.RegexParamsProvider, str string) (bool, string) {\n\tregex := params.GetRegexp()\n\tif regex == nil {\n\t\treturn true, str\n\t}\n\n\tif !regex.MatchString(str) {\n\t\treturn false, \"\"\n\t}\n\n\tsubstrings := regex.FindStringSubmatch(str)\n\tif len(substrings) == 1 {\n\t\treturn true, str\n\t}\n\n\treturn true, substrings[1]\n}", "func (v *VerbalExpression) Test(s string) bool {\n\treturn v.Regex().Match([]byte(s))\n}", "func Operator(input string) int {\n\tvar o string = \"+-*/%^\"\n\n\treturn strings.IndexAny(input, o)\n}", "func OkMatchesString(label, val, regex string, t *testing.T) {\n\tre := regexp.MustCompile(regex)\n\tif re.MatchString(val) {\n\t\tt.Logf(\"ok - %s: '%s' matches '%s'\\n\", label, val, regex)\n\t} else {\n\t\tt.Logf(\"not ok - %s: String '%s' doesn't match '%s'\", label, val, regex)\n\t\tt.Fail()\n\t}\n}", "func (g GiveCommand) Matches(str string) bool {\n\treturn giveReg.MatchString(str)\n}", "func (rule *Rule) match(bufferLeft string, bufferRight string) (bool, error) {\n\tre, err := regexp.Compile(rule.RegexpLeft)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tmatches := re.FindStringSubmatch(bufferLeft)\n\tnames := re.SubexpNames()\n\tif len(matches) != len(names) {\n\t\treturn false, nil\n\t}\n\tparamMap := map[string]string{}\n\tfor i, name := range names {\n\t\tif i != 0 && name != \"\" {\n\t\t\tparamMap[name] = matches[i]\n\t\t}\n\t}\n\trule.BufferLeft = strings.Replace(\n\t\trule.BufferLeft,\n\t\t\"[]\",\n\t\tbufferLeft,\n\t\t-1,\n\t)\n\trule.BufferRight = strings.Replace(\n\t\trule.BufferRight,\n\t\t\"[]\",\n\t\tbufferRight,\n\t\t-1,\n\t)\n\trule.paramMap = paramMap\n\treturn true, nil\n}", "func StringMatchesPattern(re *regexp.Regexp, reDesc string) String {\n\treturn func(v string) error {\n\t\tif !re.MatchString(v) {\n\t\t\treturn fmt.Errorf(\"%s does not match the pattern: %s\",\n\t\t\t\tv, reDesc)\n\t\t}\n\t\treturn nil\n\t}\n}", "func RegexMatchFunc(args ...interface{}) (interface{}, error) {\n\tname1 := args[0].(string)\n\tname2 := args[1].(string)\n\n\treturn (bool)(RegexMatch(name1, name2)), nil\n}", "func (p *Parser) parseRegexpLiteral() asti.ExpressionI {\n\n\tflags := \"\"\n\n\tval := p.curToken.Literal\n\tif strings.HasPrefix(val, \"(?\") {\n\t\tval = strings.TrimPrefix(val, \"(?\")\n\n\t\ti := 0\n\t\tfor i < len(val) {\n\n\t\t\tif val[i] == ')' {\n\n\t\t\t\tval = val[i+1:]\n\t\t\t\tbreak\n\t\t\t} else {\n\t\t\t\tflags += string(val[i])\n\t\t\t}\n\n\t\t\ti++\n\t\t}\n\t}\n\treturn &ast.RegexpLiteral{Token: p.curToken, Value: val, Flags: flags}\n}", "func Match(pattern string, b []byte) (matched bool, error string) {\n\tre, err := CompileRegexp(pattern);\n\tif err != \"\" {\n\t\treturn false, err\n\t}\n\treturn re.Match(b), \"\";\n}", "func groupOnlyMatches(group *syntax.Regexp, s string) bool {\n\tswitch group.Op {\n\tcase syntax.OpLiteral:\n\t\tfor _, r := range group.Rune {\n\t\t\tif !strings.ContainsRune(s, r) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t\treturn true\n\n\tcase syntax.OpCharClass:\n\t\tfor i := 0; i < len(group.Rune); i += 2 {\n\t\t\tlo, hi := group.Rune[i], group.Rune[i+1]\n\t\t\tfor r := lo; r <= hi; r++ {\n\t\t\t\tif !strings.ContainsRune(s, r) {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn true\n\n\tcase syntax.OpStar, syntax.OpPlus, syntax.OpRepeat, syntax.OpQuest, syntax.OpCapture:\n\t\treturn groupOnlyMatches(group.Sub[0], s)\n\n\tcase syntax.OpConcat, syntax.OpAlternate:\n\t\tfor _, sub := range group.Sub {\n\t\t\tif !groupOnlyMatches(sub, s) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\n\tdefault:\n\t\treturn false\n\t}\n\treturn true\n}", "func assertStringRegexp(t *testing.T, pattern, s string) {\n\tmatches, _ := regexp.MatchString(pattern, s)\n\n\tif !matches {\n\t\tt.Errorf(\"%s is not in format %s\", s, pattern)\n\t}\n}", "func wildcardMatch(pattern, text string) (bool, error) {\n\tpatternLen := len(pattern)\n\ttextLen := len(text)\n\tif patternLen == 0 {\n\t\treturn textLen == 0, nil\n\t}\n\n\tif pattern == \"*\" {\n\t\treturn true, nil\n\t}\n\n\tpattern = strings.ToLower(pattern)\n\ttext = strings.ToLower(text)\n\n\tmatch, err := regexp.MatchString(toRegexPattern(pattern), text)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"wildcardMatch: unable to perform regex matching: %w\", err)\n\t}\n\n\treturn match, nil\n}", "func isMatch(exp *regexp.Regexp, val interface{}) bool {\n\tswitch v := val.(type) {\n\tcase []rune:\n\t\treturn exp.MatchString(string(v))\n\tcase []byte:\n\t\treturn exp.Match(v)\n\tcase string:\n\t\treturn exp.MatchString(v)\n\tdefault:\n\t\treturn false\n\t}\n}", "func Regexp(expr string) (func(string) bool, error) {\n\tif expr == \"\" {\n\t\treturn nil, fmt.Errorf(\"empty regex expression\")\n\t}\n\n\t// add the last $ if missing (and not wildcard(?))\n\tif i := expr[len(expr)-1]; i != '$' && i != '*' {\n\t\texpr += \"$\"\n\t}\n\n\tr, err := regexp.Compile(expr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn r.MatchString, nil\n}", "func (e PiRegExp) exp() *regexp.Regexp { return regexp.MustCompile(string(e)) }", "func Match(patternStr, str string) (matched bool, err error) {\n\tif patternStr == \"\" {\n\t\treturn true, nil\n\t}\n\n\tpattern := preparePattern(patternStr)\n\tstrs, err := prepareStr(str)\n\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn match(pattern, strs)\n}", "func valid_expression(line string, numbers [4]string) bool {\n\tnumbers_string := strings.Trim(reg.ReplaceAllString(line, \" \"), \" \")\t\n\n\tfor _, num := range numbers {\n\t\tnew_string := strings.Replace(numbers_string, num, \"\", 1)\n\t\tif new_string == numbers_string {\n\t\t\treturn false\n\t\t} \n\t\tnumbers_string = new_string\n\t}\n\n\treturn true\n}", "func TestExpr(t *testing.T) {\n\tt.Run(\"valid expression compiling into RPN and JSON encoding/decoding\", func(t *testing.T) {\n\t\tentries := []testExprEntry{\n\t\t\t{\n\t\t\t\traw: \"!tasty__|delish\",\n\t\t\t\texpectedRPN: \"tasty_,!,delish,|\",\n\t\t\t\texpectedJSON: `{\"raw\":\"!tasty|delish\",\"rpn\":[{\"s\":\"tasty\",\"!\":1,\"r\":1},{\"s\":\"!\"},{\"s\":\"delish\"},{\"s\":\"|\"}],\"compiled\":true}`,\n\t\t\t\tevalRPN: []testEvalEntry{},\n\t\t\t},\n\t\t\t{\n\t\t\t\traw: \"barfoo|(foobar)\",\n\t\t\t\texpectedRPN: \"barfoo,foobar,|\",\n\t\t\t\texpectedJSON: `{\"raw\":\"barfoo|(foobar)\",\"rpn\":[{\"s\":\"barfoo\"},{\"s\":\"foobar\"},{\"s\":\"|\"}],\"compiled\":true}`,\n\t\t\t\tevalRPN: []testEvalEntry{\n\t\t\t\t\t{text: \"this is a basic example of some text foobar\", shouldMatch: true, strs: []string{\"foobar\"}},\n\t\t\t\t\t{text: \"this is a barfoo basic example of some text foo bar\", shouldMatch: true, strs: []string{\"barfoo\"}},\n\t\t\t\t\t{text: \"this is a bar foo basic example of some text foo bar\", shouldMatch: false},\n\t\t\t\t\t{text: \"this is a basic example foo of some text bar\", shouldMatch: false},\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\traw: \"((((ch?ips))))|(fish***+(((tasty))))\",\n\t\t\t\texpectedRPN: \"ch?ips,fish*,tasty,+,|\",\n\t\t\t\texpectedJSON: `{\"raw\":\"((((ch?ips))))|(fish***+(((tasty))))\",\"rpn\":[{\"s\":\"ch?ips\",\"r\":1},{\"s\":\"fish*\",\"r\":1},{\"s\":\"tasty\"},{\"s\":\"+\"},{\"s\":\"|\"}],\"compiled\":true}`,\n\t\t\t\tevalRPN: []testEvalEntry{\n\t\t\t\t\t{text: \"chips fish tasty\", shouldMatch: true, strs: []string{\"fish\", \"tasty\", \"chips\"}}, // \"fish tasty\" is not a returned match because of regex behavior\n\t\t\t\t\t{text: \"fish tasty\", shouldMatch: true, strs: []string{\"fish\", \"tasty\"}},\n\t\t\t\t\t{text: \"chiips\", shouldMatch: true, strs: []string{\"chiips\"}},\n\t\t\t\t\t{text: \"fish\", shouldMatch: false},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\n\t\tfor i, entry := range entries {\n\t\t\tt.Run(\"should all pass\", func(t *testing.T) {\n\t\t\t\ttestExprHelper(t, i, entry)\n\t\t\t})\n\t\t}\n\t})\n\n\tt.Run(\"invalid expression compiling into RPN and JSON encoding/decoding\", func(t *testing.T) {\n\t\tconst (\n\t\t\tinfixErr = SyntaxError(\"unexpected infix operator, want operand\")\n\t\t)\n\n\t\tentries := []testExprEntry{\n\t\t\t{raw: \"(hi++hi1)\", err: infixErr, expectedJSON: `{\"raw\":\"(hi++hi1)\",\"rpn\":[],\"compiled\":false}`},\n\t\t}\n\n\t\tfor i, entry := range entries {\n\t\t\tt.Run(\"should all fail\", func(t *testing.T) {\n\t\t\t\tentry.shouldFail = true\n\t\t\t\ttestExprHelper(t, i, entry)\n\t\t\t})\n\t\t}\n\t})\n\n}", "func (b ValExprBuilder) RegExp(val interface{}) BoolExprBuilder {\n\treturn b.makeComparisonExpr(astRegExp, makeValExpr(val))\n}", "func (l *ECMAScriptLexer) isRegexPossible() bool {\n\tif lastToken == nil {\n\t\t// No token has been produced yet: at the start of the input,\n\t\t// no division is possible, so a regex literal _is_ possible.\n\t\treturn true\n\t}\n\n\ttokenType := lastToken.GetTokenType()\n\tif tokenType == ECMAScriptLexerIdentifier ||\n\t\ttokenType == ECMAScriptLexerNullLiteral ||\n\t\ttokenType == ECMAScriptLexerBooleanLiteral ||\n\t\ttokenType == ECMAScriptLexerThis ||\n\t\ttokenType == ECMAScriptLexerCloseBracket ||\n\t\ttokenType == ECMAScriptLexerCloseParen ||\n\t\ttokenType == ECMAScriptLexerOctalIntegerLiteral ||\n\t\ttokenType == ECMAScriptLexerDecimalLiteral ||\n\t\ttokenType == ECMAScriptLexerHexIntegerLiteral ||\n\t\ttokenType == ECMAScriptLexerStringLiteral ||\n\t\ttokenType == ECMAScriptLexerPlusPlus ||\n\t\ttokenType == ECMAScriptLexerMinusMinus {\n\t\t// After any of the tokens above, no regex literal can follow.\n\t\treturn false\n\t}\n\n\t// In all other cases, a regex literal _is_ possible.\n\treturn true\n}", "func ShExpMatch(str, shexp string) bool {\n\tshexp = strings.Replace(shexp, \".\", \"\\\\.\", -1)\n\tshexp = strings.Replace(shexp, \"?\", \".?\", -1)\n\tshexp = strings.Replace(shexp, \"*\", \".*\", -1)\n\tmatched, err := regexp.MatchString(\"^\"+shexp+\"$\", str)\n\treturn err == nil && matched\n}", "func TestPatternMatches(t *testing.T) {\n\tmatch, _ := Matches(\"fileutils.go\", []string{\"*.go\"})\n\tif match != true {\n\t\tt.Errorf(\"failed to get a match, got %v\", match)\n\t}\n}", "func (t anyToken) Matches(r rune) bool {\n\tif t.negate {\n\t\treturn false\n\t}\n\treturn true\n}", "func TestCompileRegexp(t *testing.T) {\n\tregex, err := regexp.Compile(\"^(B|b)rian$\")\n\n\tif regex == nil {\n\t\tt.Errorf(\"Regex did not compile %v\", err)\n\t}\n}", "func (r *Go) Match(s string) bool {\n\tt := time.Now()\n\tdefer MatchHistogram.With(\"string\", s, \"duration\", \"seconds\").Observe(time.Since(t).Seconds())\n\n\treturn r.reg.MatchString(s)\n}", "func Operation(value, requiredValue string, operation oval.Operation) bool {\n\tswitch operation {\n\tcase oval.OpEquals:\n\t\treturn value == requiredValue\n\tcase oval.OpNotEquals:\n\t\treturn value != requiredValue\n\tcase oval.OpCaseInsensitiveEquals:\n\t\treturn strings.EqualFold(value, requiredValue)\n\tcase oval.OpCaseInsensitiveNotEquals:\n\t\treturn !strings.EqualFold(value, requiredValue)\n\tcase oval.OpGreaterThan:\n\t\treturn value > requiredValue\n\tcase oval.OpLessThan:\n\t\treturn value < requiredValue\n\tcase oval.OpGreaterThanOrEqual:\n\t\treturn value >= requiredValue\n\tcase oval.OpLessThanOrEqual:\n\t\treturn value <= requiredValue\n\tcase oval.OpPatternMatch:\n\t\texp, err := regexp.Compile(requiredValue)\n\t\tif err != nil {\n\t\t\treturn false\n\t\t}\n\t\treturn exp.Match([]byte(value))\n\n\tdefault:\n\t\treturn false\n\t}\n}", "func TestEval(t *testing.T) {\n\tany := `.+`\n\ttestCases := []struct {\n\t\tname string\n\t\tquery string\n\t\twantErr string\n\t\twant []values.Value\n\t}{\n\t\t{\n\t\t\tname: \"string interpolation\",\n\t\t\tquery: `\n\t\t\t\tstr = \"str\"\n\t\t\t\ting = \"ing\"\n\t\t\t\t\"str + ing = ${str+ing}\"`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewString(\"str + ing = string\"),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"string interpolation missing field\",\n\t\t\tquery: `\n\t\t\t\tr = makeRecord(o: {a: \"foo\", b: 42})\n\t\t\t\t\"r._value = ${r._value}\"`,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"string interpolation non-string type\",\n\t\t\tquery: `\n\t\t\t\tr = makeRecord(o: {a: \"foo\", b: 42})\n\t\t\t\t\"r._value = ${r.b}\"`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewString(\"r._value = 42\"),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"string interpolation wrong type\",\n\t\t\tquery: `\n\t\t\t\tr = makeRecord(o: {a: \"foo\", b: 42})\n\t\t\t\t\"r = ${r}\"`,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"call builtin function\",\n\t\t\tquery: \"six()\",\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewFloat(6.0),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"call function with fail\",\n\t\t\tquery: \"fail()\",\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"call function with duplicate args\",\n\t\t\tquery: \"plusOne(x:1.0, x:2.0)\",\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"binary expressions\",\n\t\t\tquery: `\n\t\t\tsix_value = six()\n\t\t\tnine_value = nine()\n\n\t\t\tfortyTwo() == six_value * nine_value\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"logical expressions short circuit\",\n\t\t\tquery: `\n six_value = six()\n nine_value = nine()\n\n not (fortyTwo() == six_value * nine_value) or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"function\",\n\t\t\tquery: `\n plusSix = (r) => r + six()\n plusSix(r:1.0) == 7.0 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"function block\",\n\t\t\tquery: `\n f = (r) => {\n r1 = 1.0 + r\n return (r + r1) / r\n }\n f(r:1.0) == 3.0 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"function block polymorphic\",\n\t\t\tquery: `\n f = (r) => {\n r2 = r * r\n return r2 / r\n }\n f(r:2.0) == 2.0 or fail()\n f(r:2) == 2 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"function with default param\",\n\t\t\tquery: `\n addN = (r,n=4) => r + n\n addN(r:2) == 6 or fail()\n addN(r:3,n:1) == 4 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"scope closing\",\n\t\t\tquery: `\n\t\t\tx = 5\n plusX = (r) => r + x\n plusX(r:2) == 7 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"nested scope mutations not visible outside\",\n\t\t\tquery: `\n\t\t\tx = 5\n xinc = () => {\n x = x + 1\n return x\n }\n xinc() == 6 or fail()\n x == 5 or fail()\n\t\t\t`,\n\t\t},\n\t\t// TODO(jsternberg): This test seems to not\n\t\t// infer the type constraints correctly for m.a,\n\t\t// but it doesn't fail.\n\t\t{\n\t\t\tname: \"return map from func\",\n\t\t\tquery: `\n toMap = (a,b) => ({\n a: a,\n b: b,\n })\n m = toMap(a:1, b:false)\n m.a == 1 or fail()\n not m.b or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"pipe expression\",\n\t\t\tquery: `\n\t\t\tadd = (a=<-,b) => a + b\n\t\t\tone = 1\n\t\t\tone |> add(b:2) == 3 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"ignore pipe default\",\n\t\t\tquery: `\n\t\t\tadd = (a=<-,b) => a + b\n\t\t\tadd(a:1, b:2) == 3 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"pipe expression function\",\n\t\t\tquery: `\n\t\t\tadd = (a=<-,b) => a + b\n\t\t\tsix() |> add(b:2.0) == 8.0 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"pipe builtin function\",\n\t\t\tquery: `\n\t\t\tsix() |> plusOne() == 7.0 or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"regex match\",\n\t\t\tquery: `\n\t\t\t\"abba\" =~ /^a.*a$/ or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"regex not match\",\n\t\t\tquery: `\n\t\t\t\"abc\" =~ /^a.*a$/ and fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"not regex match\",\n\t\t\tquery: `\n\t\t\t\"abc\" !~ /^a.*a$/ or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"not regex not match\",\n\t\t\tquery: `\n\t\t\t\"abba\" !~ /^a.*a$/ and fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"options metadata\",\n\t\t\tquery: `\n\t\t\toption task = {\n\t\t\t\tname: \"foo\",\n\t\t\t\trepeat: 100,\n\t\t\t}\n\t\t\ttask.name == \"foo\" or fail()\n\t\t\ttask.repeat == 100 or fail()\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"query with side effects\",\n\t\t\tquery: `sideEffect() == 0 or fail()`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewInt(0),\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"array index expression\",\n\t\t\tquery: `\n\t\t\t\ta = [1, 2, 3]\n\t\t\t\tx = a[1]\n\t\t\t\tx == 2 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"dict expression\",\n\t\t\tquery: `\n\t\t\t\tm = [\"a\" + \"b\": 0, \"c\" + \"d\": 1]\n\t\t\t\tx = get(dict: m, key: \"ab\", default: 2)\n\t\t\t\ty = get(dict: m, key: \"cd\", default: 2)\n\t\t\t\tz = get(dict: m, key: \"ef\", default: 2)\n\t\t\t\tx == 0 and y == 1 and z == 2 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"empy dictionary\",\n\t\t\tquery: `\n\t\t\t\tm0 = [:]\n\t\t\t\tm1 = insert(dict: m0, key: \"a\", value: 0)\n\t\t\t\tm2 = insert(dict: m0, key: 0, value: \"a\")\n\t\t\t\tv1 = get(dict: m1, key: \"a\", default: -1)\n\t\t\t\tv2 = get(dict: m2, key: 0, default: \"b\")\n\t\t\t\tv1 == 0 and v2 == \"a\" or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"array index expression out of bounds low\",\n\t\t\tquery: `\n\t\t\t\ta = [1, 2, 3]\n\t\t\t\ti = -1\n\t\t\t\tx = a[i]\n\t\t\t`,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"array index expression out of bounds high\",\n\t\t\tquery: `\n\t\t\t\ta = [1, 2, 3]\n\t\t\t\ti = 3\n\t\t\t\tx = a[i]\n\t\t\t`,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"array with complex index expression\",\n\t\t\tquery: `\n\t\t\t\tf = () => ({l: 0, m: 1, n: 2})\n\t\t\t\ta = [1, 2, 3]\n\t\t\t\tx = a[f().l]\n\t\t\t\ty = a[f().m]\n\t\t\t\tz = a[f().n]\n\t\t\t\tx == 1 or fail()\n\t\t\t\ty == 2 or fail()\n\t\t\t\tz == 3 or fail()\n\t\t\t`,\n\t\t},\n\t\t{\n\t\t\tname: \"short circuit logical and\",\n\t\t\tquery: `\n false and fail()\n `,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"short circuit logical or\",\n\t\t\tquery: `\n true or fail()\n `,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"no short circuit logical and\",\n\t\t\tquery: `\n true and fail()\n `,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"no short circuit logical or\",\n\t\t\tquery: `\n false or fail()\n `,\n\t\t\twantErr: any,\n\t\t},\n\t\t{\n\t\t\tname: \"conditional true\",\n\t\t\tquery: `\n\t\t\t\tif 1 != 0 then 10 else 100\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewInt(10),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"conditional false\",\n\t\t\tquery: `\n\t\t\t\tif 1 == 0 then 10 else 100\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewInt(100),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"conditional in function\",\n\t\t\tquery: `\n\t\t\t\tf = (t, c, a) => if t then c else a\n\t\t\t\t{\n\t\t\t\t\tv1: f(t: false, c: 30, a: 300),\n\t\t\t\t\tv2: f(t: true, c: \"cats\", a: \"dogs\"),\n\t\t\t\t}\n\t\t\t`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewObjectWithValues(map[string]values.Value{\n\t\t\t\t\t\"v1\": values.NewInt(300),\n\t\t\t\t\t\"v2\": values.NewString(\"cats\"),\n\t\t\t\t}),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"exists\",\n\t\t\tquery: `hasValue(o: makeRecord(o: {value: 1}))`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(true),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"exists null\",\n\t\t\tquery: `hasValue(o: makeRecord(o: {val: 2}))`,\n\t\t\twant: []values.Value{\n\t\t\t\tvalues.NewBool(false),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"invalid function parameter\",\n\t\t\tquery: `from(bucket: \"telegraf\") |> window(every: 0s)`,\n\t\t\twantErr: `error calling function \"window\" @\\d+:\\d+-\\d+:\\d+: window function requires at least one of \"every\" or \"period\" to be set and non-zero`,\n\t\t},\n\t\t{\n\t\t\t// tests that we don't nest error messages when\n\t\t\t// a function call fails and gets piped into another\n\t\t\t// function.\n\t\t\tname: \"nested function error\",\n\t\t\tquery: `from(bucket: \"telegraf\") |> window(every: 0s) |> mean()`,\n\t\t\twantErr: `error calling function \"window\" @\\d+:\\d+-\\d+:\\d+: window function requires at least one of \"every\" or \"period\" to be set and non-zero`,\n\t\t},\n\t}\n\n\tfor _, tc := range testCases {\n\t\ttc := tc\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\tsrc := prelude + tc.query\n\n\t\t\tctx, deps := dependency.Inject(context.Background(), dependenciestest.Default())\n\t\t\tdefer deps.Finish()\n\n\t\t\tsideEffects, _, err := runtime.Eval(ctx, src)\n\t\t\tif err != nil {\n\t\t\t\tif tc.wantErr == \"\" {\n\t\t\t\t\tt.Fatalf(\"unexpected error: %s\", err)\n\t\t\t\t}\n\n\t\t\t\t// We expect an error, so it should be a non-internal Flux error.\n\t\t\t\tif code := flux.ErrorCode(err); code == codes.Internal || code == codes.Unknown {\n\t\t\t\t\tt.Errorf(\"expected non-internal error code, got %s\", code)\n\t\t\t\t}\n\n\t\t\t\tre := regexp.MustCompile(tc.wantErr)\n\t\t\t\tif got := err.Error(); !re.MatchString(got) {\n\t\t\t\t\tt.Errorf(\"expected error to match pattern %q, but error was %q\", tc.wantErr, got)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t} else if tc.wantErr != \"\" {\n\t\t\t\tt.Fatal(\"expected error\")\n\t\t\t}\n\n\t\t\tvs := getSideEffectsValues(sideEffects)\n\t\t\tif tc.want != nil && !cmp.Equal(tc.want, vs, semantictest.CmpOptions...) {\n\t\t\t\tt.Fatalf(\"unexpected side effect values -want/+got: \\n%s\", cmp.Diff(tc.want, vs, semantictest.CmpOptions...))\n\t\t\t}\n\t\t})\n\t}\n}", "func RegExp(pattern string) *RegExpMatcher {\n\treturn &RegExpMatcher{Pattern: pattern}\n}", "func Regexp(expr *regexp.Regexp) Pattern {\n\treturn regexpMatch{expr}\n}", "func literal(s string) *regexp.Regexp {\n\tre := match(regexp.QuoteMeta(s))\n\n\tif _, complete := re.LiteralPrefix(); !complete {\n\t\tpanic(\"must be a literal\")\n\t}\n\n\treturn re\n}", "func literal(s string) *regexp.Regexp {\n\tre := match(regexp.QuoteMeta(s))\n\n\tif _, complete := re.LiteralPrefix(); !complete {\n\t\tpanic(\"must be a literal\")\n\t}\n\n\treturn re\n}", "func literal(s string) *regexp.Regexp {\n\tre := match(regexp.QuoteMeta(s))\n\n\tif _, complete := re.LiteralPrefix(); !complete {\n\t\tpanic(\"must be a literal\")\n\t}\n\n\treturn re\n}", "func (this *MatchString) matchStr(str string, mSrc string) bool {\n\tres, err := regexp.MatchString(mSrc, str)\n\treturn res == true && err == nil\n}", "func matchPattern(pattern, name string) (matched bool) {\n\tif pattern == \"\" {\n\t\treturn name == pattern\n\t}\n\tif pattern == \"*\" {\n\t\treturn true\n\t}\n\trName, rPattern := make([]rune, 0, len(name)), make([]rune, 0, len(pattern))\n\tfor _, r := range name {\n\t\trName = append(rName, r)\n\t}\n\tfor _, r := range pattern {\n\t\trPattern = append(rPattern, r)\n\t}\n\treturn deepMatchRune(rName, rPattern, false)\n}", "func (m Matcher) Match(pattern string, alternatives ...string) Matcher {\n\treturn m\n}", "func (re *RegexpStd) Match(b []byte) bool {\n\tpanic(\"\")\n}", "func PlusLiteral(src string) bool {\n\tidx := 0\n\tisMatching := true\n\n\tplusBoolResult54 := false\n\n\tstaticStr53 := \"a\"\n\tstaticStr53Len := len(staticStr53)\n\n\tif staticStr53Len+idx > len(src) {\n\t\tisMatching = false\n\t} else if staticStr53 == src[idx:idx+staticStr53Len] {\n\t\tidx += staticStr53Len\n\t\tisMatching = true\n\t} else {\n\t\tisMatching = false\n\t}\n\tplusIdxResult55 := idx\n\n\tif isMatching {\n\t\tfor {\n\n\t\t\tstaticStr56 := \"a\"\n\t\t\tstaticStr56Len := len(staticStr56)\n\n\t\t\tif staticStr56Len+plusIdxResult55 > len(src) {\n\t\t\t\tplusBoolResult54 = false\n\t\t\t} else if staticStr56 == src[plusIdxResult55:plusIdxResult55+staticStr56Len] {\n\t\t\t\tplusIdxResult55 += staticStr56Len\n\t\t\t\tplusBoolResult54 = true\n\t\t\t} else {\n\t\t\t\tplusBoolResult54 = false\n\t\t\t}\n\n\t\t\tif !plusBoolResult54 {\n\t\t\t\tbreak\n\t\t\t} else {\n\t\t\t\tisMatching = plusBoolResult54\n\t\t\t\tidx = plusIdxResult55\n\t\t\t}\n\t\t}\n\t}\n\n\tif idx != len(src) {\n\t\tisMatching = false\n\t}\n\n\treturn isMatching\n\t// End of PlusLiteral\n}", "func Match(pattern string, b []byte) bool {\n\tre, err := Compile(pattern)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn re.match(b)\n}", "func (m *Matcher) Match(message string) (matched bool, arg string, err error) {\n\tmatches := m.matchRegex.FindStringSubmatch(message)\n\tif matches == nil {\n\t\tmatched = false\n\t\treturn\n\t}\n\n\tmatched = true\n\n\tif len(m.arguments) == 0 {\n\t\treturn\n\t} else if len(matches) == 0 {\n\t\tlog.Fatal(\"Expected argument but did not set it up in the regex.\")\n\t}\n\n\targ = strings.TrimSpace(matches[1])\n\tif arg == \"\" {\n\t\terr = Error{name: m.name, arguments: m.arguments}\n\t}\n\treturn\n}", "func TestCheckBinaryExprStringAndNotString(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `\"abc\" &^ \"abc\"`, env,\n\t\t`invalid operation: \"abc\" &^ \"abc\" (operator &^ not defined on string)`,\n\t)\n\n}", "func match(a, b string) bool {\n\treturn strings.EqualFold(a, b)\n}", "func (v *validator) IncludesOperator(str string) bool {\n\tincludes, _ := regexp.MatchString(v.ValidOperatorExp, str)\n\treturn includes\n}", "func matchHere(regex []rune, text []rune) (result bool, err error) {\n\tif len(regex) == 0 {\n\t\treturn true, err\n\t}\n\tif regex[0] == '*' {\n\t\treturn result, invalidRegexErr\n\t}\n\tif regex[0] == '$' {\n\t\tif len(regex) > 1 {\n\t\t\treturn result, invalidRegexErr\n\t\t}\n\t\treturn len(text) == 0, err\n\t}\n\tif len(regex) > 1 && regex[1] == '*' {\n\t\treturn matchStar(regex[0], regex[2:], text)\n\t}\n\tif regex[0] == '.' || regex[0] == text[0] {\n\t\treturn matchHere(regex[1:], text[1:])\n\t}\n\treturn result, err\n}", "func (c *Command) Match(msg string) bool {\n\treturn c.Reg.MatchString(msg)\n}" ]
[ "0.6712451", "0.6439957", "0.6344449", "0.6230377", "0.6228559", "0.62159455", "0.6212523", "0.62079054", "0.6198076", "0.618296", "0.61804867", "0.61698055", "0.61658955", "0.6119233", "0.60992146", "0.60759795", "0.6074126", "0.6049856", "0.6049856", "0.6004666", "0.5965482", "0.5946833", "0.59389585", "0.5938309", "0.5936453", "0.592317", "0.592317", "0.5906203", "0.5903971", "0.58985984", "0.5890553", "0.58868456", "0.58866423", "0.5886526", "0.58638924", "0.585738", "0.5854871", "0.5849236", "0.58485514", "0.5836679", "0.5836679", "0.5836679", "0.58358276", "0.5816529", "0.58155406", "0.5794802", "0.57804626", "0.5759765", "0.57549345", "0.5751519", "0.57176685", "0.56990033", "0.5685371", "0.5685371", "0.56827193", "0.5673635", "0.56699955", "0.56684643", "0.5665756", "0.56579834", "0.56479406", "0.5647006", "0.5636057", "0.5623677", "0.5621855", "0.5618398", "0.5617027", "0.5609481", "0.5601026", "0.5598444", "0.5596198", "0.55911016", "0.5583233", "0.55787206", "0.5575652", "0.55665135", "0.55572134", "0.55567163", "0.5549807", "0.554641", "0.55375934", "0.55022824", "0.54798627", "0.5471521", "0.5452459", "0.54503036", "0.54503036", "0.54503036", "0.54498476", "0.5448998", "0.5447653", "0.54427266", "0.54289925", "0.54278666", "0.54254764", "0.54202306", "0.5419662", "0.54189825", "0.5418972", "0.54121643" ]
0.7500275
0
CompileRegex uses regexp.Compile to compile a new regexp.Regexp according to pattern.
func CompileRegex(pattern string) (*regexp.Regexp, error) { return regexp.Compile(pattern) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func CompileRegexp(pattern string) *regexp.Regexp {\n\tif pattern == \"\" {\n\t\treturn nil\n\t}\n\n\t// add ^...$ to all regexp when not given\n\tif !strings.HasPrefix(pattern, \"^\") {\n\t\tpattern = \"^\" + pattern\n\t}\n\tif !strings.HasSuffix(pattern, \"$\") {\n\t\tpattern += \"$\"\n\t}\n\n\t// make all regexp case insensitive by default\n\tif !strings.Contains(pattern, \"(?i)\") {\n\t\tpattern = \"(?i)\" + pattern\n\t}\n\n\treturn regexp.MustCompile(pattern)\n}", "func CompileRegexp(str string) (regexp *Regexp, error string) {\n\tregexp = new(Regexp);\n\tregexp.expr = str;\n\tregexp.inst = make([]instr, 0, 20);\n\terror = regexp.doParse();\n\treturn;\n}", "func Compile(input string) (*Regexp, error) {\n\treturn Compile2(input, 0)\n}", "func (r *Regexps) Compile(pattern string) (*regexp.Regexp, error) {\n\tr.Lock()\n\tdefer r.Unlock()\n\tif r.items[pattern] != nil {\n\t\treturn r.items[pattern], nil\n\t}\n\tre, err := regexp.Compile(pattern)\n\tif err == nil {\n\t\tr.items[pattern] = re\n\t}\n\treturn re, err\n}", "func getRegexpCompile(pattern string) *regexp.Regexp {\n\tr, err := regexp.Compile(pattern)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn r\n}", "func Compile(pattern string) (Matcher, error) {\n\tregex, err := syntax.Parse(pattern, syntax.Perl)\n\tif err != nil {\n\t\treturn Matcher{}, err\n\t}\n\n\tregex = optimize(regex).Simplify()\n\tm, err := compile(regex)\n\treturn Matcher{m}, err\n}", "func (p *DefaultParser) CompileRegex(fields []string) ([]*regexp.Regexp, error) {\n\tr := make([]*regexp.Regexp, len(fields))\n\tfor i, f := range fields {\n\t\t/*\n\t\t\tvar nextField string\n\t\t\tif i == len(fields)-1 {\n\t\t\t\tnextField = \"\"\n\t\t\t} else {\n\t\t\t\tnextField = fields[i+1]\n\t\t\t}\n\t\t*/\n\t\tfieldName := p.NamedGroup(f)\n\t\t// TODO(stratus): This is the foundation for possibly two\n\t\t// regexes - one for easy single line fields and another one for\n\t\t// multi-field.\n\t\tre, err := regexp.Compile(fmt.Sprintf(`(?mis)%s\\s*(?P<%s>.*?)\\n`, f, fieldName))\n\t\t//re, err := regexp.Compile(fmt.Sprintf(`(?mis)%s\\s*(?P<%s>.*?)%s`, f, fieldName, nextField))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tr[i] = re\n\t}\n\treturn r, nil\n}", "func Regexp(pattern string) (*regexp.Regexp, error) {\n\treturn pool.Compile(pattern)\n}", "func Compile(expr string) (*NRegexp, error) {\n\tre, err := regexp.Compile(expr)\n\treturn &NRegexp{Regexp: re}, err\n}", "func compile(pattern string, observeCase bool) (err error) {\n\tif observeCase {\n\t\t_, err = regexp.Compile(pattern)\n\t} else {\n\t\t_, err = regexp.Compile(\"(?i)\" + pattern)\n\t}\n\n\treturn err\n}", "func (c *TransferCounter) CompileRegex(operator string) (*regexp.Regexp, error) {\n\tvar r *regexp.Regexp\n\tvar err error\n\n\tfor _, regionOperator := range regionOperators {\n\t\tif operator == regionOperator {\n\t\t\tr, err = regexp.Compile(\".*?operator finish.*?region-id=([0-9]*).*?\" + operator + \".*?store \\\\[([0-9]*)\\\\] to \\\\[([0-9]*)\\\\].*?\")\n\t\t}\n\t}\n\n\tfor _, leaderOperator := range leaderOperators {\n\t\tif operator == leaderOperator {\n\t\t\tr, err = regexp.Compile(\".*?operator finish.*?region-id=([0-9]*).*?\" + operator + \".*?store ([0-9]*) to ([0-9]*).*?\")\n\t\t}\n\t}\n\n\tif r == nil {\n\t\terr = errors.New(\"unsupported operator. \")\n\t}\n\treturn r, err\n}", "func TestCompileRegexp(t *testing.T) {\n\tregex, err := regexp.Compile(\"^(B|b)rian$\")\n\n\tif regex == nil {\n\t\tt.Errorf(\"Regex did not compile %v\", err)\n\t}\n}", "func Compile(re *syntax.Regexp,) (*syntax.Prog, error)", "func (c Configuration) compileRegex(listofPaths []string) []*regexp.Regexp {\n\tlogger := c.logger()\n\tvar regexpObjects []*regexp.Regexp\n\tfor _, path := range listofPaths {\n\t\treg, err := regexp.Compile(path)\n\t\tif err != nil {\n\t\t\tlogger.Error().Err(err).Msgf(\"Error while compiling regex: %v\", err)\n\t\t\tcontinue\n\t\t}\n\t\tregexpObjects = append(regexpObjects, reg)\n\t}\n\treturn regexpObjects\n}", "func compile(regex string) (regslc []rune) {\n regslc = make([]rune, 0, len(regex) + 10)\n \n for _, r := range regex {\n if r == '+' {\n regslc = append(regslc, regslc[len(regslc) - 1], '*')\n } else {\n regslc = append(regslc, r)\n }\n } \n return regslc\n }", "func CompileERE(pattern string) (*regexp.Regexp, error) { return regexp.CompilePOSIX(pattern) }", "func compileRegex() {\n\tfor _, value := range filterObj.Signatures {\n\t\tr, _ := regexp.Compile(value.Regex)\n\t\tsignatureRegex = append(signatureRegex, r)\n\t}\n\tfor _, value := range filterObj.Shallows {\n\t\tr, _ := regexp.Compile(value.Regex)\n\t\tshallowRegex = append(shallowRegex, r)\n\t}\n}", "func (p *Generic) CompileRegexps(patterns map[string]string) (map[string]*regexp.Regexp, error) {\n\tresults := make(map[string]*regexp.Regexp)\n\tfor name, str := range patterns {\n\t\tre, err := regexp.Compile(str)\n\t\tif nil != err {\n\t\t\treturn results, fmt.Errorf(\"Cannot compile '%s' regexp: %s\", name, err.Error())\n\t\t}\n\t\tresults[name] = re\n\t}\n\n\treturn results, nil\n}", "func MustCompile(input string) *Regexp {\n\treturn MustCompile2(input, 0)\n}", "func MustCompile(str string) *Regexp {\n\tregexp, error := CompileRegexp(str);\n\tif error != \"\" {\n\t\tpanicln(`regexp: compiling \"`, str, `\": `, error)\n\t}\n\treturn regexp;\n}", "func MustCompile(str string) *NRegexp {\n\treturn &NRegexp{Regexp: regexp.MustCompile(str)}\n}", "func Compile(re string, lookup interface{}) (*regexp.Regexp, error) {\n\tvar regexpCompiled = regexp.MustCompile(re)\n\n\tlookupValue := reflect.ValueOf(lookup)\n\tif lookupValue.Kind() != reflect.Ptr {\n\t\treturn nil, fmt.Errorf(\"lookup must be a pointer\")\n\t}\n\n\tlookupElem := lookupValue.Elem()\n\tsubexpNames := map[string]int{}\n\tfor i, name := range regexpCompiled.SubexpNames() {\n\t\tif name == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tif cases.Title(language.English, cases.NoLower).String(name) != name {\n\t\t\treturn nil, fmt.Errorf(\"field '%s' in regexp `%s` is not capitalized\", name, re)\n\t\t}\n\t\tsubexpNames[name] = i\n\t}\n\n\tfor i := 0; i < lookupElem.NumField(); i++ {\n\t\tfield := lookupElem.Type().Field(i)\n\t\tfieldName := field.Name\n\t\tindex, foundIndex := subexpNames[fieldName]\n\t\tif !foundIndex {\n\t\t\treturn nil, fmt.Errorf(\"field '%s' is not in regexp `%s`\", fieldName, re)\n\t\t}\n\t\tlookupElem.FieldByName(fieldName).SetInt(int64(index))\n\t\tdelete(subexpNames, fieldName)\n\t}\n\n\tvar unusedNames []string\n\tfor name := range subexpNames {\n\t\tunusedNames = append(unusedNames, name)\n\t}\n\n\tif len(unusedNames) > 0 {\n\t\treturn nil, fmt.Errorf(\"regexp `%s` has named groups not found in lookup: %s\", re, strings.Join(unusedNames, \", \"))\n\t}\n\n\treturn regexpCompiled, nil\n}", "func (p *unicodePattern) Compile(patternStr string, escape byte) {\n\tp.patChars, p.patTypes = stringutil.CompilePatternInner(patternStr, escape)\n}", "func NewRegex(v string) (Value, error) {\n\trx, err := regexp.Compile(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rxValue{rx}, nil\n}", "func compileRegexps(regexps []string) ([]*regexp.Regexp, error) {\n\tvar errs []error\n\tvar compiled []*regexp.Regexp\n\n\tfor _, exp := range regexps {\n\t\tre, err := regexp.Compile(exp)\n\t\tif err == nil {\n\t\t\tcompiled = append(compiled, re)\n\t\t} else {\n\t\t\terrs = append(errs, err)\n\t\t}\n\t}\n\n\tif len(errs) > 0 {\n\t\treturn nil, errors.Join(errs...)\n\t}\n\treturn compiled, nil\n}", "func compileGlob(glob string) (*regexp.Regexp, error) {\n\tglob = strings.Replace(glob, \".\", \"\\\\.\", -1)\n\tglob = strings.Replace(glob, \"*\", \".*\", -1)\n\tglob = \"^\" + glob + \"$\"\n\treturn regexp.Compile(glob)\n}", "func NewRegexp(s string) (Regexp, error) {\n\tregex, err := regexp.Compile(s)\n\treturn Regexp{\n\t\tRegexp: regex,\n\t\toriginal: s,\n\t}, err\n}", "func NewRegexp(exp string) (*Regexp, error) {\n\tr, err := regexp.Compile(exp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Regexp{r}, nil\n}", "func compileCInstructionRegexp() *regexp.Regexp {\n\tdestRe := \"(?:([ADM]+)=)?\" // the destination part of the regexp\n\tcompRe := pipe(CComp) // the compute part of the regexp\n\tjumpRe := pipe(CJump) // the jump part of the regexp\n\ttmp := fmt.Sprintf(\"^%s%s(?:;%s)?$\", destRe, compRe, jumpRe)\n\tlog.Debug().Str(\"C Instruction Regex\", tmp)\n\treturn regexp.MustCompile(tmp)\n}", "func r(pattern string) *regexp.Regexp { return regexp.MustCompile(pattern) }", "func Regex(pattern, options string) Val {\n\tregex := primitive.Regex{Pattern: pattern, Options: options}\n\treturn Val{t: bsontype.Regex, primitive: regex}\n}", "func Regexp(expr *regexp.Regexp) Pattern {\n\treturn regexpMatch{expr}\n}", "func Convert(pattern string) (*regexp.Regexp, error) {\n\tpattern = \"^\" + metaReplacer.Replace(regexp.QuoteMeta(pattern)) + \"$\"\n\treturn regexp.Compile(pattern)\n}", "func (p *parser) regex() Node {\n\ttoken := p.expect(TokenRegex)\n\tr, err := newRegex(token.pos, token.val)\n\tif err != nil {\n\t\tp.error(err)\n\t}\n\treturn r\n}", "func NewRegexp(s string) (Regexp, error) {\n\tregex, err := regexp.Compile(\"^(?:\" + s + \")$\")\n\treturn Regexp{\n\t\tRegexp: regex,\n\t\toriginal: s,\n\t}, err\n}", "func (l *LazyRegexp) Pattern() *regexp.Regexp {\n\tif l.regexp == nil { // No need to take a lock. Worst case, we'll just compile it multiple times.\n\t\tl.regexp = regexp.MustCompile(l.pattern)\n\t}\n\treturn l.regexp\n}", "func MustCompilePOSIX(str string) *NRegexp {\n\treturn &NRegexp{Regexp: regexp.MustCompilePOSIX(str)}\n}", "func newCompiledRegexes(matchItems []v1beta1.MatchItem) (map[*v1beta1.MatchItem]*regexp.Regexp, error) {\n\tvar err error\n\tcache := make(map[*v1beta1.MatchItem]*regexp.Regexp)\n\n\tfor i := range matchItems {\n\t\tif matchItems[i].MatchRegex != nil {\n\t\t\tcache[&matchItems[i]], err = regexp.Compile(*matchItems[i].MatchRegex)\n\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn cache, nil\n}", "func CompilePOSIX(expr string) (*NRegexp, error) {\n\tre, err := regexp.CompilePOSIX(expr)\n\treturn &NRegexp{Regexp: re}, err\n}", "func CompileStrs(regStrs []string) (*regexp.Regexp, error) {\n\tif len(regStrs) == 0 {\n\t\treturn nil, fmt.Errorf(\"regStrs must be greater than or equal to 1\")\n\t}\n\tregStr := \"(\" + strings.Join(regStrs, \"|\") + \")\"\n\tre, err := regexp.Compile(regStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn re, nil\n}", "func TransformRegExp(pattern string) (string, error) {\n\n\tlog.Printf(\"runtime.Caller(1): %#+v\\n\", myutils.Slice(runtime.Caller(1))...)\n\tos.Exit(349857394)\n\n\t// if pattern == \"\" {\n\t// \treturn \"\", nil\n\t// }\n\n\tswitch pattern {\n\t// common cases that will leave as is\n\t// but frequently occurred in many\n\t// popular frameworks like react.js etc.\n\tcase _regexp_empty,\n\t\t_regexp_any_nc,\n\t\t_regexp_any_c,\n\t\t_regexp_anyplus_nc,\n\t\t_regexp_anyplus_c,\n\n\t\t_regexp_squote_nc,\n\t\t_regexp_squote_c,\n\t\t_regexp_dquote_nc,\n\t\t_regexp_dquote_c,\n\n\t\t_regexp_plus_nc,\n\t\t_regexp_plus_c,\n\n\t\t_regexp_urlenc_space_nc,\n\t\t_regexp_urlenc_space_c,\n\t\t_regexp_http_validation,\n\n\t\t_regexp_any_uppercase_nc,\n\t\t_regexp_any_uppercase_c,\n\n\t\t_regexp_is_ms_pref:\n\t\treturn pattern, nil\n\t}\n\n\t// TODO If without \\, if without (?=, (?!, then another shortcut\n\n\tparser := _RegExp_parser{\n\t\tstr: pattern,\n\t\tlength: len(pattern),\n\t\tgoRegexp: bytes.NewBuffer(make([]byte, 0, 3*len(pattern)/2)),\n\t}\n\tparser.read() // Pull in the first character\n\tparser.scan()\n\tvar err error\n\tif len(parser.errors) > 0 {\n\t\terr = parser.errors[0]\n\t}\n\tif parser.invalid {\n\t\tlog.Printf(\"Input: regexp=[%s]\\n\", pattern)\n\t\tlog.Printf(\"Output: regexp=[%s] err=[%s]\\n\", parser.goRegexp.String(), err)\n\t\treturn \"\", err\n\t}\n\n\t// Might not be re2 compatible, but is still a valid JavaScript RegExp\n\treturn parser.goRegexp.String(), err\n}", "func (mem *Mem) Compile() error {\n\tchanged := true\n\tvisited := make(map[int]bool)\n\tfor changed {\n\t\tchanged = false\n\t\tfor state := range mem.patterns {\n\t\t\tif visited[state] {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tvisited[state] = true\n\t\t\tif err := compile(mem, state); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tchanged = true\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Config) CompileRegexps() error {\n\tfor _, licenseRegexps := range c.Licenses {\n\t\tlicenseRegexp := strings.Join(licenseRegexps, \"\\n\")\n\t\tre, err := regexp.Compile(licenseRegexp)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tc.licensesRegexps = append(c.licensesRegexps, re)\n\t}\n\n\tc.accept = make([]rule, 0, len(c.Accept))\n\tfor _, rule := range c.Accept {\n\t\tc.accept = append(c.accept, accept(rule))\n\t}\n\n\tc.reject = make([]rule, 0, len(c.Reject))\n\tfor _, rule := range c.Reject {\n\t\tc.reject = append(c.reject, reject(rule))\n\t}\n\n\treturn nil\n}", "func (re *Regexp) Convert() (*goregexp.Regexp, error) {\n\treturn goregexp.Compile(re.pattern)\n}", "func NewLazyRegexp(pattern string) LazyRegexp {\n\treturn LazyRegexp{pattern, nil}\n}", "func RegExp(pattern string) *RegExpMatcher {\n\treturn &RegExpMatcher{Pattern: pattern}\n}", "func generatePatternForRegexp(pattern string) (string, error) {\n\tpattern = patternRegexp.ReplaceAllStringFunc(pattern, func(subMatch string) string {\n\t\t// The sub match string conforms the parameter pattern: `{parameter-name:regexp-expression}`.\n\t\tfoos := strings.SplitN(subMatch, \":\", 2)\n\t\tif len(foos) < 2 {\n\t\t\treturn `([^/]+)`\n\t\t} else {\n\t\t\treturn \"(\" + foos[1][0:len(foos[1])-1] + \")\"\n\t\t}\n\t})\n\t// Checking for abnormal patterns.\n\t_, err := regexp.Compile(pattern)\n\treturn pattern, err\n}", "func NewGenerator(regex string) (*Generator, error) {\n\tre, err := syntax.Parse(regex, syntax.Perl)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t//fmt.Println(\"Compiled re \", re)\n\treturn &Generator{\n\t\tre: re,\n\t\trand: rand.New(rand.NewSource(time.Now().UnixNano())),\n\t}, nil\n}", "func BuildRegex(cfg YamlConfig) *regexp.Regexp {\n\n\t// Create subPatterns slice from cfg.Definitions\n\tsubPatterns := make([]interface{}, len(cfg.Definitions))\n\tfor i, def := range cfg.Definitions {\n\t\tsubPatterns[i] = def.Pattern\n\t}\n\n\t// Interpolate subpatterns in main pattern, compile regex\n\tpattern := fmt.Sprintf(cfg.LogPattern, subPatterns...)\n\tregex := regexp.MustCompile(pattern)\n\n\treturn regex\n\n}", "func (v *VerbalExpression) Regex() *regexp.Regexp {\n\n\tif !v.compiled {\n\t\tv.regexp = regexp.MustCompile(\n\t\t\tstrings.Join([]string{\n\t\t\t\tstrings.Join(v.parts, \"\"),\n\t\t\t\t`(?` + v.getFlags() + `)`,\n\t\t\t\tv.prefixes,\n\t\t\t\tv.expression,\n\t\t\t\tv.suffixes}, \"\"))\n\t\tv.compiled = true\n\t}\n\treturn v.regexp\n}", "func (e *ExpressionWithSeverity) compile() error {\n\tr, err := regexp.Compile(e.Expression)\n\tif err != nil {\n\t\treturn err\n\t}\n\te.compiled = r\n\treturn nil\n}", "func RegexpMatch(content, pattern string) bool {\n\tre, err := pool.Compile(pattern)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn re.MatchString(content)\n}", "func buildRegexFromList(patterns []string) *regexp.Regexp {\n\tif patterns == nil || len(patterns) == 0 {\n\t\tpatterns = make([]string, 0)\n\t}\n\treturn compileOrDie(\"(?i)^(\" + strings.Join(patterns, \"|\") + \")$\")\n}", "func CompileString(in string) (Matcher, error) {\n\tif in == \"\" {\n\t\treturn Matcher{(*emptyStringMatcher)(nil)}, nil\n\t}\n\treturn Matcher{&substringMatcher{in, []byte(in)}}, nil\n}", "func MustCompile(pattern string) *Abbrev {\n\tabb, err := Compile(pattern)\n\tif err != nil {\n\t\tpanic(\"abbrev: Compile(\" + pattern + \"):\" + err.Error())\n\t}\n\treturn abb\n}", "func regex(s string) (*regexp.Regexp, error) {\n\tif rawString.MatchString(s) {\n\t\ts = fmt.Sprintf(\"^%s$\", s)\n\t}\n\treturn regexp.Compile(s)\n}", "func (p *parser) parseRegex() (rx *regexp.Regexp, err error) {\n\ti := p.i\n\tif len(p.s) < i+2 {\n\t\treturn nil, errors.New(\"expected regular expression, found EOF instead\")\n\t}\n\n\t// number of open parens or brackets;\n\t// when it becomes negative, finished parsing regex\n\topen := 0\n\nloop:\n\tfor i < len(p.s) {\n\t\tswitch p.s[i] {\n\t\tcase '(', '[':\n\t\t\topen++\n\t\tcase ')', ']':\n\t\t\topen--\n\t\t\tif open < 0 {\n\t\t\t\tbreak loop\n\t\t\t}\n\t\t}\n\t\ti++\n\t}\n\n\tif i >= len(p.s) {\n\t\treturn nil, errors.New(\"EOF in regular expression\")\n\t}\n\trx, err = regexp.Compile(p.s[p.i:i])\n\tp.i = i\n\treturn rx, err\n}", "func (p *parser) parseRegex() (rx *regexp.Regexp, err error) {\n\ti := p.i\n\tif len(p.s) < i+2 {\n\t\treturn nil, errors.New(\"expected regular expression, found EOF instead\")\n\t}\n\n\t// number of open parens or brackets;\n\t// when it becomes negative, finished parsing regex\n\topen := 0\n\nloop:\n\tfor i < len(p.s) {\n\t\tswitch p.s[i] {\n\t\tcase '(', '[':\n\t\t\topen++\n\t\tcase ')', ']':\n\t\t\topen--\n\t\t\tif open < 0 {\n\t\t\t\tbreak loop\n\t\t\t}\n\t\t}\n\t\ti++\n\t}\n\n\tif i >= len(p.s) {\n\t\treturn nil, errors.New(\"EOF in regular expression\")\n\t}\n\trx, err = regexp.Compile(p.s[p.i:i])\n\tp.i = i\n\treturn rx, err\n}", "func NewComp(r Regex) Regex {\n\treturn &comp{\n\t\tr: r,\n\t}\n}", "func MustCompile(sel string) Selector {\n\tcompiled, err := Compile(sel)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn compiled\n}", "func Regex(expr ConstString, input String) (Bool, error) {\n\tif expr.HasVariable() {\n\t\treturn nil, fmt.Errorf(\"regex requires a constant expression as its first parameter, but it has a variable parameter\")\n\t}\n\te, err := expr.Eval()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr, err := regexp.Compile(e)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn TrimBool(&regex{\n\t\tr: r,\n\t\tS: input,\n\t\texpr: e,\n\t\thash: Hash(\"regex\", expr, input),\n\t\thasVariable: input.HasVariable(),\n\t}), nil\n}", "func Match(regex string, text string) bool {\n runerx := compile(regex)\n runetxt := []rune(text)\n \n if len(runerx) > 0 && runerx[0] == '^' {\n return matchhere(runerx[1:], runetxt)\n }\n \n for {\n if matchhere(runerx, runetxt) {\n return true\n }\n if len(runetxt) == 0 {\n return false\n }\n runetxt = runetxt[1:]\n }\n }", "func RegexpString(re *syntax.Regexp,) string", "func TestPattern2(t *testing.T) {\n\tre := MustCompile(\"a$\")\n\tif !re.MatchString(\"a\") {\n\t\tt.Errorf(\"expect to match\\n\")\n\t}\n\tif re.MatchString(\"ab\") {\n\t\tt.Errorf(\"expect to mismatch\\n\")\n\t}\n}", "func RegexpParse(content, pattern string) ([]string, error) {\n\tre, err := pool.Compile(pattern)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"re:[%s] error:%+v\", pattern, err)\n\t}\n\tvar ret []string\n\tres := re.FindAllStringSubmatch(content, -1)\n\tfor i := range res {\n\t\tswitch {\n\t\tcase len(res[i]) == 1:\n\t\t\tret = append(ret, res[i][0])\n\t\tcase len(res[i]) > 1:\n\t\t\tret = append(ret, res[i][1:]...)\n\t\t}\n\t}\n\treturn ret, nil\n}", "func compile(conf Config) regexps {\n\tr := make(regexps)\n\n\tfor k, v := range conf {\n\t\tr[regexp.MustCompile(k)] = v\n\t}\n\n\treturn r\n}", "func (m *Builder) Compile() (*MultiGlob, error) {\n\tvar final *parser.Node\n\tfor _, p := range m.patterns {\n\t\tif final == nil {\n\t\t\tfinal = p\n\t\t} else {\n\t\t\tfinal = parser.Merge(final, p)\n\t\t}\n\t}\n\n\tpatterns := make(map[string]*parser.Node)\n\tfor k, v := range m.patterns {\n\t\tpatterns[k] = v\n\t}\n\n\treturn &MultiGlob{\n\t\tnode: final,\n\t\tpatterns: patterns,\n\t}, nil\n}", "func NewAccessControlDomainRegex(pattern regexp.Regexp) SubjectObjectMatcher {\n\tvar iuser, igroup = -1, -1\n\n\tfor i, group := range pattern.SubexpNames() {\n\t\tswitch group {\n\t\tcase subexpNameUser:\n\t\t\tiuser = i\n\t\tcase subexpNameGroup:\n\t\t\tigroup = i\n\t\t}\n\t}\n\n\tif iuser != -1 || igroup != -1 {\n\t\treturn AccessControlDomainRegex{Pattern: pattern, SubexpNameUser: iuser, SubexpNameGroup: igroup}\n\t}\n\n\treturn AccessControlDomainRegexBasic{Pattern: pattern}\n}", "func (vm *VM) initRegexpObject(regexp string) *RegexpObject {\n\tr, err := regexp2.Compile(regexp, 0)\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn &RegexpObject{\n\t\tBaseObj: NewBaseObject(vm.TopLevelClass(classes.RegexpClass)),\n\t\tregexp: r,\n\t}\n}", "func expandRegex(r *regexp.Regexp, tmpl string, src string) string {\n\tvar res []byte\n\tfor _, submatches := range r.FindAllStringSubmatchIndex(src, 1) {\n\t\tres = r.ExpandString(res, tmpl, src, submatches)\n\t}\n\treturn string(res)\n}", "func NewRegexpFile(r *regexp.Regexp) *File {\n\treturn NewFile(&regexpElement{Data: r})\n}", "func MustCompile(expr string) *Expr {\n\texp, err := Compile(expr)\n\tif err != nil {\n\t\treturn &Expr{s: expr, q: nopQuery{}}\n\t}\n\treturn exp\n}", "func MustNewRegexp(s string) Regexp {\n\tre, err := NewRegexp(s)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn re\n}", "func MustNewRegexp(s string) Regexp {\n\tre, err := NewRegexp(s)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn re\n}", "func Compile(sel string) (Selector, error) {\n\tcompiled, err := ParseGroup(sel)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn Selector(compiled.Match), nil\n}", "func NewSelectorRegex(s *Selector) (*SelectorRegex, error) {\n\tsr := new(SelectorRegex)\n\tvar err error\n\tsr.selector = s\n\tsr.groupRegex, err = regexp.Compile(anchorRegex(s.Gvk.Group))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr.versionRegex, err = regexp.Compile(anchorRegex(s.Gvk.Version))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr.kindRegex, err = regexp.Compile(anchorRegex(s.Gvk.Kind))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr.nameRegex, err = regexp.Compile(anchorRegex(s.Name))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr.namespaceRegex, err = regexp.Compile(anchorRegex(s.Namespace))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn sr, nil\n}", "func compile(x string) string {\n\tif x == \"\" {\n\t\treturn \"\"\n\t}\n\n\tinput := bytes.NewBufferString(x)\n\toutput := bytes.NewBufferString(\"\")\n\n\tfor {\n\t\ti, _, err := input.ReadRune()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tswitch i {\n\t\tdefault:\n\t\t\toutput.WriteRune(i)\n\t\tcase EscapeChar:\n\t\t\tcompileColorSyntax(input, output)\n\t\t}\n\t}\n\treturn output.String()\n}", "func compilePath(path spec.Path) (*regexp.Regexp, []string) {\n\tvar pathParamNames []string\n\tparts := strings.Split(string(path), \"/\")\n\tpattern := `\\A`\n\n\tfor _, part := range parts {\n\t\tif part == \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tsubmatches := pathParameterPattern.FindAllStringSubmatch(part, -1)\n\t\tif submatches == nil {\n\t\t\tpattern += `/` + part\n\t\t} else {\n\t\t\t// Special characters as defined by:\n\t\t\t//\n\t\t\t// https://tools.ietf.org/html/rfc3986#section-3.3\n\t\t\tpattern += `/(?P<` + submatches[0][1] + `>[\\w@:%-._~!$&'()*+,;=]+)`\n\t\t\tpathParamNames = append(pathParamNames, submatches[0][1])\n\t\t}\n\t}\n\n\treturn regexp.MustCompile(pattern + `\\z`), pathParamNames\n}", "func New(expr string) (*Regexp, error) {\n\treturn NewWithLimit(expr, DefaultLimit)\n}", "func (r *RegexLexer) maybeCompile() (err error) {\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\tif r.compiled {\n\t\treturn nil\n\t}\n\tfor state, rules := range r.rules {\n\t\tfor i, rule := range rules {\n\t\t\tif rule.Regexp == nil {\n\t\t\t\tpattern := \"(?:\" + rule.Pattern + \")\"\n\t\t\t\tif rule.flags != \"\" {\n\t\t\t\t\tpattern = \"(?\" + rule.flags + \")\" + pattern\n\t\t\t\t}\n\t\t\t\tpattern = `\\G` + pattern\n\t\t\t\trule.Regexp, err = regexp2.Compile(pattern, 0)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"failed to compile rule %s.%d: %s\", state, i, err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\nrestart:\n\tseen := map[LexerMutator]bool{}\n\tfor state := range r.rules {\n\t\tfor i := 0; i < len(r.rules[state]); i++ {\n\t\t\trule := r.rules[state][i]\n\t\t\tif compile, ok := rule.Mutator.(LexerMutator); ok {\n\t\t\t\tif seen[compile] {\n\t\t\t\t\treturn fmt.Errorf(\"saw mutator %T twice; this should not happen\", compile)\n\t\t\t\t}\n\t\t\t\tseen[compile] = true\n\t\t\t\tif err := compile.MutateLexer(r.rules, state, i); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\t// Process the rules again in case the mutator added/removed rules.\n\t\t\t\t//\n\t\t\t\t// This sounds bad, but shouldn't be significant in practice.\n\t\t\t\tgoto restart\n\t\t\t}\n\t\t}\n\t}\n\tr.compiled = true\n\treturn nil\n}", "func R(pattern string) *regexp.Regexp {\n\treturn regexp.MustCompile(pattern)\n}", "func CompileRules(r io.Reader, a *cli.Args) *CompiledRuleList {\n\tvar rules = &Rules{}\n\trules.LoadRules(r)\n\tvar crl CompiledRuleList\n\n\t// Build a slice of all compiled patterns which we then use with data\n\t// consumed from stdin.\n\tvar cexp *regexp.Regexp\n\tvar pattern string\n\tvar debug bool\n\tfor i, r := range rules.ListOfRules() {\n\t\tif r.PatString == MatchNoOutput {\n\t\t\tpattern = zeroInputPattern\n\t\t} else {\n\t\t\tpattern = r.PatString\n\t\t}\n\t\tif r.ERESyntax {\n\t\t\tcexp = regexp.MustCompilePOSIX(pattern)\n\t\t} else {\n\t\t\tcexp = regexp.MustCompile(pattern)\n\t\t}\n\t\t// If debug is set on individual rule, or as a command-line argument,\n\t\t// enable it on compiled rule as well.\n\t\tdebug = r.Debug || a.Debug()\n\t\tcrl.Append(CompiledRule{\n\t\t\tre: cexp,\n\t\t\tdesc: rules.RuleByIndex(i),\n\t\t\tdebug: debug})\n\t}\n\n\treturn &crl\n}", "func getRegexp(env *lisp.LEnv, v *lisp.LVal) (re *regexp.Regexp, lerr *lisp.LVal) {\n\tif v.Type == lisp.LString {\n\t\tre, err := regexp.Compile(v.Str)\n\t\tif err != nil {\n\t\t\treturn nil, invalidPatternError(env, err)\n\t\t}\n\t\treturn re, nil\n\t}\n\tif v.Type != lisp.LNative {\n\t\treturn nil, env.Errorf(\"argument is not a regexp: %v\", v.Type)\n\t}\n\tre, ok := v.Native.(*regexp.Regexp)\n\tif !ok {\n\t\treturn nil, env.Errorf(\"argument is not a regexp: %v\", v)\n\t}\n\treturn re, nil\n}", "func MustCompile(constraint string) go_version.Constraints {\n\tverCheck, err := Compile(constraint)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"cannot compile go-version constraint '%s' %s\", constraint, err))\n\t}\n\treturn verCheck\n}", "func MustCompile(url string) *Schema {\n\treturn NewCompiler().MustCompile(url)\n}", "func regExp(context interface{}, value string) (bson.RegEx, error) {\n\tidx := strings.IndexByte(value[1:], '/')\n\tif value[0] != '/' || idx == -1 {\n\t\terr := fmt.Errorf(\"Parameter %q is not a regular expression\", value)\n\t\tlog.Error(context, \"varLookup\", err, \"Regex parsing\")\n\t\treturn bson.RegEx{}, err\n\t}\n\n\tpattern := value[1 : idx+1]\n\tl := len(pattern) + 2\n\n\tvar options string\n\tif l < len(value) {\n\t\toptions = value[l:]\n\t}\n\n\treturn bson.RegEx{Pattern: pattern, Options: options}, nil\n}", "func MaskToRegex(rawMask string) (*regexp.Regexp, error) {\n\tinput := bytes.NewBufferString(rawMask)\n\n\toutput := &bytes.Buffer{}\n\toutput.WriteByte('^')\n\n\tfor {\n\t\tc, err := input.ReadByte()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\n\t\tif c == '\\\\' {\n\t\t\tc, err = input.ReadByte()\n\t\t\tif err != nil {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(\"\\\\\"))\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif c == '?' || c == '*' || c == '\\\\' {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(string(c)))\n\t\t\t} else {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(\"\\\\\" + string(c)))\n\t\t\t}\n\t\t} else if trans, ok := maskTranslations[c]; ok {\n\t\t\toutput.WriteString(trans)\n\t\t} else {\n\t\t\toutput.WriteString(regexp.QuoteMeta(string(c)))\n\t\t}\n\t}\n\n\toutput.WriteByte('$')\n\n\treturn regexp.Compile(output.String())\n}", "func NewRegexPatternSet(ctx *pulumi.Context,\n\tname string, args *RegexPatternSetArgs, opts ...pulumi.ResourceOption) (*RegexPatternSet, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.RegularExpressionList == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RegularExpressionList'\")\n\t}\n\tif args.Scope == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Scope'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource RegexPatternSet\n\terr := ctx.RegisterResource(\"aws-native:wafv2:RegexPatternSet\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func Regex(path string, r string) ([]string, error) {\n\tclearVars()\n\n\tregex, err := regexp.Compile(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = Find(path, func(s string, i os.FileInfo) {\n\t\tif regex.FindString(i.Name()) != \"\" {\n\t\t\tresults = append(results, s)\n\t\t}\n\t})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn results, nil\n}", "func RegexpSimplify(re *syntax.Regexp,) *syntax.Regexp", "func TestCompiledRegexInGoRoutine(t *testing.T) {\n\tregex, err := regexp.Compile(\"^(B|b)rian$\")\n\tif err != nil {\n\t\tt.Errorf(\"Regex did not compile %v\", err)\n\t}\n\n\tch := make(chan bool)\n\ttests := []string{\"brian\", \"Brian\"}\n\tfor _, test := range tests {\n\t\tgo func(t string) {\n\t\t\tch <- regex.MatchString(t)\n\t\t}(test)\n\t}\n\n\tfirst, second := <-ch, <-ch\n\n\tif first && second {\n\t\treturn\n\t}\n\n\tt.Error(\"String did not match\")\n}", "func NewTestGlobRegexp(glob string) (*regexp.Regexp, error) {\n\tif _, err := validateGlob(glob); err != nil {\n\t\treturn nil, err\n\t}\n\treturn compileGlob(glob)\n}", "func createGlobbingRegex(globbing string) *regexp.Regexp {\n\tif globbing == \"\" {\n\t\t// nil here as \"\" is fast-tracked elsewhere\n\t\treturn nil\n\t}\n\n\tg := regexp.QuoteMeta(globbing)\n\tg = strings.ReplaceAll(g, \"\\\\*\", \".*\")\n\tg = strings.ReplaceAll(g, \"\\\\?\", \".\")\n\t// (?i) forces case insensitive matches\n\tg = \"(?i)^\" + g + \"$\"\n\treturn regexp.MustCompile(g)\n}", "func Pattern(pattern *regexp.Regexp) dgo.Type {\n\treturn internal.PatternType(pattern)\n}", "func wildCardToRegexp(pattern string) string {\n\tvar result strings.Builder\n\tresult.WriteString(\"(?i)\")\n\n\trpattern := strings.Replace(pattern, \"%\", \".*\", -1)\n\trpattern = strings.Replace(rpattern, \"_\", \".+\", -1)\n\tresult.WriteString(rpattern)\n\n\treturn result.String()\n}", "func newPattern(pattern string) *Pattern {\n\treturn &Pattern{*newPath(pattern)}\n\n}", "func (r *RuleDefinition) Compile() error {\n\tr.rule = &rule{\n\t\tEnv: make(map[string]*regexp.Regexp, 0),\n\t}\n\n\tif r.Action != nil {\n\t\taction, err := regexp.Compile(*(r.Action))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tr.rule.Action = action\n\t}\n\n\tfor k, v := range r.Env {\n\t\treg, err := regexp.Compile(v)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tr.rule.Env[k] = reg\n\t}\n\treturn nil\n}", "func NewRegexpParser(re string) (*RegexpParser, error) {\n\tregex, err := regexp.Compile(re)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif regex.NumSubexp() == 0 {\n\t\treturn nil, errMissingCapture\n\t}\n\tnameIndex := map[int]string{}\n\tuniqueNames := map[string]struct{}{}\n\tfor i, n := range regex.SubexpNames() {\n\t\tif n != \"\" {\n\t\t\tif !model.LabelName(n).IsValid() {\n\t\t\t\treturn nil, fmt.Errorf(\"invalid extracted label name '%s'\", n)\n\t\t\t}\n\t\t\tif _, ok := uniqueNames[n]; ok {\n\t\t\t\treturn nil, fmt.Errorf(\"duplicate extracted label name '%s'\", n)\n\t\t\t}\n\t\t\tnameIndex[i] = n\n\t\t\tuniqueNames[n] = struct{}{}\n\t\t}\n\t}\n\tif len(nameIndex) == 0 {\n\t\treturn nil, errMissingCapture\n\t}\n\treturn &RegexpParser{\n\t\tregex: regex,\n\t\tnameIndex: nameIndex,\n\t\tkeys: internedStringSet{},\n\t}, nil\n}", "func StaticRegexReplace(scope *Scope, input tf.Output, pattern string, rewrite string, optional ...StaticRegexReplaceAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{\"pattern\": pattern, \"rewrite\": rewrite}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"StaticRegexReplace\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func parsePCRE(s string) (*PCRE, error) {\n\tc := strings.Count(s, \"/\")\n\tif c < 2 {\n\t\treturn nil, fmt.Errorf(\"all pcre patterns must contain at least 2 '/', found: %d\", c)\n\t}\n\n\tl := strings.LastIndex(s, \"/\")\n\tif l < 0 {\n\t\treturn nil, fmt.Errorf(\"couldn't find options in PCRE\")\n\t}\n\n\ti := strings.Index(s, \"/\")\n\tif l < 0 {\n\t\treturn nil, fmt.Errorf(\"couldn't find start of pattern\")\n\t}\n\n\treturn &PCRE{\n\t\tPattern: []byte(s[i+1 : l]),\n\t\tOptions: []byte(s[l+1:]),\n\t}, nil\n}" ]
[ "0.81953734", "0.7839732", "0.77232397", "0.74875295", "0.73487467", "0.7305067", "0.724314", "0.72297364", "0.7156587", "0.71189314", "0.7027613", "0.69368225", "0.690044", "0.68722314", "0.68648875", "0.68534964", "0.68054307", "0.66829604", "0.6646106", "0.6637849", "0.6623873", "0.6561792", "0.64774853", "0.64458734", "0.6333776", "0.62747777", "0.6242336", "0.6230166", "0.6194096", "0.6119954", "0.61013067", "0.6088221", "0.60161626", "0.6000842", "0.6000572", "0.5975564", "0.59677607", "0.5966688", "0.59140295", "0.5905143", "0.58835596", "0.58374846", "0.5804895", "0.57370126", "0.5721096", "0.57132226", "0.57091314", "0.57039773", "0.5688235", "0.56769127", "0.5654043", "0.56385756", "0.56316495", "0.5630284", "0.5628571", "0.5627111", "0.5616236", "0.5616236", "0.5615711", "0.56094795", "0.5562594", "0.5519554", "0.54976434", "0.5495736", "0.5490377", "0.5439489", "0.54204917", "0.54098773", "0.5405627", "0.5402627", "0.53969514", "0.53904176", "0.5384532", "0.5384532", "0.53664166", "0.5346022", "0.5331659", "0.53023887", "0.52888966", "0.52464986", "0.5232371", "0.5217759", "0.516251", "0.5161722", "0.51514244", "0.5143279", "0.5124198", "0.5111954", "0.5111121", "0.5094419", "0.5094396", "0.5080556", "0.50733346", "0.50635576", "0.50241697", "0.5019884", "0.5002917", "0.49725768", "0.49519998", "0.4949029" ]
0.900864
0
CompileERE uses regexp.CompilePOSIX to compile a new regexp.Regexp according to pattern. Uses POSIX extended regexp behavior.
func CompileERE(pattern string) (*regexp.Regexp, error) { return regexp.CompilePOSIX(pattern) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func CompilePOSIX(expr string) (*NRegexp, error) {\n\tre, err := regexp.CompilePOSIX(expr)\n\treturn &NRegexp{Regexp: re}, err\n}", "func CompileRegex(pattern string) (*regexp.Regexp, error) { return regexp.Compile(pattern) }", "func MustCompilePOSIX(str string) *NRegexp {\n\treturn &NRegexp{Regexp: regexp.MustCompilePOSIX(str)}\n}", "func CompileRegexp(pattern string) *regexp.Regexp {\n\tif pattern == \"\" {\n\t\treturn nil\n\t}\n\n\t// add ^...$ to all regexp when not given\n\tif !strings.HasPrefix(pattern, \"^\") {\n\t\tpattern = \"^\" + pattern\n\t}\n\tif !strings.HasSuffix(pattern, \"$\") {\n\t\tpattern += \"$\"\n\t}\n\n\t// make all regexp case insensitive by default\n\tif !strings.Contains(pattern, \"(?i)\") {\n\t\tpattern = \"(?i)\" + pattern\n\t}\n\n\treturn regexp.MustCompile(pattern)\n}", "func CompileRegexp(str string) (regexp *Regexp, error string) {\n\tregexp = new(Regexp);\n\tregexp.expr = str;\n\tregexp.inst = make([]instr, 0, 20);\n\terror = regexp.doParse();\n\treturn;\n}", "func Compile(input string) (*Regexp, error) {\n\treturn Compile2(input, 0)\n}", "func Compile(expr string) (*NRegexp, error) {\n\tre, err := regexp.Compile(expr)\n\treturn &NRegexp{Regexp: re}, err\n}", "func Compile(re *syntax.Regexp,) (*syntax.Prog, error)", "func getRegexpCompile(pattern string) *regexp.Regexp {\n\tr, err := regexp.Compile(pattern)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn r\n}", "func Compile(pattern string) (Matcher, error) {\n\tregex, err := syntax.Parse(pattern, syntax.Perl)\n\tif err != nil {\n\t\treturn Matcher{}, err\n\t}\n\n\tregex = optimize(regex).Simplify()\n\tm, err := compile(regex)\n\treturn Matcher{m}, err\n}", "func compile(regex string) (regslc []rune) {\n regslc = make([]rune, 0, len(regex) + 10)\n \n for _, r := range regex {\n if r == '+' {\n regslc = append(regslc, regslc[len(regslc) - 1], '*')\n } else {\n regslc = append(regslc, r)\n }\n } \n return regslc\n }", "func compile(pattern string, observeCase bool) (err error) {\n\tif observeCase {\n\t\t_, err = regexp.Compile(pattern)\n\t} else {\n\t\t_, err = regexp.Compile(\"(?i)\" + pattern)\n\t}\n\n\treturn err\n}", "func Regexp(pattern string) (*regexp.Regexp, error) {\n\treturn pool.Compile(pattern)\n}", "func TestCompileRegexp(t *testing.T) {\n\tregex, err := regexp.Compile(\"^(B|b)rian$\")\n\n\tif regex == nil {\n\t\tt.Errorf(\"Regex did not compile %v\", err)\n\t}\n}", "func (c Configuration) compileRegex(listofPaths []string) []*regexp.Regexp {\n\tlogger := c.logger()\n\tvar regexpObjects []*regexp.Regexp\n\tfor _, path := range listofPaths {\n\t\treg, err := regexp.Compile(path)\n\t\tif err != nil {\n\t\t\tlogger.Error().Err(err).Msgf(\"Error while compiling regex: %v\", err)\n\t\t\tcontinue\n\t\t}\n\t\tregexpObjects = append(regexpObjects, reg)\n\t}\n\treturn regexpObjects\n}", "func Compile(re string, lookup interface{}) (*regexp.Regexp, error) {\n\tvar regexpCompiled = regexp.MustCompile(re)\n\n\tlookupValue := reflect.ValueOf(lookup)\n\tif lookupValue.Kind() != reflect.Ptr {\n\t\treturn nil, fmt.Errorf(\"lookup must be a pointer\")\n\t}\n\n\tlookupElem := lookupValue.Elem()\n\tsubexpNames := map[string]int{}\n\tfor i, name := range regexpCompiled.SubexpNames() {\n\t\tif name == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tif cases.Title(language.English, cases.NoLower).String(name) != name {\n\t\t\treturn nil, fmt.Errorf(\"field '%s' in regexp `%s` is not capitalized\", name, re)\n\t\t}\n\t\tsubexpNames[name] = i\n\t}\n\n\tfor i := 0; i < lookupElem.NumField(); i++ {\n\t\tfield := lookupElem.Type().Field(i)\n\t\tfieldName := field.Name\n\t\tindex, foundIndex := subexpNames[fieldName]\n\t\tif !foundIndex {\n\t\t\treturn nil, fmt.Errorf(\"field '%s' is not in regexp `%s`\", fieldName, re)\n\t\t}\n\t\tlookupElem.FieldByName(fieldName).SetInt(int64(index))\n\t\tdelete(subexpNames, fieldName)\n\t}\n\n\tvar unusedNames []string\n\tfor name := range subexpNames {\n\t\tunusedNames = append(unusedNames, name)\n\t}\n\n\tif len(unusedNames) > 0 {\n\t\treturn nil, fmt.Errorf(\"regexp `%s` has named groups not found in lookup: %s\", re, strings.Join(unusedNames, \", \"))\n\t}\n\n\treturn regexpCompiled, nil\n}", "func r(pattern string) *regexp.Regexp { return regexp.MustCompile(pattern) }", "func TransformRegExp(pattern string) (string, error) {\n\n\tlog.Printf(\"runtime.Caller(1): %#+v\\n\", myutils.Slice(runtime.Caller(1))...)\n\tos.Exit(349857394)\n\n\t// if pattern == \"\" {\n\t// \treturn \"\", nil\n\t// }\n\n\tswitch pattern {\n\t// common cases that will leave as is\n\t// but frequently occurred in many\n\t// popular frameworks like react.js etc.\n\tcase _regexp_empty,\n\t\t_regexp_any_nc,\n\t\t_regexp_any_c,\n\t\t_regexp_anyplus_nc,\n\t\t_regexp_anyplus_c,\n\n\t\t_regexp_squote_nc,\n\t\t_regexp_squote_c,\n\t\t_regexp_dquote_nc,\n\t\t_regexp_dquote_c,\n\n\t\t_regexp_plus_nc,\n\t\t_regexp_plus_c,\n\n\t\t_regexp_urlenc_space_nc,\n\t\t_regexp_urlenc_space_c,\n\t\t_regexp_http_validation,\n\n\t\t_regexp_any_uppercase_nc,\n\t\t_regexp_any_uppercase_c,\n\n\t\t_regexp_is_ms_pref:\n\t\treturn pattern, nil\n\t}\n\n\t// TODO If without \\, if without (?=, (?!, then another shortcut\n\n\tparser := _RegExp_parser{\n\t\tstr: pattern,\n\t\tlength: len(pattern),\n\t\tgoRegexp: bytes.NewBuffer(make([]byte, 0, 3*len(pattern)/2)),\n\t}\n\tparser.read() // Pull in the first character\n\tparser.scan()\n\tvar err error\n\tif len(parser.errors) > 0 {\n\t\terr = parser.errors[0]\n\t}\n\tif parser.invalid {\n\t\tlog.Printf(\"Input: regexp=[%s]\\n\", pattern)\n\t\tlog.Printf(\"Output: regexp=[%s] err=[%s]\\n\", parser.goRegexp.String(), err)\n\t\treturn \"\", err\n\t}\n\n\t// Might not be re2 compatible, but is still a valid JavaScript RegExp\n\treturn parser.goRegexp.String(), err\n}", "func (r *Regexps) Compile(pattern string) (*regexp.Regexp, error) {\n\tr.Lock()\n\tdefer r.Unlock()\n\tif r.items[pattern] != nil {\n\t\treturn r.items[pattern], nil\n\t}\n\tre, err := regexp.Compile(pattern)\n\tif err == nil {\n\t\tr.items[pattern] = re\n\t}\n\treturn re, err\n}", "func compileGlob(glob string) (*regexp.Regexp, error) {\n\tglob = strings.Replace(glob, \".\", \"\\\\.\", -1)\n\tglob = strings.Replace(glob, \"*\", \".*\", -1)\n\tglob = \"^\" + glob + \"$\"\n\treturn regexp.Compile(glob)\n}", "func NewRegexp(exp string) (*Regexp, error) {\n\tr, err := regexp.Compile(exp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Regexp{r}, nil\n}", "func (p *Generic) CompileRegexps(patterns map[string]string) (map[string]*regexp.Regexp, error) {\n\tresults := make(map[string]*regexp.Regexp)\n\tfor name, str := range patterns {\n\t\tre, err := regexp.Compile(str)\n\t\tif nil != err {\n\t\t\treturn results, fmt.Errorf(\"Cannot compile '%s' regexp: %s\", name, err.Error())\n\t\t}\n\t\tresults[name] = re\n\t}\n\n\treturn results, nil\n}", "func compileRegexps(regexps []string) ([]*regexp.Regexp, error) {\n\tvar errs []error\n\tvar compiled []*regexp.Regexp\n\n\tfor _, exp := range regexps {\n\t\tre, err := regexp.Compile(exp)\n\t\tif err == nil {\n\t\t\tcompiled = append(compiled, re)\n\t\t} else {\n\t\t\terrs = append(errs, err)\n\t\t}\n\t}\n\n\tif len(errs) > 0 {\n\t\treturn nil, errors.Join(errs...)\n\t}\n\treturn compiled, nil\n}", "func compileCInstructionRegexp() *regexp.Regexp {\n\tdestRe := \"(?:([ADM]+)=)?\" // the destination part of the regexp\n\tcompRe := pipe(CComp) // the compute part of the regexp\n\tjumpRe := pipe(CJump) // the jump part of the regexp\n\ttmp := fmt.Sprintf(\"^%s%s(?:;%s)?$\", destRe, compRe, jumpRe)\n\tlog.Debug().Str(\"C Instruction Regex\", tmp)\n\treturn regexp.MustCompile(tmp)\n}", "func (p *DefaultParser) CompileRegex(fields []string) ([]*regexp.Regexp, error) {\n\tr := make([]*regexp.Regexp, len(fields))\n\tfor i, f := range fields {\n\t\t/*\n\t\t\tvar nextField string\n\t\t\tif i == len(fields)-1 {\n\t\t\t\tnextField = \"\"\n\t\t\t} else {\n\t\t\t\tnextField = fields[i+1]\n\t\t\t}\n\t\t*/\n\t\tfieldName := p.NamedGroup(f)\n\t\t// TODO(stratus): This is the foundation for possibly two\n\t\t// regexes - one for easy single line fields and another one for\n\t\t// multi-field.\n\t\tre, err := regexp.Compile(fmt.Sprintf(`(?mis)%s\\s*(?P<%s>.*?)\\n`, f, fieldName))\n\t\t//re, err := regexp.Compile(fmt.Sprintf(`(?mis)%s\\s*(?P<%s>.*?)%s`, f, fieldName, nextField))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tr[i] = re\n\t}\n\treturn r, nil\n}", "func compileRegex() {\n\tfor _, value := range filterObj.Signatures {\n\t\tr, _ := regexp.Compile(value.Regex)\n\t\tsignatureRegex = append(signatureRegex, r)\n\t}\n\tfor _, value := range filterObj.Shallows {\n\t\tr, _ := regexp.Compile(value.Regex)\n\t\tshallowRegex = append(shallowRegex, r)\n\t}\n}", "func buildRegexFromList(patterns []string) *regexp.Regexp {\n\tif patterns == nil || len(patterns) == 0 {\n\t\tpatterns = make([]string, 0)\n\t}\n\treturn compileOrDie(\"(?i)^(\" + strings.Join(patterns, \"|\") + \")$\")\n}", "func (c *TransferCounter) CompileRegex(operator string) (*regexp.Regexp, error) {\n\tvar r *regexp.Regexp\n\tvar err error\n\n\tfor _, regionOperator := range regionOperators {\n\t\tif operator == regionOperator {\n\t\t\tr, err = regexp.Compile(\".*?operator finish.*?region-id=([0-9]*).*?\" + operator + \".*?store \\\\[([0-9]*)\\\\] to \\\\[([0-9]*)\\\\].*?\")\n\t\t}\n\t}\n\n\tfor _, leaderOperator := range leaderOperators {\n\t\tif operator == leaderOperator {\n\t\t\tr, err = regexp.Compile(\".*?operator finish.*?region-id=([0-9]*).*?\" + operator + \".*?store ([0-9]*) to ([0-9]*).*?\")\n\t\t}\n\t}\n\n\tif r == nil {\n\t\terr = errors.New(\"unsupported operator. \")\n\t}\n\treturn r, err\n}", "func Regexp(expr *regexp.Regexp) Pattern {\n\treturn regexpMatch{expr}\n}", "func parsePCRE(s string) (*PCRE, error) {\n\tc := strings.Count(s, \"/\")\n\tif c < 2 {\n\t\treturn nil, fmt.Errorf(\"all pcre patterns must contain at least 2 '/', found: %d\", c)\n\t}\n\n\tl := strings.LastIndex(s, \"/\")\n\tif l < 0 {\n\t\treturn nil, fmt.Errorf(\"couldn't find options in PCRE\")\n\t}\n\n\ti := strings.Index(s, \"/\")\n\tif l < 0 {\n\t\treturn nil, fmt.Errorf(\"couldn't find start of pattern\")\n\t}\n\n\treturn &PCRE{\n\t\tPattern: []byte(s[i+1 : l]),\n\t\tOptions: []byte(s[l+1:]),\n\t}, nil\n}", "func R(pattern string) *regexp.Regexp {\n\treturn regexp.MustCompile(pattern)\n}", "func NewRegexp(s string) (Regexp, error) {\n\tregex, err := regexp.Compile(s)\n\treturn Regexp{\n\t\tRegexp: regex,\n\t\toriginal: s,\n\t}, err\n}", "func Convert(pattern string) (*regexp.Regexp, error) {\n\tpattern = \"^\" + metaReplacer.Replace(regexp.QuoteMeta(pattern)) + \"$\"\n\treturn regexp.Compile(pattern)\n}", "func newCompiledRegexes(matchItems []v1beta1.MatchItem) (map[*v1beta1.MatchItem]*regexp.Regexp, error) {\n\tvar err error\n\tcache := make(map[*v1beta1.MatchItem]*regexp.Regexp)\n\n\tfor i := range matchItems {\n\t\tif matchItems[i].MatchRegex != nil {\n\t\t\tcache[&matchItems[i]], err = regexp.Compile(*matchItems[i].MatchRegex)\n\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn cache, nil\n}", "func HybridGlobRegexPartCompiler(partStr string, next NextPartFn) (Matcher, error) {\n\tchildMatcher, err := next()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tswitch partStr {\n\tcase \"**\":\n\t\treturn AnyRecursiveMatcher(childMatcher), nil\n\tcase \"*\":\n\t\treturn AnyMatcher(childMatcher), nil\n\t}\n\tif strings.HasPrefix(partStr, \"^\") || strings.HasSuffix(partStr, \"$\") {\n\t\treturn parseRegexToMatcher(partStr, childMatcher)\n\t}\n\treturn GlobMatcherSm(partStr, childMatcher), nil\n}", "func generatePatternForRegexp(pattern string) (string, error) {\n\tpattern = patternRegexp.ReplaceAllStringFunc(pattern, func(subMatch string) string {\n\t\t// The sub match string conforms the parameter pattern: `{parameter-name:regexp-expression}`.\n\t\tfoos := strings.SplitN(subMatch, \":\", 2)\n\t\tif len(foos) < 2 {\n\t\t\treturn `([^/]+)`\n\t\t} else {\n\t\t\treturn \"(\" + foos[1][0:len(foos[1])-1] + \")\"\n\t\t}\n\t})\n\t// Checking for abnormal patterns.\n\t_, err := regexp.Compile(pattern)\n\treturn pattern, err\n}", "func CompileStrs(regStrs []string) (*regexp.Regexp, error) {\n\tif len(regStrs) == 0 {\n\t\treturn nil, fmt.Errorf(\"regStrs must be greater than or equal to 1\")\n\t}\n\tregStr := \"(\" + strings.Join(regStrs, \"|\") + \")\"\n\tre, err := regexp.Compile(regStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn re, nil\n}", "func (p *unicodePattern) Compile(patternStr string, escape byte) {\n\tp.patChars, p.patTypes = stringutil.CompilePatternInner(patternStr, escape)\n}", "func getRegexp(env *lisp.LEnv, v *lisp.LVal) (re *regexp.Regexp, lerr *lisp.LVal) {\n\tif v.Type == lisp.LString {\n\t\tre, err := regexp.Compile(v.Str)\n\t\tif err != nil {\n\t\t\treturn nil, invalidPatternError(env, err)\n\t\t}\n\t\treturn re, nil\n\t}\n\tif v.Type != lisp.LNative {\n\t\treturn nil, env.Errorf(\"argument is not a regexp: %v\", v.Type)\n\t}\n\tre, ok := v.Native.(*regexp.Regexp)\n\tif !ok {\n\t\treturn nil, env.Errorf(\"argument is not a regexp: %v\", v)\n\t}\n\treturn re, nil\n}", "func NewLazyRegexp(pattern string) LazyRegexp {\n\treturn LazyRegexp{pattern, nil}\n}", "func MustCompile(str string) *NRegexp {\n\treturn &NRegexp{Regexp: regexp.MustCompile(str)}\n}", "func zoektCompile(p *protocol.PatternInfo) (zoektquery.Q, error) {\n\tvar parts []zoektquery.Q\n\t// we are redoing work here, but ensures we generate the same regex and it\n\t// feels nicer than passing in a readerGrep since handle path directly.\n\tif rg, err := compile(p); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tre, err := syntax.Parse(rg.re.String(), syntax.Perl)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tparts = append(parts, &zoektquery.Regexp{\n\t\t\tRegexp: re,\n\t\t\tContent: true,\n\t\t\tCaseSensitive: !rg.ignoreCase,\n\t\t})\n\t}\n\n\tfor _, pat := range p.IncludePatterns {\n\t\tif !p.PathPatternsAreRegExps {\n\t\t\treturn nil, errors.New(\"hybrid search expects PathPatternsAreRegExps\")\n\t\t}\n\t\tre, err := syntax.Parse(pat, syntax.Perl)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tparts = append(parts, &zoektquery.Regexp{\n\t\t\tRegexp: re,\n\t\t\tFileName: true,\n\t\t\tCaseSensitive: p.PathPatternsAreCaseSensitive,\n\t\t})\n\t}\n\n\tif p.ExcludePattern != \"\" {\n\t\tif !p.PathPatternsAreRegExps {\n\t\t\treturn nil, errors.New(\"hybrid search expects PathPatternsAreRegExps\")\n\t\t}\n\t\tre, err := syntax.Parse(p.ExcludePattern, syntax.Perl)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tparts = append(parts, &zoektquery.Not{Child: &zoektquery.Regexp{\n\t\t\tRegexp: re,\n\t\t\tFileName: true,\n\t\t\tCaseSensitive: p.PathPatternsAreCaseSensitive,\n\t\t}})\n\t}\n\n\treturn zoektquery.Simplify(zoektquery.NewAnd(parts...)), nil\n}", "func NewRegexp(s string) (Regexp, error) {\n\tregex, err := regexp.Compile(\"^(?:\" + s + \")$\")\n\treturn Regexp{\n\t\tRegexp: regex,\n\t\toriginal: s,\n\t}, err\n}", "func RegexpString(re *syntax.Regexp,) string", "func MustCompile(str string) *Regexp {\n\tregexp, error := CompileRegexp(str);\n\tif error != \"\" {\n\t\tpanicln(`regexp: compiling \"`, str, `\": `, error)\n\t}\n\treturn regexp;\n}", "func (l *LazyRegexp) Pattern() *regexp.Regexp {\n\tif l.regexp == nil { // No need to take a lock. Worst case, we'll just compile it multiple times.\n\t\tl.regexp = regexp.MustCompile(l.pattern)\n\t}\n\treturn l.regexp\n}", "func NewComp(r Regex) Regex {\n\treturn &comp{\n\t\tr: r,\n\t}\n}", "func RegexpSimplify(re *syntax.Regexp,) *syntax.Regexp", "func RegExp(pattern string) *RegExpMatcher {\n\treturn &RegExpMatcher{Pattern: pattern}\n}", "func (vm *VM) initRegexpObject(regexp string) *RegexpObject {\n\tr, err := regexp2.Compile(regexp, 0)\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn &RegexpObject{\n\t\tBaseObj: NewBaseObject(vm.TopLevelClass(classes.RegexpClass)),\n\t\tregexp: r,\n\t}\n}", "func Regex(pattern, options string) Val {\n\tregex := primitive.Regex{Pattern: pattern, Options: options}\n\treturn Val{t: bsontype.Regex, primitive: regex}\n}", "func rawPattern(s string) *regexp.Regexp {\n\treturn regexp.MustCompile(regexp.QuoteMeta(s))\n}", "func RE(str string) func() *regexp.Regexp {\n\treturn sync.OnceValue(func() *regexp.Regexp {\n\t\treturn regexp.MustCompile(str)\n\t})\n}", "func (c *Config) CompileRegexps() error {\n\tfor _, licenseRegexps := range c.Licenses {\n\t\tlicenseRegexp := strings.Join(licenseRegexps, \"\\n\")\n\t\tre, err := regexp.Compile(licenseRegexp)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tc.licensesRegexps = append(c.licensesRegexps, re)\n\t}\n\n\tc.accept = make([]rule, 0, len(c.Accept))\n\tfor _, rule := range c.Accept {\n\t\tc.accept = append(c.accept, accept(rule))\n\t}\n\n\tc.reject = make([]rule, 0, len(c.Reject))\n\tfor _, rule := range c.Reject {\n\t\tc.reject = append(c.reject, reject(rule))\n\t}\n\n\treturn nil\n}", "func (e PiRegExp) exp() *regexp.Regexp { return regexp.MustCompile(string(e)) }", "func CompileString(in string) (Matcher, error) {\n\tif in == \"\" {\n\t\treturn Matcher{(*emptyStringMatcher)(nil)}, nil\n\t}\n\treturn Matcher{&substringMatcher{in, []byte(in)}}, nil\n}", "func NewRegexpSet(patterns []string) (retVal *RegexpSet, err error) {\n\tfor _, pattern := range patterns {\n\t\t_, err := regexp.Compile(pattern)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"Invalid regular pattern: '%v'.\\n\", pattern)\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn &RegexpSet{patterns}, nil\n}", "func compilePath(path spec.Path) (*regexp.Regexp, []string) {\n\tvar pathParamNames []string\n\tparts := strings.Split(string(path), \"/\")\n\tpattern := `\\A`\n\n\tfor _, part := range parts {\n\t\tif part == \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tsubmatches := pathParameterPattern.FindAllStringSubmatch(part, -1)\n\t\tif submatches == nil {\n\t\t\tpattern += `/` + part\n\t\t} else {\n\t\t\t// Special characters as defined by:\n\t\t\t//\n\t\t\t// https://tools.ietf.org/html/rfc3986#section-3.3\n\t\t\tpattern += `/(?P<` + submatches[0][1] + `>[\\w@:%-._~!$&'()*+,;=]+)`\n\t\t\tpathParamNames = append(pathParamNames, submatches[0][1])\n\t\t}\n\t}\n\n\treturn regexp.MustCompile(pattern + `\\z`), pathParamNames\n}", "func MustCompile(input string) *Regexp {\n\treturn MustCompile2(input, 0)\n}", "func New(expr string) (*Regexp, error) {\n\treturn NewWithLimit(expr, DefaultLimit)\n}", "func (re *Regexp) Convert() (*goregexp.Regexp, error) {\n\treturn goregexp.Compile(re.pattern)\n}", "func RegexpParse(content, pattern string) ([]string, error) {\n\tre, err := pool.Compile(pattern)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"re:[%s] error:%+v\", pattern, err)\n\t}\n\tvar ret []string\n\tres := re.FindAllStringSubmatch(content, -1)\n\tfor i := range res {\n\t\tswitch {\n\t\tcase len(res[i]) == 1:\n\t\t\tret = append(ret, res[i][0])\n\t\tcase len(res[i]) > 1:\n\t\t\tret = append(ret, res[i][1:]...)\n\t\t}\n\t}\n\treturn ret, nil\n}", "func NewRegexpFile(r *regexp.Regexp) *File {\n\treturn NewFile(&regexpElement{Data: r})\n}", "func expandRegex(r *regexp.Regexp, tmpl string, src string) string {\n\tvar res []byte\n\tfor _, submatches := range r.FindAllStringSubmatchIndex(src, 1) {\n\t\tres = r.ExpandString(res, tmpl, src, submatches)\n\t}\n\treturn string(res)\n}", "func NewRegexpMatcher(terms ...*regexp.Regexp) (RegexpMatcher, error) {\n\tif len(terms) == 0 {\n\t\treturn matchNothingRegexpMatcher, nil\n\t}\n\n\t// combine expressions into one using OR, i.e.\n\t// [RE1, RE2, ..., REn] -> (RE1)|(RE2)|...|(REn)\n\tbuf := bytes.NewBuffer([]byte(`(\\A`))\n\tsep := []byte(`\\z)|(\\A`)\n\n\tfor _, term := range terms {\n\t\treBytes := []byte(term.String())\n\t\t// strip leading beginning-of-line matchers, as they are already included into the combined expression\n\t\treBytes = bytes.TrimPrefix(bytes.TrimLeft(reBytes, \"^\"), []byte(`\\A`))\n\t\t// strip trailing end-of-line matchers, as they are already included into the combined expression\n\t\treBytes = bytes.TrimSuffix(bytes.TrimRight(reBytes, \"$\"), []byte(`\\z`))\n\n\t\tbuf.Write(reBytes)\n\t\tbuf.Write(sep)\n\t}\n\tbuf.Truncate(buf.Len() - len(sep)) // trim trailing separator\n\tbuf.WriteString(`\\z)`)\n\n\tcombined := buf.String()\n\n\tre, err := regexp.Compile(combined)\n\tif err != nil {\n\t\treturn matchNothingRegexpMatcher, fmt.Errorf(\"malformed regexp %q: %s\", combined, err)\n\t}\n\n\treturn RegexpMatcher{re}, nil\n}", "func NewGenerator(regex string) (*Generator, error) {\n\tre, err := syntax.Parse(regex, syntax.Perl)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t//fmt.Println(\"Compiled re \", re)\n\treturn &Generator{\n\t\tre: re,\n\t\trand: rand.New(rand.NewSource(time.Now().UnixNano())),\n\t}, nil\n}", "func Globexp(glob string) *regexp.Regexp {\n\tvar re bytes.Buffer\n\n\tre.WriteString(\"^\")\n\n\ti, inGroup, L := 0, false, len(glob)\n\n\tfor i < L {\n\t\tr, w := utf8.DecodeRuneInString(glob[i:])\n\n\t\tswitch r {\n\t\tdefault:\n\t\t\tre.WriteRune(r)\n\n\t\tcase '\\\\', '$', '^', '+', '.', '(', ')', '=', '!', '|':\n\t\t\tre.WriteRune('\\\\')\n\t\t\tre.WriteRune(r)\n\n\t\tcase '/':\n\t\t\t// TODO optimize later, string could be long\n\t\t\trest := glob[i:]\n\t\t\tre.WriteRune('/')\n\t\t\tif strings.HasPrefix(rest, \"/**/\") {\n\t\t\t\tre.WriteString(zeroOrMoreDirectories)\n\t\t\t\tw *= 4\n\t\t\t} else if rest == \"/**\" {\n\t\t\t\tre.WriteString(\".*\")\n\t\t\t\tw *= 3\n\t\t\t}\n\n\t\tcase '?':\n\t\t\tre.WriteRune('.')\n\n\t\tcase '[', ']':\n\t\t\tre.WriteRune(r)\n\n\t\tcase '{':\n\t\t\tif i < L-1 {\n\t\t\t\tif glob[i+1:i+2] == \"{\" {\n\t\t\t\t\tre.WriteString(\"\\\\{\")\n\t\t\t\t\tw *= 2\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tinGroup = true\n\t\t\tre.WriteRune('(')\n\n\t\tcase '}':\n\t\t\tif inGroup {\n\t\t\t\tinGroup = false\n\t\t\t\tre.WriteRune(')')\n\t\t\t} else {\n\t\t\t\tre.WriteRune('}')\n\t\t\t}\n\n\t\tcase ',':\n\t\t\tif inGroup {\n\t\t\t\tre.WriteRune('|')\n\t\t\t} else {\n\t\t\t\tre.WriteRune('\\\\')\n\t\t\t\tre.WriteRune(r)\n\t\t\t}\n\n\t\tcase '*':\n\t\t\trest := glob[i:]\n\t\t\tif strings.HasPrefix(rest, \"**/\") {\n\t\t\t\tre.WriteString(zeroOrMoreDirectories)\n\t\t\t\tw *= 3\n\t\t\t} else {\n\t\t\t\tre.WriteString(anyRune)\n\t\t\t}\n\t\t}\n\n\t\ti += w\n\t}\n\n\tre.WriteString(\"$\")\n\t//log.Printf(\"regex string %s\", re.String())\n\treturn regexp.MustCompile(re.String())\n}", "func NewRegex(v string) (Value, error) {\n\trx, err := regexp.Compile(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rxValue{rx}, nil\n}", "func TestPattern2(t *testing.T) {\n\tre := MustCompile(\"a$\")\n\tif !re.MatchString(\"a\") {\n\t\tt.Errorf(\"expect to match\\n\")\n\t}\n\tif re.MatchString(\"ab\") {\n\t\tt.Errorf(\"expect to mismatch\\n\")\n\t}\n}", "func MaskToRegex(rawMask string) (*regexp.Regexp, error) {\n\tinput := bytes.NewBufferString(rawMask)\n\n\toutput := &bytes.Buffer{}\n\toutput.WriteByte('^')\n\n\tfor {\n\t\tc, err := input.ReadByte()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\n\t\tif c == '\\\\' {\n\t\t\tc, err = input.ReadByte()\n\t\t\tif err != nil {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(\"\\\\\"))\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif c == '?' || c == '*' || c == '\\\\' {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(string(c)))\n\t\t\t} else {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(\"\\\\\" + string(c)))\n\t\t\t}\n\t\t} else if trans, ok := maskTranslations[c]; ok {\n\t\t\toutput.WriteString(trans)\n\t\t} else {\n\t\t\toutput.WriteString(regexp.QuoteMeta(string(c)))\n\t\t}\n\t}\n\n\toutput.WriteByte('$')\n\n\treturn regexp.Compile(output.String())\n}", "func (e *ExpressionWithSeverity) compile() error {\n\tr, err := regexp.Compile(e.Expression)\n\tif err != nil {\n\t\treturn err\n\t}\n\te.compiled = r\n\treturn nil\n}", "func (v *VerbalExpression) Regex() *regexp.Regexp {\n\n\tif !v.compiled {\n\t\tv.regexp = regexp.MustCompile(\n\t\t\tstrings.Join([]string{\n\t\t\t\tstrings.Join(v.parts, \"\"),\n\t\t\t\t`(?` + v.getFlags() + `)`,\n\t\t\t\tv.prefixes,\n\t\t\t\tv.expression,\n\t\t\t\tv.suffixes}, \"\"))\n\t\tv.compiled = true\n\t}\n\treturn v.regexp\n}", "func CompileRules(r io.Reader, a *cli.Args) *CompiledRuleList {\n\tvar rules = &Rules{}\n\trules.LoadRules(r)\n\tvar crl CompiledRuleList\n\n\t// Build a slice of all compiled patterns which we then use with data\n\t// consumed from stdin.\n\tvar cexp *regexp.Regexp\n\tvar pattern string\n\tvar debug bool\n\tfor i, r := range rules.ListOfRules() {\n\t\tif r.PatString == MatchNoOutput {\n\t\t\tpattern = zeroInputPattern\n\t\t} else {\n\t\t\tpattern = r.PatString\n\t\t}\n\t\tif r.ERESyntax {\n\t\t\tcexp = regexp.MustCompilePOSIX(pattern)\n\t\t} else {\n\t\t\tcexp = regexp.MustCompile(pattern)\n\t\t}\n\t\t// If debug is set on individual rule, or as a command-line argument,\n\t\t// enable it on compiled rule as well.\n\t\tdebug = r.Debug || a.Debug()\n\t\tcrl.Append(CompiledRule{\n\t\t\tre: cexp,\n\t\t\tdesc: rules.RuleByIndex(i),\n\t\t\tdebug: debug})\n\t}\n\n\treturn &crl\n}", "func wildCardToRegexp(pattern string) string {\n\tvar result strings.Builder\n\tfor i, literal := range strings.Split(pattern, \"*\") {\n\n\t\t// Replace * with .*\n\t\tif i > 0 {\n\t\t\tresult.WriteString(\".*\")\n\t\t}\n\n\t\t// Quote any regular expression meta characters in the\n\t\t// literal text.\n\t\tresult.WriteString(regexp.QuoteMeta(literal))\n\t}\n\treturn result.String()\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func (p *parser) regex() Node {\n\ttoken := p.expect(TokenRegex)\n\tr, err := newRegex(token.pos, token.val)\n\tif err != nil {\n\t\tp.error(err)\n\t}\n\treturn r\n}", "func NewRegexPatternSet(ctx *pulumi.Context,\n\tname string, args *RegexPatternSetArgs, opts ...pulumi.ResourceOption) (*RegexPatternSet, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.RegularExpressionList == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RegularExpressionList'\")\n\t}\n\tif args.Scope == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Scope'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource RegexPatternSet\n\terr := ctx.RegisterResource(\"aws-native:wafv2:RegexPatternSet\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func regex(s string) (*regexp.Regexp, error) {\n\tif rawString.MatchString(s) {\n\t\ts = fmt.Sprintf(\"^%s$\", s)\n\t}\n\treturn regexp.Compile(s)\n}", "func regExp(context interface{}, value string) (bson.RegEx, error) {\n\tidx := strings.IndexByte(value[1:], '/')\n\tif value[0] != '/' || idx == -1 {\n\t\terr := fmt.Errorf(\"Parameter %q is not a regular expression\", value)\n\t\tlog.Error(context, \"varLookup\", err, \"Regex parsing\")\n\t\treturn bson.RegEx{}, err\n\t}\n\n\tpattern := value[1 : idx+1]\n\tl := len(pattern) + 2\n\n\tvar options string\n\tif l < len(value) {\n\t\toptions = value[l:]\n\t}\n\n\treturn bson.RegEx{Pattern: pattern, Options: options}, nil\n}", "func NewTestGlobRegexp(glob string) (*regexp.Regexp, error) {\n\tif _, err := validateGlob(glob); err != nil {\n\t\treturn nil, err\n\t}\n\treturn compileGlob(glob)\n}", "func buildResourceRegexp(s string) (*regexp.Regexp, error) {\n\thash := strings.Split(s, \":\")\n\tfor i, v := range hash {\n\t\tif v == \"\" || v == \"*\" {\n\t\t\thash[i] = \".*\"\n\t\t}\n\t}\n\treturn regexp.Compile(strings.Join(hash, \":\"))\n}", "func createGlobbingRegex(globbing string) *regexp.Regexp {\n\tif globbing == \"\" {\n\t\t// nil here as \"\" is fast-tracked elsewhere\n\t\treturn nil\n\t}\n\n\tg := regexp.QuoteMeta(globbing)\n\tg = strings.ReplaceAll(g, \"\\\\*\", \".*\")\n\tg = strings.ReplaceAll(g, \"\\\\?\", \".\")\n\t// (?i) forces case insensitive matches\n\tg = \"(?i)^\" + g + \"$\"\n\treturn regexp.MustCompile(g)\n}", "func buildSearchTerm(term string) *regexp.Regexp {\n\tvar ret *regexp.Regexp\n\tvar regex string\n\n\t// --regex\n\tif opts.Regex {\n\t\t// use search term as regex\n\t\tregex = term\n\t} else {\n\t\t// use search term as normal string, escape it for regex usage\n\t\tregex = regexp.QuoteMeta(term)\n\t}\n\n\t// --ignore-case\n\tif opts.CaseInsensitive {\n\t\tregex = \"(?i:\" + regex + \")\"\n\t}\n\n\t// --verbose\n\tif opts.Verbose {\n\t\tlogMessage(fmt.Sprintf(\"Using regular expression: %s\", regex))\n\t}\n\n\t// --regex-posix\n\tif opts.RegexPosix {\n\t\tret = regexp.MustCompilePOSIX(regex)\n\t} else {\n\t\tret = regexp.MustCompile(regex)\n\t}\n\n\treturn ret\n}", "func wildCardToRegexp(pattern string) string {\n\tvar result strings.Builder\n\tresult.WriteString(\"(?i)\")\n\n\trpattern := strings.Replace(pattern, \"%\", \".*\", -1)\n\trpattern = strings.Replace(rpattern, \"_\", \".+\", -1)\n\tresult.WriteString(rpattern)\n\n\treturn result.String()\n}", "func regexAfterInit(mrb *oruby.MrbState, self oruby.Value) oruby.MrbValue {\n\treg := mrb.Data(self).(*regexp.Regexp)\n\tsource := reg.String()\n\toptions := 0\n\n\t_ = mrb.IVSet(self, mrb.Intern(\"@source\"), mrb.Value(source))\n\n\tif strings.HasPrefix(source, \"(?m\") {\n\t\toptions = options | Multiline\n\t}\n\tif strings.HasPrefix(source, \"(?i\") {\n\t\toptions = options | IgnoreCase\n\t}\n\tif strings.HasPrefix(source, \"(?im\") {\n\t\toptions = options | Multiline | IgnoreCase\n\t}\n\n\t_ = mrb.IVSet(self, mrb.Intern(\"@options\"), mrb.Value(options))\n\treturn self\n}", "func TestCompiledRegexInGoRoutine(t *testing.T) {\n\tregex, err := regexp.Compile(\"^(B|b)rian$\")\n\tif err != nil {\n\t\tt.Errorf(\"Regex did not compile %v\", err)\n\t}\n\n\tch := make(chan bool)\n\ttests := []string{\"brian\", \"Brian\"}\n\tfor _, test := range tests {\n\t\tgo func(t string) {\n\t\t\tch <- regex.MatchString(t)\n\t\t}(test)\n\t}\n\n\tfirst, second := <-ch, <-ch\n\n\tif first && second {\n\t\treturn\n\t}\n\n\tt.Error(\"String did not match\")\n}", "func (m *Builder) Compile() (*MultiGlob, error) {\n\tvar final *parser.Node\n\tfor _, p := range m.patterns {\n\t\tif final == nil {\n\t\t\tfinal = p\n\t\t} else {\n\t\t\tfinal = parser.Merge(final, p)\n\t\t}\n\t}\n\n\tpatterns := make(map[string]*parser.Node)\n\tfor k, v := range m.patterns {\n\t\tpatterns[k] = v\n\t}\n\n\treturn &MultiGlob{\n\t\tnode: final,\n\t\tpatterns: patterns,\n\t}, nil\n}", "func NewMatcher(pats []string) (*Matcher, error) {\n\tif len(pats) == 1 && strings.HasPrefix(pats[0], \"(\") && strings.HasSuffix(pats[0], \")\") {\n\t\treturn compileExpr(pats[0][1 : len(pats[0])-1])\n\t}\n\treturn compileGlobs(pats)\n}", "func RegexFromGlob(glob string) string {\n\treturn RegexFromGlobWithOptions(glob, Options{\n\t\tExtended: true,\n\t\tGlobStar: true,\n\t\tDelimiter: '/',\n\t})\n}", "func compile(conf Config) regexps {\n\tr := make(regexps)\n\n\tfor k, v := range conf {\n\t\tr[regexp.MustCompile(k)] = v\n\t}\n\n\treturn r\n}", "func MustNewRegexp(s string) Regexp {\n\tre, err := NewRegexp(s)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn re\n}", "func MustNewRegexp(s string) Regexp {\n\tre, err := NewRegexp(s)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn re\n}", "func NewWithLimit(expr string, size uint) (*Regexp, error) {\n\tparsed, err := syntax.Parse(expr, syntax.Perl)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn NewParsedWithLimit(expr, parsed, size)\n}", "func Compile(sel string) (Selector, error) {\n\tcompiled, err := ParseGroup(sel)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn Selector(compiled.Match), nil\n}", "func ProjectRegexp() *regexp.Regexp {\n\treturn regexp.MustCompile(\"^projects/\" + NameRegex + \"$\")\n}", "func Match(regex string, text string) bool {\n runerx := compile(regex)\n runetxt := []rune(text)\n \n if len(runerx) > 0 && runerx[0] == '^' {\n return matchhere(runerx[1:], runetxt)\n }\n \n for {\n if matchhere(runerx, runetxt) {\n return true\n }\n if len(runetxt) == 0 {\n return false\n }\n runetxt = runetxt[1:]\n }\n }", "func RegexFromGlobWithOptions(glob string, config Options) string {\n\treStr := \"\"\n\n\tdelimiter := '/'\n\tif config.Delimiter != 0 {\n\t\tdelimiter = config.Delimiter\n\t}\n\n\tdelimiterOutsideClass, delimiterInsideClass := escapeDelimiter(delimiter)\n\n\tinGroup := false\n\n\tglobLen := len(glob)\n\n\tfor i := 0; i < globLen; i++ {\n\t\tc := glob[i]\n\n\t\tswitch c {\n\t\tcase '/', '$', '^', '+', '.', '(', ')', '=', '!', '|':\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '?':\n\t\t\tif config.Extended {\n\t\t\t\treStr += \".\"\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '[', ']':\n\t\t\tif config.Extended {\n\t\t\t\treStr += string(c)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '{':\n\t\t\tif config.Extended {\n\t\t\t\tinGroup = true\n\t\t\t\treStr += \"(\"\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '}':\n\t\t\tif config.Extended {\n\t\t\t\tinGroup = false\n\t\t\t\treStr += \")\"\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase ',':\n\t\t\tif inGroup {\n\t\t\t\treStr += \"|\"\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '*':\n\t\t\t// Move over all consecutive \"*\"'s.\n\t\t\t// Also store the previous and next characters\n\t\t\tvar nextChar, prevChar rune\n\t\t\tif i > 0 {\n\t\t\t\tprevChar = rune(glob[i-1])\n\t\t\t}\n\t\t\tstarCount := 1\n\t\t\tfor i < globLen-1 && glob[i+1] == '*' {\n\t\t\t\tstarCount++\n\t\t\t\ti++\n\t\t\t}\n\n\t\t\tif i < globLen-1 {\n\t\t\t\tnextChar = rune(glob[i+1])\n\t\t\t}\n\n\t\t\tif !config.GlobStar {\n\t\t\t\t// globstar is disabled, so treat any number of \"*\" as one\n\t\t\t\treStr += \".*\"\n\t\t\t} else {\n\t\t\t\t// globstar is enabled, so determine if this is a globstar segment\n\t\t\t\tisGlobstar := starCount > 1 && // multiple \"*\"'s\n\t\t\t\t\t(prevChar == delimiter || prevChar == 0) && // from the start of the segment\n\t\t\t\t\t(nextChar == delimiter || nextChar == 0) // to the end of the segment\n\n\t\t\t\tif isGlobstar {\n\t\t\t\t\t// it's a globstar, so match zero or more path segments\n\t\t\t\t\treStr += \"(?:(?:[^\" + delimiterInsideClass + \"]*(?:\" + delimiterOutsideClass + \"|$))*)\"\n\t\t\t\t\ti++ // move over the delimiter\n\t\t\t\t} else {\n\t\t\t\t\t// it's not a globstar, so only match one path segment\n\t\t\t\t\treStr += \"(?:[^\" + delimiterInsideClass + \"]*)\"\n\t\t\t\t}\n\t\t\t}\n\n\t\tdefault:\n\t\t\treStr += string(c)\n\t\t}\n\t}\n\n\treturn \"^\" + reStr + \"$\"\n}", "func StructuralPatToRegexpQuery(pattern string, shortcircuit bool) (zoektquery.Q, error) {\n\tsubstrings := matchHoleRegexp.Split(pattern, -1)\n\tvar children []zoektquery.Q\n\tvar pieces []string\n\tfor _, s := range substrings {\n\t\tpiece := regexp.QuoteMeta(s)\n\t\tonMatchWhitespace := lazyregexp.New(`[\\s]+`)\n\t\tpiece = onMatchWhitespace.ReplaceAllLiteralString(piece, `[\\s]+`)\n\t\tpieces = append(pieces, piece)\n\t}\n\n\tif len(pieces) == 0 {\n\t\treturn &zoektquery.Const{Value: true}, nil\n\t}\n\tvar rs string\n\tif shortcircuit {\n\t\t// As a shortcircuit, do not match across newlines of structural search pieces.\n\t\trs = \"(\" + strings.Join(pieces, \").*?(\") + \")\"\n\t} else {\n\t\trs = \"(\" + strings.Join(pieces, \")(.|\\\\s)*?(\") + \")\"\n\t}\n\tre, _ := syntax.Parse(rs, syntax.ClassNL|syntax.PerlX|syntax.UnicodeGroups)\n\tchildren = append(children, &zoektquery.Regexp{\n\t\tRegexp: re,\n\t\tCaseSensitive: true,\n\t\tContent: true,\n\t})\n\treturn &zoektquery.And{Children: children}, nil\n}" ]
[ "0.73134357", "0.67773414", "0.67725855", "0.6715364", "0.6700544", "0.66973156", "0.6678918", "0.63899606", "0.63447237", "0.63398224", "0.61929864", "0.6109555", "0.6099", "0.6074681", "0.6041667", "0.6004279", "0.589885", "0.5801142", "0.5786491", "0.5763816", "0.56644255", "0.5663318", "0.56456065", "0.56442845", "0.56427246", "0.5629208", "0.5590656", "0.5529395", "0.5527368", "0.5501303", "0.549223", "0.54762876", "0.5471823", "0.54419637", "0.5425465", "0.53897667", "0.53824365", "0.5366259", "0.53605074", "0.5333118", "0.53247845", "0.53158873", "0.53016907", "0.5276074", "0.5238953", "0.52332664", "0.520839", "0.518953", "0.51875377", "0.515439", "0.51533216", "0.5092116", "0.5085098", "0.50748914", "0.504452", "0.5022292", "0.5012514", "0.49986887", "0.49918112", "0.4988224", "0.49685335", "0.49455956", "0.49400082", "0.49339983", "0.48854947", "0.4860072", "0.48513535", "0.48481926", "0.48462072", "0.48312518", "0.4820656", "0.48060614", "0.47895306", "0.47795776", "0.47688773", "0.47688773", "0.47688773", "0.47641373", "0.47609437", "0.47525227", "0.47316507", "0.47306448", "0.47162136", "0.47153997", "0.47006616", "0.46885127", "0.4659328", "0.46498123", "0.4647607", "0.4645188", "0.46387938", "0.46356666", "0.46332642", "0.46332642", "0.46277815", "0.46262473", "0.4621981", "0.46106094", "0.460696", "0.46016786" ]
0.91923267
0
QuoteRegex uses regexp.QuoteMeta to returnsoperand with all regex metachars escaped.
func QuoteRegex(operand string) string { return regexp.QuoteMeta(operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func quote(s string) string {\n\treturn regexp.QuoteMeta(s)\n}", "func Quote(operand string) string { return strconv.Quote(operand) }", "func RegexpSlashQuote(sb *strings.Builder, str string) {\n\tutil.WriteByte(sb, '/')\n\tfor _, c := range str {\n\t\tswitch c {\n\t\tcase '\\t':\n\t\t\tutil.WriteString(sb, `\\t`)\n\t\tcase '\\n':\n\t\t\tutil.WriteString(sb, `\\n`)\n\t\tcase '\\r':\n\t\t\tutil.WriteString(sb, `\\r`)\n\t\tcase '/':\n\t\t\tutil.WriteString(sb, `\\/`)\n\t\tcase '\\\\':\n\t\t\tutil.WriteString(sb, `\\\\`)\n\t\tdefault:\n\t\t\tif c < 0x20 {\n\t\t\t\tutil.Fprintf(sb, `\\u{%X}`, c)\n\t\t\t} else {\n\t\t\t\tutil.WriteRune(sb, c)\n\t\t\t}\n\t\t}\n\t}\n\tutil.WriteByte(sb, '/')\n}", "func QuoteMeta(s string) string {\n\tvar buf bytes.Buffer\n\n\tfor _, ch := range s {\n\t\tswitch ch {\n\t\tcase '.', '+', '\\\\', '(', '$', ')', '[', '^', ']', '*', '?':\n\t\t\tbuf.WriteRune('\\\\')\n\t\t}\n\n\t\tbuf.WriteRune(ch)\n\t}\n\n\treturn buf.String()\n}", "func rawPattern(s string) *regexp.Regexp {\n\treturn regexp.MustCompile(regexp.QuoteMeta(s))\n}", "func TestMatchStringWithQuoteMeta(t *testing.T) {\n\tpattern, str := regexp.QuoteMeta(\"[foo]\"), \"[foo]\"\n\n\tif match, err := regexp.MatchString(pattern, str); match != true {\n\t\tt.Errorf(\"MatchString did not match %q %v\", str, err)\n\t}\n}", "func Unquote(operand string) (unquoted string, err error) { return strconv.Unquote(operand) }", "func (c CredentialService) GetRegexToEscape() string {\n\treturn c.RegexToEscape\n}", "func compile(regex string) (regslc []rune) {\n regslc = make([]rune, 0, len(regex) + 10)\n \n for _, r := range regex {\n if r == '+' {\n regslc = append(regslc, regslc[len(regslc) - 1], '*')\n } else {\n regslc = append(regslc, r)\n }\n } \n return regslc\n }", "func ShellQuote(s string) string {\n\tif len(s) > 0 && !strings.ContainsAny(s, shellmeta) {\n\t\treturn s // fast path for common case\n\t}\n\tdouble := strings.Contains(s, \"\\\"\")\n\tsingle := strings.Contains(s, \"'\")\n\tif double && single {\n\t\tif shellreplacer == nil {\n\t\t\tpairs := make([]string, len(shellmeta)*2)\n\t\t\tfor i := 0; i < len(shellmeta); i++ {\n\t\t\t\tpairs[i*2] = string(shellmeta[i])\n\t\t\t\tpairs[i*2+1] = \"\\\\\" + string(shellmeta[i])\n\t\t\t}\n\t\t\tshellreplacer = strings.NewReplacer(pairs...)\n\t\t}\n\t\treturn shellreplacer.Replace(s)\n\t} else if single {\n\t\t// use double quotes, but be careful of $\n\t\treturn \"\\\"\" + strings.Replace(s, \"$\", \"\\\\$\", -1) + \"\\\"\"\n\t} else {\n\t\t// use single quotes\n\t\treturn \"'\" + s + \"'\"\n\t}\n\tpanic(\"unreachable code\")\n}", "func WildcardToRegex(wildcard string) string {\n\tvar b strings.Builder\n\tb.WriteByte('^')\n\tfor {\n\t\tidx := strings.IndexByte(wildcard, '*')\n\t\tif idx < 0 {\n\t\t\tbreak\n\t\t}\n\t\tb.WriteString(regexp.QuoteMeta(wildcard[:idx]))\n\t\tb.WriteString(\"(.*)\")\n\t\twildcard = wildcard[idx+1:]\n\t}\n\tb.WriteString(regexp.QuoteMeta(wildcard))\n\tb.WriteByte('$')\n\treturn b.String()\n}", "func (self *T) mQUOTE() {\r\n \r\n \r\n\t\t_type := T_QUOTE\r\n\t\t_channel := antlr3rt.DEFAULT_TOKEN_CHANNEL\r\n\t\t// C:/dev/antlr.github/antlr/runtime/Go/antlr/test/T.g:11:6: ( '\\\"' )\r\n\t\t// C:/dev/antlr.github/antlr/runtime/Go/antlr/test/T.g:11:9: '\\\"'\r\n\t\t{\r\n\t\tself.MatchChar('\"') \r\n\r\n\r\n\t\t}\r\n\r\n\t\tself.State().SetType( _type )\r\n\t\tself.State().SetChannel( _channel )\r\n}", "func wildCardToRegexp(pattern string) string {\n\tvar result strings.Builder\n\tfor i, literal := range strings.Split(pattern, \"*\") {\n\n\t\t// Replace * with .*\n\t\tif i > 0 {\n\t\t\tresult.WriteString(\".*\")\n\t\t}\n\n\t\t// Quote any regular expression meta characters in the\n\t\t// literal text.\n\t\tresult.WriteString(regexp.QuoteMeta(literal))\n\t}\n\treturn result.String()\n}", "func (e PiRegExp) exp() *regexp.Regexp { return regexp.MustCompile(string(e)) }", "func RegexFromGlobWithOptions(glob string, config Options) string {\n\treStr := \"\"\n\n\tdelimiter := '/'\n\tif config.Delimiter != 0 {\n\t\tdelimiter = config.Delimiter\n\t}\n\n\tdelimiterOutsideClass, delimiterInsideClass := escapeDelimiter(delimiter)\n\n\tinGroup := false\n\n\tglobLen := len(glob)\n\n\tfor i := 0; i < globLen; i++ {\n\t\tc := glob[i]\n\n\t\tswitch c {\n\t\tcase '/', '$', '^', '+', '.', '(', ')', '=', '!', '|':\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '?':\n\t\t\tif config.Extended {\n\t\t\t\treStr += \".\"\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '[', ']':\n\t\t\tif config.Extended {\n\t\t\t\treStr += string(c)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '{':\n\t\t\tif config.Extended {\n\t\t\t\tinGroup = true\n\t\t\t\treStr += \"(\"\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '}':\n\t\t\tif config.Extended {\n\t\t\t\tinGroup = false\n\t\t\t\treStr += \")\"\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase ',':\n\t\t\tif inGroup {\n\t\t\t\treStr += \"|\"\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treStr += \"\\\\\" + string(c)\n\n\t\tcase '*':\n\t\t\t// Move over all consecutive \"*\"'s.\n\t\t\t// Also store the previous and next characters\n\t\t\tvar nextChar, prevChar rune\n\t\t\tif i > 0 {\n\t\t\t\tprevChar = rune(glob[i-1])\n\t\t\t}\n\t\t\tstarCount := 1\n\t\t\tfor i < globLen-1 && glob[i+1] == '*' {\n\t\t\t\tstarCount++\n\t\t\t\ti++\n\t\t\t}\n\n\t\t\tif i < globLen-1 {\n\t\t\t\tnextChar = rune(glob[i+1])\n\t\t\t}\n\n\t\t\tif !config.GlobStar {\n\t\t\t\t// globstar is disabled, so treat any number of \"*\" as one\n\t\t\t\treStr += \".*\"\n\t\t\t} else {\n\t\t\t\t// globstar is enabled, so determine if this is a globstar segment\n\t\t\t\tisGlobstar := starCount > 1 && // multiple \"*\"'s\n\t\t\t\t\t(prevChar == delimiter || prevChar == 0) && // from the start of the segment\n\t\t\t\t\t(nextChar == delimiter || nextChar == 0) // to the end of the segment\n\n\t\t\t\tif isGlobstar {\n\t\t\t\t\t// it's a globstar, so match zero or more path segments\n\t\t\t\t\treStr += \"(?:(?:[^\" + delimiterInsideClass + \"]*(?:\" + delimiterOutsideClass + \"|$))*)\"\n\t\t\t\t\ti++ // move over the delimiter\n\t\t\t\t} else {\n\t\t\t\t\t// it's not a globstar, so only match one path segment\n\t\t\t\t\treStr += \"(?:[^\" + delimiterInsideClass + \"]*)\"\n\t\t\t\t}\n\t\t\t}\n\n\t\tdefault:\n\t\t\treStr += string(c)\n\t\t}\n\t}\n\n\treturn \"^\" + reStr + \"$\"\n}", "func (r *IPRule) GrepEscaped() (string, string) {\n\treturn strings.Replace(r.SourceIP.String(), \".\", `\\.`, 3), strings.Replace(r.RouteTableName, \".\", `\\.`, -1)\n}", "func literal(s string) *regexp.Regexp {\n\tre := match(regexp.QuoteMeta(s))\n\n\tif _, complete := re.LiteralPrefix(); !complete {\n\t\tpanic(\"must be a literal\")\n\t}\n\n\treturn re\n}", "func literal(s string) *regexp.Regexp {\n\tre := match(regexp.QuoteMeta(s))\n\n\tif _, complete := re.LiteralPrefix(); !complete {\n\t\tpanic(\"must be a literal\")\n\t}\n\n\treturn re\n}", "func literal(s string) *regexp.Regexp {\n\tre := match(regexp.QuoteMeta(s))\n\n\tif _, complete := re.LiteralPrefix(); !complete {\n\t\tpanic(\"must be a literal\")\n\t}\n\n\treturn re\n}", "func (b ValExprBuilder) RegExp(val interface{}) BoolExprBuilder {\n\treturn b.makeComparisonExpr(astRegExp, makeValExpr(val))\n}", "func escape(input string) string {\n\treturn strings.ReplaceAll(strings.ReplaceAll(input, `\\`, `\\\\`), \"'\", `\\'`)\n}", "func Regex(pattern, options string) Val {\n\tregex := primitive.Regex{Pattern: pattern, Options: options}\n\treturn Val{t: bsontype.Regex, primitive: regex}\n}", "func escapeQuotes(s string) string {\n\treturn quoteEscaper.Replace(s)\n}", "func escapeQuotes(s string) string {\n\treturn quoteEscaper.Replace(s)\n}", "func RegexpSimplify(re *syntax.Regexp,) *syntax.Regexp", "func wildCardToRegexp(pattern string) string {\n\tvar result strings.Builder\n\tresult.WriteString(\"(?i)\")\n\n\trpattern := strings.Replace(pattern, \"%\", \".*\", -1)\n\trpattern = strings.Replace(rpattern, \"_\", \".+\", -1)\n\tresult.WriteString(rpattern)\n\n\treturn result.String()\n}", "func (p *Parser) parseRegexpLiteral() asti.ExpressionI {\n\n\tflags := \"\"\n\n\tval := p.curToken.Literal\n\tif strings.HasPrefix(val, \"(?\") {\n\t\tval = strings.TrimPrefix(val, \"(?\")\n\n\t\ti := 0\n\t\tfor i < len(val) {\n\n\t\t\tif val[i] == ')' {\n\n\t\t\t\tval = val[i+1:]\n\t\t\t\tbreak\n\t\t\t} else {\n\t\t\t\tflags += string(val[i])\n\t\t\t}\n\n\t\t\ti++\n\t\t}\n\t}\n\treturn &ast.RegexpLiteral{Token: p.curToken, Value: val, Flags: flags}\n}", "func getMagicQuotesRuntime(ctx phpv.Context, args []*phpv.ZVal) (*phpv.ZVal, error) {\n\treturn phpv.ZBool(false).ZVal(), nil\n}", "func lexOpQuoted(l *lexer) lexFn {\n\tomitSpaces(l)\n\n\tl.acceptRun(_OpValueRunes)\n\tl.emit(OPV_QUOTED)\n\n\treturn lexText\n}", "func QuoteMeta(s string) string {\n\t// // A byte loop is correct because all metacharacters are ASCII.\n\t// var i int\n\t// for i = 0; i < len(s); i++ {\n\t// \tif special(s[i]) {\n\t// \t\tbreak\n\t// \t}\n\t// }\n\t// // No meta characters found, so return original string.\n\t// if i >= len(s) {\n\t// \treturn s\n\t// }\n\n\t// b := make([]byte, 2*len(s)-i)\n\t// copy(b, s[:i])\n\t// j := i\n\t// for ; i < len(s); i++ {\n\t// \tif special(s[i]) {\n\t// \t\tb[j] = '\\\\'\n\t// \t\tj++\n\t// \t}\n\t// \tb[j] = s[i]\n\t// \tj++\n\t// }\n\t// return string(b[:j])\n\tpanic(\"\")\n}", "func TransformRegExp(pattern string) (string, error) {\n\n\tlog.Printf(\"runtime.Caller(1): %#+v\\n\", myutils.Slice(runtime.Caller(1))...)\n\tos.Exit(349857394)\n\n\t// if pattern == \"\" {\n\t// \treturn \"\", nil\n\t// }\n\n\tswitch pattern {\n\t// common cases that will leave as is\n\t// but frequently occurred in many\n\t// popular frameworks like react.js etc.\n\tcase _regexp_empty,\n\t\t_regexp_any_nc,\n\t\t_regexp_any_c,\n\t\t_regexp_anyplus_nc,\n\t\t_regexp_anyplus_c,\n\n\t\t_regexp_squote_nc,\n\t\t_regexp_squote_c,\n\t\t_regexp_dquote_nc,\n\t\t_regexp_dquote_c,\n\n\t\t_regexp_plus_nc,\n\t\t_regexp_plus_c,\n\n\t\t_regexp_urlenc_space_nc,\n\t\t_regexp_urlenc_space_c,\n\t\t_regexp_http_validation,\n\n\t\t_regexp_any_uppercase_nc,\n\t\t_regexp_any_uppercase_c,\n\n\t\t_regexp_is_ms_pref:\n\t\treturn pattern, nil\n\t}\n\n\t// TODO If without \\, if without (?=, (?!, then another shortcut\n\n\tparser := _RegExp_parser{\n\t\tstr: pattern,\n\t\tlength: len(pattern),\n\t\tgoRegexp: bytes.NewBuffer(make([]byte, 0, 3*len(pattern)/2)),\n\t}\n\tparser.read() // Pull in the first character\n\tparser.scan()\n\tvar err error\n\tif len(parser.errors) > 0 {\n\t\terr = parser.errors[0]\n\t}\n\tif parser.invalid {\n\t\tlog.Printf(\"Input: regexp=[%s]\\n\", pattern)\n\t\tlog.Printf(\"Output: regexp=[%s] err=[%s]\\n\", parser.goRegexp.String(), err)\n\t\treturn \"\", err\n\t}\n\n\t// Might not be re2 compatible, but is still a valid JavaScript RegExp\n\treturn parser.goRegexp.String(), err\n}", "func quote(s string) string {\n\tquoted := strings.Builder{}\n\tquoted.WriteRune('\"')\n\tfor _, c := range s {\n\t\tif strings.ContainsRune(charsToQuote, c) {\n\t\t\tquoted.WriteRune('\\\\')\n\t\t}\n\t\tquoted.WriteRune(c)\n\t}\n\tquoted.WriteRune('\"')\n\treturn quoted.String()\n}", "func regex(s string) (*regexp.Regexp, error) {\n\tif rawString.MatchString(s) {\n\t\ts = fmt.Sprintf(\"^%s$\", s)\n\t}\n\treturn regexp.Compile(s)\n}", "func extractAmavisPCREExpression(content string) string{\n patterns := []string{`(?m)^qr'(?P<result>.*)'\\w+$`,\n `(?m)^qr'(?P<result>.*)'$`}\n result := []byte{}\n for i := 0; (i<len(patterns) && string(result) == \"\"); i++ {\n pattern := regexp.MustCompile(patterns[i])\n if pattern.MatchString(content){\n template := \"$result\"\n for _, submatches := range pattern.FindAllStringSubmatchIndex(content, -1) {\n // Apply the captured submatches to the template and append the output\n // to the result.\n result = pattern.ExpandString(result, template, content, submatches)\n }\n }\n }\n return string(result)\n}", "func LiteralQuoteEscape(quote rune, literal string) string {\n\tif len(literal) > 1 {\n\t\tquoteb := byte(quote)\n\t\tif literal[0] == quoteb && literal[len(literal)-1] == quoteb {\n\t\t\t// Already escaped??\n\t\t\treturn literal\n\t\t}\n\t}\n\tvar buf bytes.Buffer\n\tLiteralQuoteEscapeBuf(&buf, quote, literal)\n\treturn buf.String()\n}", "func (v *VerbalExpression) Regex() *regexp.Regexp {\n\n\tif !v.compiled {\n\t\tv.regexp = regexp.MustCompile(\n\t\t\tstrings.Join([]string{\n\t\t\t\tstrings.Join(v.parts, \"\"),\n\t\t\t\t`(?` + v.getFlags() + `)`,\n\t\t\t\tv.prefixes,\n\t\t\t\tv.expression,\n\t\t\t\tv.suffixes}, \"\"))\n\t\tv.compiled = true\n\t}\n\treturn v.regexp\n}", "func Quote(s string) Node {\n\tends := `\"`\n\tswitch {\n\tcase !strings.Contains(s, `\"`):\n\t\tends = `\"`\n\tcase !strings.Contains(s, `'`):\n\t\tends = `'`\n\tcase !strings.Contains(s, \"`\"):\n\t\tends = \"`\"\n\tdefault:\n\t\ts = strings.ReplaceAll(s, `\"`, `\\\"`)\n\t}\n\n\treturn Node{ast.Quote{Val: ends + s + ends}}\n}", "func (s *SymbolStore) rpathEscaped(rpath, basepath string) []string {\n\tbasedir := filepath.Dir(basepath)\n\tif strings.Contains(rpath, \"$ORIGIN\") {\n\t\trpath = strings.Replace(rpath, \"$ORIGIN\", basedir, -1)\n\t}\n\n\tret := []string{rpath}\n\n\tif strings.Contains(rpath, \"$LIB\") {\n\t\tfor _, l := range s.rlibDirs {\n\t\t\tret = append(ret, strings.Replace(rpath, \"$LIB\", l, -1))\n\t\t}\n\t}\n\treturn ret\n}", "func (r *Renderer) Escaped(txt string) {\n\tr.printf(\"%s\", html.EscapeString(txt))\n}", "func MaskToRegex(rawMask string) (*regexp.Regexp, error) {\n\tinput := bytes.NewBufferString(rawMask)\n\n\toutput := &bytes.Buffer{}\n\toutput.WriteByte('^')\n\n\tfor {\n\t\tc, err := input.ReadByte()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\n\t\tif c == '\\\\' {\n\t\t\tc, err = input.ReadByte()\n\t\t\tif err != nil {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(\"\\\\\"))\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif c == '?' || c == '*' || c == '\\\\' {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(string(c)))\n\t\t\t} else {\n\t\t\t\toutput.WriteString(regexp.QuoteMeta(\"\\\\\" + string(c)))\n\t\t\t}\n\t\t} else if trans, ok := maskTranslations[c]; ok {\n\t\t\toutput.WriteString(trans)\n\t\t} else {\n\t\t\toutput.WriteString(regexp.QuoteMeta(string(c)))\n\t\t}\n\t}\n\n\toutput.WriteByte('$')\n\n\treturn regexp.Compile(output.String())\n}", "func doubleQuoteEscape(line string) string {\n\tconst doubleQuoteSpecialChars = \"\\\\\\n\\r\\\"!$`\"\n\tfor _, c := range doubleQuoteSpecialChars {\n\t\ttoReplace := \"\\\\\" + string(c)\n\t\tif c == '\\n' {\n\t\t\ttoReplace = `\\n`\n\t\t}\n\t\tif c == '\\r' {\n\t\t\ttoReplace = `\\r`\n\t\t}\n\t\tline = strings.ReplaceAll(line, string(c), toReplace)\n\t}\n\treturn line\n}", "func (p *Parser) quote() *Expr {\n\treturn Cons(atomExpr(tokQuote), Cons(p.List(), nil))\n}", "func (b *bot) RegisterRegexCmd(p Plugin, kind Kind, r *regexp.Regexp, resp ResponseHandler) {\n\tnewResp := func(req Request) bool {\n\t\tif !req.Msg.Command {\n\t\t\treturn false\n\t\t}\n\t\treturn resp(req)\n\t}\n\tb.RegisterRegex(p, kind, r, newResp)\n}", "func createGlobbingRegex(globbing string) *regexp.Regexp {\n\tif globbing == \"\" {\n\t\t// nil here as \"\" is fast-tracked elsewhere\n\t\treturn nil\n\t}\n\n\tg := regexp.QuoteMeta(globbing)\n\tg = strings.ReplaceAll(g, \"\\\\*\", \".*\")\n\tg = strings.ReplaceAll(g, \"\\\\?\", \".\")\n\t// (?i) forces case insensitive matches\n\tg = \"(?i)^\" + g + \"$\"\n\treturn regexp.MustCompile(g)\n}", "func (i *Info) GetSelStringRegex() string {\n\n\tvar qOps string\n\tops := i.GetQueryOps()\n\tfor _, op := range ops {\n\t\tswitch op {\n\t\tcase \"EQ\":\n\t\t\tqOps = qOps + \"EQ|eq|\"\n\t\tcase \"NE\":\n\t\t\tqOps = qOps + \"NE|ne|\"\n\t\tcase \"LIKE\":\n\t\t\tqOps = qOps + \"LIKE|like|\"\n\t\tcase \"LT\":\n\t\t\tqOps = qOps + \"LT|lt|\"\n\t\tcase \"LE\":\n\t\t\tqOps = qOps + \"LE|le|\"\n\t\tcase \"GT\":\n\t\t\tqOps = qOps + \"GT|gt|\"\n\t\tcase \"GE\":\n\t\t\tqOps = qOps + \"GE|ge|\"\n\t\tdefault:\n\t\t\t// ignore all others\n\t\t}\n\t}\n\tif len(qOps) > 0 {\n\t\tqOps = superCleanString(qOps)\n\t\tqOps = strings.TrimSuffix(qOps, \"|\")\n\t\t// [(]+(?:EQ|eq|NE|ne|LIKE|like)+[ ']+[a-zA-Z0-9_]+[')]+\n\t\tqOps = \"[(]+(?:\" + qOps + \")+[ ']+[a-zA-Z0-9_]+[')]+\"\n\t\treturn qOps\n\t}\n\treturn \"\"\n}", "func RegexpString(re *syntax.Regexp,) string", "func Regexp(pattern string) (*regexp.Regexp, error) {\n\treturn pool.Compile(pattern)\n}", "func GetQueryRe(query string) (*regexp.Regexp, error) {\n\treturn regexp.Compile(\"(?i)\" + regexp.QuoteMeta(query))\n}", "func escape(token string) string {\n\ttoken = strings.Replace(token, \"~\", \"~0\", -1)\n\ttoken = strings.Replace(token, \"/\", \"~1\", -1)\n\treturn url.PathEscape(token)\n}", "func CanBackquote(s string) bool {}", "func (c *Client) Quote(v interface{}) Field {\n\tvar b []byte\n\tvar cp bool\n\tswitch s := v.(type) {\n\tcase string:\n\t\tb = []byte(s)\n\tcase []byte:\n\t\tb, cp = s, true\n\tcase fmt.Stringer:\n\t\tb = []byte(s.String())\n\tdefault:\n\t\treturn nil\n\t}\n\tif q := QuoteBytes(b, false); q != nil {\n\t\treturn string(q)\n\t} else if cp {\n\t\tb = append([]byte(nil), b...)\n\t}\n\treturn NewLiteral(b)\n}", "func lexQuote(l *lexer) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.next(); r != eof && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated quoted string\")\n\t\tcase '\"':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(itemString)\n\treturn lexInsideAction\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func expression(res ...*regexp.Regexp) *regexp.Regexp {\n\tvar s string\n\tfor _, re := range res {\n\t\ts += re.String()\n\t}\n\n\treturn match(s)\n}", "func regExp(context interface{}, value string) (bson.RegEx, error) {\n\tidx := strings.IndexByte(value[1:], '/')\n\tif value[0] != '/' || idx == -1 {\n\t\terr := fmt.Errorf(\"Parameter %q is not a regular expression\", value)\n\t\tlog.Error(context, \"varLookup\", err, \"Regex parsing\")\n\t\treturn bson.RegEx{}, err\n\t}\n\n\tpattern := value[1 : idx+1]\n\tl := len(pattern) + 2\n\n\tvar options string\n\tif l < len(value) {\n\t\toptions = value[l:]\n\t}\n\n\treturn bson.RegEx{Pattern: pattern, Options: options}, nil\n}", "func (i *Info) GetSelBoolRegex() string {\n\n\tvar qOps string\n\tops := i.GetQueryOps()\n\tfor _, op := range ops {\n\t\tswitch op {\n\t\tcase \"EQ\":\n\t\t\tqOps = qOps + \"EQ|eq|\"\n\t\tcase \"NE\":\n\t\t\tqOps = qOps + \"NE|ne|\"\n\t\tdefault:\n\t\t\t// ignore all others\n\t\t}\n\t}\n\tif len(qOps) > 0 {\n\t\tqOps = superCleanString(qOps)\n\t\tqOps = strings.TrimSuffix(qOps, \"|\")\n\t\t// [(]+(?:EQ|eq)+[ ']+(?:true|TRUE|false|FALSE)+[')]+\n\t\tqOps = \"[(]+(?:\" + qOps + \")+[ ']+(?:true|TRUE|false|FALSE)+[')]+\"\n\t\treturn qOps\n\t}\n\treturn \"\"\n}", "func StructuralPatToRegexpQuery(pattern string, shortcircuit bool) (zoektquery.Q, error) {\n\tsubstrings := matchHoleRegexp.Split(pattern, -1)\n\tvar children []zoektquery.Q\n\tvar pieces []string\n\tfor _, s := range substrings {\n\t\tpiece := regexp.QuoteMeta(s)\n\t\tonMatchWhitespace := lazyregexp.New(`[\\s]+`)\n\t\tpiece = onMatchWhitespace.ReplaceAllLiteralString(piece, `[\\s]+`)\n\t\tpieces = append(pieces, piece)\n\t}\n\n\tif len(pieces) == 0 {\n\t\treturn &zoektquery.Const{Value: true}, nil\n\t}\n\tvar rs string\n\tif shortcircuit {\n\t\t// As a shortcircuit, do not match across newlines of structural search pieces.\n\t\trs = \"(\" + strings.Join(pieces, \").*?(\") + \")\"\n\t} else {\n\t\trs = \"(\" + strings.Join(pieces, \")(.|\\\\s)*?(\") + \")\"\n\t}\n\tre, _ := syntax.Parse(rs, syntax.ClassNL|syntax.PerlX|syntax.UnicodeGroups)\n\tchildren = append(children, &zoektquery.Regexp{\n\t\tRegexp: re,\n\t\tCaseSensitive: true,\n\t\tContent: true,\n\t})\n\treturn &zoektquery.And{Children: children}, nil\n}", "func quote(str string, quote byte) string {\n\tif strings.IndexByte(str, '\\n') < 0 {\n\t\tbuf := []byte{quote}\n\t\tbuf = appendEscaped(buf, str, quote, true)\n\t\tbuf = append(buf, quote)\n\t\treturn string(buf)\n\t}\n\tbuf := []byte{quote, quote, quote}\n\tbuf = append(buf, multiSep...)\n\tbuf = appendEscapeMulti(buf, str, quote)\n\tbuf = append(buf, quote, quote, quote)\n\treturn string(buf)\n}", "func ShellEscape(input string) string {\n\tif input == \"\" {\n\t\treturn \"''\"\n\t}\n\n\tvar sb strings.Builder\n\tsb.Grow(len(input) * 2)\n\n\tescape := false\n\tfor _, c := range []byte(input) {\n\t\tmode := modeTable[c]\n\t\tswitch mode {\n\t\tcase lit:\n\t\t\tsb.WriteByte(c)\n\t\tcase quo:\n\t\t\tsb.WriteByte(c)\n\t\t\tescape = true\n\t\tcase \"\":\n\t\t\tsb.Write([]byte{'\\\\', 'x', hextable[c>>4], hextable[c&0x0f]})\n\t\t\tescape = true\n\t\tdefault:\n\t\t\tsb.WriteString(string(mode))\n\t\t\tescape = true\n\t\t}\n\t}\n\n\tif escape {\n\t\treturn \"$'\" + sb.String() + \"'\"\n\t}\n\n\treturn sb.String()\n}", "func Regexp(expr *regexp.Regexp) Pattern {\n\treturn regexpMatch{expr}\n}", "func (n NamespaceStar) Quote() string { return `'` + n.Value() + `'` }", "func tokenize(infix string) []interface{} {\n\n\tvar tokens []interface{}\n\n\tvar s string\n\tvar escapeStr string\n\tappendToS := false\n\twantsToEscape := false\n\tnegate := false\n\tignoreCase := len(infix) >= 4 && strings.HasPrefix(infix, \"(?i)\")\n\t// all an input of \"(?i)\" + regex which will make it case insensitive.\n\t// put the infix string to lower case to eliminate differences between cases.\n\tif ignoreCase {\n\t\tinfix = infix[4:] // remove the (?i) on the string\n\t}\n\n\tfor i, r := range infix {\n\n\t\tif wantsToEscape { // there was a backslash escape the character\n\t\t\tescapeStr += string(r)\n\t\t\twantsToEscape = false\n\t\t\tswitch r {\n\t\t\tcase 'd': // \\d\n\t\t\t\ttokens = append(tokens, digitToken{val: `\\d`, negate: negate})\n\t\t\tcase 'w': // \\w\n\t\t\t\ttokens = append(tokens, wordToken{val: `\\w`, negate: negate})\n\t\t\tcase 's':\n\t\t\t\ttokens = append(tokens, spaceToken{val: `\\s`, negate: negate})\n\t\t\tdefault: // it's an escaped character\n\t\t\t\ttokens = append(tokens, characterClassToken{val: string(r), negate: negate, caseInsensitive: ignoreCase})\n\t\t\t}\n\t\t\tnegate = false\n\n\t\t\t// handle implicit concatenation\n\t\t\tatEnd := i == len(infix)-1\n\t\t\tif !atEnd && !isExplicitOperator(infix[i+1]) && !isClosingBracket(infix[i+1]) {\n\t\t\t\tif !isOpeningBracket(r) && r != '|' {\n\t\t\t\t\ttokens = append(tokens, characterClassToken{val: \".\", negate: false})\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tcontinue\n\t\t}\n\n\t\tif r == '\\\\' { // want to escape a character\n\t\t\twantsToEscape = true\n\t\t\tescapeStr = \"\"\n\t\t\tcontinue\n\t\t}\n\n\t\t// if we should invert the match\n\t\tif r == '^' {\n\t\t\tnegate = true\n\t\t\tcontinue\n\t\t}\n\n\t\t// if it's an underscore (and isn't being escaped), add a token that will match any character\n\t\tif r == '_' {\n\t\t\ttokens = append(tokens, anyToken{val: \"_\", negate: negate})\n\n\t\t\t// handle implicit concatenation\n\t\t\tatEnd := i == len(infix)-1\n\t\t\tif !atEnd && !isExplicitOperator(infix[i+1]) && !isClosingBracket(infix[i+1]) {\n\t\t\t\tif !isOpeningBracket(r) && r != '|' {\n\t\t\t\t\ttokens = append(tokens, characterClassToken{val: \".\", negate: false})\n\t\t\t\t}\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tstartingClass, endingClass := r == '[', r == ']'\n\n\t\t// don't want to append the last element in a token\n\t\tif appendToS && !endingClass {\n\t\t\ts += string(r) // add as a single character of a multi character token\n\t\t} else if !startingClass && !endingClass { // add the single character as a token\n\t\t\ttokens = append(tokens, characterClassToken{string(r), negate, ignoreCase})\n\n\t\t\t// handle implicit concatenation\n\t\t\tatEnd := i == len(infix)-1\n\t\t\tif !atEnd && !isExplicitOperator(infix[i+1]) && !isClosingBracket(infix[i+1]) {\n\t\t\t\tif !isOpeningBracket(r) && r != '|' {\n\t\t\t\t\ttokens = append(tokens, characterClassToken{val: \".\", negate: false})\n\t\t\t\t}\n\t\t\t}\n\t\t\tnegate = false\n\t\t}\n\n\t\tif startingClass { // we're going to start a multi character token\n\t\t\tappendToS = true\n\t\t} else if endingClass { // reached end of character class\n\t\t\ttokens = append(tokens, characterClassToken{val: s, negate: negate}) // add the full string as a single token\n\t\t\tnegate = false\n\n\t\t\tatEnd := i == len(infix)-1\n\t\t\tif !atEnd && !isExplicitOperator(infix[i+1]) {\n\t\t\t\t// only don't add implicit concat if the next character is an explicit operator\n\t\t\t\ttokens = append(tokens, characterClassToken{\".\", false, ignoreCase})\n\t\t\t}\n\n\t\t\ts = \"\"\n\t\t\tappendToS = false // stop building up character class\n\n\t\t}\n\t}\n\n\treturn tokens\n}", "func expandRegex(r *regexp.Regexp, tmpl string, src string) string {\n\tvar res []byte\n\tfor _, submatches := range r.FindAllStringSubmatchIndex(src, 1) {\n\t\tres = r.ExpandString(res, tmpl, src, submatches)\n\t}\n\treturn string(res)\n}", "func QuoteRune(r rune) string {}", "func DummyQueryEscape(s string) string {\n\treturn s\n}", "func Regex(expr ConstString, input String) (Bool, error) {\n\tif expr.HasVariable() {\n\t\treturn nil, fmt.Errorf(\"regex requires a constant expression as its first parameter, but it has a variable parameter\")\n\t}\n\te, err := expr.Eval()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr, err := regexp.Compile(e)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn TrimBool(&regex{\n\t\tr: r,\n\t\tS: input,\n\t\texpr: e,\n\t\thash: Hash(\"regex\", expr, input),\n\t\thasVariable: input.HasVariable(),\n\t}), nil\n}", "func NewQuotes(market *Market, profile *Profile) *Quotes {\n\tre, err := regexp.Compile(\"_\\\\w+\")\n\tif err != nil {\n\t\t//TODO\n\t\tfmt.Print(\"error when create regex\")\n\t}\n\treturn &Quotes{\n\t\tmarket: market,\n\t\tprofile: profile,\n\t\tre: re,\n\t\terrors: ``,\n\t}\n}", "func NewRegex(v string) (Value, error) {\n\trx, err := regexp.Compile(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rxValue{rx}, nil\n}", "func backslashEscape(s string, set []byte) string {\n\tvar buf bytes.Buffer\n\tfor _, b := range []byte(s) {\n\t\tif b == '\\\\' || bytes.IndexByte(set, b) != -1 {\n\t\t\tbuf.WriteByte('\\\\')\n\t\t}\n\t\tbuf.WriteByte(b)\n\t}\n\treturn buf.String()\n}", "func Quote(b []byte, mark byte) []byte {\n\tfor i, j := 0, len(b); i < j; i++ {\n\t\tif b[i] == mark || b[i] == '\\\\' {\n\t\t\tb = append(b, 0)\n\t\t\tcopy(b[i+1:], b[i:])\n\t\t\tb[i] = '\\\\'\n\t\t\ti++\n\t\t\tj++\n\t\t}\n\t}\n\tb = append(b, 0, 0)\n\tcopy(b[1:], b)\n\tb[0] = mark\n\tb[len(b)-1] = mark\n\treturn b\n}", "func Globexp(glob string) *regexp.Regexp {\n\tvar re bytes.Buffer\n\n\tre.WriteString(\"^\")\n\n\ti, inGroup, L := 0, false, len(glob)\n\n\tfor i < L {\n\t\tr, w := utf8.DecodeRuneInString(glob[i:])\n\n\t\tswitch r {\n\t\tdefault:\n\t\t\tre.WriteRune(r)\n\n\t\tcase '\\\\', '$', '^', '+', '.', '(', ')', '=', '!', '|':\n\t\t\tre.WriteRune('\\\\')\n\t\t\tre.WriteRune(r)\n\n\t\tcase '/':\n\t\t\t// TODO optimize later, string could be long\n\t\t\trest := glob[i:]\n\t\t\tre.WriteRune('/')\n\t\t\tif strings.HasPrefix(rest, \"/**/\") {\n\t\t\t\tre.WriteString(zeroOrMoreDirectories)\n\t\t\t\tw *= 4\n\t\t\t} else if rest == \"/**\" {\n\t\t\t\tre.WriteString(\".*\")\n\t\t\t\tw *= 3\n\t\t\t}\n\n\t\tcase '?':\n\t\t\tre.WriteRune('.')\n\n\t\tcase '[', ']':\n\t\t\tre.WriteRune(r)\n\n\t\tcase '{':\n\t\t\tif i < L-1 {\n\t\t\t\tif glob[i+1:i+2] == \"{\" {\n\t\t\t\t\tre.WriteString(\"\\\\{\")\n\t\t\t\t\tw *= 2\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tinGroup = true\n\t\t\tre.WriteRune('(')\n\n\t\tcase '}':\n\t\t\tif inGroup {\n\t\t\t\tinGroup = false\n\t\t\t\tre.WriteRune(')')\n\t\t\t} else {\n\t\t\t\tre.WriteRune('}')\n\t\t\t}\n\n\t\tcase ',':\n\t\t\tif inGroup {\n\t\t\t\tre.WriteRune('|')\n\t\t\t} else {\n\t\t\t\tre.WriteRune('\\\\')\n\t\t\t\tre.WriteRune(r)\n\t\t\t}\n\n\t\tcase '*':\n\t\t\trest := glob[i:]\n\t\t\tif strings.HasPrefix(rest, \"**/\") {\n\t\t\t\tre.WriteString(zeroOrMoreDirectories)\n\t\t\t\tw *= 3\n\t\t\t} else {\n\t\t\t\tre.WriteString(anyRune)\n\t\t\t}\n\t\t}\n\n\t\ti += w\n\t}\n\n\tre.WriteString(\"$\")\n\t//log.Printf(\"regex string %s\", re.String())\n\treturn regexp.MustCompile(re.String())\n}", "func compileRegex() {\n\tfor _, value := range filterObj.Signatures {\n\t\tr, _ := regexp.Compile(value.Regex)\n\t\tsignatureRegex = append(signatureRegex, r)\n\t}\n\tfor _, value := range filterObj.Shallows {\n\t\tr, _ := regexp.Compile(value.Regex)\n\t\tshallowRegex = append(shallowRegex, r)\n\t}\n}", "func (p *Parser) lexQuote(l *lex.Lexer) lex.StateFn {\n\t// lexQuote is called for ', \", and `.\n\tif l.Next() != '\"' {\n\t\treturn l.Errorf(\"only support double-quoted strings\")\n\t}\n\tl.Ignore()\n\nloop:\n\tfor {\n\t\tswitch l.Next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.Next(); r != lex.EOF && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase lex.EOF, '\\n':\n\t\t\treturn l.Errorf(\"unterminated quoted string\")\n\t\tcase '\"':\n\t\t\tbreak loop\n\t\t}\n\t}\n\tl.Dec(1)\n\tl.Emit(typeString)\n\tl.Inc(1)\n\tl.Ignore()\n\treturn p.lexInsideAction\n}", "func lexQuote(l *lexer) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.next(); r != token.Eof && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase token.Eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated quoted string\")\n\t\tcase '\"':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(token.ItemString)\n\treturn lexInsideList\n}", "func (g *Generator) escape() string {\n\treturn html.EscapeString(g.Input)\n}", "func (p *parser) regex() Node {\n\ttoken := p.expect(TokenRegex)\n\tr, err := newRegex(token.pos, token.val)\n\tif err != nil {\n\t\tp.error(err)\n\t}\n\treturn r\n}", "func RegExp(pattern string) *RegExpMatcher {\n\treturn &RegExpMatcher{Pattern: pattern}\n}", "func tcEscape(input string) string {\n\tvar buf bytes.Buffer\n\tfor {\n\t\tnextSpecial := strings.IndexAny(input, \"'\\n\\r|[]\")\n\t\tswitch nextSpecial {\n\t\tcase -1:\n\t\t\tif buf.Len() == 0 {\n\t\t\t\treturn input\n\t\t\t}\n\t\t\treturn buf.String() + input\n\t\tcase 0:\n\t\tdefault:\n\t\t\tbuf.WriteString(input[:nextSpecial])\n\t\t}\n\t\tswitch input[nextSpecial] {\n\t\tcase '\\'':\n\t\t\tbuf.WriteString(\"|'\")\n\t\tcase '\\n':\n\t\t\tbuf.WriteString(\"|n\")\n\t\tcase '\\r':\n\t\t\tbuf.WriteString(\"|r\")\n\t\tcase '|':\n\t\t\tbuf.WriteString(\"||\")\n\t\tcase '[':\n\t\t\tbuf.WriteString(\"|[\")\n\t\tcase ']':\n\t\t\tbuf.WriteString(\"|]\")\n\t\t}\n\t\tinput = input[nextSpecial+1:]\n\t}\n}", "func appendQuote(dest []byte, s string, escapeHTML bool) []byte {\n\tdest = append(dest, '\"')\n\tstart := 0\n\tfor i := 0; i < len(s); {\n\t\tif b := s[i]; b < utf8.RuneSelf {\n\t\t\tif htmlSafeSet[b] || (!escapeHTML && safeSet[b]) {\n\t\t\t\ti++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif start < i {\n\t\t\t\tdest = append(dest, s[start:i]...)\n\t\t\t}\n\t\t\tdest = append(dest, '\\\\')\n\t\t\tswitch b {\n\t\t\tcase '\\\\', '\"':\n\t\t\t\tdest = append(dest, b)\n\t\t\tcase '\\n':\n\t\t\t\tdest = append(dest, 'n')\n\t\t\tcase '\\r':\n\t\t\t\tdest = append(dest, 'r')\n\t\t\tcase '\\t':\n\t\t\t\tdest = append(dest, 't')\n\t\t\tdefault:\n\t\t\t\t// This encodes bytes < 0x20 except for \\t, \\n and \\r.\n\t\t\t\t// If escapeHTML is set, it also escapes <, >, and &\n\t\t\t\t// because they can lead to security holes when\n\t\t\t\t// user-controlled strings are rendered into JSON\n\t\t\t\t// and served to some browsers.\n\t\t\t\tdest = append(dest, []byte(`u00`)...)\n\t\t\t\tdest = append(dest, hex[b>>4])\n\t\t\t\tdest = append(dest, hex[b&0xF])\n\t\t\t}\n\t\t\ti++\n\t\t\tstart = i\n\t\t\tcontinue\n\t\t}\n\t\tc, size := utf8.DecodeRuneInString(s[i:])\n\t\tif c == utf8.RuneError && size == 1 {\n\t\t\tif start < i {\n\t\t\t\tdest = append(dest, s[start:i]...)\n\t\t\t}\n\t\t\tdest = append(dest, []byte(`\\ufffd`)...)\n\t\t\ti += size\n\t\t\tstart = i\n\t\t\tcontinue\n\t\t}\n\t\t// U+2028 is LINE SEPARATOR.\n\t\t// U+2029 is PARAGRAPH SEPARATOR.\n\t\t// They are both technically valid characters in JSON strings,\n\t\t// but don't work in JSONP, which has to be evaluated as JavaScript,\n\t\t// and can lead to security holes there. It is valid JSON to\n\t\t// escape them, so we do so unconditionally.\n\t\t// See http://timelessrepo.com/json-isnt-a-javascript-subset for discussion.\n\t\tif c == '\\u2028' || c == '\\u2029' {\n\t\t\tif start < i {\n\t\t\t\tdest = append(dest, s[start:i]...)\n\t\t\t}\n\t\t\tdest = append(dest, []byte(`\\u202`)...)\n\t\t\tdest = append(dest, hex[c&0xF])\n\t\t\ti += size\n\t\t\tstart = i\n\t\t\tcontinue\n\t\t}\n\t\ti += size\n\t}\n\tif start < len(s) {\n\t\tdest = append(dest, s[start:]...)\n\t}\n\tdest = append(dest, '\"')\n\treturn dest\n}", "func appendShellSafeQuote(buf []byte, s string) []byte {\n\tbuf = append(buf, '\"')\n\tfor width := 0; len(s) > 0; s = s[width:] {\n\t\tr := rune(s[0])\n\t\twidth = 1\n\t\tif r >= utf8.RuneSelf {\n\t\t\tr, width = utf8.DecodeRuneInString(s)\n\t\t}\n\t\tif width == 1 {\n\t\t\tswitch r {\n\t\t\tcase utf8.RuneError:\n\t\t\t\tbuf = append(buf, '\\\\')\n\t\t\t\tbuf = append(buf, '0'+s[0]>>6)\n\t\t\t\tbuf = append(buf, '0'+((s[0]>>3)&7))\n\t\t\t\tbuf = append(buf, '0'+(s[0]&7))\n\t\t\t// Stuff which should be escaped\n\t\t\tcase '\\\\':\n\t\t\t\tbuf = append(buf, `\\\\`...)\n\t\t\tcase '\"':\n\t\t\t\tbuf = append(buf, `\\\"`...)\n\t\t\tcase '$':\n\t\t\t\tbuf = append(buf, `\\$`...)\n\t\t\tdefault:\n\t\t\t\tbuf = append(buf, byte(r))\n\t\t\t}\n\t\t} else {\n\t\t\tvar runeTmp [utf8.UTFMax]byte\n\t\t\tn := utf8.EncodeRune(runeTmp[:], r)\n\t\t\tbuf = append(buf, runeTmp[:n]...)\n\t\t}\n\t}\n\treturn append(buf, '\"')\n}", "func TestRsyncEscapeCommand(t *testing.T) {\n\t// the strings to run through rsyncEscapeCommand\n\tstringsToEscape := []string{\n\t\t`thisshouldnotgetescapedorquoted`,\n\t\t`this should get quoted for spaces`,\n\t\t`this\" should get escaped and quoted`,\n\t\t`\"this should get escaped and quoted\"`,\n\t\t`this\\ should get quoted`,\n\t\t`this' should get quoted`,\n\t}\n\t// this is how the strings should be escaped by rsyncEscapeCommand\n\tstringsShouldMatch := []string{\n\t\t`thisshouldnotgetescapedorquoted`,\n\t\t`\"this should get quoted for spaces\"`,\n\t\t`\"this\"\" should get escaped and quoted\"`,\n\t\t`\"\"\"this should get escaped and quoted\"\"\"`,\n\t\t`\"this\\ should get quoted\"`,\n\t\t`\"this' should get quoted\"`,\n\t}\n\n\tescapedStrings := rsyncEscapeCommand(stringsToEscape)\n\n\tfor key, val := range escapedStrings {\n\t\tif val != stringsShouldMatch[key] {\n\t\t\tt.Errorf(\"%v did not match %v\", val, stringsShouldMatch[key])\n\t\t}\n\t}\n}", "func (regex *Regex) String() string {\n\treturn regex.raw\n}", "func (i *Info) GetSelNumberRegex() string {\n\n\tvar qOps string\n\tops := i.GetQueryOps()\n\tfor _, op := range ops {\n\t\tswitch op {\n\t\tcase \"EQ\":\n\t\t\tqOps = qOps + \"EQ|eq|\"\n\t\tcase \"NE\":\n\t\t\tqOps = qOps + \"NE|ne|\"\n\t\tcase \"LT\":\n\t\t\tqOps = qOps + \"LT|lt|\"\n\t\tcase \"LE\":\n\t\t\tqOps = qOps + \"LE|le|\"\n\t\tcase \"GT\":\n\t\t\tqOps = qOps + \"GT|gt|\"\n\t\tcase \"GE\":\n\t\t\tqOps = qOps + \"GE|ge|\"\n\t\tdefault:\n\t\t\t// ignore all others\n\t\t}\n\t}\n\tif len(qOps) > 0 {\n\t\tqOps = superCleanString(qOps)\n\t\tqOps = strings.TrimSuffix(qOps, \"|\")\n\t\t// [(]+(?:EQ|eq|LT|lt|LE|le|GT|gt|GE|ge)+[ ]+[0-9._]+[)]+\n\n\t\tif i.IsFloatFieldType() {\n\t\t\tqOps = \"[(]+(?:\" + qOps + \")+[ ]+[0-9._]+[)]+\"\n\t\t\treturn qOps\n\t\t}\n\n\t\tif i.IsUIntFieldType() {\n\t\t\tqOps = \"[(]+(?:\" + qOps + \")+[ ]+[0-9_]+[)]+\"\n\t\t\treturn qOps\n\t\t}\n\n\t\tif i.IsIntFieldType() {\n\t\t\tqOps = \"[(]+(?:\" + qOps + \")+[ ]+[0-9_-]+[)]+\"\n\t\t\treturn qOps\n\t\t}\n\t}\n\treturn \"\"\n}", "func getMagicQuotesGpc(ctx phpv.Context, args []*phpv.ZVal) (*phpv.ZVal, error) {\n\treturn phpv.ZBool(false).ZVal(), nil\n}", "func Escape(target string, query string) string {\n\treturn strings.Replace(target, query, fmt.Sprintf(\"\\\\%s\", query), -1)\n}", "func escape(s interface{}) string {\n\tif s, ok := s.(string); ok {\n\t\t// s = strings.Replace(s, `\"`, `'`, -1)\n\t\t// s = strings.Replace(s, `\\`, `\\\\`, -1)\n\t\ts = strings.Replace(s, `\\.`, `.`, -1)\n\t\treturn s\n\t}\n\treturn fmt.Sprintf(\"%s\", s)\n}", "func Convert(pattern string) (*regexp.Regexp, error) {\n\tpattern = \"^\" + metaReplacer.Replace(regexp.QuoteMeta(pattern)) + \"$\"\n\treturn regexp.Compile(pattern)\n}", "func lexRawQuote(l *reader) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.next(); r != eof && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated quoted string\")\n\t\tcase '`':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(itemString)\n\treturn lexPunctuation\n}", "func lexQuote(l *lexer) stateFn {\n\tfor {\n\t\tswitch l.next() {\n\t\t//case '\\\\':\n\t\t//\tif r := l.next(); r != eof && r != '\\n' {\n\t\t//\t\tbreak\n\t\t//\t}\n\t\t//\tfallthrough\n\t\tcase eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated quoted string\")\n\t\tcase '\"':\n\t\t\tl.emit(itemStrLit)\n\t\t\treturn lexSchema\n\t\t}\n\t}\n}", "func Tokenize(s string) (e shared.Equation, err error) {\n e.Arguments = extractArguements(s)\n e.Operators = extractOperators(s)\n return\n}", "func quoteLiteral(s string) string {\n\tvar b strings.Builder\n\tb.Grow(len(s)*2 + 3)\n\n\tb.WriteRune('E')\n\tb.WriteRune('\\'')\n\n\thasSlash := false\n\tfor _, c := range s {\n\t\tif c == '\\\\' {\n\t\t\tb.WriteString(`\\\\`)\n\t\t\thasSlash = true\n\t\t} else if c == '\\'' {\n\t\t\tb.WriteString(`''`)\n\t\t} else {\n\t\t\tb.WriteRune(c)\n\t\t}\n\t}\n\n\tb.WriteRune('\\'')\n\n\ts = b.String()\n\tif !hasSlash {\n\t\t// remove unnecessary E at the beginning\n\t\treturn s[1:]\n\t}\n\treturn s\n}", "func Ror2QueryEscape(s string) string {\n\tbuf := new(strings.Builder)\n\tfor _, c := range []byte(s) {\n\t\tif c == ' ' {\n\t\t\thexEscape(buf, ' ')\n\t\t} else if _, ok := unescapedQueryCharacters[c]; ok {\n\t\t\tbuf.WriteByte(c)\n\t\t} else {\n\t\t\thexEscape(buf, c)\n\t\t}\n\t}\n\treturn buf.String()\n}", "func quote(dest *bytes.Buffer, bs []byte) {\n\tdest.WriteByte('\"')\n\tfor _, b := range bs {\n\t\tif b == '\\n' {\n\t\t\tdest.WriteString(`\\n`)\n\t\t\tcontinue\n\t\t}\n\t\tif b == '\\\\' {\n\t\t\tdest.WriteString(`\\\\`)\n\t\t\tcontinue\n\t\t}\n\t\tif b == '\"' {\n\t\t\tdest.WriteString(`\\\"`)\n\t\t\tcontinue\n\t\t}\n\t\tif (b >= 32 && b <= 126) || b == '\\t' {\n\t\t\tdest.WriteByte(b)\n\t\t\tcontinue\n\t\t}\n\t\tfmt.Fprintf(dest, \"\\\\x%02x\", b)\n\t}\n\tdest.WriteByte('\"')\n}", "func RegexReplace(scope *Scope, input tf.Output, pattern tf.Output, rewrite tf.Output, optional ...RegexReplaceAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"RegexReplace\",\n\t\tInput: []tf.Input{\n\t\t\tinput, pattern, rewrite,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func escape(s []byte) string {\n\ta := string(s)\n\tb := strconv.Quote(a)\n\tb = b[1 : len(b)-1]\n\tif a != b {\n\t\tb += \" (esc)\"\n\t}\n\treturn b\n}", "func lexQuote(l *reader) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase '\\\\':\n\t\t\tif r := l.next(); r != eof && r != '\\n' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tcase eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated quoted string\")\n\t\tcase '\"':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(itemString)\n\treturn lexPunctuation\n}", "func EscapeAllQuotes(s string) string {\n return strings.Replace(s, \"\\\"\", \"\\\\\\\"\", -1)\n}", "func lexRawQuote(l *lexer) stateFn {\nLoop:\n\tfor {\n\t\tswitch l.next() {\n\t\tcase eof, '\\n':\n\t\t\treturn l.errorf(\"unterminated raw quoted string\")\n\t\tcase '`':\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tl.emit(itemRawString)\n\treturn lexInsideAction\n}", "func (c *TransferCounter) CompileRegex(operator string) (*regexp.Regexp, error) {\n\tvar r *regexp.Regexp\n\tvar err error\n\n\tfor _, regionOperator := range regionOperators {\n\t\tif operator == regionOperator {\n\t\t\tr, err = regexp.Compile(\".*?operator finish.*?region-id=([0-9]*).*?\" + operator + \".*?store \\\\[([0-9]*)\\\\] to \\\\[([0-9]*)\\\\].*?\")\n\t\t}\n\t}\n\n\tfor _, leaderOperator := range leaderOperators {\n\t\tif operator == leaderOperator {\n\t\t\tr, err = regexp.Compile(\".*?operator finish.*?region-id=([0-9]*).*?\" + operator + \".*?store ([0-9]*) to ([0-9]*).*?\")\n\t\t}\n\t}\n\n\tif r == nil {\n\t\terr = errors.New(\"unsupported operator. \")\n\t}\n\treturn r, err\n}" ]
[ "0.64418626", "0.6150245", "0.5972613", "0.58816785", "0.57087374", "0.56048954", "0.53677803", "0.5344235", "0.5316042", "0.52762926", "0.52702194", "0.52489674", "0.52244294", "0.5212476", "0.5168041", "0.5153413", "0.51419055", "0.51419055", "0.51419055", "0.514183", "0.51043487", "0.5097339", "0.50573415", "0.50573415", "0.5031865", "0.5027085", "0.50255096", "0.5018072", "0.50115526", "0.49844447", "0.4958622", "0.4950498", "0.4922168", "0.4913996", "0.49119103", "0.48928663", "0.48626697", "0.4852796", "0.48475406", "0.48367205", "0.4830755", "0.48245454", "0.48137307", "0.48064333", "0.47981", "0.47838938", "0.4763014", "0.475221", "0.47500622", "0.47485295", "0.4745372", "0.4744368", "0.47367832", "0.47367832", "0.47367832", "0.47234756", "0.4703576", "0.47028562", "0.4691269", "0.46879753", "0.46856442", "0.4671631", "0.46700054", "0.4666477", "0.46531942", "0.4642766", "0.46415287", "0.46391085", "0.46350798", "0.461278", "0.4609492", "0.46071804", "0.46038976", "0.45950988", "0.45939007", "0.4584366", "0.45837322", "0.45795518", "0.45780528", "0.4571784", "0.45700437", "0.45696965", "0.45695186", "0.4560198", "0.45583746", "0.45547083", "0.4552843", "0.45446008", "0.45346767", "0.45342216", "0.4523233", "0.45111403", "0.45079708", "0.45072222", "0.45050192", "0.44960794", "0.44941646", "0.44792446", "0.44604045", "0.44511902" ]
0.8789912
0
/ Encoding Base64Encode uses base64.StdEncoding to encode operand.
func Base64Encode(operand string) string { return base64.StdEncoding.EncodeToString([]byte(operand)) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func base64Encode(value string) string {\n\treturn base64.StdEncoding.EncodeToString([]byte(value))\n}", "func runtimeEncodeBase64(ic *interop.Context) error {\n\tsrc := ic.VM.Estack().Pop().Bytes()\n\tresult := base64.StdEncoding.EncodeToString(src)\n\tic.VM.Estack().PushVal([]byte(result))\n\treturn nil\n}", "func Base64Encode(vm *otto.Otto) {\n\tvm.Set(\"base64encode\", func(call otto.FunctionCall) otto.Value {\n\t\ta0 := call.Argument(0)\n\t\tif !a0.IsString() {\n\t\t\tfmt.Println(\"ERROR\", \"base64encode(string)\")\n\t\t\treturn otto.Value{}\n\t\t}\n\t\ts, err := a0.ToString()\n\t\tif err != nil {\n\t\t\tfmt.Println(\"ERROR\", err)\n\t\t\treturn otto.Value{}\n\t\t}\n\t\ts = base64.StdEncoding.EncodeToString([]byte(s))\n\t\tv, err := vm.ToValue(s)\n\t\tif err != nil {\n\t\t\tfmt.Println(\"ERROR\", err)\n\t\t\treturn otto.Value{}\n\t\t}\n\t\treturn v\n\t})\n}", "func base64Encode(src interface{}) (string, error) {\n\tif src == nil {\n\t\treturn \"\", consts.ErrNilInterface\n\t}\n\tsrcMarshal, err := json.Marshal(src)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tsrcString := string(srcMarshal)\n\t// TODO maybe use Trim\n\treturn strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), \"=\"), nil\n}", "func Base64Encode(src string) string {\n\treturn base64.StdEncoding.EncodeToString([]byte(src))\n}", "func base64encode(src []byte) string {\n\tvar buf bytes.Buffer\n\tencoder := base64.NewEncoder(base64.StdEncoding, &buf)\n\tencoder.Write(src)\n\tencoder.Close()\n\treturn buf.String()\n}", "func encodeBase64(source []byte) string {\n\t// make a byte slice just big enough for the result of the encode operation\n\tdest := make([]byte, base64.StdEncoding.EncodedLen(len(source)))\n\t// encode it\n\tbase64.StdEncoding.Encode(dest, source)\n\t// convert this byte buffer to a string\n\treturn bytes.NewBuffer(dest).String()\n}", "func base64Encode(b []byte) string {\n\treturn strings.TrimSuffix(base64.URLEncoding.EncodeToString(b), \"==\")\n}", "func base64encode(v string) string {\n\treturn base64.StdEncoding.EncodeToString([]byte(v))\n}", "func base64Encode(toEncode []byte) string {\n\treturn base64.StdEncoding.EncodeToString(toEncode)\n}", "func base64Encode(src string) string {\n\treturn strings.\n\t\tTrimRight(base64.URLEncoding.\n\t\t\tEncodeToString([]byte(src)), \"=\")\n}", "func EncodeBase64(value string) string {\n\treturn base64.StdEncoding.EncodeToString([]byte(value))\n}", "func Base64Encode(str string) string {\n\treturn base64.StdEncoding.EncodeToString([]byte(str))\n}", "func EncodeAsBase64(e interface{}) string {\n\tvar b bytes.Buffer\n\tenc := gob.NewEncoder(&b)\n\tif err := enc.Encode(e); err != nil {\n\t\tpanic(err)\n\t}\n\tbytes := b.Bytes()\n\tsha := sha1.New()\n\tsha.Write(bytes)\n\tbytes = sha.Sum(nil)\n\tb64 := base64.StdEncoding.EncodeToString(bytes)\n\treturn b64\n}", "func EncodeBase64(bytes []byte) string {\n\treturn base64.URLEncoding.EncodeToString(bytes)\n}", "func Base64Encode(raw []byte) string {\n\treturn base64.URLEncoding.EncodeToString(raw)\n}", "func Encode64(inputStr string) string {\n\treturn base64.StdEncoding.EncodeToString([]byte(inputStr))\n}", "func Base64Encode(text string) string {\n\treturn base64.StdEncoding.EncodeToString([]byte(text))\n}", "func __b64encode(out *[]byte, src *[]byte, mode int)", "func Base64Encode(input []byte) []byte {\n\tenc := base64.StdEncoding\n\tencLength := enc.EncodedLen(len(input))\n\toutput := make([]byte, encLength)\n\tenc.Encode(output, input)\n\treturn output\n}", "func base64Encode(data string) string {\n\t// Check whether the data is already Base64 encoded; don't double-encode\n\tif isBase64Encoded(data) {\n\t\treturn data\n\t}\n\t// data has not been encoded encode and return\n\treturn base64.StdEncoding.EncodeToString([]byte(data))\n}", "func Base64Encode(src []byte) []byte {\n\treturn DefaultEncoding.Base64Encode(src)\n}", "func Base64Encoder(asBytes Bytes16) string {\n\treturn base64.StdEncoding.EncodeToString(asBytes[:])\n}", "func Base64Encode(b []byte) []byte {\r\n\tbuf := make([]byte, base64.RawURLEncoding.EncodedLen(len(b)))\r\n\tbase64.RawURLEncoding.Encode(buf, b)\r\n\treturn buf\r\n}", "func (enc *Base64Encoding) Base64Encode(src []byte) []byte {\n\tn := enc.Encoding.EncodedLen(len(src))\n\tdst := make([]byte, n)\n\tenc.Encoding.Encode(dst, src)\n\tfor n > 0 && dst[n-1] == '=' {\n\t\tn--\n\t}\n\treturn dst[:n]\n}", "func base64Encode(src []byte) string {\n\tbuf := make([]byte, base64EncodedLen(len(src)))\n\n\t// mehhhhh actually base64 encoding is a bit more involved\n\t// and it feels like not a good use of time to implement it myself right now,\n\t// I should come back to it. Basically you take 3 bytes of input,\n\t// and then break it into 4 groups of 6 bits, and then encode that\n\t// to produce 4 bytes of base64\n\tbase64.StdEncoding.Encode(buf, src)\n\treturn string(buf)\n}", "func Base64(bin []byte) string {\n\treturn base64.StdEncoding.EncodeToString(bin)\n}", "func EncodeBase64(scope *Scope, input tf.Output, optional ...EncodeBase64Attr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"EncodeBase64\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func tob64(in []byte) string {\n\treturn base64.RawURLEncoding.EncodeToString(in)\n}", "func Base64(str []byte, alphaSet ...string) string {\n\tencoding := base64.StdEncoding\n\tif len(alphaSet) != 0 && alphaSet[0] != \"\" {\n\t\tencoding = base64.NewEncoding(alphaSet[0])\n\t}\n\treturn encoding.EncodeToString(str)\n}", "func Encode(in []byte) (string, error) {\n\treturn b64.StdEncoding.EncodeToString(in), nil\n}", "func EncodeBase64(data []byte) (out []byte) {\n\tout = make([]byte, Encoding.EncodedLen(len(data)))\n\tEncoding.Encode(out, data)\n\treturn\n}", "func B64Encode(data []byte) string {\n\tenc := base64.StdEncoding.EncodeToString(data)\n\treturn enc\n}", "func Base64(data []byte) string {\n\treturn base64.StdEncoding.EncodeToString(data)\n}", "func Base64Transaction(tx []byte) (bs string) {\n jtx, _ := json.Marshal(tx)\n bs = base64.StdEncoding.EncodeToString(jtx)\n return bs\n}", "func EncryptBase64(data []byte) string {\n\treturn base64.StdEncoding.EncodeToString(data)\n}", "func B64EncodeByteToStr(inputBytes []byte) string {\r\n\treturn base64.StdEncoding.EncodeToString(inputBytes)\r\n}", "func base64Str(b []byte) string {\n\treturn base64.StdEncoding.EncodeToString(b)\n}", "func Encode(value []byte) []byte {\n var length int = len(value)\n encoded := make([]byte, base64.URLEncoding.EncodedLen(length))\n base64.URLEncoding.Encode(encoded, value)\n return encoded\n}", "func encode(value []byte) []byte {\n\tencoded := make([]byte, base64.URLEncoding.EncodedLen(len(value)))\n\tbase64.URLEncoding.Encode(encoded, value)\n\treturn encoded\n}", "func Encode(src []byte) string {\n\treturn base64.RawURLEncoding.EncodeToString(src)\n}", "func BASE64EncodeString(str string) string {\n\treturn base64.StdEncoding.EncodeToString([]byte(str))\n}", "func EncodeToBase64(ext string, value []byte) string {\n\tn := 5 + 8 + len(ext) + base64.StdEncoding.EncodedLen(len(value))\n\thas := strings.HasPrefix(ext, \"image/\")\n\tif !has {\n\t\tn += 6\n\t}\n\tbuilder := strings.Builder{}\n\tbuilder.Grow(n)\n\tbuilder.WriteString(\"data:\")\n\tif !has {\n\t\tbuilder.WriteString(\"image/\")\n\t}\n\tbuilder.WriteString(ext)\n\tbuilder.WriteString(\";base64,\")\n\tbuilder.WriteString(base64.StdEncoding.EncodeToString(value))\n\treturn builder.String()\n}", "func AuthenticateToBase64(auth *Authenticate) (string, error) {\n\tjs, err := json.Marshal(auth)\n\tif err != nil {\n\t\treturn \"\", nil\n\t}\n\n\treturn b64.StdEncoding.EncodeToString(js), nil\n}", "func encodeAuthToBase64(authConfig types.AuthConfig) (string, error) {\n\tbuf, err := json.Marshal(authConfig)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn base64.URLEncoding.EncodeToString(buf), nil\n}", "func tob64S(s string) string {\n\treturn base64.RawURLEncoding.EncodeToString([]byte(s))\n}", "func ToBase64Query(s *string) *string {\n\treturn String(base64.StdEncoding.EncodeToString([]byte(StringValue(s))))\n}", "func joseBase64UrlEncode(b []byte) string {\n\treturn strings.TrimRight(base64.URLEncoding.EncodeToString(b), \"=\")\n}", "func BytesToBase64(b []byte) string {\n\treturn base64.URLEncoding.EncodeToString(b)\n}", "func B64Encrypt(qrToken string) string {\n\tsalt := \"/qr-g3nerat0r/\"\n\tdata := salt + qrToken\n\treturn base64.StdEncoding.EncodeToString([]byte(data))\n}", "func (s Sig) Base64Encoded() (string, error) {\n\tsig, ok := s.Layer.Annotations[CosignSigKey]\n\tif !ok {\n\t\treturn \"\", errors.New(\"cosign signature not found in the layer annotations\")\n\t}\n\tif _, err := encoding.DecodeString(sig); err != nil {\n\t\treturn \"\", fmt.Errorf(\"invalid base64 encoded signature: %w\", err)\n\t}\n\treturn sig, nil\n}", "func Base64Encoded() (string, error) {\n\tbyt, err := generateBytes()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tencoded := base64.StdEncoding.EncodeToString(byt)\n\treturn encoded, nil\n}", "func HashBase64(input string) string {\n\thash512 := sha512.Sum512([]byte(input))\n\n\treturn base64.StdEncoding.EncodeToString(hash512[:])\n}", "func (kp *Full) SignBase64(input []byte) (string, error) {\n\tsig, err := kp.Sign(input)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn base64.StdEncoding.EncodeToString(sig), nil\n}", "func (p *policy) Base64() string {\n\treturn base64.StdEncoding.EncodeToString([]byte(p.String()))\n}", "func Base64MimeEncoder(b string) (m string) {\n\n\tm = base64.StdEncoding.EncodeToString([]byte(b))\n\tthe_len := len(m)\n\n\tif (the_len <= maxLen) {\n\t\treturn m\n\t}\n\n\tnew_m := []byte(m)\n\n\t// set the slice capacity to the slice len + each newline delimiters\n\tm1 := make([]byte, 0, the_len+(len(delimiter)*int(the_len/maxLen)))\n\tii := 0\n\tfor i := 0; i < int(the_len/maxLen); i++ {\n\t\tm1 = append(m1, new_m[i*maxLen:(i+1)*maxLen]...)\n\t\tm1 = append(m1, delimiter...)\n\t\tii++\n\t}\n\tm1 = append(m1, new_m[ii*maxLen:the_len]...)\n\tm = string(m1)\n\treturn m\n}", "func (c *publicKey) Base64() (string, error) {\n\tb, err := c.Raw()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn base64.StdEncoding.EncodeToString(b), nil\n}", "func (fnEncodeBase64) Eval(params ...interface{}) (interface{}, error) {\n\ts1, err := coerce.ToBytes(params[0])\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"encodeBase64 function first parameter [%+v] must be []byte\", params[0])\n\t}\n\n\tebuf := make([]byte, base64.StdEncoding.EncodedLen(len(s1)))\n\tbase64.StdEncoding.Encode(ebuf, s1)\n\treturn ebuf, err\n}", "func Encode(src []byte) []byte {\n\tdst := make([]byte, base64.StdEncoding.EncodedLen(len(src)))\n\tbase64.StdEncoding.Encode(dst, src)\n\treturn dst\n}", "func (s Signature) Base64() ([]byte, error) {\n\treturn jose.Base64Encode(s), nil\n}", "func toBase64Json(data any) (string, error) {\n\tif data == nil {\n\t\treturn \"\", nil\n\t}\n\tb, err := json.Marshal(data)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn base64.URLEncoding.EncodeToString(b), nil\n}", "func Base64(n int) (ss string, err error) {\n\tbs, err := Bytes(n)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tss = base64.StdEncoding.EncodeToString(bs)\n\n\treturn\n}", "func Base64(n int) string { return String(n, Base64Chars) }", "func Encode(data []byte) string {\n\t// Factorio\n\tvar b bytes.Buffer\n\tencoder := base64.NewEncoder(base64.StdEncoding, &b)\n\tcompress := zlib.NewWriter(encoder)\n\tcompress.Write(data)\n\tcompress.Close()\n\tencoder.Close()\n\treturn b.String()\n}", "func (me TCryptoBinary) ToXsdtBase64Binary() xsdt.Base64Binary { return xsdt.Base64Binary(me) }", "func EncodeBlessingsBase64(blessings security.Blessings) (string, error) {\n\tif blessings.IsZero() {\n\t\treturn \"\", fmt.Errorf(\"no blessings found\")\n\t}\n\tstr, err := base64urlVomEncode(blessings)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"base64url-vom encoding failed: %v\", err)\n\t}\n\treturn str, nil\n}", "func Base64UrlSafeEncode(source []byte) string {\n\tbyteArr := base64.StdEncoding.EncodeToString(source)\n\tsafeUrl := strings.Replace(byteArr, \"/\", \"_\", -1)\n\tsafeUrl = strings.Replace(safeUrl, \"+\", \"-\", -1)\n\tsafeUrl = strings.Replace(safeUrl, \"=\", \"\", -1)\n\treturn safeUrl\n}", "func EncodeBase64(savePath string, fileNama string) string {\n\tfile, err := os.Open(savePath + fileNama)\n\tdefer file.Close()\n\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\n\tfi, _ := file.Stat() // interface\n\tsize := fi.Size() // file size\n\n\tdata := make([]byte, size)\n\tfile.Read(data)\n\n\treturn base64.StdEncoding.EncodeToString(data)\n}", "func BASE64UrlEncodeString(str string) string {\n\tstr = HmacMd5Encode(str)\n\treturn base64.URLEncoding.EncodeToString([]byte(str))\n}", "func (me TDigestValueType) ToXsdtBase64Binary() xsdt.Base64Binary { return xsdt.Base64Binary(me) }", "func Encode(value string) string {\n\tencoded := make([]byte, base64.URLEncoding.EncodedLen(len(value)))\n\tbase64.URLEncoding.Encode(encoded, []byte(value))\n\treturn string(encoded)\n}", "func writeBase64(data []byte, partWriter io.Writer) error {\n\tbufsiz := base64.StdEncoding.EncodedLen(len(data))\n\tbuffer := make([]byte, bufsiz)\n\tbase64.StdEncoding.Encode(buffer, data)\n\t_, err := partWriter.Write(buffer)\n\n\treturn err\n}", "func EncodePublicKeyBase64(key security.PublicKey) (string, error) {\n\tbuf, err := key.MarshalBinary()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn base64.URLEncoding.EncodeToString(buf), nil\n}", "func (a *AttributeEndorsement) ToBase64() string {\n\treturn base64.StdEncoding.EncodeToString(a.ToBytes())\n}", "func EncodeToken64(token string) string {\n\trefreshToken := base64.StdEncoding.EncodeToString([]byte(token))\n\treturn refreshToken\n}", "func Encode(part []byte) string {\n\treturn strings.TrimRight(base64.URLEncoding.EncodeToString(part), \"=\")\n}", "func EncryptToBase64(cryptKey, text string) (string, error) {\n\tencrypter, err := newDefaultBase64(cryptKey)\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn encrypter.Encrypt(text)\n}", "func (cs CryptoService) EncryptBase64(plain []byte) (base64Cipher []byte, err error) {\n\tcipher, err := cs.Encrypt(plain)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbase64Cipher = make([]byte, base64.RawStdEncoding.EncodedLen(len(cipher)))\n\tbase64.RawStdEncoding.Encode(base64Cipher, cipher)\n\n\treturn\n}", "func ExportAsBase64String(insta *goinsta.Instagram) (string, error) {\n\tbytes, err := ExportAsBytes(insta)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tsEnc := base64.StdEncoding.EncodeToString(bytes)\n\treturn sEnc, nil\n}", "func Base64EncodeMap(m Map) string {\n\tdata, err := json.Marshal(m)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tencoded := base64.URLEncoding.EncodeToString(data)\n\treturn encoded\n}", "func (me TCryptoBinary) String() string { return xsdt.Base64Binary(me).String() }", "func EcbEncryptBase64(key, src string) (string, error) {\n\tkeyBytes, err := getKeyBytes(key)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tsrcBytes, err := base64.StdEncoding.DecodeString(src)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tencryptedBytes, err := EcbEncrypt(keyBytes, srcBytes)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn base64.StdEncoding.EncodeToString(encryptedBytes), nil\n}", "func EncodeString(data []byte) string {\n\treturn base64.StdEncoding.EncodeToString(data)\n}", "func EncodeToString(src []byte) string {\n\treturn base64.StdEncoding.EncodeToString(src)\n}", "func (c *Cursor) ToBase64() (string, error) {\n\tb, err := json.Marshal(c)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn base64.URLEncoding.EncodeToString(b), nil\n}", "func EncodeUUIDBase64(id *uuid.UUID) string {\n\tif id == nil {\n\t\treturn \"\"\n\t}\n\treturn base64.RawStdEncoding.EncodeToString(id.Bytes())\n}", "func (n *NameTransform) B64EncodeToString(src []byte) string {\n\treturn n.B64.EncodeToString(src)\n}", "func (n *NameTransform) B64EncodeToString(src []byte) string {\n\treturn n.B64.EncodeToString(src)\n}", "func Base64ToBytes(h string) []byte {\n\ts, err := base64.URLEncoding.DecodeString(h)\n\tif err != nil {\n\t\tfmt.Errorf(\"faild to convert Base64ToBytes(%s) with error : %s\", h, err.Error())\n\t\treturn []byte(\"\")\n\t}\n\treturn s\n}", "func Base64Decode(operand string) (string, error) {\n\tbytes, err := base64.StdEncoding.DecodeString(operand)\n\treturn string(bytes), err\n}", "func HexToBase64(hex []byte) string {\n\tencoder := base64.StdEncoding\n\treturn encoder.EncodeToString(hex)\n}", "func (objID *ObjID) B64() (string){\n\treturn base64.StdEncoding.EncodeToString(objID.Key)\n}", "func _encode_map(bitmap []byte) string {\n\treturn base64.StdEncoding.EncodeToString(bitmap)\n}", "func (a I2PAddr) Base64() string {\n\treturn string(a)\n}", "func ConvertToBase64Str(key []byte) string {\n\treturn base64.StdEncoding.EncodeToString(key)\n}", "func EncodeBase64Reader(out io.Writer, src io.Reader) (err error) {\n\tencoder := base64.NewEncoder(Encoding, out)\n\t_, err = io.Copy(encoder, src)\n\tencoder.Close()\n\treturn\n}", "func EncodeEscape(b []byte) []byte {\r\n\tbuf := make([]byte, base64.RawURLEncoding.EncodedLen(len(b))+2)\r\n\tbuf[0] = '\"'\r\n\tbase64.RawURLEncoding.Encode(buf[1:], b)\r\n\tbuf[len(buf)-1] = '\"'\r\n\treturn buf\r\n}", "func (store *SessionCookieStore) encode(src []byte) string {\n\tbuf := make([]byte, base64.URLEncoding.EncodedLen(len(src)))\n\tbase64.URLEncoding.Encode(buf, src)\n\tfor {\n\t\tif buf[len(buf)-1] != '=' {\n\t\t\tbreak\n\t\t}\n\t\tbuf = buf[:len(buf)-1]\n\t}\n\treturn string(buf)\n}", "func NewBase64Encoding(alphabet string) *Base64Encoding {\n\treturn &Base64Encoding{\n\t\tAlphabet: alphabet,\n\t\tEncoding: base64.NewEncoding(alphabet),\n\t}\n}", "func (cb *codedBuffer) encodeFixed64(x uint64) error {\n\tcb.buf = append(cb.buf,\n\t\tuint8(x),\n\t\tuint8(x>>8),\n\t\tuint8(x>>16),\n\t\tuint8(x>>24),\n\t\tuint8(x>>32),\n\t\tuint8(x>>40),\n\t\tuint8(x>>48),\n\t\tuint8(x>>56))\n\treturn nil\n}" ]
[ "0.7592985", "0.75759953", "0.75514597", "0.7529357", "0.7516748", "0.7492097", "0.74903625", "0.7457392", "0.7451023", "0.74448234", "0.7409036", "0.74026495", "0.7399233", "0.7367904", "0.7363434", "0.7363157", "0.7342584", "0.73423576", "0.7318747", "0.7312583", "0.73022944", "0.7276614", "0.7259247", "0.7223415", "0.7206804", "0.71969515", "0.71731466", "0.71425724", "0.71077865", "0.7064677", "0.70579726", "0.70457", "0.70018756", "0.694825", "0.6860284", "0.6855229", "0.68492615", "0.68359846", "0.68281144", "0.67736554", "0.67603284", "0.6758718", "0.67411417", "0.6686122", "0.66825193", "0.66821355", "0.66705203", "0.6642625", "0.6600131", "0.6595059", "0.6572322", "0.65610594", "0.6551542", "0.6509952", "0.6502291", "0.65015626", "0.64968467", "0.649252", "0.64909583", "0.64407057", "0.6432506", "0.64260393", "0.64158416", "0.64156264", "0.63738245", "0.63611907", "0.6308741", "0.6301959", "0.6298336", "0.62636936", "0.62635815", "0.6247483", "0.6241469", "0.6241198", "0.6225476", "0.6215076", "0.62024784", "0.6197937", "0.6192336", "0.6184812", "0.6164213", "0.61482185", "0.61441606", "0.6123397", "0.6109202", "0.61088645", "0.60985136", "0.60985136", "0.6075283", "0.6071101", "0.60574186", "0.60512346", "0.6040169", "0.6036258", "0.6023754", "0.60166264", "0.60043913", "0.60002583", "0.5992909", "0.59888864" ]
0.8992932
0
Base64Decode uses base64.StdEncoding to decode operand.
func Base64Decode(operand string) (string, error) { bytes, err := base64.StdEncoding.DecodeString(operand) return string(bytes), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Base64Decode(src []byte) ([]byte, error) {\n\treturn DefaultEncoding.Base64Decode(src)\n}", "func base64Decode(s string) ([]byte, error) {\n\t// add back missing padding\n\tswitch len(s) % 4 {\n\tcase 2:\n\t\ts += \"==\"\n\tcase 3:\n\t\ts += \"=\"\n\t}\n\treturn base64.URLEncoding.DecodeString(s)\n}", "func base64Decode(s string) ([]byte, error) {\n\t// add back missing padding\n\tswitch len(s) % 4 {\n\tcase 2:\n\t\ts += \"==\"\n\tcase 3:\n\t\ts += \"=\"\n\t}\n\treturn base64.URLEncoding.DecodeString(s)\n}", "func base64Decode(b string) string {\n\tdata, err := base64.StdEncoding.DecodeString(b)\n\tif err != nil {\n\t\treturn string(b)\n\t}\n\treturn string(data)\n}", "func base64Decode(value string) (string, error) {\n\tres, err := base64.StdEncoding.DecodeString(value)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(res), nil\n}", "func Base64Decode(encoded string) (string, error) {\n\tresult, err := base64.StdEncoding.DecodeString(encoded)\n\treturn string(result), err\n}", "func (enc *Base64Encoding) Base64Decode(src []byte) ([]byte, error) {\n\tnumOfEquals := 4 - (len(src) % 4)\n\tfor i := 0; i < numOfEquals; i++ {\n\t\tsrc = append(src, '=')\n\t}\n\tdst := make([]byte, enc.Encoding.DecodedLen(len(src)))\n\tn, err := enc.Encoding.Decode(dst, src)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn dst[:n], nil\n}", "func Base64Decode(vm *otto.Otto) {\n\tvm.Set(\"base64decode\", func(call otto.FunctionCall) otto.Value {\n\t\ta0 := call.Argument(0)\n\t\tif !a0.IsString() {\n\t\t\tfmt.Println(\"ERROR\", \"base64decode(string)\")\n\t\t\treturn otto.Value{}\n\t\t}\n\t\ts, err := a0.ToString()\n\t\tif err != nil {\n\t\t\tfmt.Println(\"ERROR\", err)\n\t\t\treturn otto.Value{}\n\t\t}\n\t\tsb, err := base64.StdEncoding.DecodeString(s)\n\t\tif err != nil {\n\t\t\tfmt.Println(\"ERROR\", err)\n\t\t\treturn otto.Value{}\n\t\t}\n\t\tv, err := vm.ToValue(string(sb))\n\t\tif err != nil {\n\t\t\tfmt.Println(\"ERROR\", err)\n\t\t\treturn otto.Value{}\n\t\t}\n\t\treturn v\n\t})\n}", "func Base64Decode(input []byte) []byte {\n\tdec := base64.StdEncoding\n\tdecLength := dec.DecodedLen(len(input))\n\toutput := make([]byte, decLength)\n\tn, err := dec.Decode(output, input)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif n < decLength {\n\t\toutput = output[:n]\n\t}\n\treturn output\n}", "func Base64Decode(encoded string) ([]byte, error) {\n\treturn base64.URLEncoding.DecodeString(encoded)\n}", "func Base64Decode(src string) (string, error) {\n\tif len(src) == 0 {\n\t\treturn \"\", fmt.Errorf(\"cannot decode empty string, occurred in sxutil package\")\n\t}\n\tdata, err := base64.StdEncoding.DecodeString(src)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(data), nil\n}", "func Base64Decode(input string) (string, error) {\n\tdata, err := base64.StdEncoding.DecodeString(input)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tdecoded := string(data)\n\treturn decoded, nil\n}", "func base64Decode(src string) (string, error) {\n\tif l := len(src) % 4; l > 0 {\n\t\tsrc += strings.Repeat(\"=\", 4-l)\n\t}\n\tdecoded, err := base64.URLEncoding.DecodeString(src)\n\tif err != nil {\n\t\terrMsg := fmt.Errorf(\"Decoding Error %s\", err)\n\t\treturn \"\", errMsg\n\t}\n\treturn string(decoded), nil\n}", "func base64Decode(src string) (string, error) {\n\tif strings.TrimSpace(src) == \"\" {\n\t\treturn \"\", consts.ErrEmptyString\n\t}\n\tif l := len(src) % 4; l > 0 {\n\t\tsrc += strings.Repeat(\"=\", 4-l)\n\t}\n\tdecoded, err := base64.URLEncoding.DecodeString(src)\n\tif err != nil {\n\t\terrMsg := fmt.Errorf(\"decoding error %s\", err)\n\t\treturn \"\", errMsg\n\t}\n\treturn string(decoded), nil\n}", "func Base64Decode(b []byte) ([]byte, error) {\r\n\tbuf := make([]byte, base64.RawURLEncoding.DecodedLen(len(b)))\r\n\tn, err := base64.RawURLEncoding.Decode(buf, b)\r\n\treturn buf[:n], err\r\n}", "func base64decode(v string) string {\n\tdata, err := base64.StdEncoding.DecodeString(v)\n\tif err != nil {\n\t\tLogger.Println(\"[ERROR] Failed decoding base64 encoded string\", err)\n\t\treturn err.Error()\n\t}\n\treturn string(data)\n}", "func runtimeDecodeBase64(ic *interop.Context) error {\n\tsrc := ic.VM.Estack().Pop().String()\n\tresult, err := base64.StdEncoding.DecodeString(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tic.VM.Estack().PushVal(result)\n\treturn nil\n}", "func DecodeBase64(source string) ([]byte, error) {\n\tout, err := base64.StdEncoding.DecodeString(source)\n\treturn out, err\n}", "func B64Decode(data string) []byte {\n\tdec, err := base64.StdEncoding.DecodeString(data)\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\treturn dec\n}", "func __b64decode(out *[]byte, src unsafe.Pointer, len int, mode int) (ret int)", "func DecodeBase64(scope *Scope, input tf.Output) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"DecodeBase64\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func DecodeBase64(s string) []byte { \n bytes, err := base64.URLEncoding.DecodeString(s) \n if err != nil { \n \tfmt.Println(\"ERROR: There was an error decoding string (my_server.go: DecodeBase64)\")\n \tfmt.Println(err)\n \treturn []byte{}\n } \n return bytes \n}", "func Base64Decode(src string) (string, error) {\n if l := len(src) % 4; l > 0 {\n src += strings.Repeat(\"=\", 4-l)\n }\n decoded, err := base64.URLEncoding.DecodeString(src)\n if err != nil {\n errMsg := fmt.Errorf(\"Error al desencodear %s\", err)\n return \"\", errMsg\n }\n return string(decoded), nil\n}", "func BASE64DecodeString(str string) string {\n\tresult, _ := base64.StdEncoding.DecodeString(str)\n\treturn string(result)\n}", "func DecodeBase64(value string) (string, error) {\n\tdecoded, err := base64.StdEncoding.DecodeString(value)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(decoded), nil\n}", "func Decode(in string) ([]byte, error) {\n\to, err := b64.StdEncoding.DecodeString(in)\n\tif err != nil {\n\t\t// maybe it's in the URL variant?\n\t\to, err = b64.URLEncoding.DecodeString(in)\n\t\tif err != nil {\n\t\t\t// ok, just give up...\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn o, nil\n}", "func DecodeBase64(encoded []byte) (out []byte, err error) {\n\tout = make([]byte, Encoding.DecodedLen(len(encoded)))\n\t_, err = Encoding.Decode(out, encoded)\n\treturn\n}", "func (b *baseSemanticUTF8Base64) Decode() string {\n\treturn b.decoded\n}", "func fnBase64Decode(ctx Context, doc *JDoc, params []string) interface{} {\n\tstats := ctx.Value(EelTotalStats).(*ServiceStats)\n\tif params == nil || len(params) != 1 {\n\t\tctx.Log().Error(\"error_type\", \"func_base64decode\", \"op\", \"base64decode\", \"cause\", \"wrong_number_of_parameters\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, SyntaxError{fmt.Sprintf(\"wrong number of parameters in call to base64decode function\"), \"base64decode\", params})\n\t\treturn \"\"\n\t}\n\n\tbs, err := base64.StdEncoding.DecodeString(extractStringParam(params[0]))\n\tif err != nil {\n\t\tctx.Log().Error(\"error_type\", \"func_base64decode\", \"op\", \"base64decode\", \"cause\", \"error_decode\", \"params\", params, \"error\", err.Error())\n\t\tstats.IncErrors()\n\t\tAddError(ctx, RuntimeError{err.Error(), \"base64decode\", params})\n\t\treturn \"\"\n\t}\n\n\treturn string(bs)\n}", "func base64DecodeBytes(encodedBytes []byte) []byte {\n\tdecodedLen := base64.StdEncoding.DecodedLen(len(encodedBytes))\n\tdecodedBytes := make([]byte, decodedLen)\n\tnumBytesRead := 0\n\tnumBytesWritten := 0\n\tfor numBytesRead+3 < len(encodedBytes) {\n\t\tn, _ := base64.StdEncoding.Decode(decodedBytes[numBytesWritten:], encodedBytes[numBytesRead:])\n\t\tnumBytesWritten += n\n\t\tnumBytesRead += base64.StdEncoding.EncodedLen(n)\n\t}\n\treturn decodedBytes[:numBytesWritten]\n}", "func Decode(data []byte) ([]byte, error) {\n\tvar (\n\t\tsrc = make([]byte, base64.StdEncoding.DecodedLen(len(data)))\n\t\tn, err = base64.StdEncoding.Decode(src, data)\n\t)\n\tif err != nil {\n\t\terr = gerror.Wrap(err, `base64.StdEncoding.Decode failed`)\n\t}\n\treturn src[:n], err\n}", "func decode(k *KeyValue) {\n\tif k.Encoding == \"binary\" {\n\t\tdecoded, err := base64.StdEncoding.DecodeString(k.Data)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error decoding base64 key/value\")\n\t\t}\n\t\tk.Data = string(decoded)\n\t}\n}", "func TestBase64Decode(t *testing.T) {\n\twant := \"Hello World\"\n\tg, _ := phpfuncs.Base64Decode(\"SGVsbG8gV29ybGQ=\")\n\tCoverCheck(t, g, want)\n}", "func DecodeBase64(img string) (string, []byte, error) {\n\tss := strings.Split(img, \",\")\n\tif len(ss) != 2 {\n\t\treturn \"\", nil, errors.New(\"invalid base64 image data\")\n\t}\n\n\ttp := strings.TrimSuffix(strings.TrimPrefix(ss[0], \"data:\"), \";base64\")\n\tbv, err := base64.StdEncoding.DecodeString(ss[1])\n\treturn tp, bv, err\n}", "func DecodeB64(msg []byte) ([]byte, error) {\n\tb64 := base64.StdEncoding\n\tencoded := bytes.TrimSpace(msg)\n\trest := make([]byte, b64.DecodedLen(len(encoded)))\n\tif n, err := b64.Decode(rest, encoded); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\trest = rest[:n]\n\t}\n\treturn rest, nil\n}", "func decodeBase64Upload(fileUpload string) ([]byte, error) {\n\t// Split the file uploads into the various parts\n\tfileParts := strings.Split(fileUpload, \";\")\n\tif len(fileParts) != 3 {\n\t\terr := errors.New(\"Error parsing the uploaded file\")\n\t\tlog.Error(err.Error())\n\t\treturn []byte{}, err\n\t}\n\t// [0] - file:[filename of uploaded file]\n\t// [1] - data:[data type (text/plain)]\n\t// [2] - base64,[data]\n\n\t// Decode the base64 file content\n\tdecodedBytes, err := base64.StdEncoding.DecodeString(fileParts[2][7:])\n\tif err != nil {\n\t\tlog.WithField(\"error\", err).Error(\"Error parsing the base64 file contents\")\n\t\treturn []byte{}, err\n\t}\n\n\t// Return the decoded bytes\n\treturn decodedBytes, nil\n}", "func B64DecodeStrToByte(inputString string) ([]byte, error) {\r\n\tdecoded, err := base64.StdEncoding.DecodeString(inputString)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\treturn decoded, nil\r\n}", "func Decode(value []byte) ([]byte, error) {\n var length int = len(value)\n decoded := make([]byte, base64.URLEncoding.DecodedLen(length))\n\n n, err := base64.URLEncoding.Decode(decoded, value)\n if err != nil {\n return nil, err\n }\n return decoded[:n], nil\n}", "func (n *NameTransform) B64DecodeString(s string) ([]byte, error) {\n\treturn n.B64.DecodeString(s)\n}", "func (n *NameTransform) B64DecodeString(s string) ([]byte, error) {\n\treturn n.B64.DecodeString(s)\n}", "func (i Base64Image) Decode() (string, error) {\n\tdecoded, err := base64.StdEncoding.DecodeString(string(i))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(decoded), nil\n}", "func fromb64(in string) string {\n\ts, err := base64.RawURLEncoding.DecodeString(in)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn string(s)\n}", "func Decode(in string, obj interface{}) {\n\tb, err := base64.StdEncoding.DecodeString(in)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = json.Unmarshal(b, obj)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func Decode(in string, obj interface{}) {\n\tb, err := base64.StdEncoding.DecodeString(in)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = json.Unmarshal(b, obj)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func DecodeString(data string) ([]byte, error) {\n\treturn base64.StdEncoding.DecodeString(data)\n}", "func Decode(in string, obj interface{}) error {\n\tb, err := base64.StdEncoding.DecodeString(in)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = json.Unmarshal(b, obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func vaultDecode(data interface{}) ([]byte, error) {\n\tencoded, ok := data.(string)\n\tif !ok {\n\t\treturn nil, errors.New(\"Received non-string data\")\n\t}\n\n\treturn base64.StdEncoding.DecodeString(prefixRegex.ReplaceAllString(encoded, \"\"))\n}", "func Base64UrlDecode(s string) ([]byte, error) {\n\tpad := len(s) % 4\n\tfor pad > 0 {\n\t\ts += \"=\"\n\t\tpad--\n\t}\n\n\treturn base64.StdEncoding.DecodeString(urlDecodeRe.Replace(s))\n}", "func (cs CryptoService) DecryptBase64(base64Cipher []byte) (plain []byte, err error) {\n\tcipher := make([]byte, base64.RawStdEncoding.DecodedLen(len(base64Cipher)))\n\t_, err = base64.RawStdEncoding.Decode(cipher, base64Cipher)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn cs.Decrypt(cipher)\n}", "func DecodeBase64Reader(out io.Writer, src io.Reader) (err error) {\n\tdecoder := base64.NewDecoder(Encoding, src)\n\t_, err = io.Copy(out, decoder)\n\treturn\n}", "func DecodeToken64(token string) (string, error) {\n\trefreshToken, err := base64.StdEncoding.DecodeString(token)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn \"\", errors.New(\"Error decoding access token\")\n\t}\n\n\treturn string(refreshToken), nil\n}", "func EcbDecryptBase64(key, src string) (string, error) {\n\tkeyBytes, err := getKeyBytes(key)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tencryptedBytes, err := base64.StdEncoding.DecodeString(src)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tdecryptedBytes, err := EcbDecrypt(keyBytes, encryptedBytes)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn base64.StdEncoding.EncodeToString(decryptedBytes), nil\n}", "func (ev Vars) Base64(key string, defaults ...[]byte) []byte {\n\tif value, hasValue := ev[key]; hasValue && len(value) > 0 {\n\t\tresult, _ := util.Base64.Decode(value)\n\t\treturn result\n\t}\n\tif len(defaults) > 0 {\n\t\treturn defaults[0]\n\t}\n\treturn nil\n}", "func vaultDecode(data interface{}) ([]byte, error) {\n\tencoded, ok := data.(string)\n\tif !ok {\n\t\treturn nil, errors.New(\"Received non-string data\")\n\t}\n\treturn base64.StdEncoding.DecodeString(strings.TrimPrefix(encoded, vaultV1DataPrefix))\n}", "func Decode(s string, l int) ([]byte, error) {\n\tr, err := base64.RawURLEncoding.DecodeString(s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(r) != l {\n\t\treturn nil, fmt.Errorf(\"base64: wrong length %d (expecting %d): %s\", 2*len(r), 2*l, s)\n\t}\n\treturn r, nil\n}", "func decodeAesBase64ByDynamics(src, key, iv string) ([]byte, error) {\n //base64\n s1, err := base64.URLEncoding.DecodeString(src)\n if err != nil {\n logrus.Error(\"decode base64 error : \", err)\n return nil, err\n }\n\n //aes128\n out, err := decodeAesCbcByDynamics(s1, key, iv)\n if err != nil {\n logrus.Error(\"decode aes cbc error : \", err)\n return nil, err\n }\n\n return out, nil\n}", "func decodeBase62(encStr string) string {\n\treturn string(noPad62Encoding.DecodeToBigInt(encStr).Bytes())\n}", "func decodeFromBase64(encodedXdr string) *b.TransactionEnvelopeBuilder {\n // Unmarshall from base64 encoded XDR format\n var decoded xdr.TransactionEnvelope\n e := xdr.SafeUnmarshalBase64(encodedXdr, &decoded)\n if e != nil {\n log.Fatal(e)\n }\n\n // convert to TransactionEnvelopeBuilder\n txEnvelopeBuilder := b.TransactionEnvelopeBuilder{E: &decoded}\n txEnvelopeBuilder.Init()\n\n return &txEnvelopeBuilder\n}", "func decodeFromBase64(encodedXdr string) *b.TransactionEnvelopeBuilder {\n // Unmarshall from base64 encoded XDR format\n var decoded xdr.TransactionEnvelope\n e := xdr.SafeUnmarshalBase64(encodedXdr, &decoded)\n if e != nil {\n log.Fatal(e)\n }\n\n // convert to TransactionEnvelopeBuilder\n txEnvelopeBuilder := b.TransactionEnvelopeBuilder{E: &decoded}\n txEnvelopeBuilder.Init()\n\n return &txEnvelopeBuilder\n}", "func DecryptFromBase64(cryptKey, text string) (string, error) {\n\tdecrypter, err := newDefaultBase64(cryptKey)\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn decrypter.Decrypt(text)\n}", "func decodeFromBase64(encodedXdr string) *b.TransactionEnvelopeBuilder {\n\t// Unmarshall from base64 encoded XDR format\n\tvar decoded xdr.TransactionEnvelope\n\te := xdr.SafeUnmarshalBase64(encodedXdr, &decoded)\n\tif e != nil {\n\t\tlog.Fatal(e)\n\t}\n\n\t// convert to TransactionEnvelopeBuilder\n\ttxEnvelopeBuilder := b.TransactionEnvelopeBuilder{E: &decoded}\n\ttxEnvelopeBuilder.Init()\n\n\treturn &txEnvelopeBuilder\n}", "func convertBase64(input string) (strfmt.Base64, error) {\n\ttemp, err := formats.Parse(\"byte\", input)\n\tif err != nil {\n\t\treturn strfmt.Base64{}, err\n\t}\n\treturn *temp.(*strfmt.Base64), nil\n}", "func (store *SessionCookieStore) decode(src string) ([]byte, error) {\n\tsize := len(src)\n\trem := (4 - size%4) % 4\n\tbuf := make([]byte, size+rem)\n\tcopy(buf, src)\n\tfor i := 0; i < rem; i++ {\n\t\tbuf[size+i] = '='\n\t}\n\tn, err := base64.URLEncoding.Decode(buf, buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn buf[:n], nil\n}", "func DecodeString(encode, content string) (string, error) {\n\tif strings.EqualFold(\"base64\", encode) {\n\t\tdecode, err := base64.StdEncoding.DecodeString(content)\n\t\treturn string(decode), err\n\t}\n\n\treturn content, nil\n}", "func tob64(in []byte) string {\n\treturn base64.RawURLEncoding.EncodeToString(in)\n}", "func Decode(in string, obj interface{}) {\n\tb, err := base64.StdEncoding.DecodeString(in)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tif compress {\n\t\tb = unzip(b)\n\t}\n\n\terr = json.Unmarshal(b, obj)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func DecodeUUIDBase64(in string) *uuid.UUID {\n\tif in == \"\" {\n\t\treturn nil\n\t}\n\tstr, err := base64.RawStdEncoding.DecodeString(in)\n\tif err != nil {\n\t\treturn nil\n\t}\n\tid := uuid.FromBytesOrNil([]byte(str))\n\tif id != uuid.Nil {\n\t\treturn &id\n\t}\n\treturn nil\n}", "func Decode(part string) ([]byte, error) {\n\tif l := len(part) % 4; l > 0 {\n\t\tpart += strings.Repeat(\"=\", 4-l)\n\t}\n\n\treturn base64.URLEncoding.DecodeString(part)\n}", "func (d *Decoder) Decode(input []byte) ([]byte, Encoding) {\n\tif len(input) == 0 {\n\t\treturn []byte{}, None\n\t}\n\n\tunmarshalled := &protodec.Empty{}\n\n\tif d.proto {\n\t\tif err := proto.Unmarshal(input, unmarshalled); err == nil {\n\t\t\t// TODO: remove control characters (unfortunately, they are all valid strings here)\n\t\t\treturn []byte(unmarshalled.String()), Proto\n\t\t}\n\t}\n\n\tif d.bitDec {\n\t\tbyteIn := strings.Trim(string(input), \"[]\") // [32 87 111 114 108 100] -> 32 87 111 114 108 100\n\n\t\tif b, err := Base2AsBytes(byteIn); err == nil {\n\t\t\treturn b, Bit\n\t\t}\n\t}\n\n\t// byte before hex, hex might contains letters, which are not valid in byte dec\n\tif d.byteDec {\n\t\tbyteIn := strings.Trim(string(input), \"[]\") // [32 87 111 114 108 100] -> 32 87 111 114 108 100\n\n\t\tif b, err := Base10AsBytes(byteIn); err == nil {\n\t\t\treturn b, Byte\n\t\t}\n\t}\n\n\t// hex after byte\n\tif d.hex {\n\t\thexIn := strings.TrimSpace(string(input)) // e.g. new line\n\t\thexIn = strings.TrimPrefix(hexIn, \"0x\") // hex prefix\n\t\thexIn = strings.Replace(hexIn, \" \", \"\", -1) // bd b2 3d bc 20 e2 8c 98 -> bdb23dbc20e28c98\n\n\t\tif b, err := hex.DecodeString(hexIn); err == nil {\n\t\t\treturn b, Hex\n\t\t}\n\t}\n\n\t// TODO: many false-positives. Decodes it when no base64 was given.\n\t// Keep it as one of the last decodings.\n\tif d.base64 {\n\t\tif b, err := base64.StdEncoding.DecodeString(strings.TrimSpace(string(input))); err == nil {\n\t\t\treturn b, Base64\n\t\t}\n\t}\n\n\treturn input, None\n}", "func Decode(value string) (string, error) {\n\tvalue = value\n\tdecoded := make([]byte, base64.URLEncoding.DecodedLen(len(value)))\n\tb, err := base64.URLEncoding.Decode(decoded, []byte(value))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(decoded[:b]), nil\n}", "func DecodeString(s string) string {\n\tif !IsEncoded(s) {\n\t\treturn s\n\t}\n\tdecodedBytes, e := base64.StdEncoding.DecodeString(s)\n\tif e != nil {\n\t\tlog.Println(\"Unexpected decoding error:\", e)\n\t\treturn s\n\t}\n\treturn string(decodedBytes)\n}", "func DecodeCursor(val *string) (string, error) {\n\tif val == nil {\n\t\treturn \"\", nil\n\t}\n\n\tdecoded, err := base64.StdEncoding.DecodeString(*val)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(decoded), nil\n}", "func DecodeCursor(val *string) (string, error) {\n\tif val == nil {\n\t\treturn \"\", nil\n\t}\n\n\tdecoded, err := base64.StdEncoding.DecodeString(*val)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(decoded), nil\n}", "func Base64(bin []byte) string {\n\treturn base64.StdEncoding.EncodeToString(bin)\n}", "func unmarshalProtoBase64(base64Str string, msg proto.Message) error {\n\tb, _ := base64.StdEncoding.DecodeString(base64Str)\n\treturn proto.Unmarshal(b, msg)\n}", "func Base64Decode() GetOption {\n\treturn func(o *getOptions) {\n\t\to.base64Decode = true\n\t}\n}", "func loadPrivateKeyBase64(base64key string) (*rsa.PrivateKey, error) {\n\tkeybytes, err := base64.StdEncoding.DecodeString(base64key)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"base64 decode failed, error=%s\\n\", err.Error())\n\t}\n\n\tprivatekey, err := x509.ParsePKCS1PrivateKey(keybytes)\n\tif err != nil {\n\t\treturn nil, errors.New(\"parse private key error!\")\n\t}\n\n\treturn privatekey, nil\n}", "func DecodeToString(payload string) string {\n\tsDec, _ := b64.StdEncoding.DecodeString(payload)\n\treturn string(sDec)\n}", "func base64Str(b []byte) string {\n\treturn base64.StdEncoding.EncodeToString(b)\n}", "func ConvertBase64StrToBytes(key string) ([]byte, error) {\n\treturn base64.StdEncoding.DecodeString(key)\n}", "func ReadBase64File(filename string) (decoded []byte, err error) {\n\tdata, err := ioutil.ReadFile(filename)\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\tb64Text := strings.Replace(string(data), \"\\n\", \"\", -1)\n\tdecoded, err = b64.StdEncoding.DecodeString(b64Text)\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\treturn decoded, nil\n}", "func (cb *codedBuffer) decodeFixed64() (x uint64, err error) {\n\t// x, err already 0\n\ti := cb.index + 8\n\tif i < 0 || i > len(cb.buf) {\n\t\terr = io.ErrUnexpectedEOF\n\t\treturn\n\t}\n\tcb.index = i\n\n\tx = uint64(cb.buf[i-8])\n\tx |= uint64(cb.buf[i-7]) << 8\n\tx |= uint64(cb.buf[i-6]) << 16\n\tx |= uint64(cb.buf[i-5]) << 24\n\tx |= uint64(cb.buf[i-4]) << 32\n\tx |= uint64(cb.buf[i-3]) << 40\n\tx |= uint64(cb.buf[i-2]) << 48\n\tx |= uint64(cb.buf[i-1]) << 56\n\treturn\n}", "func DecodeBlessingsBase64(encoded string) (security.Blessings, error) {\n\tvar b security.Blessings\n\tif err := base64urlVomDecode(encoded, &b); err != nil {\n\t\treturn security.Blessings{}, fmt.Errorf(\"failed to decode %v: %v\", encoded, err)\n\t}\n\treturn b, nil\n}", "func AuthenticateToBase64(auth *Authenticate) (string, error) {\n\tjs, err := json.Marshal(auth)\n\tif err != nil {\n\t\treturn \"\", nil\n\t}\n\n\treturn b64.StdEncoding.EncodeToString(js), nil\n}", "func base64DecodedWriter(output io.Writer) io.WriteCloser {\n\treadPipe, writePipe := io.Pipe()\n\tgo func() {\n\t\tdecoder := base64.NewDecoder(base64.StdEncoding, readPipe)\n\t\t_, err := io.Copy(output, decoder)\n\t\tif err != nil {\n\t\t\treadPipe.CloseWithError(err)\n\t\t\treturn\n\t\t}\n\t\terr = readPipe.Close()\n\t\tif err != nil {\n\t\t\tpanic(\"readPipe must not return error on close\")\n\t\t}\n\t}()\n\n\treturn writePipe\n}", "func decodeBase64DERCert(certStr string) (*x509.Certificate, error) {\n\t// decode base64\n\tderBytes, err := base64.StdEncoding.DecodeString(certStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// parse the DER-encoded certificate\n\treturn x509.ParseCertificate(derBytes)\n}", "func DeserializeProtoBase64(s string, pb proto.Message) error {\n\tb, err := base64.RawStdEncoding.DecodeString(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn proto.Unmarshal(b, pb)\n}", "func decode(id *ID, src []byte) {\n\tencoder.Decode(id[:], src)\n}", "func (s *hashingServer) Decode(ctx context.Context, request *proto.HashingDecodeRequest) (*proto.HashingDecodeResponse, error) {\n\thashedText := request.GetHashedText()\n\n\tbytePlainText, err := base64.StdEncoding.DecodeString(hashedText)\n\n\tif err == nil {\n\t\treturn &proto.HashingDecodeResponse{PlainText: string(bytePlainText)}, nil\n\t} else {\n\t\treturn &proto.HashingDecodeResponse{PlainText: err.Error()}, nil\n\t}\n}", "func _decode_map(mapstr string) []byte {\n\tbitset, err := base64.StdEncoding.DecodeString(mapstr)\n\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\treturn bitset\n}", "func decrypt(encoded string, key []byte) (string, error) {\n\tcipherText, err := base64.URLEncoding.DecodeString(encoded)\n\tif err != nil {\n\t\treturn encoded, err\n\t}\n\tblock, err := aes.NewCipher(key)\n\tif err != nil {\n\t\treturn encoded, err\n\t}\n\tif len(cipherText) < aes.BlockSize {\n\t\terr = errors.New(\"ciphertext block size is too short\")\n\t\treturn encoded, err\n\t}\n\tiv := cipherText[:aes.BlockSize]\n\tcipherText = cipherText[aes.BlockSize:]\n\tstream := cipher.NewCFBDecrypter(block, iv)\n\tstream.XORKeyStream(cipherText, cipherText)\n\tdecoded := string(cipherText)\n\n\t// By design decrypt with incorrect key must end up with the value\n\tif strings.Index(decoded, anchor) != 0 {\n\t\treturn encoded, nil\n\t}\n\n\tdecoded = strings.Replace(decoded, anchor, \"\", 1) // remove anchor from string\n\treturn decoded, nil\n}", "func Base64ToBytes(h string) []byte {\n\ts, err := base64.URLEncoding.DecodeString(h)\n\tif err != nil {\n\t\tfmt.Errorf(\"faild to convert Base64ToBytes(%s) with error : %s\", h, err.Error())\n\t\treturn []byte(\"\")\n\t}\n\treturn s\n}", "func Base64(s string) bool {\n\tif len(s) == 0 {\n\t\treturn false\n\t}\n\t_, err := base64.StdEncoding.DecodeString(s)\n\n\treturn err == nil\n}", "func DecodeString(in string, objs ...interface{}) (err error) {\n\tbbuf, err := base64.URLEncoding.DecodeString(in)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = DecodeBytes(bbuf, objs...)\n\treturn\n}", "func (c *Cursor) FromBase64(b64 string) (err error) {\n\tcursor, err := base64.URLEncoding.DecodeString(b64)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn json.Unmarshal(cursor, c)\n}", "func DecodeByDynamics(src, key, iv string) ([]byte, error) {\n return decodeAesBase64ByDynamics(src, key, iv)\n}", "func newBaseSemanticUTF8Base64(encoded string, additionalTypeValue uint64) *baseSemanticUTF8Base64 {\n\tdecodedBuf, err := base64.StdEncoding.DecodeString(encoded)\n\tif err != nil {\n\t\tlog.Errorf(\"Error decoding base64 encoded string [%s] due to %s\", encoded, err)\n\t\treturn nil\n\t}\n\treturn &baseSemanticUTF8Base64{\n\t\tbaseSemantic: baseSemantic{\n\t\t\tbaseDataItem: baseDataItem{\n\t\t\t\tmajorType: MajorTypeSemantic,\n\t\t\t},\n\t\t\tadditionalTypeValue: additionalTypeValue,\n\t\t},\n\t\tencoded: encoded,\n\t\tdecoded: string(decodedBuf),\n\t}\n}", "func decode(buf []byte) uint64 {\n\tif buf == nil {\n\t\treturn 0\n\t}\n\treturn binary.BigEndian.Uint64(buf)\n}", "func Base64(data []byte) string {\n\treturn base64.StdEncoding.EncodeToString(data)\n}", "func Base64(n int) string { return String(n, Base64Chars) }" ]
[ "0.8132559", "0.80681306", "0.80681306", "0.80534303", "0.8034052", "0.795179", "0.7944728", "0.7894075", "0.7866363", "0.7828444", "0.7825174", "0.78025174", "0.7749249", "0.7738054", "0.77317834", "0.7590619", "0.7585206", "0.7550624", "0.7548784", "0.7532218", "0.7496599", "0.7474966", "0.7450695", "0.7438541", "0.7434491", "0.7358501", "0.73132825", "0.7133919", "0.71241695", "0.7050279", "0.7018499", "0.7015069", "0.7008012", "0.69847435", "0.68706787", "0.684428", "0.6815146", "0.6807844", "0.67674154", "0.67674154", "0.67488956", "0.6722791", "0.66628695", "0.66628695", "0.6651596", "0.66447484", "0.6637045", "0.6606432", "0.65956485", "0.65955436", "0.65572", "0.65379715", "0.6475436", "0.6469988", "0.6458273", "0.6449026", "0.6433299", "0.64283174", "0.64283174", "0.64185154", "0.6399053", "0.63932645", "0.6392756", "0.63754976", "0.6275668", "0.62708354", "0.6263957", "0.62455237", "0.62325233", "0.62088215", "0.61551297", "0.615028", "0.615028", "0.61296386", "0.6110522", "0.6094514", "0.60754484", "0.6056627", "0.6031098", "0.6030612", "0.6012054", "0.59876394", "0.59864175", "0.59809047", "0.5968245", "0.5957578", "0.59546244", "0.5893557", "0.58881116", "0.5869071", "0.5852691", "0.58411425", "0.58354056", "0.5820008", "0.5817639", "0.5806457", "0.5799794", "0.5780901", "0.5777918", "0.5773127" ]
0.86120623
0
ParseBool uses strconv.ParseBool to parse operand as a bool.
func ParseBool(operand string) (value bool, err error) { return strconv.ParseBool(operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ParseBool(str string) bool {\n\tb, _ := strconv.ParseBool(str)\n\treturn b\n}", "func ParseBool(str string) (bool, error) {\n\tif str == \"on\" {\n\t\treturn true, nil\n\t}\n\tif str == \"off\" {\n\t\treturn false, nil\n\t}\n\treturn strconv.ParseBool(str)\n}", "func ParseBool(s string) (bool, error) {\n\tswitch s {\n\tdefault:\n\t\tb, err := strconv.ParseBool(s)\n\t\tif err != nil {\n\t\t\treturn b, errz.Err(err)\n\t\t}\n\t\treturn b, nil\n\tcase \"1\", \"yes\", \"Yes\", \"YES\", \"y\", \"Y\":\n\t\treturn true, nil\n\tcase \"0\", \"no\", \"No\", \"NO\", \"n\", \"N\":\n\t\treturn false, nil\n\t}\n}", "func ParseBool(str string) (bool, error) {}", "func ParseBool(str string) (val bool, isBool bool) {\n\t// Note: Not using strconv.ParseBool because I want it a bit looser (any casing) and to allow yes/no/off/on values.\n\tlstr := strings.ToLower(strings.TrimSpace(str))\n\tswitch lstr {\n\tcase \"false\", \"f\", \"0\", \"no\", \"n\", \"off\":\n\t\tisBool = true\n\tcase \"true\", \"t\", \"1\", \"yes\", \"y\", \"on\":\n\t\tval = true\n\t\tisBool = true\n\t}\n\treturn\n}", "func parseBool(str string) (value bool, err error) {\n\tswitch str {\n\tcase \"1\", \"t\", \"T\", \"true\", \"TRUE\", \"True\", \"YES\", \"yes\", \"Yes\", \"ON\", \"on\", \"On\":\n\t\treturn true, nil\n\tcase \"0\", \"f\", \"F\", \"false\", \"FALSE\", \"False\", \"NO\", \"no\", \"No\", \"OFF\", \"off\", \"Off\":\n\t\treturn false, nil\n\t}\n\treturn false, fmt.Errorf(\"parsing \\\"%s\\\": invalid syntax\", str)\n}", "func parseBool(asString string) (bool, error) {\n\tswitch asString {\n\tcase \"true\":\n\t\treturn true, nil\n\tcase \"false\":\n\t\treturn false, nil\n\tdefault:\n\t\treturn false, fmt.Errorf(\"could not parse %q as a bool\", asString)\n\t}\n}", "func ParseBool(val interface{}) (value bool, err error) {\n\tif val != nil {\n\t\tswitch v := val.(type) {\n\t\tcase bool:\n\t\t\treturn v, nil\n\t\tcase string:\n\t\t\tswitch v {\n\t\t\tcase \"1\", \"t\", \"T\", \"true\", \"TRUE\", \"True\", \"YES\", \"yes\", \"Yes\", \"Y\", \"y\", \"ON\", \"on\", \"On\":\n\t\t\t\treturn true, nil\n\t\t\tcase \"0\", \"f\", \"F\", \"false\", \"FALSE\", \"False\", \"NO\", \"no\", \"No\", \"N\", \"n\", \"OFF\", \"off\", \"Off\":\n\t\t\t\treturn false, nil\n\t\t\t}\n\t\tcase int8, int32, int64:\n\t\t\tstrV := fmt.Sprintf(\"%s\", v)\n\t\t\tif strV == \"1\" {\n\t\t\t\treturn true, nil\n\t\t\t} else if strV == \"0\" {\n\t\t\t\treturn false, nil\n\t\t\t}\n\t\tcase float64:\n\t\t\tif v == 1 {\n\t\t\t\treturn true, nil\n\t\t\t} else if v == 0 {\n\t\t\t\treturn false, nil\n\t\t\t}\n\t\t}\n\t\treturn false, fmt.Errorf(\"parsing %q: invalid syntax\", val)\n\t}\n\treturn false, fmt.Errorf(\"parsing <nil>: invalid syntax\")\n}", "func ParseBool(q url.Values, name string) (bool, bool, error) {\n\tstringVal := q.Get(name)\n\tif stringVal == \"\" {\n\t\treturn false, false, nil\n\t}\n\n\tval, err := strconv.ParseBool(stringVal)\n\tif err != nil {\n\t\treturn false, false, trace.BadParameter(\n\t\t\t\"'%v': expected 'true' or 'false', got %v\", name, stringVal)\n\t}\n\treturn val, true, nil\n}", "func ParseBoolean(s string) Boolean {\n\tif s == \"\" {\n\t\treturn NullBoolean()\n\t}\n\n\treturn NewBoolean(s == \"true\")\n}", "func ParseBool(key string) (bool, error) {\n\tval := os.Getenv(key)\n\n\tif val == \"\" {\n\t\treturn false, notFoundError(key, \"ParseBool\")\n\t}\n\n\tparsedVal, err := strconv.ParseBool(val)\n\tif err != nil {\n\t\treturn false, &EnvError{\"ParseBool\", key, err}\n\t}\n\n\treturn parsedVal, err\n}", "func ParseBoolP(cmd *cobra.Command, name string) (*bool, error) {\n\tflagRaw, err := cmd.Flags().GetString(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar flagVal *bool\n\tss, err := strconv.ParseBool(flagRaw)\n\tif err != nil && flagRaw != \"\" {\n\t\treturn nil, err\n\t}\n\n\tif flagRaw != \"\" && err == nil {\n\t\treturn &ss, nil\n\t}\n\n\treturn flagVal, nil\n}", "func parseBoolFromString(content string, aggErr *AggregateError) bool {\n result, err := strconv.ParseBool(content)\n if err != nil {\n aggErr.Append(err)\n }\n return result\n}", "func (f flagBool) Parse(value string) interface{} {\n\tswitch value {\n\tcase \"1\", \"t\", \"T\", \"true\", \"TRUE\", \"True\", \"y\", \"Y\", \"yes\", \"YES\", \"Yes\":\n\t\treturn true\n\t}\n\treturn false\n}", "func ParseToBool(data string) (bool, error) {\n\treturn strconv.ParseBool(data)\n}", "func mustParseBool(str string) bool {\n\tb, err := strconv.ParseBool(str)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn b\n}", "func parseBool(content []byte, aggErr *AggregateError) bool {\n result, err := strconv.ParseBool(string(content))\n if err != nil {\n aggErr.Append(err)\n }\n return result\n}", "func parseBoolean(s *scanner) (bool, error) {\n\tif s.eof() || s.data[s.off] != '?' {\n\t\treturn false, &UnmarshalError{s.off, ErrInvalidBooleanFormat}\n\t}\n\ts.off++\n\n\tif s.eof() {\n\t\treturn false, &UnmarshalError{s.off, ErrInvalidBooleanFormat}\n\t}\n\n\tswitch s.data[s.off] {\n\tcase '0':\n\t\ts.off++\n\n\t\treturn false, nil\n\tcase '1':\n\t\ts.off++\n\n\t\treturn true, nil\n\t}\n\n\treturn false, &UnmarshalError{s.off, ErrInvalidBooleanFormat}\n}", "func ParseFlagBool(args []string) (bool, int, error) {\n\tif strings.ContainsAny(args[0], \"= \") {\n\t\tparts := strings.SplitN(args[0], \"=\", 2)\n\t\tif len(parts) == 1 {\n\t\t\tparts = strings.SplitN(args[0], \" \", 2)\n\t\t}\n\t\tif len(parts) == 2 {\n\t\t\tval, isBool := ParseBool(parts[1])\n\t\t\tif !isBool {\n\t\t\t\treturn false, 0, fmt.Errorf(\"invalid %s bool value: [%s]\", parts[0], parts[1])\n\t\t\t}\n\t\t\treturn val, 0, nil\n\t\t}\n\t\treturn false, 0, fmt.Errorf(\"unable to split flag and value from string: [%s]\", args[0])\n\t}\n\tif len(args) > 1 {\n\t\tval, isBool := ParseBool(args[1])\n\t\tif isBool {\n\t\t\treturn val, 1, nil\n\t\t}\n\t}\n\treturn true, 0, nil\n}", "func (s *Str) Bool() bool {\n\tval, err := strconv.ParseBool(s.val)\n\tif err != nil {\n\t\ts.err = err\n\t}\n\treturn val\n}", "func ParseQueryBool(param string, request *http.Request, params imageserver.Params) error {\n\ts := request.URL.Query().Get(param)\n\tif s == \"\" {\n\t\treturn nil\n\t}\n\tb, err := strconv.ParseBool(s)\n\tif err != nil {\n\t\treturn newParseTypeParamError(param, \"bool\", err)\n\t}\n\tparams.Set(param, b)\n\treturn nil\n}", "func (p Parser) Bool(ctx context.Context) (*bool, error) {\n\tvalue, err := p.Source.String(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif value == nil {\n\t\treturn nil, nil\n\t}\n\n\tparsed, err := stringutil.ParseBool(*value)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &parsed, nil\n}", "func (this *OptionBool) Parse(arg string) error {\n argint, err := strconv.Atoi(arg)\n if err != nil {\n return err\n }\n\n //this.opt_storage = argint!=0;\n var storage *bool\n storage = this.opt_storage.(*bool)\n\n *storage = argint != 0\n\n return nil\n}", "func parseBoolEx(repr string) (value bool, err error) {\n\tif value, err = strconv.ParseBool(repr); err != nil {\n\t\tswitch repr {\n\t\tcase \"y\", \"yes\", \"YES\", \"Yes\":\n\t\t\treturn true, nil\n\t\tcase \"n\", \"no\", \"NO\", \"No\":\n\t\t\treturn false, nil\n\t\t}\n\t}\n\n\treturn\n}", "func BoolConverter(str string, target reflect.Value) (ok bool) {\n\tb, err := strconv.ParseBool(str)\n\tif err != nil {\n\t\treturn\n\t}\n\n\ttarget.SetBool(b)\n\treturn true\n}", "func (p *Parser) parseBoolean() asti.ExpressionI {\n\treturn &ast.Boolean{Token: p.curToken, Value: p.curTokenIs(tokentype.TRUE)}\n}", "func (p *Parser) parseBoolean() ast.Expression {\n\treturn &ast.Boolean{Token: p.cur, Value: p.cur.Is(token.TRUE)}\n}", "func (f *flag) Bool() bool {\n\tvalue, err := strconv.ParseBool(f.String())\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn value\n}", "func DeserializeBool(d []byte) (Bool, error) {\n\tif len(d) != 1 {\n\t\treturn false, errors.New(\"deserialize bool: invalid length\")\n\t}\n\tif d[0] == 0 {\n\t\treturn false, nil\n\t} else if d[0] == 1 {\n\t\treturn true, nil\n\t} else {\n\t\treturn false, errors.New(\"deserialize bool: invalid value\")\n\t}\n}", "func Bool(value interface{}) bool {\r\n\ts := String(value)\r\n\tb, _ := strconv.ParseBool(s)\r\n\treturn b\r\n}", "func Bool(val string) error {\n\tif strings.EqualFold(val, \"true\") || strings.EqualFold(val, \"false\") {\n\t\treturn nil\n\t}\n\n\treturn fmt.Errorf(\"invalid bool value '%s', can be only 'true' or 'false'\", val)\n}", "func AsBool(v string) bool {\n\tif v == \"\" {\n\t\treturn false\n\t}\n\tout, _ := apiutils.ParseBool(v)\n\treturn out\n}", "func ParseBoolEnv(envKey string, defaultValue bool) (bool, error) {\n\tvalue := os.Getenv(envKey)\n\tif value == \"\" {\n\t\treturn defaultValue, nil\n\t}\n\tb, err := strconv.ParseBool(value)\n\tif err != nil {\n\t\treturn false, errors.Wrapf(err, \"failed to parse %s\", envKey)\n\t}\n\treturn b, nil\n}", "func (p *Parser) Bool(description string) *bool {\n\tv := false\n\top := p.newOpt(description, false)\n\top.valueType = boolValue\n\top.value = &v\n\treturn &v\n}", "func StrToBool(s string) (bool, error) {\n\tclean := strings.TrimSpace(s)\n\n\tif regexp.MustCompile(`(?i)^(1|yes|true|y|t)$`).MatchString(clean) {\n\t\treturn true, nil\n\t}\n\n\tif regexp.MustCompile(`(?i)^(0|no|false|n|f)$`).MatchString(clean) {\n\t\treturn false, nil\n\t}\n\n\treturn false, fmt.Errorf(\"cannot convert string value '%s' into a boolean\", clean)\n}", "func ParseBooleanDefaultFalse(s string) Boolean {\n\tif s == \"\" {\n\t\treturn NewBoolean(false)\n\t}\n\n\treturn NewBoolean(s == \"true\")\n}", "func (o BoolObj) Parse() ([][]string, error) {\n\treturn [][]string{\n\t\t{string(*o.Prefix)},\n\t\t{strconv.FormatBool(o.Val)},\n\t}, nil\n}", "func ConvertToBool(value string) bool {\n\tboolValue, err := strconv.ParseBool(value)\n\tif err != nil {\n\t\tutilsDiagnostics.ConvertToBoolErr(err, value)\n\t\treturn false\n\t}\n\treturn boolValue\n}", "func ParseBooleanDefaultTrue(s string) Boolean {\n\tif s == \"\" {\n\t\treturn NewBoolean(true)\n\t}\n\n\treturn NewBoolean(s == \"true\")\n}", "func (_Node *NodeFilterer) ParseLogBool(log types.Log) (*NodeLogBool, error) {\n\tevent := new(NodeLogBool)\n\tif err := _Node.contract.UnpackLog(event, \"LogBool\", log); err != nil {\n\t\treturn nil, err\n\t}\n\tevent.Raw = log\n\treturn event, nil\n}", "func StringToBool(s string, def bool) bool {\n\tv, err := strconv.ParseBool(s)\n\tif err != nil {\n\t\tlog.Warnf(\"failed to parse bool value: %s\", s)\n\t\treturn def\n\t}\n\treturn v\n}", "func (p *Parser) GetBool(path ...string) (bool, error) {\n\tval, err := p.GetString(path...)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif val == \"true\" {\n\t\treturn true, nil\n\t}\n\tif val == \"false\" {\n\t\treturn false, nil\n\t}\n\treturn false, ErrBoolParse(val)\n}", "func ReadBool(prompt string) bool {\n\tres := ReadString(prompt)\n\tb, err := strconv.ParseBool(res)\n\tif err != nil {\n\t\tlog.Fatal(\"could not parse string into bool:\", res)\n\t}\n\treturn b\n}", "func isTrue(s string) bool {\n\tv, _ := strconv.ParseBool(s)\n\treturn v\n}", "func AssertBool(s string) bool {\n\tb, err := strconv.ParseBool(s)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn b\n}", "func (_LvRecording *LvRecordingFilterer) ParseLogBool(log types.Log) (*LvRecordingLogBool, error) {\n\tevent := new(LvRecordingLogBool)\n\tif err := _LvRecording.contract.UnpackLog(event, \"LogBool\", log); err != nil {\n\t\treturn nil, err\n\t}\n\tevent.Raw = log\n\treturn event, nil\n}", "func ValBool(k string, d bool, p map[string]string) (v bool, err error) {\n\n\tbStr, ok := p[k]\n\tif !ok {\n\t\tv = d\n\t\treturn\n\t}\n\n\ttErr := fmt.Errorf(\"invalid value for the parameter %s\", k)\n\tb, err := strconv.ParseBool(bStr)\n\tif err != nil {\n\t\terr = tErr\n\t\treturn\n\t}\n\n\tv = b\n\n\treturn\n}", "func ToBoolean(str string) (bool, error) {\n\tres, err := strconv.ParseBool(str)\n\tif err != nil {\n\t\tres = false\n\t}\n\treturn res, err\n}", "func ToBoolean(str string) (bool, error) {\n\tres, err := strconv.ParseBool(str)\n\tif err != nil {\n\t\tres = false\n\t}\n\treturn res, err\n}", "func StrToBool(s string) bool {\n\tif s == \"да\" {\n\t\treturn true\n\t}\n\n\tif v, err := strconv.ParseBool(s); err == nil {\n\t\treturn v\n\t}\n\n\treturn false\n}", "func CastBool(val interface{}) (bool, bool) {\n\tswitch val.(type) {\n\tcase bool:\n\t\treturn val.(bool), true\n\tcase int:\n\t\treturn val.(int) != 0, true\n\tcase int8:\n\t\treturn val.(int8) != 0, true\n\tcase int16:\n\t\treturn val.(int16) != 0, true\n\tcase int32:\n\t\treturn val.(int32) != 0, true\n\tcase int64:\n\t\treturn val.(int64) != 0, true\n\tcase uint:\n\t\treturn val.(uint) != 0, true\n\tcase uint8:\n\t\treturn val.(uint8) != 0, true\n\tcase uint16:\n\t\treturn val.(uint16) != 0, true\n\tcase uint32:\n\t\treturn val.(uint32) != 0, true\n\tcase uint64:\n\t\treturn val.(uint64) != 0, true\n\tcase float32:\n\t\treturn val.(float32) != 0, true\n\tcase float64:\n\t\treturn val.(float64) != 0, true\n\tcase string:\n\t\tif bval, err := strconv.ParseBool(val.(string)); err != nil {\n\t\t\tif fval, ok := CastFloat(val.(string)); ok {\n\t\t\t\treturn fval != 0, true\n\t\t\t}\n\t\t\treturn false, false\n\t\t} else {\n\t\t\treturn bval, true\n\t\t}\n\t}\n\treturn false, false\n}", "func parseNullableBoolFromString(content string, aggErr *AggregateError) *bool {\n if len(content) == 0 {\n return nil\n }\n result := parseBoolFromString(content, aggErr)\n return &result\n}", "func (r *Response) Bool() (bool, error) {\n\treturn strconv.ParseBool(r.String())\n}", "func UnmarshalBool(src []byte) bool {\n\treturn src[0] != 0\n}", "func (_LvRecordableStream *LvRecordableStreamFilterer) ParseLogBool(log types.Log) (*LvRecordableStreamLogBool, error) {\n\tevent := new(LvRecordableStreamLogBool)\n\tif err := _LvRecordableStream.contract.UnpackLog(event, \"LogBool\", log); err != nil {\n\t\treturn nil, err\n\t}\n\tevent.Raw = log\n\treturn event, nil\n}", "func (v *VarSet) Bool(name string) (bool, error) {\n\tval, ok := v.load(name)\n\tif !ok {\n\t\treturn false, nil\n\t}\n\trv, err := strconv.ParseBool(val)\n\tif err != nil {\n\t\treturn false, &varNotParsableError{name: name}\n\t}\n\treturn rv, nil\n}", "func (f *FlagSet) Bool(name string) bool {\n\tvalue := f.String(name)\n\tif value != \"\" {\n\t\tval, err := strconv.ParseBool(value)\n\t\tif err != nil {\n\t\t\treturn false\n\t\t}\n\t\treturn val\n\t}\n\treturn false\n}", "func atob(str string) (value bool, err error) {\n\tv, err := strconv.ParseBool(str)\n\tif err == nil {\n\t\treturn v, nil\n\t}\n\n\tswitch str {\n\tcase \"y\", \"Y\", \"yes\", \"YES\", \"Yes\":\n\t\treturn true, nil\n\tcase \"n\", \"N\", \"no\", \"NO\", \"No\":\n\t\treturn false, nil\n\t}\n\n\t// Check extra characters, if any.\n\tboolExtra, ok := ExtraBoolString[str]\n\tif ok {\n\t\treturn boolExtra, nil\n\t}\n\n\treturn false, err // Return error of 'strconv.Atob'\n}", "func flagToBool(f string) bool {\n\tif f == \"true\" {\n\t\treturn true\n\t} else {\n\t\treturn false\n\t}\n}", "func convertToBoolLiteral(token string) (InstExpr, bool) {\n\tp := Position{0, 0, 0}\n\tswitch token {\n\tcase \"true\":\n\t\treturn InstTruePredicate{p}, false\n\tcase \"false\":\n\t\treturn InstTruePredicate{p}, false\n\tdefault:\n\t\treturn InstFalsePredicate{p}, true\n\t}\n}", "func GetBool(v interface{}) bool {\n\tswitch result := v.(type) {\n\tcase bool:\n\t\treturn result\n\tdefault:\n\t\tif d := GetString(v); d != \"\" {\n\t\t\tvalue, _ := strconv.ParseBool(d)\n\t\t\treturn value\n\t\t}\n\t}\n\treturn false\n}", "func BoolToBool(bool_ bool) bool {\n\treturn bool_\n}", "func (lx *lexer) isBool() bool {\r\n\tstr := strings.ToLower(lx.input[lx.start:lx.pos])\r\n\treturn str == \"true\" || str == \"false\" ||\r\n\t\tstr == \"on\" || str == \"off\" ||\r\n\t\tstr == \"yes\" || str == \"no\"\r\n}", "func (f *Form) Bool(param string, defaultValue bool) bool {\n\tvals, ok := f.values[param]\n\tif !ok {\n\t\treturn defaultValue\n\t}\n\tparamVal, err := strconv.ParseBool(vals[0])\n\tif err != nil {\n\t\tf.err = err\n\t\treturn defaultValue\n\t}\n\treturn paramVal\n}", "func Bool(key string, def bool) bool {\n\tif s := String(key, \"\"); s != \"\" {\n\t\tif d, err := strconv.ParseBool(s); err == nil {\n\t\t\treturn d\n\t\t} else {\n\t\t\tLog(key, err)\n\t\t}\n\t}\n\treturn def\n}", "func Bool(name string, defaultValue bool) bool {\n\tif strVal, ok := os.LookupEnv(name); ok {\n\t\tif res, err := strconv.ParseBool(strVal); err == nil {\n\t\t\treturn res\n\t\t}\n\t}\n\n\treturn defaultValue\n}", "func Bool(v string) (bool, bool, error) {\n\tswitch os.Getenv(v) {\n\tcase \"true\":\n\t\treturn true, true, nil\n\tcase \"false\":\n\t\treturn false, true, nil\n\tcase \"\":\n\t\treturn false, false, nil\n\tdefault:\n\t\treturn false, false, fmt.Errorf(\"%s must be 'true' or 'false'\", v)\n\t}\n}", "func Bool(i interface{}) bool {\n\tif i == nil {\n\t\treturn false\n\t}\n\tif v, ok := i.(bool); ok {\n\t\treturn v\n\t}\n\tif s := String(i); s != \"\" && s != \"0\" && s != \"false\" && s != \"off\" {\n\t\treturn true\n\t}\n\treturn false\n}", "func GetBool(json []byte, path ...string) (bool, error) {\n\tval, err := GetString(json, path...)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif val == \"true\" {\n\t\treturn true, nil\n\t}\n\tif val == \"false\" {\n\t\treturn false, nil\n\t}\n\treturn false, ErrBoolParse(val)\n}", "func Bool(v interface{}) *bool {\n\tswitch v.(type) {\n\tcase bool:\n\t\tval := v.(bool)\n\t\treturn &val\n\tcase int, uint, int32, int16, int8, int64, uint32, uint16, uint8, uint64, float32, float64:\n\t\tval, err := strconv.Atoi(fmt.Sprintf(\"%v\", v))\n\t\tif err != nil {\n\t\t\texception.Err(err, 500).Ctx(M{\"v\": v}).Throw()\n\t\t}\n\t\tres := false\n\t\tif val != 0 {\n\t\t\tres = true\n\t\t}\n\t\treturn &res\n\tdefault:\n\t\tval := fmt.Sprintf(\"%v\", v)\n\t\tres := false\n\t\tif val != \"\" {\n\t\t\tres = true\n\t\t}\n\t\treturn &res\n\t}\n}", "func (_Content *ContentFilterer) ParseLogBool(log types.Log) (*ContentLogBool, error) {\n\tevent := new(ContentLogBool)\n\tif err := _Content.contract.UnpackLog(event, \"LogBool\", log); err != nil {\n\t\treturn nil, err\n\t}\n\tevent.Raw = log\n\treturn event, nil\n}", "func GetBool(name string) bool {\n\t//params, err := url.ParseQuery(r.URL.RawQuery)\n\t//if err != nil {\n\t//\treturn false\n\t//}\n\n\t//value, ok := params[name]\n\t//if !ok {\n\t//\treturn false\n\t//}\n\n\tstrValue := strings.Join([]string{\"\", \"\"}, \"\")\n\tif strValue == \"\" {\n\t\treturn true\n\t}\n\n\tboolValue, err := strconv.ParseBool(strValue)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn boolValue\n}", "func FromBool(v bool) JSON {\n\tif v {\n\t\treturn TrueJSONValue\n\t}\n\treturn FalseJSONValue\n}", "func (k *Key) Bool() (bool, error) {\n\treturn parseBool(k.String())\n}", "func (g GitPushOptions) Bool(key string, def bool) bool {\n\tif val, ok := g[key]; ok {\n\t\tif b, err := strconv.ParseBool(val); err == nil {\n\t\t\treturn b\n\t\t}\n\t}\n\treturn def\n}", "func (nvp *NameValues) Bool(name string) (bool, bool) {\n\tvalue, _ := nvp.String(name)\n\treturn (value == \"true\" || value == \"yes\" || value == \"1\" || value == \"-1\" || value == \"on\"), true\n}", "func Bool(flag string, value bool, description string) *bool {\n\tvar v bool\n\tBoolVar(&v, flag, value, description)\n\treturn &v\n}", "func Bool(name string) (bool, error) {\n\tv, err := getenv(name)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn strconv.ParseBool(v)\n}", "func Str2Bool(v interface{}) (t bool) {\n\tvar i = 0\n\n\tswitch v.(type) {\n\tcase string:\n\t\ti, _ = strconv.Atoi(v.(string))\n\n\tcase int:\n\t\ti = v.(int)\n\n\tcase bool:\n\t\tif v.(bool) == true {\n\t\t\ti = 1\n\t\t} else {\n\t\t\ti = 0\n\t\t}\n\t}\n\n\tif i > 0 {\n\t\tt = true\n\t}\n\n\treturn\n}", "func ReadBool(r io.Reader) (bool, error) {\n\td := make([]byte, 1, 1)\n\t_, err := io.ReadFull(r, d)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tswitch uint8(d[0]) {\n\tcase 1:\n\t\treturn true, nil\n\tcase 0:\n\t\treturn false, nil\n\tdefault:\n\t\treturn false, ErrorInvalidValue\n\t}\n}", "func NewBool(b bool) *Value {\n\treturn &Value{b, Bool}\n}", "func CastToBool(i interface{}) (bool, error) {\n\tstr := CastOrEmpty(i)\n\treturn strconv.ParseBool(str)\n}", "func NewBool(value bool) *Value {\n\treturn &Value{kind: kindBool, boolContent: value}\n}", "func (_LvStreamRightsHolder *LvStreamRightsHolderFilterer) ParseLogBool(log types.Log) (*LvStreamRightsHolderLogBool, error) {\n\tevent := new(LvStreamRightsHolderLogBool)\n\tif err := _LvStreamRightsHolder.contract.UnpackLog(event, \"LogBool\", log); err != nil {\n\t\treturn nil, err\n\t}\n\tevent.Raw = log\n\treturn event, nil\n}", "func ExampleBool() {\n\n\t// Bool conversion from other bool values will be returned without\n\t// modification.\n\tfmt.Println(conv.Bool(true))\n\tfmt.Println(conv.Bool(false))\n\n\t// Bool conversion from strings consider the following values true:\n\t// \"t\", \"T\", \"true\", \"True\", \"TRUE\",\n\t// \t \"y\", \"Y\", \"yes\", \"Yes\", \"YES\", \"1\"\n\t//\n\t// It considers the following values false:\n\t// \"f\", \"F\", \"false\", \"False\", \"FALSE\",\n\t// \"n\", \"N\", \"no\", \"No\", \"NO\", \"0\"\n\tfmt.Println(conv.Bool(\"T\"))\n\tfmt.Println(conv.Bool(\"False\"))\n\n\t// Bool conversion from other supported types will return true unless it is\n\t// the zero value for the given type.\n\tfmt.Println(conv.Bool(int64(123)))\n\tfmt.Println(conv.Bool(int64(0)))\n\tfmt.Println(conv.Bool(time.Duration(123)))\n\tfmt.Println(conv.Bool(time.Duration(0)))\n\tfmt.Println(conv.Bool(time.Now()))\n\tfmt.Println(conv.Bool(time.Time{}))\n\n\t// All other types will return false.\n\tfmt.Println(conv.Bool(struct{ string }{\"\"}))\n\n\t// Output:\n\t// true <nil>\n\t// false <nil>\n\t// true <nil>\n\t// false <nil>\n\t// true <nil>\n\t// false <nil>\n\t// true <nil>\n\t// false <nil>\n\t// true <nil>\n\t// false <nil>\n\t// false cannot convert struct { string }{string:\"\"} (type struct { string }) to bool\n}", "func validateBoolParam(ctx *HttpContext, param *HttpParam) {\n\n\tparam.Raw = retrieveParamValue(ctx, param).(string)\n\n\tif len(param.Raw) == 0 && param.Required {\n\t\tappendInvalidErrorCode(ctx, param)\n\t\treturn\n\t}\n\n\tif len(param.Raw) == 0 { return }\n\n\tif val, err := strconv.ParseBool(param.Raw); err != nil {\n\t\tappendInvalidErrorCode(ctx, param)\n\t} else {\n\t\tparam.setPresentValue(val)\n\t}\n}", "func getBoolVal(input string) bool {\n\tinput = strings.ToLower(input)\n\tif input == \"yes\" || input == \"true\" {\n\t\treturn true\n\t}\n\treturn false\n}", "func (cfg *Config) Bool(name string) (bool, error) {\n\tv, ok := cfg.findLast(name)\n\tif !ok {\n\t\treturn false, fmt.Errorf(\"config %s: not found\", name)\n\t}\n\tif v == nil {\n\t\t// No equals sign, which implies true.\n\t\treturn true, nil\n\t}\n\tb, ok := parseBool(v)\n\tif !ok {\n\t\treturn false, fmt.Errorf(\"config %s: cannot parse %q as a bool\", name, v)\n\t}\n\treturn b, nil\n}", "func StringToBool(str String) Bool {\n\tv := &stringToBool{from: str}\n\tstr.AddListener(v)\n\treturn v\n}", "func Bool(v interface{}, defaults ...bool) (b bool) {\n\tswitch tv := v.(type) {\n\tcase nil:\n\t\tif 1 < len(defaults) {\n\t\t\tb = defaults[1]\n\t\t}\n\tcase bool:\n\t\tb = tv\n\tcase string:\n\t\tvar err error\n\t\tif 1 < len(defaults) {\n\t\t\tb = defaults[1]\n\t\t} else if b, err = strconv.ParseBool(tv); err != nil {\n\t\t\tif 0 < len(defaults) {\n\t\t\t\tb = defaults[0]\n\t\t\t}\n\t\t}\n\tcase gen.Bool:\n\t\tb = bool(tv)\n\tcase gen.String:\n\t\tvar err error\n\t\tif 1 < len(defaults) {\n\t\t\tb = defaults[1]\n\t\t} else if b, err = strconv.ParseBool(string(tv)); err != nil {\n\t\t\tif 0 < len(defaults) {\n\t\t\t\tb = defaults[0]\n\t\t\t}\n\t\t}\n\tdefault:\n\t\tif 0 < len(defaults) {\n\t\t\tb = defaults[0]\n\t\t}\n\t}\n\treturn\n}", "func Bool(key string) (value bool, err error) {\n\treturn strconv.ParseBool(Get(key))\n}", "func ConvertToBool(value interface{}) (bool, bool) {\n\tif v, ok := value.(bool); ok {\n\t\treturn v, ok\n\t}\n\n\t// try converting \"true\" \"false\"\n\tif v, ok := value.(string); ok {\n\t\tif strings.ToLower(v) == \"true\" {\n\t\t\treturn true, true\n\t\t} else if strings.ToLower(v) == \"false\" {\n\t\t\treturn false, true\n\t\t}\n\t}\n\n\t// try convert as number\n\tv, ok := ConvertToInt8(value)\n\tif ok {\n\t\tif v == 1 {\n\t\t\treturn true, true\n\t\t} else if v == 0 {\n\t\t\treturn false, true\n\t\t}\n\t}\n\n\treturn false, false\n}", "func (fa formulaArg) ToBool() formulaArg {\n\tvar b bool\n\tvar err error\n\tswitch fa.Type {\n\tcase ArgString:\n\t\tb, err = strconv.ParseBool(fa.String)\n\t\tif err != nil {\n\t\t\treturn newErrorFormulaArg(formulaErrorVALUE, err.Error())\n\t\t}\n\tcase ArgNumber:\n\t\tif fa.Boolean && fa.Number == 1 {\n\t\t\tb = true\n\t\t}\n\t}\n\treturn newBoolFormulaArg(b)\n}", "func parseNullableBool(content []byte, aggErr *AggregateError) *bool {\n if len(content) == 0 {\n return nil\n }\n result := parseBool(content, aggErr)\n return &result\n}", "func Bool(name string) bool {\n\treturn strings.EqualFold(String(name), \"true\")\n}", "func typeConvertBool(i interface{}) bool {\n\tif i == nil {\n\t\treturn false\n\t}\n\tif v, ok := i.(bool); ok {\n\t\treturn v\n\t}\n\tif s := typeConvertString(i); s != \"\" && s != \"0\" && s != \"false\" && s != \"off\" {\n\t\treturn true\n\t}\n\treturn false\n}", "func Bool(b *bool, title, id, class string, valid Validator) (jquery.JQuery, error) {\n\tj := jq(\"<input>\").AddClass(ClassPrefix + \"-bool\").AddClass(class)\n\tj.SetAttr(\"type\", \"checkbox\")\n\tj.SetAttr(\"title\", title).SetAttr(\"id\", id)\n\tj.SetProp(\"checked\", *b)\n\tj.SetData(\"prev\", *b)\n\tj.Call(jquery.CHANGE, func(event jquery.Event) {\n\t\tval := event.Target.Get(\"checked\").String()\n\t\tbNew, e := strconv.ParseBool(val)\n\t\tif e != nil {\n\t\t\t// Theorectially impossible\n\t\t\tpanic(fmt.Sprintf(\"value '%s' has invalid type, expected bool\", val))\n\t\t}\n\t\tif valid != nil && !valid.Validate(bNew) {\n\t\t\tbNew = j.Data(\"prev\").(bool)\n\t\t\tj.SetProp(\"checked\", bNew)\n\t\t}\n\t\t*b = bNew\n\t\tj.SetData(\"prev\", bNew)\n\t})\n\treturn j, nil\n}", "func (v *Value) AsBool(dv bool) bool {\n\tif v.IsUndefined() {\n\t\treturn dv\n\t}\n\tswitch tv := v.raw.(type) {\n\tcase string:\n\t\tb, err := strconv.ParseBool(tv)\n\t\tif err != nil {\n\t\t\treturn dv\n\t\t}\n\t\treturn b\n\tcase int:\n\t\treturn tv == 1\n\tcase float64:\n\t\treturn tv == 1.0\n\tcase bool:\n\t\treturn tv\n\tcase time.Time:\n\t\treturn tv.UnixNano() > 0\n\tcase time.Duration:\n\t\treturn tv.Nanoseconds() > 0\n\t}\n\treturn dv\n}", "func ToBool(v interface{}, def bool) bool {\r\n\tif b, ok := v.(bool); ok {\r\n\t\treturn b\r\n\t}\r\n\tif i, ok := v.(int); ok {\r\n\t\treturn i > 0\r\n\t}\r\n\tif i, ok := v.(float64); ok {\r\n\t\treturn i > 0\r\n\t}\r\n\tif i, ok := v.(float32); ok {\r\n\t\treturn i > 0\r\n\t}\r\n\tif ss, ok := v.([]string); ok {\r\n\t\tv = ss[0]\r\n\t}\r\n\tif s, ok := v.(string); ok {\r\n\t\tif s == \"on\" {\r\n\t\t\treturn true\r\n\t\t}\r\n\t\tif s == \"off\" || s == \"\" {\r\n\t\t\treturn false\r\n\t\t}\r\n\t\tif b, err := strconv.ParseBool(s); err == nil {\r\n\t\t\treturn b\r\n\t\t}\r\n\t}\r\n\r\n\treturn def\r\n\r\n}", "func (me StringData) toBoolean() bool {\n\tif b, err := strconv.ParseBool(me.val); err != nil {\n\t\treturn b\n\t}\n\treturn false\n}" ]
[ "0.87281555", "0.8482916", "0.84199905", "0.83869666", "0.8338142", "0.8029671", "0.8000297", "0.797828", "0.7886067", "0.76948595", "0.7463733", "0.73488927", "0.73259276", "0.73199666", "0.7224256", "0.7078567", "0.70561355", "0.70538497", "0.70096534", "0.6897837", "0.6872132", "0.68547046", "0.6813949", "0.67767596", "0.67498904", "0.6690416", "0.66801584", "0.66245246", "0.6618238", "0.654166", "0.65407103", "0.64977396", "0.6476334", "0.6444846", "0.63952625", "0.63470906", "0.63280076", "0.63122827", "0.6281464", "0.62742704", "0.62565434", "0.62197787", "0.6194626", "0.6175407", "0.6173472", "0.6170293", "0.61675256", "0.616637", "0.616637", "0.61423135", "0.6132946", "0.61320126", "0.61227506", "0.6121923", "0.6121472", "0.61137193", "0.6098659", "0.60847294", "0.60644823", "0.60457087", "0.6042368", "0.6034545", "0.60306436", "0.6019492", "0.60193706", "0.6015828", "0.6011349", "0.59843236", "0.5947753", "0.5947328", "0.59431374", "0.59215575", "0.5901222", "0.5900243", "0.58926666", "0.5881693", "0.5872805", "0.58632827", "0.5859304", "0.5854019", "0.5834267", "0.58304983", "0.58301824", "0.58259135", "0.5817926", "0.5815218", "0.5812025", "0.58070946", "0.5805372", "0.5801636", "0.57936245", "0.57902825", "0.57851744", "0.577905", "0.5774631", "0.57643616", "0.5763901", "0.57625115", "0.57154036", "0.5711507" ]
0.89225525
0
ParseInt uses strconv.ParseInt with base == 0 (auto) and bitsize = 32 to parse operand as an int.
func ParseInt(operand string) (value int, err error) { i64, err := strconv.ParseInt(operand, 0, 32) return int(i64), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ParseInt(s String, base Int) Int {\n\tresult, _ := strconv.ParseInt(string(s), int(base), 0)\n\treturn Int(result)\n}", "func ParseInt(key string, base int, bitSize int) (int64, error) {\n\tfnParseInt := \"ParseInt\"\n\n\tval := os.Getenv(key)\n\n\tif val == \"\" {\n\t\treturn 0, notFoundError(key, fnParseInt)\n\t}\n\n\tparsedVal, err := strconv.ParseInt(val, base, bitSize)\n\n\tif err != nil {\n\t\treturn 0, &EnvError{fnParseInt, key, err}\n\t}\n\n\treturn parsedVal, err\n}", "func main() {\n\t// strconv.ParseInt(input, decimal/binary, bits)\n\tval, _ := strconv.ParseInt(os.Args[1], 10, 8)\n\tfmt.Println(\"int8 value is:\", int8(val))\n\n\tval, _ = strconv.ParseInt(os.Args[2], 10, 16)\n\tfmt.Println(\"int16 value is:\", int16(val))\n\n\tval, _ = strconv.ParseInt(os.Args[3], 10, 32)\n\tfmt.Println(\"int32 value is:\", int32(val))\n\n\tval, _ = strconv.ParseInt(os.Args[4], 10, 64)\n\tfmt.Println(\"int64 value is:\", int64(val))\n\n\tval, _ = strconv.ParseInt(os.Args[5], 2, 8)\n\tfmt.Printf(\"%s is: %d\\n\", os.Args[5], int8(val))\n}", "func ParseToInt(data string) (int64, error) {\n\treturn strconv.ParseInt(data, 0, 64)\n}", "func ParseInt(str string) int {\n\ti, _ := strconv.Atoi(str)\n\treturn i\n}", "func ParseInt(s string) (int, error) {\n\tif s = strings.TrimSpace(s); s != \"\" {\n\t\treturn strconv.Atoi(s)\n\t}\n\treturn 0, nil\n}", "func (t *numberUtils) ParseInt(s string) int {\n\tparsed, err := strconv.ParseInt(s, 0, 0)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn int(parsed)\n}", "func (p *parser) ParseInt(v string) int {\n\n\tif p.err != nil {\n\t\treturn 0\n\t}\n\n\tvalue, err := strconv.Atoi(v)\n\tif err != nil {\n\t\tp.err = err\n\t\treturn 0\n\t}\n\n\treturn value\n}", "func parseIntEx(s string, bitSize int) (int64, error) {\n\tif s[0:2] == \"0b\" {\n\t\treturn strconv.ParseInt(s[2:], 2, bitSize)\n\t} else {\n\t\treturn strconv.ParseInt(s, 0, bitSize)\n\t}\n}", "func parseIntegerValue(input string) (res int) {\n\tvar err error\n\n\tres, err = strconv.Atoi(input)\n\tif err != nil {\n\t\t// negative values must be dropped here\n\t\treturn\n\t}\n\treturn\n}", "func ParseInt(s string) (int64, error) {\n\tisNegative := strings.HasPrefix(s, \"-\")\n\tif isNegative {\n\t\ts = s[1:]\n\t}\n\tsum, err := ParseUint(s)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tif isNegative {\n\t\tif sum > -math.MinInt64 {\n\t\t\treturn 0, ErrOverflow\n\t\t}\n\t\treturn -int64(sum), nil\n\t}\n\tif sum > math.MaxInt64 {\n\t\treturn 0, ErrOverflow\n\t}\n\treturn int64(sum), nil\n}", "func parseInt(b []byte, t reflect.Type) (int64, []byte, error) {\n\tvar value int64\n\tvar count int\n\n\tif len(b) == 0 {\n\t\treturn 0, b, syntaxError(b, \"cannot decode integer from an empty input\")\n\t}\n\n\tif b[0] == '-' {\n\t\tconst max = math.MinInt64\n\t\tconst lim = max / 10\n\n\t\tif len(b) == 1 {\n\t\t\treturn 0, b, syntaxError(b, \"cannot decode integer from '-'\")\n\t\t}\n\n\t\tif len(b) > 2 && b[1] == '0' && '0' <= b[2] && b[2] <= '9' {\n\t\t\treturn 0, b, syntaxError(b, \"invalid leading character '0' in integer\")\n\t\t}\n\n\t\tfor _, d := range b[1:] {\n\t\t\tif !(d >= '0' && d <= '9') {\n\t\t\t\tif count == 0 {\n\t\t\t\t\tb, err := inputError(b, t)\n\t\t\t\t\treturn 0, b, err\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif value < lim {\n\t\t\t\treturn 0, b, unmarshalOverflow(b, t)\n\t\t\t}\n\n\t\t\tvalue *= 10\n\t\t\tx := int64(d - '0')\n\n\t\t\tif value < (max + x) {\n\t\t\t\treturn 0, b, unmarshalOverflow(b, t)\n\t\t\t}\n\n\t\t\tvalue -= x\n\t\t\tcount++\n\t\t}\n\n\t\tcount++\n\t} else {\n\t\tconst max = math.MaxInt64\n\t\tconst lim = max / 10\n\n\t\tif len(b) > 1 && b[0] == '0' && '0' <= b[1] && b[1] <= '9' {\n\t\t\treturn 0, b, syntaxError(b, \"invalid leading character '0' in integer\")\n\t\t}\n\n\t\tfor _, d := range b {\n\t\t\tif !(d >= '0' && d <= '9') {\n\t\t\t\tif count == 0 {\n\t\t\t\t\tb, err := inputError(b, t)\n\t\t\t\t\treturn 0, b, err\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tx := int64(d - '0')\n\n\t\t\tif value > lim {\n\t\t\t\treturn 0, b, unmarshalOverflow(b, t)\n\t\t\t}\n\n\t\t\tif value *= 10; value > (max - x) {\n\t\t\t\treturn 0, b, unmarshalOverflow(b, t)\n\t\t\t}\n\n\t\t\tvalue += x\n\t\t\tcount++\n\t\t}\n\t}\n\n\tif count < len(b) {\n\t\tswitch b[count] {\n\t\tcase '.', 'e', 'E': // was this actually a float?\n\t\t\tv, r, err := parseNumber(b)\n\t\t\tif err != nil {\n\t\t\t\tv, r = b[:count+1], b[count+1:]\n\t\t\t}\n\t\t\treturn 0, r, unmarshalTypeError(v, t)\n\t\t}\n\t}\n\n\treturn value, b[count:], nil\n}", "func ParseInt(buf []byte) (int, int, error) {\n\tif len(buf) < 2 || buf[0]>>4 != 0xc {\n\t\treturn 0, 0, errors.New(\"can't parse int, invalid byte buf\")\n\t}\n\n\tvar i, n int\n\n\tswitch buf[0] {\n\tcase 0xc8:\n\t\ti = int(int8(buf[1]))\n\t\tn = 2\n\tcase 0xc9:\n\t\ti = int(int16(binary.BigEndian.Uint16(buf[1:3])))\n\t\tn = 3\n\tcase 0xca:\n\t\ti = int(int32(binary.BigEndian.Uint32(buf[1:5])))\n\t\tn = 5\n\tcase 0xcb:\n\t\ti = int(int64(binary.BigEndian.Uint64(buf[1:9])))\n\t\tn = 9\n\t}\n\n\treturn i, n, nil\n}", "func Int(s string) int {\n\tif s == \"\" {\n\t\treturn 0\n\t}\n\n\tresult, err := strconv.ParseInt(s, 10, 64)\n\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn int(result)\n}", "func ParseInt(c CFG.JobConfig,\n\tinstance *INST.Instance, member, message string, defaultVal int) error {\n\tstr, _ := reflections.GetField(c, member)\n\tif str == \"\" {\n\t\treflections.SetField(instance, member, defaultVal)\n\t} else if num, err := strconv.Atoi(str.(string)); err != nil {\n\t\treturn fmt.Errorf(message, c)\n\t} else {\n\t\treflections.SetField(instance, member, num)\n\t}\n\treturn nil\n}", "func (p *parser) parseInteger() (int, error) {\n\ti := p.i\n\tstart := i\n\tfor i < len(p.s) && '0' <= p.s[i] && p.s[i] <= '9' {\n\t\ti++\n\t}\n\tif i == start {\n\t\treturn 0, errors.New(\"expected integer, but didn't find it\")\n\t}\n\tp.i = i\n\n\tval, err := strconv.Atoi(p.s[start:i])\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn val, nil\n}", "func (p *parser) parseInteger() (int, error) {\n\ti := p.i\n\tstart := i\n\tfor i < len(p.s) && '0' <= p.s[i] && p.s[i] <= '9' {\n\t\ti++\n\t}\n\tif i == start {\n\t\treturn 0, errors.New(\"expected integer, but didn't find it\")\n\t}\n\tp.i = i\n\n\tval, err := strconv.Atoi(p.s[start:i])\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn val, nil\n}", "func ParseInt(b []byte) (int64, error) {\n\tvar val int64\n\n\tif len(b) == 0 {\n\t\treturn 0, errorInvalidUint64(b)\n\t}\n\n\tif b[0] == '-' {\n\t\tconst max = Int64Min\n\t\tconst lim = max / 10\n\n\t\tif b = b[1:]; len(b) == 0 {\n\t\t\treturn 0, errorInvalidUint64(b)\n\t\t}\n\n\t\tfor _, d := range b {\n\t\t\tif !(d >= '0' && d <= '9') {\n\t\t\t\treturn 0, errorInvalidInt64(b)\n\t\t\t}\n\n\t\t\tif val < lim {\n\t\t\t\treturn 0, errorOverflowInt64(b)\n\t\t\t}\n\n\t\t\tval *= 10\n\t\t\tx := int64(d - '0')\n\n\t\t\tif val < (max + x) {\n\t\t\t\treturn 0, errorOverflowInt64(b)\n\t\t\t}\n\n\t\t\tval -= x\n\t\t}\n\t} else {\n\t\tconst max = Int64Max\n\t\tconst lim = max / 10\n\n\t\tfor _, d := range b {\n\t\t\tif !(d >= '0' && d <= '9') {\n\t\t\t\treturn 0, errorInvalidInt64(b)\n\t\t\t}\n\t\t\tx := int64(d - '0')\n\n\t\t\tif val > lim {\n\t\t\t\treturn 0, errorOverflowInt64(b)\n\t\t\t}\n\n\t\t\tif val *= 10; val > (max - x) {\n\t\t\t\treturn 0, errorOverflowInt64(b)\n\t\t\t}\n\n\t\t\tval += x\n\t\t}\n\t}\n\n\treturn val, nil\n}", "func ParseInt32(strval string) (int32, error) {\n\tval, err := parseInt(strval, 32)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int32(val), nil\n}", "func ParseInt(text string, defaultValue int) int {\n\tif text == \"\" {\n\t\treturn defaultValue\n\t}\n\n\tnum, err := strconv.Atoi(text)\n\tif err != nil {\n\t\treturn defaultValue\n\t}\n\treturn num\n}", "func ParseInteger(s string) (Number, error) {\n\tif res, ok := new(big.Int).SetString(s, 0); ok {\n\t\tif res.IsInt64() {\n\t\t\treturn Integer(res.Int64()), nil\n\t\t}\n\t\treturn (*BigInt)(res), nil\n\t}\n\treturn nil, fmt.Errorf(ErrExpectedInteger, s)\n}", "func ParseInt(from []byte) Key {\n\tv, err := strconv.ParseInt(string(from), 10, 64)\n\tif err != nil {\n\t\treturn NilKey{}\n\t} else {\n\t\treturn IntKey(v)\n\t}\n}", "func parseInt(str string) int {\n\ti, err := strconv.Atoi(str)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn i\n}", "func (t *numberUtils) ParseInt32(s string) int32 {\n\tparsed, err := strconv.ParseInt(s, 0, 32)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn int32(parsed)\n}", "func safeParseInt(s string, fallback int) int {\n\trslt, error := strconv.Atoi(s)\n\tif error == nil {\n\t\treturn rslt\n\t}\n\n\treturn fallback\n}", "func asInt(param string) (int64, error) {\n\ti, err := strconv.ParseInt(param, 0, 32)\n\tif err != nil {\n\t\treturn 0, validator.ErrBadParameter\n\t}\n\treturn i, nil\n}", "func ParseInt32(length uint32, data []byte) uint32 {\n\treturn ParseUint32(length, data)\n}", "func (p *parser) parseInt(annotations []Symbol, base intBase) Int {\n\ttext := p.next().Val\n\t// An empty slice of bytes is not a valid int, so we're going to make the assumption\n\t// that we can check the first element of the text slice.\n\treturn Int{annotations: annotations, isSet: true, base: base, isNegative: text[0] == '-', text: text}\n}", "func ParseStringToInt(str string) (int, error) {\n\ti64, err := strconv.ParseInt(str, 10, 0)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int(i64), nil\n}", "func parseBase128Int(bytes []byte, initOffset int) (ret, offset int, err error) {\n\toffset = initOffset\n\tvar ret64 int64\n\tfor shifted := 0; offset < len(bytes); shifted++ {\n\t\t// 5 * 7 bits per byte == 35 bits of data\n\t\t// Thus the representation is either non-minimal or too large for an int32\n\t\tif shifted == 5 {\n\t\t\terr = StructuralError{\"base 128 integer too large\"}\n\t\t\treturn\n\t\t}\n\t\tret64 <<= 7\n\t\tb := bytes[offset]\n\t\tret64 |= int64(b & 0x7f)\n\t\toffset++\n\t\tif b&0x80 == 0 {\n\t\t\tret = int(ret64)\n\t\t\t// Ensure that the returned value fits in an int on all platforms\n\t\t\tif ret64 > math.MaxInt32 {\n\t\t\t\terr = StructuralError{\"base 128 integer too large\"}\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t}\n\terr = SyntaxError{\"truncated base 128 integer\"}\n\treturn\n}", "func strToInt(input string) int {\n\toutput, _ := strconv.Atoi(input)\n\treturn output\n}", "func (v Value) Int(bitSize int) (int64, error) {\n\ts, err := v.getIntStr()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tn, err := strconv.ParseInt(s, 10, bitSize)\n\tif err != nil {\n\t\treturn 0, v.newError(\"%v\", err)\n\t}\n\treturn n, nil\n}", "func intFromString(s string) int {\n\tvalue := strings.Replace(s, \",\", \"\", -1)\n\tvalue = strings.TrimSpace(value)\n\tv, err := strconv.ParseInt(value, 10, 64)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn 0\n\t}\n\treturn int(v)\n}", "func ParseInt(val interface{}) (int, error) {\n\tnumber, err := ParseFloat64(val)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int(number), nil\n}", "func StrToInt(s string) int {\n\tif n, err := strconv.Atoi(s); err == nil {\n\t\treturn n\n\t}\n\treturn 0\n}", "func (input *PuzzleInput) ParseAsInt() int {\n\tnum, err := strconv.Atoi(string(input.Data))\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn num\n}", "func parseIntFromString(content string, aggErr *AggregateError) int {\n result, err := strconv.Atoi(content)\n if err != nil {\n aggErr.Append(err)\n }\n return result\n}", "func String2Int(s string) int {\n\ti, _ := strconv.Atoi(s)\n\treturn i\n}", "func parseInt(s string, base int) (Value, error) {\n\tvar n int64\n\tvar err error\n\tvar cutoff, maxVal int64\n\tvar sign bool\n\ti := 0\n\n\tif len(s) < 1 {\n\t\terr = strconv.ErrSyntax\n\t\tgoto Error\n\t}\n\n\tswitch s[0] {\n\tcase '-':\n\t\tsign = true\n\t\ts = s[1:]\n\tcase '+':\n\t\ts = s[1:]\n\t}\n\n\tif len(s) < 1 {\n\t\terr = strconv.ErrSyntax\n\t\tgoto Error\n\t}\n\n\t// Look for hex prefix.\n\tif s[0] == '0' && len(s) > 1 && (s[1] == 'x' || s[1] == 'X') {\n\t\tif base == 0 || base == 16 {\n\t\t\tbase = 16\n\t\t\ts = s[2:]\n\t\t}\n\t}\n\n\tswitch {\n\tcase len(s) < 1:\n\t\terr = strconv.ErrSyntax\n\t\tgoto Error\n\n\tcase 2 <= base && base <= 36:\n\t// valid base; nothing to do\n\n\tcase base == 0:\n\t\t// Look for hex prefix.\n\t\tswitch {\n\t\tcase s[0] == '0' && len(s) > 1 && (s[1] == 'x' || s[1] == 'X'):\n\t\t\tif len(s) < 3 {\n\t\t\t\terr = strconv.ErrSyntax\n\t\t\t\tgoto Error\n\t\t\t}\n\t\t\tbase = 16\n\t\t\ts = s[2:]\n\t\tdefault:\n\t\t\tbase = 10\n\t\t}\n\n\tdefault:\n\t\terr = errors.New(\"invalid base \" + strconv.Itoa(base))\n\t\tgoto Error\n\t}\n\n\t// Cutoff is the smallest number such that cutoff*base > maxInt64.\n\t// Use compile-time constants for common cases.\n\tswitch base {\n\tcase 10:\n\t\tcutoff = math.MaxInt64/10 + 1\n\tcase 16:\n\t\tcutoff = math.MaxInt64/16 + 1\n\tdefault:\n\t\tcutoff = math.MaxInt64/int64(base) + 1\n\t}\n\n\tmaxVal = math.MaxInt64\n\tfor ; i < len(s); i++ {\n\t\tif n >= cutoff {\n\t\t\t// n*base overflows\n\t\t\treturn parseLargeInt(float64(n), s[i:], base, sign)\n\t\t}\n\t\tv := digitVal(s[i])\n\t\tif v >= base {\n\t\t\tbreak\n\t\t}\n\t\tn *= int64(base)\n\n\t\tn1 := n + int64(v)\n\t\tif n1 < n || n1 > maxVal {\n\t\t\t// n+v overflows\n\t\t\treturn parseLargeInt(float64(n)+float64(v), s[i+1:], base, sign)\n\t\t}\n\t\tn = n1\n\t}\n\n\tif i == 0 {\n\t\terr = strconv.ErrSyntax\n\t\tgoto Error\n\t}\n\n\tif sign {\n\t\tn = -n\n\t}\n\treturn intToValue(n), nil\n\nError:\n\treturn _NaN, err\n}", "func unmarshalStringyInt(b []byte) (i int64, err error) {\n\ts := \"\"\n\ts, err = unmarshalStringyValue(b)\n\tif err != nil {\n\t\treturn\n\t}\n\ti, err = strconv.ParseInt(s, 10, 64)\n\treturn\n}", "func Str2Int(str string) (int, error) {\n\treturn strconv.Atoi(str)\n}", "func StringToInt(s string) (int64, error) {\n\treturn strconv.ParseInt(s, 10, 64)\n}", "func StringToInt(s string) int {\n\ti, _ := strconv.Atoi(s)\n\treturn i\n}", "func strToInt(str string) int {\n\t// Remove space\n\tif strings.Contains(str, \" \") {\n\t\tstr = strings.ReplaceAll(str, \" \", \"\")\n\t}\n\ti, err := strconv.Atoi(str)\n\tif err != nil {\n\t\tErrorLog(\"cannot parse string to int: %s\", str)\n\t\treturn 0\n\t}\n\tInfoLogV1(\"strToInt %d\", i)\n\treturn i\n}", "func StringToInt(s string) int {\n\tintValue, _ := strconv.Atoi(s)\n\treturn intValue\n}", "func parseInt(s string) (string, int, bool) {\n\ts = strings.TrimSpace(s)\n\tindex := 0\n\tnum := 0\n\tneg := false\n\tif s[0] == '-' {\n\t\tneg = true\n\t\tindex++\n\t}\n\tfor ; s[index] >= '0' && s[index] <= '9'; index++ {\n\t\tnum *= 10\n\t\tnum += int(s[index] - '0')\n\t}\n\tif neg {\n\t\tnum *= -1\n\t}\n\t// Did s start with a number?\n\tif index == 0 || (index == 1 && neg) {\n\t\treturn s, 0, false\n\t}\n\treturn s[index:], num, true\n}", "func toInt(rawInt string) int {\n\tparsed, _ := strconv.Atoi(strings.Replace(strings.Replace(rawInt, \"$\", \"\", -1), \",\", \"\", -1))\n\treturn parsed\n}", "func parseInt(content []byte, aggErr *AggregateError) int {\n result, err := strconv.Atoi(string(content))\n if err != nil {\n aggErr.Append(err)\n }\n return result\n}", "func NewIntFromString(s string) (res Int, ok bool) {\n\ti, ok := newIntegerFromString(s)\n\tif !ok {\n\t\treturn\n\t}\n\t// Check overflow\n\tif i.BitLen() > 255 {\n\t\tok = false\n\t\treturn\n\t}\n\treturn Int{i}, true\n}", "func ParseTinyInt(b byte) (int, error) {\n\tif b > 0x7f {\n\t\treturn 0, errors.New(\"expected tiny-int\")\n\t}\n\treturn int(b), nil\n}", "func toInt(str string) (int64, error) {\n\tres, err := strconv.ParseInt(str, 0, 64)\n\tif err != nil {\n\t\tres = 0\n\t}\n\treturn res, err\n}", "func Int(a string) int {\n\tnum, err := strconv.Atoi(a)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn num\n}", "func stringToInt(s string) (int64, NumberType) {\n\tn, f, tp := StringToNumber(s)\n\tswitch tp {\n\tcase IsInt:\n\t\treturn n, IsInt\n\tcase IsFloat:\n\t\treturn FloatToInt(f)\n\t}\n\treturn 0, NaN\n}", "func mustParseInt(expr string) (uint, error) {\n\tnum, err := strconv.Atoi(expr)\n\tif err != nil {\n\t\treturn 0, fmt.Errorf(\"Failed to parse int from %s: %s\", expr, err)\n\t}\n\tif num < 0 {\n\t\treturn 0, fmt.Errorf(\"Negative number (%d) not allowed: %s\", num, expr)\n\t}\n\n\treturn uint(num), nil\n}", "func (p Parser) Int(ctx context.Context) (*int, error) {\n\tvalue, err := p.Source.String(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif value == nil {\n\t\treturn nil, nil\n\t}\n\tparsed, err := strconv.Atoi(*value)\n\tif err != nil {\n\t\treturn nil, ex.New(err)\n\t}\n\treturn &parsed, nil\n}", "func StringAsInteger(s string) int {\n\tif s == \"\" {\n\t\treturn 0\n\t}\n\n\tif i, err := strconv.ParseInt(s, 10, 32); err == nil {\n\t\treturn int(i)\n\t}\n\n\treturn 0\n}", "func atoi(s string) (int, bool) {\n\tsLen := len(s)\n\tif intSize == 32 && (0 < sLen && sLen < 10) || intSize == 64 && (0 < sLen && sLen < 19) {\n\t\tn := 0\n\t\tfor _, ch := range []byte(s) {\n\t\t\tch -= '0'\n\t\t\tif ch > 9 {\n\t\t\t\treturn 0, false\n\t\t\t}\n\t\t\tn = n*10 + int(ch)\n\t\t}\n\n\t\treturn n, true\n\t}\n\n\treturn 0, false\n}", "func coerceToInt(s string) int {\n\tn, err := strconv.Atoi(s)\n\tif err != nil {\n\t\tpanic(errors.Wrapf(err, \"invalid number %q\", s))\n\t}\n\treturn n\n}", "func ParseInt(value int) []int {\n\tstr := strconv.Itoa(value)\n\tvar result []int\n\tfor i := 0; i < len(str); i++ {\n\t\tvalue, err := strconv.Atoi(str[i : i+1])\n\t\tif err != nil {\n\t\t\tprintln(\"FEHLER\")\n\t\t}\n\t\tresult = append(result, value)\n\t}\n\treturn result\n}", "func Intval(s string) int {\n\n\tif s == \"\" {\n\t\treturn 0\n\t}\n\n\ti, err := strconv.Atoi(s)\n\tif err != nil {\n\t\treturn 0\n\t}\n\treturn i\n}", "func (s *Str) Int() int {\n\tval, err := strconv.Atoi(s.val)\n\tif err != nil {\n\t\ts.err = err\n\t}\n\treturn val\n}", "func (r *memcachedScraper) parseInt(key, value string) (int64, bool) {\n\ti, err := strconv.ParseInt(value, 10, 64)\n\tif err != nil {\n\t\tr.logInvalid(\"int\", key, value)\n\t\treturn 0, false\n\t}\n\treturn i, true\n}", "func Str2int(Num string)int{\n s, _ := strconv.Atoi(Num)\n return s\n}", "func parseInteger(part expreduceapi.Ex) (value int64, isInteger bool) {\n\tinteger, isInteger := part.(*atoms.Integer)\n\tif isInteger {\n\t\treturn integer.Val.Int64(), true\n\t}\n\treturn 0, false\n}", "func parseInts(in string) (int, int) {\n\tparts := strings.Split(in, \"-\")\n\tif len(parts) != 2 {\n\t\tpanic(in)\n\t}\n\n\ti, err := strconv.Atoi(parts[0])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tj, err := strconv.Atoi(parts[1])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn i, j\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func atoi(in string) int {\n\tn, err := strconv.Atoi(in)\n\tmust(err)\n\treturn n\n}", "func ParseInt8(strval string) (int8, error) {\n\tval, err := parseInt(strval, 8)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int8(val), nil\n}", "func mustParseInt32(val string) int32 {\n\tvv, _ := strconv.ParseInt(val, 10, 32)\n\treturn int32(vv)\n}", "func ParseQueryInt(param string, request *http.Request, params imageserver.Params) error {\n\ts := request.URL.Query().Get(param)\n\tif s == \"\" {\n\t\treturn nil\n\t}\n\ti, err := strconv.Atoi(s)\n\tif err != nil {\n\t\treturn newParseTypeParamError(param, \"string\", err)\n\t}\n\tparams.Set(param, i)\n\treturn nil\n}", "func ParseInt16(strval string) (int16, error) {\n\tval, err := parseInt(strval, 16)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int16(val), nil\n}", "func parseInstruction(n *big.Int) (int, int, int, int) {\n\tinstruction := n.String()\n\tfor len(instruction) < 5 {\n\t\tinstruction = \"0\" + instruction\n\t}\n\topcode, err := strconv.Atoi(instruction[3:])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfirstMode, err := strconv.Atoi(instruction[2:3])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tsecondMode, err := strconv.Atoi(instruction[1:2])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tthirdMode, err := strconv.Atoi(instruction[0:1])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn opcode, firstMode, secondMode, thirdMode\n}", "func NewIntFromString(s string) *big.Int {\n\tv, ok := new(big.Int).SetString(s, 10) //nolint:gomnd\n\tif !ok {\n\t\tpanic(fmt.Sprintf(\"Bad base 10 string %s\", s))\n\t}\n\treturn v\n}", "func ParseBigInt(str string) (_ *cells.BinaryCell, err error) {\n\tbytes, err := hex.DecodeString(str)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn cells.New(OpUint64, bytes), nil\n}", "func intFromStr(s string) (int, bool) {\n\tvar (\n\t\tfoundNum bool\n\t\tr int\n\t)\n\tfor i := 0; i < len(s); i++ {\n\t\tif s[i] >= '0' && s[i] <= '9' {\n\t\t\tfoundNum = true\n\t\t\tr = r*10 + int(s[i]-'0')\n\t\t} else if foundNum {\n\t\t\treturn r, true\n\t\t}\n\t}\n\tif foundNum {\n\t\treturn r, true\n\t}\n\treturn 0, false\n}", "func Int(name string) (int, error) {\n\ti, err := strconv.Atoi(String(name))\n\n\tif err != nil {\n\t\treturn 0, fmt.Errorf(\"failed to decode input %q as int: %w\", name, err)\n\t}\n\n\treturn i, nil\n}", "func atoi(s string) int {\n\td, _ := strconv.Atoi(s)\n\treturn d\n}", "func atoi(s string) int {\n\td, _ := strconv.Atoi(s)\n\treturn d\n}", "func myAtoi(str string) int {\n\t// Trim white space\n\ti := 0\n\tfor i < len(str) && str[i] == ' ' {\n\t\ti++\n\t}\n\n\tvar negative bool\n\t// Check for negative/positive sign\n\tif i < len(str) && (str[i] == '-' || str[i] == '+') {\n\t\tnegative = (str[i] == '-')\n\t\ti++\n\t}\n\n\tvar result int\n\tfor ; i < len(str); i++ {\n\t\tif str[i] < '0' || str[i] > '9' {\n\t\t\tbreak\n\t\t}\n\n\t\t// Use MaxInt64 on a 64-bit OS\n\t\tif result > math.MaxInt32/10 || (result == math.MaxInt32/10 && str[i]-'0' > math.MaxInt32%10) {\n\t\t\tif negative {\n\t\t\t\treturn math.MinInt32\n\t\t\t}\n\t\t\treturn math.MaxInt32\n\t\t}\n\n\t\tresult = (result * 10) + int(str[i]-'0')\n\t}\n\n\tif negative {\n\t\treturn 0 - result\n\t}\n\n\treturn result\n}", "func (p *Parser) parseIntegerLiteral() asti.ExpressionI {\n\tlit := &ast.IntegerLiteral{Token: p.curToken}\n\n\tvar value int64\n\tvar err error\n\n\tif strings.HasPrefix(p.curToken.Literal, \"0b\") {\n\t\tvalue, err = strconv.ParseInt(p.curToken.Literal[2:], 2, 64)\n\t} else if strings.HasPrefix(p.curToken.Literal, \"0x\") {\n\t\tvalue, err = strconv.ParseInt(p.curToken.Literal[2:], 16, 64)\n\t} else {\n\t\tvalue, err = strconv.ParseInt(p.curToken.Literal, 10, 64)\n\t}\n\n\tif err != nil {\n\t\tp.AddError(\"could not parse %q as integer\", p.curToken.Literal)\n\t\treturn nil\n\t}\n\tlit.Value = value\n\treturn lit\n}", "func (this *OptionInt) Parse(arg string) error {\n argint, err := strconv.Atoi(arg)\n if err != nil {\n return err\n }\n\n //this.opt_storage = argint;\n var storage *int\n storage = this.opt_storage.(*int)\n\n *storage = argint\n\n return nil\n}", "func Str2Int(v interface{}) (i int) {\n\tswitch v.(type) {\n\tcase string:\n\t\ti, _ = strconv.Atoi(v.(string))\n\n\tcase int:\n\t\ti = v.(int)\n\t}\n\n\treturn\n}", "func convInt(n string) int {\r\n\tpnt, err := strconv.Atoi(n)\r\n\tif err != nil {\r\n\t\treturn -1\r\n\t}\r\n\treturn pnt\r\n}", "func parseInt(str string, separator string) (ret int, remainder string, err error ) {\n remainder = str\n index := strings.Index(str, separator)\n if index == 0 {\n return ret, remainder, fmt.Errorf(\"separtor %s first component of string %s\", separator, str)\n }\n if index < 0 {\n /* no separator found */\n remainder = \"\"\n ret, err = strconv.Atoi(str)\n return ret, remainder, err\n }\n\n intStr := str[0:index]\n if index == len(str) -1 {\n remainder = \"\"\n } else {\n remainder = str[index+1:]\n }\n ret, err = strconv.Atoi(intStr)\n return ret, remainder, err\n}", "func hex2int(hexStr string) int {\n\tval, _ := strconv.ParseInt(hexStr, 16, 64)\n\treturn int(val)\n}", "func Int(str string) bool {\n\tif len(str) == 0 {\n\t\treturn true\n\t}\n\t_, err := strconv.Atoi(str)\n\n\treturn err == nil\n}", "func StringToInt32(s string) int32 {\n\ti, _ := strconv.ParseInt(s, 10, 32)\n\treturn int32(i)\n}", "func ParseNumber(s string, bitSize int) (Number, error) {\n\tif strings.HasPrefix(s, \"-\") {\n\t\tif n, err := ParseNumber(s[1:], bitSize); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\treturn n.Neg(), nil\n\t\t}\n\t}\n\n\tvar f big.Float\n\tif strings.HasPrefix(s, \"0x\") || strings.HasPrefix(s, \"0b\") || !strings.HasPrefix(s, \"0\") {\n\t\t_, _, err := f.Parse(s, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tbinString := \"0b\"\n\t\toctalToBin := map[string]string{\n\t\t\t\"0\": \"000\",\n\t\t\t\"1\": \"001\",\n\t\t\t\"2\": \"010\",\n\t\t\t\"3\": \"011\",\n\t\t\t\"4\": \"100\",\n\t\t\t\"5\": \"101\",\n\t\t\t\"6\": \"110\",\n\t\t\t\"7\": \"111\",\n\t\t}\n\t\tfor _, ch := range s[1:] {\n\t\t\tif bin, ok := octalToBin[string(ch)]; ok {\n\t\t\t\tbinString += bin\n\t\t\t} else {\n\t\t\t\treturn nil, errors.New(\"unknown octal digit: \" + string(ch))\n\t\t\t}\n\t\t}\n\t\treturn ParseNumber(binString, bitSize)\n\t}\n\tvar i big.Int\n\tf.Int(&i)\n\treturn NewNumber(&i, bitSize), nil\n}", "func String2Int(v string) int {\n\treturn int(String2Int64(v))\n}", "func ParseUint(s string, base int, bitSize int) (uint64, error) {}", "func ParseBigInt(s string, base int) (*big.Int, error) {\n\tif base < 0 || base > 16 {\n\t\treturn nil, errors.New(\"ParseBigInt: invalid base\")\n\t}\n\ts = strings.Replace(s, \"\\n\", \"\", -1)\n\tz, ok := new(big.Int).SetString(s, base)\n\tif !ok {\n\t\treturn nil, errors.New(\"ParseBigInt: invalid string\")\n\t}\n\treturn z, nil\n}" ]
[ "0.7732646", "0.747722", "0.74192244", "0.73542285", "0.730553", "0.7294096", "0.72713816", "0.7267639", "0.7195607", "0.7017008", "0.6923082", "0.6920135", "0.691729", "0.6908659", "0.69058967", "0.6884467", "0.6884467", "0.6877954", "0.68716604", "0.68492323", "0.6831126", "0.68308127", "0.6829783", "0.67838967", "0.67735064", "0.6752803", "0.66816956", "0.66680384", "0.6640054", "0.6637033", "0.6631468", "0.66197175", "0.66123635", "0.65978897", "0.65706086", "0.65467316", "0.65101427", "0.6466294", "0.646037", "0.64508295", "0.6442048", "0.6436632", "0.6417362", "0.63985246", "0.6380138", "0.6377606", "0.63604224", "0.6354254", "0.6344032", "0.63413626", "0.6333261", "0.6327643", "0.63036585", "0.6295177", "0.62901235", "0.628096", "0.6264508", "0.6259207", "0.6258579", "0.6245018", "0.6242011", "0.623773", "0.62375295", "0.6235368", "0.62353206", "0.62165105", "0.62165105", "0.62165105", "0.62165105", "0.62165105", "0.62165105", "0.62165105", "0.62165105", "0.62165105", "0.62165105", "0.62165105", "0.6209281", "0.6193788", "0.6170923", "0.6169693", "0.6146307", "0.61451864", "0.61299783", "0.6127932", "0.61266357", "0.61201745", "0.61201745", "0.6096255", "0.6094741", "0.6083225", "0.6048149", "0.6040686", "0.6032362", "0.60150355", "0.60058296", "0.5985584", "0.59855556", "0.59722006", "0.5956047", "0.59459186" ]
0.8082527
0
ParseFloat uses strconv.ParseFloat with bitsize == 64 to parse operand as a float.
func ParseFloat(operand string) (f float64, err error) { return strconv.ParseFloat(operand, 64) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ParseFloat(s string, bitSize int) (float64, error) {\n\treturn strconv.ParseFloat(s, bitSize)\n}", "func ParseFloat(s string, base int, prec uint, mode RoundingMode) (f *Float, b int, err error) {}", "func ParseFloat(s string, base int, prec uint, mode big.RoundingMode,) (*big.Float, int, error)", "func ParseToFloat(data string) (float64, error) {\n\treturn strconv.ParseFloat(data, 0)\n}", "func FloatParse(z *big.Float, s string, base int) (*big.Float, int, error)", "func ParseFloat(key string, bitSize int) (float64, error) {\n\tfnParseFloat := \"ParseFloat\"\n\n\tval := os.Getenv(key)\n\n\tif val == \"\" {\n\t\treturn 0, notFoundError(key, fnParseFloat)\n\t}\n\n\tparsedVal, err := strconv.ParseFloat(val, bitSize)\n\n\tif err != nil {\n\t\treturn 0, &EnvError{fnParseFloat, key, err}\n\t}\n\n\treturn parsedVal, err\n}", "func ParseFloatDef(s string, def float64) float64 {\n\tif f, err := strconv.ParseFloat(s, 32); err == nil {\n\t\treturn f\n\t}\n\treturn def\n}", "func ParseFloat(s string) TermT {\n\tcs := C.CString(s)\n\tdefer C.free(unsafe.Pointer(cs))\n\treturn TermT(C.yices_parse_float(cs))\n}", "func (this *OptionFloat) Parse(arg string) error {\n argfloat, err := strconv.ParseFloat(arg, 64)\n if err != nil {\n return err\n }\n\n //this.opt_storage = float64(argfloat);\n var storage *float64\n storage = this.opt_storage.(*float64)\n\n *storage = float64(argfloat)\n\n return nil\n}", "func ParseFloat64(str string) float64 {\n\tf, _ := strconv.ParseFloat(str, 32)\n\treturn f\n}", "func (c *Currency) ParseFloat(s string) error {\n\tf, err := strconv.ParseFloat(s, 64)\n\tif err != nil {\n\t\tif log.IsTrace() {\n\t\t\tlog.Trace(\"money.Currency.ParseFloat\", \"err\", err, \"arg\", s, \"currency\", c)\n\t\t}\n\t\treturn log.Error(\"money.Currency.ParseFloat.ParseFloat\", \"err\", err, \"arg\", s)\n\t}\n\tc.Valid = true\n\t*c = c.Setf(f)\n\treturn nil\n}", "func Parse(s string) (float64, error) {\n\treturn strconv.ParseFloat(s, 64)\n}", "func ParseQueryFloat(param string, request *http.Request, params imageserver.Params) error {\n\ts := request.URL.Query().Get(param)\n\tif s == \"\" {\n\t\treturn nil\n\t}\n\tf, err := strconv.ParseFloat(s, 64)\n\tif err != nil {\n\t\treturn newParseTypeParamError(param, \"float\", err)\n\t}\n\tparams.Set(param, f)\n\treturn nil\n}", "func ParseFloat(from []byte) Key {\n\tv, err := strconv.ParseFloat(string(from), 64)\n\tif err != nil {\n\t\treturn NilKey{}\n\t} else {\n\t\treturn FloatKey(v)\n\t}\n}", "func AsF(str string) (float64, bool) {\n\tres, err := strconv.ParseFloat(str, 64)\n\treturn res, err == nil\n}", "func (v Value) Float(bitSize int) (float64, error) {\n\tif v.typ != Number {\n\t\treturn 0, v.newError(\"%s is not a number\", v.Raw())\n\t}\n\tf, err := strconv.ParseFloat(v.Raw(), bitSize)\n\tif err != nil {\n\t\treturn 0, v.newError(\"%v\", err)\n\t}\n\treturn f, nil\n}", "func ParseFloat64(s string, defaultIfEmpty float64) (n float64, err error) {\n\tif len(s) == 0 {\n\t\treturn defaultIfEmpty, nil\n\t}\n\n\tn, err = strconv.ParseFloat(s, 64)\n\n\treturn\n}", "func ParseFloat64(s string) (float64, error) {\n\tif s = strings.TrimSpace(s); s != \"\" {\n\t\treturn strconv.ParseFloat(s, 64)\n\t}\n\treturn 0.0, nil\n}", "func ParseFloatPrefix(s string, bitSize int) (float64, int, error)", "func ParseFloat64(val interface{}) (float64, error) {\n\tswitch val.(type) {\n\tcase nil:\n\t\treturn 1, nil\n\tcase json.Number:\n\t\treturn val.(json.Number).Float64()\n\tdefault:\n\t\treturn -1, fmt.Errorf(\"unable to casting number %v (type %T)\", val, val)\n\t}\n}", "func (v Value) ParseFloat64() (val float64, err error) {\n\treturn strconv.ParseFloat(v.String(), 64)\n}", "func toFloat(str string) (float64, error) {\n\tres, err := strconv.ParseFloat(str, 64)\n\tif err != nil {\n\t\tres = 0.0\n\t}\n\treturn res, err\n}", "func LooksLikeAFloat(inputStr string) (matched bool, value float64) {\n\tmatched = floatRe.MatchString(inputStr)\n\tif matched {\n\t\tvalue, _ = strconv.ParseFloat(inputStr, 64)\n\t\treturn true, value\n\t}\n\treturn\n}", "func IsFloat(str string) bool {\n\treturn str != \"\" && rxFloat.MatchString(str)\n}", "func StringToFloat(param string) float64 {\n\tval, _ := strconv.ParseFloat(param, 10)\n\treturn val\n}", "func readFloat(s []byte) (mantissa uint64, exp int, neg, trunc, ok bool) {\n\tconst uint64digits = 19\n\ti := 0\n\n\t// optional sign\n\tif i >= len(s) {\n\t\treturn\n\t}\n\tswitch {\n\tcase s[i] == '+':\n\t\ti++\n\tcase s[i] == '-':\n\t\tneg = true\n\t\ti++\n\t}\n\n\t// digits\n\tsawdot := false\n\tsawdigits := false\n\tnd := 0\n\tndMant := 0\n\tdp := 0\n\tfor ; i < len(s); i++ {\n\t\tswitch c := s[i]; true {\n\t\tcase c == '.':\n\t\t\tif sawdot {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tsawdot = true\n\t\t\tdp = nd\n\t\t\tcontinue\n\n\t\tcase '0' <= c && c <= '9':\n\t\t\tsawdigits = true\n\t\t\tif c == '0' && nd == 0 { // ignore leading zeros\n\t\t\t\tdp--\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tnd++\n\t\t\tif ndMant < uint64digits {\n\t\t\t\tmantissa *= 10\n\t\t\t\tmantissa += uint64(c - '0')\n\t\t\t\tndMant++\n\t\t\t} else if s[i] != '0' {\n\t\t\t\ttrunc = true\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tif !sawdigits {\n\t\treturn\n\t}\n\tif !sawdot {\n\t\tdp = nd\n\t}\n\n\t// optional exponent moves decimal point.\n\t// if we read a very large, very long number,\n\t// just be sure to move the decimal point by\n\t// a lot (say, 100000). it doesn't matter if it's\n\t// not the exact number.\n\tif i < len(s) && (s[i] == 'e' || s[i] == 'E') {\n\t\ti++\n\t\tif i >= len(s) {\n\t\t\treturn\n\t\t}\n\t\tesign := 1\n\t\tif s[i] == '+' {\n\t\t\ti++\n\t\t} else if s[i] == '-' {\n\t\t\ti++\n\t\t\tesign = -1\n\t\t}\n\t\tif i >= len(s) || s[i] < '0' || s[i] > '9' {\n\t\t\treturn\n\t\t}\n\t\te := 0\n\t\tfor ; i < len(s) && '0' <= s[i] && s[i] <= '9'; i++ {\n\t\t\tif e < 10000 {\n\t\t\t\te = e*10 + int(s[i]) - '0'\n\t\t\t}\n\t\t}\n\t\tdp += e * esign\n\t}\n\n\tif i != len(s) {\n\t\treturn\n\t}\n\n\texp = dp - ndMant\n\tok = true\n\treturn\n\n}", "func parseFloat64(s string) (float64, bool) {\n\tf, err := strconv.ParseFloat(strings.TrimSpace(s), 64)\n\tif err != nil {\n\t\treturn 0, false\n\t}\n\treturn f, true\n}", "func convertStringToFloat(str string) float64 {\n\tconst float64Bitsize = 64\n\tconvertedString, err := strconv.ParseFloat(str, float64Bitsize)\n\t// Store error in string array which will be checked in main function later to see if there is a need to exit\n\tif err != nil {\n\t\tlog.Error(\"String to float error: %v\", err)\n\t\terrorsConversion = append(errorsConversion, err)\n\t}\n\treturn convertedString\n}", "func (p *Parser) parseFloatLiteral() asti.ExpressionI {\n\tflo := &ast.FloatLiteral{Token: p.curToken}\n\tvalue, err := strconv.ParseFloat(p.curToken.Literal, 64)\n\tif err != nil {\n\t\tp.AddError(\"could not parse %q as float\", p.curToken.Literal)\n\t\treturn nil\n\t}\n\tflo.Value = value\n\treturn flo\n}", "func stringToFloat(num string) *float64 {\n\tif num == nan {\n\t\treturn nil\n\t}\n\n\tf, err := strconv.ParseFloat(num, 64)\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\treturn &f\n}", "func IsFloat(val any) bool {\n\tif val == nil {\n\t\treturn false\n\t}\n\n\tswitch rv := val.(type) {\n\tcase float32, float64:\n\t\treturn true\n\tcase string:\n\t\treturn rv != \"\" && rxFloat.MatchString(rv)\n\t}\n\treturn false\n}", "func Floatfrombits(tk obj.Token, args []oop.VarDef) oop.Val {\n\tval := args[0].Val\n\tif val.Type != oop.Int {\n\t\tval.Data = 0.0\n\t}\n\tb := uint64(val.Data.(float64))\n\treturn oop.Val{Data: *(*float64)(unsafe.Pointer(&b)), Type: oop.Float}\n}", "func FloatConverter(str string, target reflect.Value) (ok bool) {\n\tf, err := strconv.ParseFloat(str, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\n\ttarget.SetFloat(f)\n\treturn true\n}", "func requireOneFloat(parse func(f float64) (Command, error)) Parser {\n\treturn requireNParams(1, func(tokens []string) (Command, error) {\n\t\tf, err := strconv.ParseFloat(tokens[0], 64)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Invalid parameter type: %w\", err)\n\t\t}\n\t\treturn parse(f)\n\t})\n}", "func toFloat(rawFloat string) float64 {\n\tparsed, _ := strconv.ParseFloat(strings.Replace(strings.Replace(strings.Replace(rawFloat, \"$\", \"\", -1), \",\", \"\", -1), \"%\", \"\", -1), 64)\n\treturn parsed\n}", "func (p *Parser) Float(description string, defaultValue ...float64) *float64 {\n\tvar v float64\n\tif len(defaultValue) > 0 {\n\t\tv = defaultValue[0]\n\t} else if strings.HasSuffix(description, \"]\") {\n\t\tif idx := strings.LastIndex(description, \"[\"); idx != 1 {\n\t\t\tvar err error\n\t\t\tv, err = strconv.ParseFloat(description[idx+1:len(description)-1], 64)\n\t\t\tif err != nil {\n\t\t\t\texit(\"optparse: could not parse default value from: %s\", description)\n\t\t\t}\n\t\t}\n\t}\n\top := p.newOpt(description, true)\n\top.valueType = floatValue\n\top.value = &v\n\treturn &v\n}", "func TestFloat(tst *testing.T) {\n\n\t// Test bool\n\tf, err := StringToFloat(\"1.256898\")\n\tbrtesting.AssertEqual(tst, err, nil, \"StringToFloat failed\")\n\tbrtesting.AssertEqual(tst, f, 1.256898, \"StringToFloat failed\")\n\tf, err = StringToFloat(\"go-bedrock\")\n\tbrtesting.AssertNotEqual(tst, err, nil, \"StringToFloat failed\")\n}", "func parseFloatValue(input string) (res float64) {\n\tvar err error\n\n\tres, err = strconv.ParseFloat(input, 64)\n\tif err != nil {\n\t\t// negative values must be dropped here\n\t\treturn\n\t}\n\n\tif res < 0 {\n\t\treturn 0\n\t}\n\treturn\n}", "func parseUnsignedFixed(value string, decimalPlaces uint8) (unsignedFixed, error) {\n\tval := strings.TrimSpace(value)\n\trex, _ := regexp.Compile(REX_FLOAT)\n\n\t/*\n\t * Check if regular expression compiles.\n\t */\n\tif rex == nil {\n\t\treturn unsignedFixed{}, fmt.Errorf(\"Failed to compile regular expression: '%s'\", REX_FLOAT)\n\t} else {\n\t\tmatches := rex.MatchString(val)\n\n\t\t/*\n\t\t * Check syntax of number.\n\t\t */\n\t\tif !matches {\n\t\t\treturn unsignedFixed{}, fmt.Errorf(\"Value '%s' does not match regular expression '%s'.\", val, REX_FLOAT)\n\t\t} else {\n\t\t\tvi := uint64(0)\n\t\t\texp := uint8(0)\n\t\t\tdot := false\n\t\t\tfail := false\n\n\t\t\t/*\n\t\t\t * Iterate over the bytes in the string.\n\t\t\t */\n\t\t\tfor _, c := range []byte(value) {\n\n\t\t\t\t/*\n\t\t\t\t * Do this as long as we're not in failure mode.\n\t\t\t\t */\n\t\t\t\tif !fail {\n\n\t\t\t\t\t/*\n\t\t\t\t\t * Check if we have to read more digits.\n\t\t\t\t\t */\n\t\t\t\t\tif !dot || (exp < decimalPlaces) {\n\n\t\t\t\t\t\t/*\n\t\t\t\t\t\t * Handle decimal digit.\n\t\t\t\t\t\t */\n\t\t\t\t\t\tif ('0' <= c) && (c <= '9') {\n\n\t\t\t\t\t\t\t/*\n\t\t\t\t\t\t\t * Handle overflow before multiplication.\n\t\t\t\t\t\t\t */\n\t\t\t\t\t\t\tif vi >= LOWER_BEFORE_SHIFT {\n\t\t\t\t\t\t\t\tvi = math.MaxUint64\n\t\t\t\t\t\t\t\tfail = true\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\tvi *= 10\n\n\t\t\t\t\t\t\t\t/*\n\t\t\t\t\t\t\t\t * If we already read decimal dot, increment exponent.\n\t\t\t\t\t\t\t\t */\n\t\t\t\t\t\t\t\tif dot {\n\t\t\t\t\t\t\t\t\texp++\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tdigit := c - '0'\n\t\t\t\t\t\t\tdigit64 := uint64(digit)\n\t\t\t\t\t\t\tvip := vi + digit64\n\n\t\t\t\t\t\t\t/*\n\t\t\t\t\t\t\t * Handle overflow before addition.\n\t\t\t\t\t\t\t */\n\t\t\t\t\t\t\tif vip < vi {\n\t\t\t\t\t\t\t\tvi = math.MaxUint64\n\t\t\t\t\t\t\t\tfail = true\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\tvi = vip\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\t/*\n\t\t\t\t\t\t * Handle dot.\n\t\t\t\t\t\t */\n\t\t\t\t\t\tif c == '.' {\n\t\t\t\t\t\t\tdot = true\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t/*\n\t\t\t * Shift the number to the right amount of decimal places.\n\t\t\t */\n\t\t\tfor exp < decimalPlaces {\n\n\t\t\t\t/*\n\t\t\t\t * Do this as long as we're not in failure mode.\n\t\t\t\t */\n\t\t\t\tif !fail {\n\n\t\t\t\t\t/*\n\t\t\t\t\t * Handle overflow before multiplication.\n\t\t\t\t\t */\n\t\t\t\t\tif vi >= LOWER_BEFORE_SHIFT {\n\t\t\t\t\t\tvi = math.MaxUint64\n\t\t\t\t\t\tfail = true\n\t\t\t\t\t} else {\n\t\t\t\t\t\tvi *= 10\n\t\t\t\t\t\texp++\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t/*\n\t\t\t * Check if parsing failed.\n\t\t\t */\n\t\t\tif fail {\n\t\t\t\treturn unsignedFixed{}, fmt.Errorf(\"%s\", \"Parsing failed due to arithmetic overflow.\")\n\t\t\t} else {\n\n\t\t\t\t/*\n\t\t\t\t * Create unsigned fixed-point number.\n\t\t\t\t */\n\t\t\t\tfx := unsignedFixed{\n\t\t\t\t\tmantissa: vi,\n\t\t\t\t\texponent: exp,\n\t\t\t\t}\n\n\t\t\t\treturn fx, nil\n\t\t\t}\n\n\t\t}\n\n\t}\n\n}", "func (f FormField) Float() float64 {\n\tif result, err := strconv.ParseFloat(f.Value, 64); err == nil {\n\t\treturn result\n\t}\n\treturn 0.0\n}", "func EvalFloat(exp string, vars map[string]interface{}) (float64, error) {\n\ttokens, err := tokenize(exp)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\trpn, err := postfix(tokens, vars)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn eval(rpn, vars)\n}", "func StrToFloat(s string) float64 {\n\tif n, err := strconv.ParseFloat(s, 64); err == nil {\n\t\treturn n\n\t}\n\treturn float64(0.0)\n}", "func strToFloat(str string) (float64, error) {\n\tnumber, err := strconv.ParseFloat(str, 64)\n\tif err == nil {\n\t\treturn number, nil\n\t}\n\treturn -1, fmt.Errorf(`invalid parameter '%v' was found in the provided csv. \n\t\tMake sure the csv contain only valid float numbers`, str)\n}", "func (p *parser) parseFloat(annotations []Symbol) Float {\n\treturn Float{annotations: annotations, isSet: true, text: p.next().Val}\n}", "func makeFloat(val string) float64 {\n\tb2, _ := strconv.ParseFloat(val, 64)\n\n\treturn b2\n}", "func ToF(str string) float64 {\n\tval, err := strconv.ParseFloat(str, 64)\n\tL.IsError(err, str)\n\treturn val\n}", "func parseFloat(emp []string, key, fileType string) (float64, error) {\n\tvalueStr := emp[headersMap[fileType][key]]\n\tif valueStr == \"\" {\n\t\treturn 0.0, nil\n\t} else {\n\t\tvalueStr = strings.Trim(valueStr, \" \")\n\t\tvalueStr = strings.Replace(valueStr, \",\", \".\", 1)\n\t\tif n := strings.Count(valueStr, \".\"); n > 1 {\n\t\t\tvalueStr = strings.Replace(valueStr, \".\", \"\", n-1)\n\t\t}\n\t}\n\treturn strconv.ParseFloat(valueStr, 64)\n}", "func parseFloat64FromString(content string, aggErr *AggregateError) float64 {\n result, err := strconv.ParseFloat(content, 64)\n if err != nil {\n aggErr.Append(err)\n }\n return result\n}", "func mustParseFloat64(s string) float64 {\n\tf64, err := strconv.ParseFloat(s, 64)\n\tif err != nil {\n\t\tcatlog.Fatalf(\"Could not convert float64: %v\", err)\n\t}\n\n\treturn f64\n}", "func StringAsFloat(s string, decimalSeparator, thousandsSeparator rune) float64 {\n\tif s == \"\" {\n\t\treturn 0.0\n\t}\n\n\tconst maxLength = 20\n\n\tif len([]rune(s)) > maxLength {\n\t\ts = s[0:maxLength]\n\t}\n\n\ts = strings.Replace(s, string(thousandsSeparator), \"\", -1)\n\n\ts = strings.Replace(s, string(decimalSeparator), \".\", -1)\n\n\tif f, err := strconv.ParseFloat(s, 64); err == nil {\n\t\treturn f\n\t}\n\n\treturn 0.0\n}", "func Float(str string) bool {\n\t_, err := strconv.ParseFloat(str, 0)\n\treturn err == nil\n}", "func (s *ss) convertFloat(str string, n int) float64 {\n\tif p := indexRune(str, 'p'); p >= 0 {\n\t\t// Atof doesn't handle power-of-2 exponents,\n\t\t// but they're easy to evaluate.\n\t\tf, err := strconv.ParseFloat(str[:p], n)\n\t\tif err != nil {\n\t\t\t// Put full string into error.\n\t\t\tif e, ok := err.(*strconv.NumError); ok {\n\t\t\t\te.Num = str\n\t\t\t}\n\t\t\ts.error(err)\n\t\t}\n\t\tm, err := strconv.Atoi(str[p+1:])\n\t\tif err != nil {\n\t\t\t// Put full string into error.\n\t\t\tif e, ok := err.(*strconv.NumError); ok {\n\t\t\t\te.Num = str\n\t\t\t}\n\t\t\ts.error(err)\n\t\t}\n\t\treturn math.Ldexp(f, m)\n\t}\n\tf, err := strconv.ParseFloat(str, n)\n\tif err != nil {\n\t\ts.error(err)\n\t}\n\treturn f\n}", "func (p *numberPacker) mustUnpackFloat(n packedNumber) float64 {\n\tnum := p.unpack(n)\n\tif *num == zero {\n\t\treturn 0.0\n\t}\n\tif num.IsInteger() {\n\t\tpanic(\"not a float\")\n\t}\n\treturn num.Float\n}", "func (r *memcachedScraper) parseFloat(key, value string) (float64, bool) {\n\ti, err := strconv.ParseFloat(value, 64)\n\tif err != nil {\n\t\tr.logInvalid(\"float\", key, value)\n\t\treturn 0, false\n\t}\n\treturn i, true\n}", "func (v Value) Float() float64 {\n\tswitch v.Typ {\n\tdefault:\n\t\tf, _ := strconv.ParseFloat(v.String(), 64)\n\t\treturn f\n\tcase ':':\n\t\treturn float64(v.IntegerV)\n\t}\n}", "func Float(val string) (out *big.Float, err error) {\n\tvalue, ret := new(big.Float).SetString(val)\n\tif !ret {\n\t\terr = fmt.Errorf(\"invalid va\")\n\t\treturn\n\t}\n\treturn value, err\n}", "func ToFloat(str string) (float64, error) {\n\tres, err := strconv.ParseFloat(str, 64)\n\tif err != nil {\n\t\tres = 0.0\n\t}\n\treturn res, err\n}", "func ToFloat(str string) (float64, error) {\n\tres, err := strconv.ParseFloat(str, 64)\n\tif err != nil {\n\t\tres = 0.0\n\t}\n\treturn res, err\n}", "func parseFloat64(content []byte, aggErr *AggregateError) float64 {\n result, err := strconv.ParseFloat(string(content), 64)\n if err != nil {\n aggErr.Append(err)\n }\n return result\n}", "func parseFloat32(lineTokens []string) (float32, error) {\n\tif len(lineTokens) < 2 {\n\t\treturn 0, fmt.Errorf(`unsupported syntax for \"%s\"; expected 1 argument; got %d`, lineTokens[0], len(lineTokens)-1)\n\t}\n\n\tval, err := strconv.ParseFloat(lineTokens[1], 32)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn float32(val), nil\n}", "func parseFloat64(str string, dest *float64) error {\n\tstr = strings.TrimSpace(str)\n\tif len(str) == 0 {\n\t\t*dest = 0\n\t\treturn nil\n\t}\n\n\tstr = strings.Replace(str, \",\", \"\", -1) // Remove commas\n\tval, err := strconv.ParseFloat(str, 64)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*dest = val\n\treturn nil\n}", "func ParseFloat32(str string) float32 {\n\tf, err := strconv.ParseFloat(str, 32)\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn float32(f)\n}", "func (f *Flags) Float(spec string, p *float64, name, usage string) {\n\tf.addOption(spec, name, usage, func(name, value string) error {\n\t\tf, err := strconv.ParseFloat(value, 64)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"invalid %s argument '%s'\", name, value)\n\t\t}\n\t\t*p = f\n\t\treturn nil\n\t})\n}", "func (a ASTNode) Float() float64 {\n\tif a.t != tval {\n\t\tpanic(ConfErr{a.pos, errors.New(\"Not a basic value\")})\n\t}\n\tv, err := strconv.ParseFloat(a.val.(string), 64)\n\tif err != nil {\n\t\tpanic(ConfErr{a.pos, err})\n\t}\n\treturn v\n}", "func strToFloat(str string, typeAbbreviation string) float64 {\n\t// Remove space\n\tif strings.Contains(str, \" \") {\n\t\tstr = strings.ReplaceAll(str, \" \", \"\")\n\t}\n\tstr = strings.ReplaceAll(str, typeAbbreviation, \"\")\n\tf, err := strconv.ParseFloat(str, 10)\n\tif err != nil {\n\t\tErrorLog(\"cannot parse string to float: %s\", str)\n\t\treturn 0\n\t}\n\tInfoLogV1(\"strToFloat %f\", f)\n\treturn f\n}", "func TestRoundTripFloat(t *testing.T) {\n\tf := func(flt float64) bool {\n\t\tflt = math.Abs(flt)\n\t\tneeded := fmt.Sprintf(\"%f\", flt)\n\t\tgotten := TestExpr(needed)\n\t\tresult, _ := strconv.ParseFloat(gotten[\"value\"].(string), 64)\n\t\treturn flt == result\n\t}\n\n\tif err := quick.Check(f, nil); err != nil {\n\t\tt.Error(err)\n\t}\n}", "func NewFloat(f float64, valid bool) Float {\n\treturn Float{\n\t\tNullFloat64: sql.NullFloat64{\n\t\t\tFloat64: f,\n\t\t\tValid: valid,\n\t\t},\n\t}\n}", "func ValidFloatStr(val string) bool {\n\tvar validFloat = regexp.MustCompile(`^[-+]?([0-9]+(\\.[0-9]+)?)$`)\n\treturn validFloat.MatchString(val)\n}", "func reformatFloat(i string) (o string, err error) {\n\tf, err := strconv.ParseFloat(i, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\to = fmt.Sprintf(\"%f\", f)\n\n\treturn\n}", "func (m MetricType) resolveFloat(val interface{}) interface{} {\n\t_, isFloat64 := val.(float64)\n\tif isFloat64 && m == FloatType {\n\t\treturn float32(val.(float64))\n\t}\n\n\treturn val\n}", "func Float64(f *frm.Field, inp ...string) {\n\tnum, err := strconv.ParseFloat(strings.TrimSpace(inp[0]), 64)\n\tf.Value = num\n\tif err != nil {\n\t\t//Return error if input string failed to convert.\n\t\tf.Err = err.Error()\n\t\treturn\n\t}\n\n\tif !f.Required && num == 0 {\n\t\t//f.ValueFloat64 is zero by default so assigning zero isn't required\n\t\treturn\n\t}\n\n\tif f.Min != nil && num < f.Min.(float64) || f.Max != nil && num > f.Max.(float64) {\n\t\tf.Err = fmt.Sprintf(\"Must be between %v and %v.\", f.Min, f.Max)\n\t\treturn\n\t}\n\n\tif rem := toFixed(math.Mod(num, float64(f.Step)), 6); rem != 0 {\n\t\tf.Err = fmt.Sprintf(\"Please enter a valid value. The two nearest values are %v and %v.\", num-rem, num-rem+float64(f.Step))\n\t}\n}", "func ConvertStringToFloat(str string) float64 {\n\ts, err := strconv.ParseFloat(str, 64)\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn s\n}", "func DecodeFloatAscending(buf []byte, tmp []byte) ([]byte, float64, error) {\n\t// Handle the simplistic cases first.\n\tswitch buf[0] {\n\tcase floatNaN, floatNaNDesc:\n\t\treturn buf[1:], math.NaN(), nil\n\tcase floatInfinity:\n\t\treturn buf[1:], math.Inf(1), nil\n\tcase floatNegativeInfinity:\n\t\treturn buf[1:], math.Inf(-1), nil\n\tcase floatZero:\n\t\treturn buf[1:], 0, nil\n\t}\n\ttmp = tmp[len(tmp):cap(tmp)]\n\tidx := bytes.IndexByte(buf, floatTerminator)\n\tif idx == -1 {\n\t\treturn nil, 0, util.Errorf(\"did not find terminator %#x in buffer %#x\", floatTerminator, buf)\n\t}\n\tswitch {\n\tcase buf[0] == floatNegLarge:\n\t\t// Negative large.\n\t\te, m, tmp2 := decodeLargeNumber(true, buf[:idx+1], tmp)\n\t\treturn buf[idx+1:], makeFloatFromMandE(true, e, m, tmp2), nil\n\tcase buf[0] > floatNegLarge && buf[0] <= floatNegMedium:\n\t\t// Negative medium.\n\t\te, m, tmp2 := decodeMediumNumber(true, buf[:idx+1], tmp)\n\t\treturn buf[idx+1:], makeFloatFromMandE(true, e, m, tmp2), nil\n\tcase buf[0] == floatNegSmall:\n\t\t// Negative small.\n\t\te, m, tmp2 := decodeSmallNumber(true, buf[:idx+1], tmp)\n\t\treturn buf[idx+1:], makeFloatFromMandE(true, e, m, tmp2), nil\n\tcase buf[0] == floatPosLarge:\n\t\t// Positive large.\n\t\te, m, tmp2 := decodeLargeNumber(false, buf[:idx+1], tmp)\n\t\treturn buf[idx+1:], makeFloatFromMandE(false, e, m, tmp2), nil\n\tcase buf[0] >= floatPosMedium && buf[0] < floatPosLarge:\n\t\t// Positive medium.\n\t\te, m, tmp2 := decodeMediumNumber(false, buf[:idx+1], tmp)\n\t\treturn buf[idx+1:], makeFloatFromMandE(false, e, m, tmp2), nil\n\tcase buf[0] == floatPosSmall:\n\t\t// Positive small.\n\t\te, m, tmp2 := decodeSmallNumber(false, buf[:idx+1], tmp)\n\t\treturn buf[idx+1:], makeFloatFromMandE(false, e, m, tmp2), nil\n\tdefault:\n\t\treturn nil, 0, util.Errorf(\"unknown prefix of the encoded byte slice: %q\", buf)\n\t}\n}", "func TestReadFloatUnsupported(t *testing.T) {\n\ttestPatterns := []struct {\n\t\tcbor []byte\n\t\terr error\n\t}{\n\t\t{\n\t\t\t// 0.0\n\t\t\t[]byte{0xf9, 0x00, 0x00},\n\t\t\tUnsupportedTypeReadError,\n\t\t},\n\t\t{\n\t\t\t// -0.0\n\t\t\t[]byte{0xf9, 0x80, 0x00},\n\t\t\tUnsupportedTypeReadError,\n\t\t},\n\t\t{\n\t\t\t// 65504.0\n\t\t\t[]byte{0xf9, 0x7b, 0xff},\n\t\t\tUnsupportedTypeReadError,\n\t\t},\n\t}\n\tfor i := range testPatterns {\n\t\tcborDecoderHarnessExpectErr(t, testPatterns[i].cbor, testPatterns[i].err)\n\t}\n}", "func mustParseFloat64(val string) float64 {\n\tvv, _ := strconv.ParseFloat(val, 64)\n\treturn vv\n}", "func parseFloat(s string) (float64, error) {\n\n\tnegative := s[0] == '-'\n\tif negative {\n\t\ts = s[1:]\n\t}\n\n\ttotal := float64(0)\n\n\tdecimalFound := false\n\tj := 0\n\n\tfor i := len(s) - 1; i >= 0; i-- {\n\n\t\tif !decimalFound && s[i] == '.' {\n\t\t\tdecimalFound = true\n\t\t\ttotal *= math.Pow(10, -float64(j))\n\t\t\tj = 0\n\t\t\tcontinue\n\t\t}\n\n\t\tdigit := float64(s[i] - '0')\n\n\t\tif digit > 9 {\n\t\t\treturn 0, StringNotValid\n\t\t}\n\n\t\ttotal += digit * math.Pow(10, float64(j))\n\t\tj++\n\t}\n\n\tif negative {\n\t\ttotal = -total\n\t}\n\n\treturn total, nil\n}", "func ConvFloat(string string) float64 {\n\tfloat, err := strconv.ParseFloat(string, 64)\n\tif err != nil {\n\t\treturn 0\n\t}\n\treturn float\n}", "func FloatUnmarshalText(z *big.Float, text []byte) error", "func setFloat(data [2]string, f func(float64)) error {\n\tval, err := strconv.ParseFloat(strings.TrimSpace(data[1]), 64)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"code %s: %s\", data[0], err.Error())\n\t}\n\tf(val)\n\treturn nil\n}", "func parseFloats(str string) ([]float64, error) {\n\tvar (\n\t\tfields = strings.Fields(str)\n\t\tfloat_parameters = make([]float64, len(fields))\n\t\terr error\n\t)\n\tfor i, v := range fields {\n\t\tfloat_parameters[i], err = strconv.ParseFloat(v, 64)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn float_parameters, nil\n}", "func GetFloat(fixControlMap map[uint64]string, key uint64) (value float64, exists bool, parseErr error) {\n\tif fixControlMap == nil {\n\t\treturn 0, false, nil\n\t}\n\trawValue, ok := fixControlMap[key]\n\tif !ok {\n\t\treturn 0, false, nil\n\t}\n\t// The same as tidbOptFloat64 in sessionctx/variable.\n\tvalue, parseErr = strconv.ParseFloat(rawValue, 64)\n\treturn value, true, parseErr\n}", "func AsFloat(r *http.Request, key string, decimalSeparator rune) float64 {\n\tif err := r.ParseForm(); err != nil {\n\t\tlog.Println(r.RequestURI, err)\n\t\treturn 0\n\t}\n\n\ts := r.FormValue(key)\n\n\tif s == \"\" {\n\t\ts = r.URL.Query().Get(key)\n\t}\n\n\tif s == \"\" {\n\t\tparams := mux.Vars(r)\n\t\ts = params[key]\n\t}\n\n\tif s == \"\" {\n\t\treturn 0\n\t}\n\n\tthousandSeparator := ','\n\n\tif decimalSeparator == ',' {\n\t\tthousandSeparator = '.'\n\t}\n\n\ts = strings.ReplaceAll(s, string(thousandSeparator), \"\")\n\n\tf, _ := strconv.ParseFloat(s, 64)\n\n\treturn f\n}", "func (c *Cell) Float() (float64, error) {\n\tf, err := strconv.ParseFloat(c.Value, 64)\n\tif err != nil {\n\t\treturn math.NaN(), err\n\t}\n\treturn f, nil\n}", "func lexFloat(lx *lexer) stateFn {\r\n\tr := lx.next()\r\n\tif unicode.IsDigit(r) {\r\n\t\treturn lexFloat\r\n\t}\r\n\r\n\t// Not a digit, if its another '.', need to see if we falsely assumed a float.\r\n\tif r == '.' {\r\n\t\treturn lexIPAddr\r\n\t}\r\n\r\n\tlx.backup()\r\n\tlx.emit(itemFloat)\r\n\treturn lx.pop()\r\n}", "func Str2float(x string) float64 {\n\tk, _ := strconv.ParseFloat(x, 64)\n\treturn k\n}", "func atof(s string) float64 {\n\tif s == \"*\" {\n\t\treturn math.NaN()\n\t}\n\n\tf, err := strconv.ParseFloat(strings.TrimSpace(s), 64)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn f\n}", "func ReadFloat(buffer []byte, offset int) float32 {\n bits := ReadUInt32(buffer, offset)\n return math.Float32frombits(bits)\n}", "func parseNullableFloat64FromString(content string, aggErr *AggregateError) *float64 {\n if len(content) == 0 {\n return nil\n }\n result := parseFloat64FromString(content, aggErr)\n return &result\n}", "func poolSyncTimeParseFloat64(e string) float64 {\n\tnum, err := strconv.ParseFloat(e, 64)\n\tif err != nil {\n\t\treturn 0\n\n\t}\n\treturn num\n}", "func (p Parser) Float64(ctx context.Context) (*float64, error) {\n\tvalue, err := p.Source.String(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif value == nil {\n\t\treturn nil, nil\n\t}\n\tparsed, err := strconv.ParseFloat(*value, 64)\n\tif err != nil {\n\t\treturn nil, ex.New(err)\n\t}\n\treturn &parsed, nil\n}", "func ParseFloat32LE(byteSlice []byte) (float32, error) {\n\tvar floatValue float32\n\tbyteSlice = byteSlice[:binary.Size(floatValue)]\n\tbuf := bytes.NewReader(byteSlice)\n\terr := binary.Read(buf, binary.LittleEndian, &floatValue)\n\treturn floatValue, err\n}", "func parseNullableFloat64(content []byte, aggErr *AggregateError) *float64 {\n if len(content) == 0 {\n return nil\n }\n result := parseFloat64(content, aggErr)\n return &result\n}", "func NewFromFloat64(x float64) (f Float, exact bool) {\n\tintRep := math.Float64bits(x)\n\tsign := intRep&0x8000000000000000 != 0\n\texp := intRep & 0x7FF0000000000000 >> 52\n\tmant := intRep & 0xFFFFFFFFFFFFF\n\tleftMant := mant & 0xFFFFFFFFFFFF0 >> 4\n\tvar a uint64\n\tb := mant & 0xF << 60\n\n\tswitch exp {\n\t// 0b11111111\n\tcase 0x7FF:\n\t\t// NaN or Inf\n\t\tif mant == 0 {\n\t\t\t// +-Inf\n\t\t\ta = 0x7FFF000000000000\n\t\t\tif sign {\n\t\t\t\ta = 0xFFFF000000000000\n\t\t\t}\n\t\t\treturn Float{a: a, b: b}, true\n\t\t}\n\t\t// +-NaN\n\n\t\ta = 0\n\t\tif sign {\n\t\t\ta = 0x8000000000000000\n\t\t}\n\t\ta = a | 0x7FFF000000000000\n\n\t\tnewMant := leftMant\n\t\ta |= newMant\n\n\t\treturn Float{a: a, b: b}, true\n\t\t// 0b00000000\n\tcase 0x0:\n\t\tif mant == 0 {\n\t\t\t// +-Zero\n\t\t\tvar a uint64\n\t\t\tif sign {\n\t\t\t\ta = 0x8000000000000000\n\t\t\t}\n\t\t\treturn Float{a: a, b: b}, true\n\t\t}\n\t}\n\n\tif sign {\n\t\ta = 0x8000000000000000\n\t}\n\n\tnewExp := (exp - 1023 + 16383) << 48\n\ta |= newExp\n\n\ta |= leftMant\n\n\treturn Float{a: a, b: b}, true\n}", "func TestStrToFloat(t *testing.T) {\n\tstrList := []StrToFloatTest{\n\t\t{\"1\", 1.0},\n\t\t{\"2,234.5 \", 2234.5},\n\t\t{\" 3,345,456.123\", 3345456.123},\n\t\t{\"-9234.43\", -9234.43},\n\t\t{\"asd\", 0},\n\t}\n\n\tfor i, str := range strList {\n\t\tnumFloat := StrToFloat(str.NumStr)\n\t\tif !reflect.DeepEqual(str.NumFloat, numFloat) {\n\t\t\tt.Errorf(\"StrToFloat(%v) failed: expected %v got %v\", i, str.NumFloat, numFloat)\n\t\t}\n\t}\n}", "func (c *Config) Float(key string) (float64, error) {\n\treturn strconv.ParseFloat(c.data[key], 64)\n}", "func StringToFloatWithFormat(str String, format string) Float {\n\tif format == \"%f\" { // Same as not using custom format.\n\t\treturn StringToFloat(str)\n\t}\n\n\tv := &stringToFloat{from: str, format: format}\n\tstr.AddListener(v)\n\treturn v\n}", "func ParseFloat64OrReturnBadRequest(w http.ResponseWriter, s string, defaultIfEmpty float64) (n float64, ok bool) {\n\tif len(s) == 0 {\n\t\treturn defaultIfEmpty, true\n\t}\n\n\tn, err := strconv.ParseFloat(s, 64)\n\tif err != nil {\n\t\tWriteErrorResponse(w, http.StatusBadRequest, err.Error())\n\t\treturn n, false\n\t}\n\n\treturn n, true\n}", "func (f FloatField) Clean(value string) (interface{}, ValidationError) {\n\tcleaned_value, err := strconv.ParseFloat(strings.TrimSpace(value), 64)\n\tif err != nil {\n\t\treturn nil, errors.New(\"The value must be a valid float.\")\n\t}\n\treturn cleaned_value, nil\n}", "func Floatbits(tk obj.Token, args []oop.VarDef) oop.Val {\n\tval := args[0].Val\n\tif val.Type != oop.Int && val.Type != oop.Float {\n\t\tval.Data = 0.0\n\t}\n\tf := val.Data.(float64)\n\treturn oop.Val{Data: float64(*(*uint64)(unsafe.Pointer(&f))), Type: oop.Int}\n}", "func TestCheckBinaryExprStringEqlFloat(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `\"abc\" == 2.0`, env,\n\t\t`cannot convert \"abc\" to type float64`,\n\t\t`invalid operation: \"abc\" == 2 (mismatched types string and float64)`,\n\t)\n\n}" ]
[ "0.8577388", "0.8037701", "0.7888861", "0.765797", "0.7510848", "0.74987507", "0.7497898", "0.73149884", "0.7054141", "0.70316344", "0.701707", "0.6983013", "0.6963574", "0.6880399", "0.68307865", "0.6796332", "0.6754063", "0.6752347", "0.6695293", "0.66873264", "0.6636752", "0.66087174", "0.6510508", "0.65018946", "0.64746255", "0.64416724", "0.6425898", "0.6393121", "0.633758", "0.6331348", "0.63294154", "0.6324411", "0.6322836", "0.6315532", "0.6292848", "0.62702036", "0.6268311", "0.6268212", "0.6241636", "0.62191254", "0.6211236", "0.6199394", "0.61826766", "0.6181979", "0.61644065", "0.61609507", "0.615349", "0.614455", "0.614016", "0.6137789", "0.6137409", "0.6125915", "0.61129487", "0.611075", "0.6109008", "0.6089158", "0.60642904", "0.60642904", "0.6061297", "0.6060087", "0.6054829", "0.60435104", "0.603616", "0.60172415", "0.599831", "0.59892017", "0.5987356", "0.5983885", "0.5971634", "0.5940027", "0.5935271", "0.5929239", "0.5916048", "0.5913515", "0.5910103", "0.5897847", "0.588875", "0.58856875", "0.5883996", "0.5882833", "0.5876646", "0.58732176", "0.586167", "0.5860703", "0.585966", "0.5857884", "0.5855691", "0.5852543", "0.5848121", "0.5845311", "0.5832736", "0.58112013", "0.5807011", "0.58063734", "0.58052796", "0.5795747", "0.57779205", "0.57395387", "0.572746", "0.57271355" ]
0.8547908
1
ParseURL uses url.Parse to parse operand as a url.
func ParseURL(operand string) (url *url.URL, err error) { return url.Parse(operand) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Parse(url string) *URL {\n\tresult := &URL{}\n\tresult.Input = url\n\tresult.Scheme, result.Opaque, result.Query, result.Fragment = splitMainPart(url)\n\tresult.Authority, result.Path = splitPath(result.Opaque)\n\tresult.UserInfo, result.Host, result.Port = splitAuthority(result.Authority)\n\tresult.Path = setDefaultPath(result.Path)\n\tresult.Scheme = setDefaultScheme(result.Scheme)\n\tresult.Port = setDefaultPort(result.Port, result.Scheme)\n\treturn result\n}", "func ParseURL(u string) (URLParts, error) {\n\treturn sas.ParseURL(u)\n}", "func ParseURL(url string) (*mgo.DialInfo, error) {\n\treturn mgo.ParseURL(url)\n}", "func ParseURL(str string) (*url.URL, error) {\n\treturn url.Parse(str)\n}", "func ParseURL(raw string) (*URL, error) {\n\tu, e := url.Parse(raw)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\treturn &URL{\n\t\tURL: *u,\n\t}, nil\n}", "func Parse(rawurl string) (*URL, error) {\n\tu, err := url.Parse(rawurl)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !u.IsAbs() {\n\t\treturn nil, errors.New(\"not absolute URL\")\n\t}\n\n\t// url.UserInfo (username:password) set nil\n\tu.User = nil\n\t// delete fragment\n\tu.Fragment = \"\"\n\n\tparsed := &URL{\n\t\tURL: u,\n\t}\n\n\tif strings.HasPrefix(u.Scheme, \"https\") || strings.HasPrefix(u.Scheme, \"ftps\") {\n\t\tparsed.TLS = true\n\t}\n\n\t// Delete port number\n\tindex := strings.Index(u.Host, \":\")\n\tif index != -1 {\n\t\tu.Host = u.Host[:index]\n\t}\n\n\tparsed.LocalScheme = isLocalScheme(u.Scheme)\n\tparsed.Origin = u.Scheme + \"://\" + u.Host + \"/\"\n\n\treturn parsed, nil\n}", "func parseUrl(rawurl string) (string, error) {\n\t_, err := url.Parse(rawurl)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn rawurl, nil\n}", "func parseURL(url string) string {\n\t_, err := strconv.Atoi(url)\n\tif err == nil {\n\t\t// If the input is just a number, assume it's a port number\n\t\turl = \"localhost:\" + url\n\t}\n\n\tif strings.HasPrefix(url, \"/\") {\n\t\t// If the input starts with a /, assume it's a relative path\n\t\turl = \"localhost\" + url\n\t}\n\n\tif !strings.HasPrefix(url, \"http://\") && !strings.HasPrefix(url, \"https://\") {\n\t\t// Add the protocol if it's not already there\n\t\turl = \"http://\" + url\n\t}\n\n\treturn url\n}", "func ParseURL(u string) (*AbtabURL, error) {\n\turl, err := url.Parse(u)\n\tif err != nil {\n\t\treturn nil, AbtabError{Message: fmt.Sprintf(\"Error: invalid url: '%s' :: %s\", u, err), CausedBy: err}\n\t}\n\n\treturn &AbtabURL{\n\t\tOriginalUrl: u,\n\t\tUrl: url,\n\t}, nil\n}", "func Parse(rawurl string) (u *url.URL, err error) {\n\tparsers := []Parser{\n\t\tParseTransport,\n\t\tParseScp,\n\t\tParseLocal,\n\t}\n\n\t// Apply each parser in turn; if the parser succeeds, accept its\n\t// result and return.\n\tfor _, p := range parsers {\n\t\tu, err = p(rawurl)\n\t\tif err == nil {\n\t\t\treturn u, err\n\t\t}\n\t}\n\n\t// It's unlikely that none of the parsers will succeed, since\n\t// ParseLocal is very forgiving.\n\treturn new(url.URL), fmt.Errorf(\"failed to parse %q\", rawurl)\n}", "func (f *FastURL) Parse(url []byte) error {\n\treturn Parse(f, url)\n}", "func ParseURL(uri string) *url.URL {\n\tif !strings.Contains(uri, \"://\") && !strings.HasPrefix(uri, \"//\") {\n\t\turi = \"//\" + uri\n\t}\n\n\turl, err := url.Parse(uri)\n\tif err != nil {\n\t\tlog.Printf(\"could not parse url %q: %v\\n\", uri, err)\n\t\treturn nil\n\t}\n\n\tif url.Scheme == \"\" {\n\t\turl.Scheme = \"http\"\n\t}\n\treturn url\n}", "func ParseURL(base string, paths ...string) string {\n\tpath := coalesce(paths...)\n\turl := path\n\n\tif !strings.HasPrefix(url, \"http\") && base != \"\" {\n\t\tif !strings.HasSuffix(base, \"/\") {\n\t\t\tbase = base + \"/\"\n\t\t}\n\n\t\tif strings.HasPrefix(path, \"/\") {\n\t\t\tpath = path[1:]\n\t\t}\n\n\t\turl = base + path\n\t}\n\n\treturn url\n}", "func URLParse(data []byte) (url *URL, err error) {\n\tcs, p, pe, eof := 0, 0, len(data), len(data)\n\tmark := 0\n\turl = new(URL)\n\n\t// this buffer is so we can unescape while we roll\n\tvar hex byte\n\tbuf := make([]byte, len(data))\n\tamt := 0\n\n\t%%{\n\t\taction mark { mark = p }\n\t\taction str_start { amt = 0 }", "func ParseURL(s string) *url.URL {\n\tu, err := url.Parse(s)\n\tif err != nil {\n\t\tlog.Printf(\"Cannot parse URL: %v\\n\", s)\n\t\treturn nil\n\t}\n\tlog.Printf(\"Parsed URL: %v\\n\", u)\n\treturn u\n}", "func ParseUrl(url string) (*url.URL, error) {\n\ttriedToFix := false\n\tfor {\n\t\tparsedUrl, err := giturls.Parse(url)\n\t\tif err != nil {\n\t\t\treturn parsedUrl, err\n\t\t}\n\n\t\tswitch parsedUrl.Scheme {\n\t\tcase \"http\":\n\t\t\tfallthrough\n\t\tcase \"https\":\n\t\t\treturn parsedUrl, nil\n\t\tcase \"ssh\":\n\t\t\treturn parsedUrl, nil\n\t\tcase \"file\":\n\t\t\tif triedToFix {\n\t\t\t\treturn parsedUrl, fmt.Errorf(\"cannot handle url format %v\", url)\n\t\t\t}\n\n\t\t\turl = fixUrl(url)\n\t\t\ttriedToFix = true\n\t\tdefault:\n\t\t\treturn parsedUrl, fmt.Errorf(\"cannot handle url format %v\", url)\n\t\t}\n\t}\n}", "func parseURL(rawurl string, defaultPort bool) (u *url.URL, err error) {\n\tu, err = url.Parse(rawurl)\n\tif err != nil {\n\t\treturn\n\t}\n\tif u.Host == \"\" {\n\t\terr = ErrNoHostURL\n\t\treturn\n\t}\n\tif u.Scheme != \"\" {\n\t\terr = ErrProtocolRelativeURL\n\t\treturn\n\t}\n\tif strings.HasSuffix(u.Path, \"/\") {\n\t\terr = ErrTrailingSlashURL\n\t\treturn\n\t}\n\tif defaultPort && u.Port() == \"\" {\n\t\tu.Host += \":\" + strconv.Itoa(int(DefaultPort))\n\t}\n\treturn\n}", "func ParseURL(rawurl string) (*URL, error) {\n\tif !hasSchemePattern.MatchString(rawurl) && scpLikeURLPattern.MatchString(rawurl) {\n\t\tmatched := scpLikeURLPattern.FindStringSubmatch(rawurl)\n\t\tuser := matched[1]\n\t\thost := matched[2]\n\t\tpath := matched[3]\n\n\t\treturn &URL{\n\t\t\tProtocol: \"ssh\",\n\t\t\tHost: host,\n\t\t\tOwner: user,\n\t\t\tRepoPath: trimPrefix(path),\n\t\t}, nil\n\t}\n\n\tparsed, err := url.Parse(rawurl)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\towner := \"\"\n\tif parsed.User != nil {\n\t\towner = parsed.User.Username()\n\t}\n\n\treturn &URL{\n\t\tProtocol: parsed.Scheme,\n\t\tHost: parsed.Host,\n\t\tOwner: owner,\n\t\tRepoPath: trimPrefix(parsed.EscapedPath()),\n\t}, nil\n}", "func parseURL(path, signature string, params *url.Values) string {\n\treturn \"\"\n}", "func ParseURL(raw string) (*URL, error) {\n\tstandardUrl, host, port, err := url.ParseURL(raw, getDefaultPort)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar u URL\n\tu.Scheme, err = ParseSchemeType(standardUrl.Scheme)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif u.Host = host; u.Host == \"\" {\n\t\treturn nil, fmt.Errorf(\"missing host\")\n\t}\n\n\tif u.Port = port; port == -1 {\n\t\treturn nil, fmt.Errorf(\"missing port\")\n\t}\n\n\tproto, err := parseProto(u.Scheme, standardUrl.RawQuery)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tu.Proto = proto\n\n\treturn &u, nil\n}", "func ParseUrl(ur string) (Nodes, error) {\n\tresp, err := http.Get(ur)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\treturn Parse(resp.Body)\n}", "func Parse(uri string) (URL, error) {\n\tif uri == \"\" {\n\t\treturn nil, ErrInvalidURL\n\t}\n\tif strings.HasPrefix(uri, \"git@\") {\n\t\tif !gitURLParser.MatchString(uri) {\n\t\t\treturn nil, ErrInvalidURL\n\t\t}\n\n\t\tmatches := gitURLParser.FindStringSubmatch(uri)\n\t\tif len(matches) != 3 {\n\t\t\treturn nil, ErrInvalidURL\n\t\t}\n\n\t\treturn gitURL{\n\t\t\tdomain: matches[1],\n\t\t\tpath: matches[2],\n\t\t}, nil\n\t}\n\tif strings.HasPrefix(uri, \"http://\") || strings.HasPrefix(uri, \"https://\") {\n\t\tu, err := url.Parse(uri)\n\t\tif err != nil {\n\t\t\treturn gitURL{}, nil\n\t\t}\n\t\tmatches := gitPathParser.FindStringSubmatch(u.Path)\n\t\tif len(matches) != 2 {\n\t\t\treturn nil, ErrInvalidURL\n\t\t}\n\n\t\treturn gitURL{\n\t\t\tdomain: u.Hostname(),\n\t\t\tpath: matches[1],\n\t\t}, nil\n\t}\n\treturn gitURL{}, nil\n}", "func ParseURL(s string) (*URL, error) {\n\tu, err := url.Parse(s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif u.Scheme != \"rtsp\" && u.Scheme != \"rtsps\" {\n\t\treturn nil, fmt.Errorf(\"unsupported scheme '%s'\", u.Scheme)\n\t}\n\n\treturn (*URL)(u), nil\n}", "func parseURL(rawurl string) (scheme string, address string, path string, err error) {\n\turi, err := url.Parse(rawurl)\n\tif err != nil {\n\t\treturn uri.Scheme, uri.Host, uri.Path, err\n\t}\n\n\tscheme = uri.Scheme\n\n\tswitch uri.Scheme {\n\tcase \"unix\":\n\t\tresult := strings.Split(uri.Path, \";\")\n\t\taddress = result[0]\n\t\tif len(result) > 1 {\n\t\t\tpath = result[1]\n\t\t}\n\tdefault:\n\t\taddress = uri.Host\n\t\tpath = uri.Path\n\t}\n\n\treturn\n}", "func ParseURL(url string) (*atom.Feed, string, error) {\r\n\tbyteValue, err := getContent(url)\r\n\tif err != nil {\r\n\t\treturn nil, \"\", err\r\n\t}\r\n\r\n\tdecoder := xml.NewDecoder(strings.NewReader(string(byteValue)))\r\n\tdecoder.CharsetReader = charset.NewReaderLabel\r\n\tfeed := atom.Feed{}\r\n\terr = decoder.Decode(&feed)\r\n\tif err != nil {\r\n\t\treturn nil, \"\", err\r\n\t}\r\n\r\n\treturn &feed, string(byteValue), nil\r\n}", "func ParseUrl(url string)(ip string,port int,path string,err error){\n\t// ipex := regexp.MustCompile(`http://(.*?)(:(\\d{2,4}))?(/.*)`)\n\tegex := regexp.MustCompile(`http://(\\d+\\.\\d+\\.\\d+\\.\\d+)(:(\\d{2,4}))?(/.*)`)\n\tregex := egex.FindAllStringSubmatch(url,-1)\n\tif len(regex) != 1 {\n\t\treturn ip,port,path,errors.New(\"url is invalid\")\n\t}\n\turlPar := regex[0]\n\tvar oport string\n\tip,oport,path = urlPar[1],urlPar[3],urlPar[4]\n\tlog.Println(urlPar)\n\tif oport == \"\" {\n\t\toport = \"80\"\n\t}\n\tport,err = strconv.Atoi(oport) \n\treturn\n}", "func ParseUrl(url string) (bool, string, string) {\n\tvar beginPos, endPos int\n\tvar host, uri string\n\tisHttps := strings.HasPrefix(url, \"https://\")\n\tif isHttps {\n\t\tbeginPos += 8\n\t}\n\tif strings.HasPrefix(url, \"http://\") {\n\t\tbeginPos += 7\n\t}\n\tendPos = strings.Index(url[beginPos:], \"/\")\n\tif endPos > 0 {\n\t\thost = url[beginPos : beginPos+endPos]\n\t\turiPos := strings.Index(url[beginPos+endPos:], \"?\")\n\t\tif uriPos > 0 {\n\t\t\turi = url[beginPos+endPos : beginPos+endPos+uriPos]\n\t\t} else {\n\t\t\turi = url[beginPos+endPos:]\n\t\t}\n\t} else {\n\t\thost = url[beginPos:]\n\t\turi = \"/\"\n\t}\n\treturn isHttps, host, uri\n}", "func Parse(f *FastURL, url []byte) error {\n\treturn parse(f, url, parseOption{\n\t\tParseProtocol: true,\n\t})\n}", "func URL(str string) *url.URL {\r\n\tparsed, err := url.Parse(str)\r\n\tif err != nil {\r\n\t\tpanic(err)\r\n\t}\r\n\treturn parsed\r\n}", "func ParseURL(path string) []string {\n\tvar parsedURL []string\n\tparsedURL = append(parsedURL, path)\n\n\tif strings.HasPrefix(path, \"/\") {\n\t\tpath = path[1:]\n\t}\n\tif strings.HasSuffix(path, \"/\") {\n\t\tpath = path[:(len(path) - 1)]\n\t}\n\tparsedURL = append(parsedURL, strings.Split(path, \"/\")...)\n\treturn parsedURL\n}", "func ParseURL(url string) (bucket string, key string, err error) {\n\tif strings.HasPrefix(url, s3URIPrefix) {\n\t\treturn parseS3URI(url)\n\t}\n\treturn parseHTTPURI(url)\n}", "func parseURL(pageURL string, relativeURL string) string {\n\tbase, err := url.Parse(pageURL)\n\tcheckErr(err)\n\trelative, err := url.Parse(relativeURL)\n\tcheckErr(err)\n\treturn base.ResolveReference(relative).String()\n}", "func Parse(uriStr string) (*URI, error) {\n\tu, err := url.Parse(uriStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !u.IsAbs() {\n\t\treturn nil, &url.Error{Op: \"gituri.Parse\", URL: uriStr, Err: errors.New(\"sourcegraph URI must be absolute\")}\n\t}\n\treturn &URI{*u}, nil\n}", "func url_parser(url_string string) url.Values{\n\tparams, err := url.ParseQuery(url_string)\n\tif err != nil{\n\t\tfmt.Println(err)\n\t}\n\treturn params\n}", "func (c *URL) Parse() string {\n\tEmpty := URL{}\n\tif *c != Empty {\n\t\tif c.Title == \"\" {\n\t\t\tc.Title = c.URL\n\t\t}\n\t\treturn fmt.Sprintf(`\"title\": \"%v\", \"url\": \"%v\" `, c.Title, c.URL)\n\t}\n\treturn \"\"\n}", "func (r *repoURL) parse(input string) error {\n\tvar t string\n\tc := strings.Split(input, \"/\")\n\tswitch l := len(c); {\n\t// owner/repo format is not supported for gitlab, it's github-only\n\tcase l == 2:\n\t\treturn sce.WithMessage(sce.ErrScorecardInternal, fmt.Sprintf(\"gitlab repo must specify host: %s\", input))\n\tcase l >= 3:\n\t\tt = input\n\t}\n\n\t// Allow skipping scheme for ease-of-use, default to https.\n\tif !strings.Contains(t, \"://\") {\n\t\tt = \"https://\" + t\n\t}\n\n\tu, e := url.Parse(t)\n\tif e != nil {\n\t\treturn sce.WithMessage(sce.ErrScorecardInternal, fmt.Sprintf(\"url.Parse: %v\", e))\n\t}\n\n\tconst splitLen = 2\n\tsplit := strings.SplitN(strings.Trim(u.Path, \"/\"), \"/\", splitLen)\n\tif len(split) != splitLen {\n\t\treturn sce.WithMessage(sce.ErrorInvalidURL, fmt.Sprintf(\"%v. Expected full repository url\", input))\n\t}\n\n\tr.scheme, r.host, r.owner, r.project = u.Scheme, u.Host, split[0], split[1]\n\treturn nil\n}", "func ParseURL(jsFileURL string, fileContents string) (string, error) {\n\tmatch := re.FindStringSubmatch(fileContents)\n\tif len(match) < 2 {\n\t\treturn \"\", nil\n\t}\n\trel := match[2]\n\n\tu, err := url.Parse(rel)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tbase, err := url.Parse(jsFileURL)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn base.ResolveReference(u).String(), nil\n}", "func (c *crawlerDS) parseURL(uri string) (bool, string) {\n\tpg, err := url.Parse(uri)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif pg.Scheme == \"https\" || pg.Scheme == \"http\" {\n\t\treturn true, pg.Host\n\t}\n\treturn false, \"\"\n\n}", "func (f *FastURL) ParseWithoutProtocol(url []byte) error {\n\treturn ParseWithoutProtocol(f, url)\n}", "func parseGitURL(gitURL string) (*url.URL, error) {\n\tgitURL = strings.TrimPrefix(gitURL, gitSSHExplicit)\n\tgitURL = strings.TrimSuffix(gitURL, gitSuffix)\n\tswitch {\n\tcase strings.HasPrefix(gitURL, gitSSHImplicit):\n\t\tgitURL = strings.TrimPrefix(gitURL, gitSSHImplicit)\n\t\tgitURL = strings.Replace(gitURL, \":\", \"/\", 1)\n\t\tgitURL = \"https://\" + gitURL\n\n\tcase strings.HasPrefix(gitURL, gitHTTPS):\n\t\t// nothing to do\n\n\tdefault:\n\t\treturn nil, errors.Errorf(\"unknown git remote URL specification: %q\", gitURL)\n\t}\n\n\treturn url.Parse(gitURL)\n}", "func parseURL(ctx context.Context, idpUrl string) (string, error) {\n\tlog := log.Logger(ctx, \"internal.utils.oidc\", \"parseURL\")\n\tresp, err := url.Parse(idpUrl)\n\tif err != nil {\n\t\tlog.Error(err, \"unable to parse the idp url\")\n\t\treturn \"\", err\n\t}\n\n\tif resp.Scheme != \"https\" {\n\t\tlog.Error(errors.New(\"OIDC IDP url must start with https\"), \"OIDC IDP url must start with https\", \"obtained\", resp.Scheme)\n\t\treturn \"\", err\n\t}\n\n\tport := resp.Port()\n\n\tif resp.Port() == \"\" {\n\t\tport = \"443\"\n\t}\n\thostName := fmt.Sprintf(\"%s:%s\", resp.Host, port)\n\tlog.Info(\"url parsed successfully\", \"hostName\", hostName)\n\treturn hostName, nil\n}", "func Parse(urlValues url.Values, target interface{}) error {\n\treturn DefaultParser.Parse(urlValues, target)\n}", "func ParseURL(serverFlag string) (*UrlRes, int, string) {\n\n\tpURL := &UrlRes{serverFlag, \"\", \"\"}\n\n\t// the URL golang Parse method has the limitation that when we pass in a host that is a string\n\t// and not an ip, without the protocol scheme, it mis-interprets the url string. For such cases we\n\t// need to explicitely make sure that we are missing a protocol scheme.\n\n\t// If no protocol exists, then append http:// as default protocol.\n\n\tif !strings.HasPrefix(strings.ToLower(serverFlag), \"https://\") &&\n\t\t!strings.HasPrefix(strings.ToLower(serverFlag), \"http://\") &&\n\t\t!strings.HasPrefix(strings.ToLower(serverFlag), \"couchbase://\") &&\n\t\t!strings.HasPrefix(strings.ToLower(serverFlag), \"couchbases://\") {\n\t\t//There is something else wrong and we need to throw an error.\n\t\tserverFlag = \"http://\" + serverFlag\n\t}\n\n\t//Parse the url\n\tparsedURL, err := url.Parse(serverFlag)\n\tif err != nil {\n\t\treturn pURL, errors.INVALID_URL, err.Error()\n\t}\n\n\tif parsedURL.Host == \"\" {\n\t\treturn pURL, errors.INVALID_URL, INVALIDHOST\n\t}\n\n\t// Check if the input url is a DNS SRV\n\t_, addr, err := net.LookupSRV(parsedURL.Scheme, \"tcp\", parsedURL.Hostname())\n\tif err == nil {\n\t\t// It is a DNS SRV .. Has couchbase or couchbases as a scheme\n\t\tparsedURL.Host = addr[0].Target\n\t}\n\n\t// We now have a valid URL. Check if we have a port\n\t_, portNo, err := net.SplitHostPort(parsedURL.Host)\n\n\t// couchbase:// and couchbases:// will represent http:// ... :8091 and\n\t// https:// ... 18091 respectively. If the port is specified along with\n\t// the scheme for this case, we throw an error.\n\n\tif parsedURL.Hostname() != \"\" {\n\t\tparsedURL.Host = parsedURL.Hostname()\n\t}\n\n\tif portNo == \"\" {\n\t\tif strings.ToLower(parsedURL.Scheme) == \"couchbase\" || strings.ToLower(parsedURL.Scheme) == \"couchbases\" {\n\n\t\t\tif strings.ToLower(parsedURL.Scheme) == \"couchbase\" {\n\t\t\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, \"8091\")\n\t\t\t\tparsedURL.Scheme = \"http\"\n\n\t\t\t} else {\n\t\t\t\tparsedURL.Scheme = \"https\"\n\t\t\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, \"18091\")\n\t\t\t}\n\n\t\t} else {\n\t\t\tif strings.ToLower(parsedURL.Scheme) == \"http\" {\n\t\t\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, \"8091\")\n\n\t\t\t} else if strings.ToLower(parsedURL.Scheme) == \"https\" {\n\t\t\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, \"18091\")\n\t\t\t}\n\t\t}\n\t} else {\n\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, portNo)\n\t\t// Cannot give port with couchbase:// couchbases://\n\t\tif strings.ToLower(parsedURL.Scheme) == \"couchbase\" || strings.ToLower(parsedURL.Scheme) == \"couchbases\" {\n\t\t\treturn pURL, errors.INVALID_URL, INVALIDPORT\n\t\t} else {\n\t\t\tif err != nil {\n\t\t\t\treturn pURL, errors.INVALID_URL, err.Error()\n\t\t\t}\n\t\t}\n\t}\n\n\tpURL.Password, _ = parsedURL.User.Password()\n\tpURL.Username = parsedURL.User.Username()\n\tpURL.ServerFlag = parsedURL.String()\n\n\treturn pURL, 0, \"\"\n}", "func parseServerURL(rawURL string) (*url.URL, error) {\n\tuu, err := url.Parse(rawURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif uu.Scheme == \"\" {\n\t\treturn nil, fmt.Errorf(\"url %s missing scheme\", rawURL)\n\t}\n\tif uu.Path != \"\" && uu.Path != \"/\" {\n\t\treturn nil, fmt.Errorf(\"url %s should not contain path\", rawURL)\n\t}\n\treturn uu, nil\n}", "func (s *URLParser) Parse(url *url.URL) (*Source, error) {\n\turlPath := url.Path\n\tif urlPath == \"\" {\n\t\treturn nil, errors.New(\"Empty url\")\n\t}\n\tp := strings.SplitN(urlPath[1:], \"/\", 2)\n\thash := p[0]\n\tif hash == \"\" {\n\t\treturn nil, errors.New(\"Empty hash\")\n\t}\n\tsourceType := \"default\"\n\tfor _, v := range s.configs.GetMods() {\n\t\tif hash == v {\n\t\t\tsourceType = v\n\t\t\tbreak\n\t\t}\n\t}\n\tsourceName := s.configs.GetMod(sourceType).Name\n\tif sourceType == \"default\" && !checkHash(hash) {\n\t\treturn nil, errors.New(fmt.Sprintf(\"Wrong hash=%s\", hash))\n\t}\n\tpath := \"/\"\n\tif len(p) > 1 {\n\t\tpath += p[1]\n\t}\n\t// path = filepath.Clean(path)\n\tnewPath, mod, err := s.extractMod(path)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to extract mod from path=%s\", path)\n\t}\n\toriginPath := newPath\n\tvar tempMod *Mod\n\tfor {\n\t\toriginPath, tempMod, err = s.extractMod(originPath)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"failed to extract mod from path=%s\", path)\n\t\t}\n\t\tif tempMod == nil {\n\t\t\tbreak\n\t\t}\n\t}\n\tss := &Source{\n\t\tInfoHash: hash,\n\t\tPath: newPath,\n\t\tOriginPath: originPath,\n\t\tToken: url.Query().Get(\"token\"),\n\t\tApiKey: url.Query().Get(\"api-key\"),\n\t\tQuery: url.RawQuery,\n\t\tType: sourceType,\n\t\tName: sourceName,\n\t\tMod: mod,\n\t}\n\treturn ss, nil\n}", "func parseURL(baseurl, ref string) string {\n\tbase, _ := url.Parse(baseurl)\n\tu, err := url.Parse(ref)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn base.ResolveReference(u).String()\n}", "func parseNotifyURL(u string) (notifyURL *model.NotifyURL, err error) {\n\tvar parsedURL *url.URL\n\tparsedURL, err = url.Parse(u)\n\tif err != nil {\n\t\treturn\n\t}\n\tnotifyURL = &model.NotifyURL{\n\t\tRawURL: u,\n\t\tSchema: parsedURL.Scheme,\n\t\tHost: parsedURL.Host,\n\t\tPath: parsedURL.Path,\n\t\tQuery: parsedURL.Query(),\n\t}\n\treturn\n}", "func MustParseURL(target string) *url.URL {\n\tu, err := url.Parse(target)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Error parsing target(%s): %v\", target, err))\n\t}\n\treturn u\n}", "func ParseEndpointURL(url string, options ...Option) (Endpoint, error) {\n\tu, err := neturl.Parse(url)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"parse url [%s]: %v\", url, err)\n\t}\n\treturn NewEndpoint(u.Scheme, net.JoinHostPort(u.Hostname(), u.Port()), options...), nil\n}", "func (uri *URI) Parse(isConnect bool, reqURI []byte) {\n\turi.Reset()\n\turi.isConnect = isConnect\n\turi.full = reqURI\n\tif len(reqURI) == 0 {\n\t\treturn\n\t}\n\tfragmentIndex := bytes.IndexByte(reqURI, '#')\n\tif fragmentIndex >= 0 {\n\t\turi.fragments = reqURI[fragmentIndex:]\n\t\turi.parseWithoutFragments(reqURI[:fragmentIndex])\n\t} else {\n\t\turi.parseWithoutFragments(reqURI)\n\t}\n\tif !isConnect && len(uri.path) == 0 {\n\t\turi.path = []byte(\"/\")\n\t}\n\tif isConnect {\n\t\turi.scheme = uri.scheme[:0]\n\t\turi.path = uri.path[:0]\n\t\turi.queries = uri.queries[:0]\n\t\turi.fragments = uri.fragments[:0]\n\t}\n\turi.hostInfo.ParseHostWithPort(string(uri.host), isConnect)\n}", "func ParseFromURL(url string) (*Node, error) {\n\tresponse, err := http.Get(url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttokenizer := html.NewTokenizer(response.Body)\n\treturn dive(tokenizer), nil\n}", "func safeParse(rawURL string) (*dburl.URL, error) {\n\tparsed, err := dburl.Parse(rawURL)\n\tif err != nil {\n\t\tif uerr := new(url.Error); errors.As(err, &uerr) {\n\t\t\treturn nil, uerr.Err\n\t\t}\n\t\treturn nil, errors.New(\"invalid URL\")\n\t}\n\treturn parsed, nil\n}", "func ParseURL(url string) Request {\n\tr := new(Request)\n\turlSections := strings.Split(url, \"/\")\n\n\tfor _, section := range urlSections {\n\t\t//first check if match table name\n\t\tmatchTableName, err := regexp.MatchString(\"^table=\", section)\n\t\tcheck(err)\n\n\t\t//then check fields\n\t\tmatchFields, err := regexp.MatchString(\"^fields=\", section)\n\t\tcheck(err)\n\n\t\tif matchTableName {\n\t\t\tif r.TableName == \"\" {\n\t\t\t\tr.TableName = section[6:]\n\t\t\t} else {\n\t\t\t\terr := errors.New(\"Error: multiple table name requests defined.\")\n\t\t\t\tcheck(err)\n\t\t\t}\n\t\t} else if matchFields {\n\t\t\tif r.Fields == \"\" {\n\t\t\t\tr.Fields = section[7:]\n\t\t\t} else {\n\t\t\t\terr := errors.New(\"Error: multiple fields defined.\")\n\t\t\t\tcheck(err)\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn *r\n}", "func (mr *MockRaindropRepositoryMockRecorder) ParseURL(ctx, url interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ParseURL\", reflect.TypeOf((*MockRaindropRepository)(nil).ParseURL), ctx, url)\n}", "func (p *Parser) Parse(urlValues url.Values, target interface{}) error {\n\tif urlValues == nil {\n\t\treturn ErrInvalidURLValues\n\t}\n\ttargetValue := reflect.ValueOf(target)\n\tif targetValue.Kind() != reflect.Ptr || targetValue.IsNil() {\n\t\treturn ErrNonPointerTarget\n\t}\n\n\ttargetElement := targetValue.Elem()\n\ttargetType := targetElement.Type()\n\n\tfor i := 0; i < targetType.NumField(); i++ {\n\t\tif err := p.ParseField(targetType.Field(i), targetElement.Field(i), urlValues); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func ParseEndpoint(endpoint string) (*url.URL, error) {\n\tendpoint = FormatEndpoint(endpoint)\n\n\tu, err := url.Parse(endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn u, nil\n}", "func MustParseURL(s string) *URL {\n\tu, err := ParseURL(s)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn u\n}", "func MustParseURL(rawURL string) *url.URL {\n\turl, err := url.Parse(rawURL)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn url\n}", "func ParseTransport(rawurl string) (*url.URL, error) {\n\tu, err := url.Parse(rawurl)\n\tif err == nil && !Transports.Valid(u.Scheme) {\n\t\terr = fmt.Errorf(\"scheme %q is not a valid transport\", u.Scheme)\n\t}\n\tif u != nil && u.User == nil {\n\t\tu.User = url.User(\"\")\n\t}\n\treturn u, err\n}", "func ParseUrl(urlstring string) map[string][]string{\n values,err := url.ParseQuery(urlstring);\n if(err != nil){\n panic(\"boom\")\n }\n return values;\n}", "func ParseLocal(rawurl string) (*url.URL, error) {\n\treturn &url.URL{\n\t\tScheme: \"file\",\n\t\tUser: url.User(\"\"),\n\t\tHost: \"\",\n\t\tPath: rawurl,\n\t}, nil\n}", "func ParseWithoutProtocol(f *FastURL, url []byte) error {\n\treturn parse(f, url, parseOption{\n\t\tParseProtocol: false,\n\t})\n}", "func (u *Parser) parse(value interface{}) (interface{}, error) {\n\tswitch m := value.(type) {\n\tcase string:\n\t\treturn parseURI(m)\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"type '%T' cannot be parsed as URI\", value)\n\t}\n}", "func ParseChannelUrl(cUrl string) *LinkUrl {\n\t//logger.Log(fmt.Sprintf(\"Entering LinkUrl:ParseChannelUrl w/ URL string as '%s'\", cUrl))\n\tif cUrl == \"\" {\n\t\treturn DefaultLinkUrl()\n\t}\n\n\t//logger.Log(fmt.Sprintf(\"Returning LinkUrl:ParseChannelUrl w/ URL string as '%s'\", cUrl))\n\treturn NewLinkUrl(cUrl)\n}", "func TestHelperParseURL(t *testing.T) {\n\ttests := []struct {\n\t\turl string\n\t\texpectedURL string\n\t\terr error\n\t}{\n\t\t{url: \"foobar.docker.io\", expectedURL: \"//foobar.docker.io\"},\n\t\t{url: \"foobar.docker.io:2376\", expectedURL: \"//foobar.docker.io:2376\"},\n\t\t{url: \"//foobar.docker.io:2376\", expectedURL: \"//foobar.docker.io:2376\"},\n\t\t{url: \"http://foobar.docker.io:2376\", expectedURL: \"http://foobar.docker.io:2376\"},\n\t\t{url: \"https://foobar.docker.io:2376\", expectedURL: \"https://foobar.docker.io:2376\"},\n\t\t{url: \"https://foobar.docker.io:2376/some/path\", expectedURL: \"https://foobar.docker.io:2376/some/path\"},\n\t\t{url: \"https://foobar.docker.io:2376/some/other/path?foo=bar\", expectedURL: \"https://foobar.docker.io:2376/some/other/path\"},\n\t\t{url: \"/foobar.docker.io\", err: errors.New(\"no hostname in URL\")},\n\t\t{url: \"ftp://foobar.docker.io:2376\", err: errors.New(\"unsupported scheme: ftp\")},\n\t}\n\n\tfor _, te := range tests {\n\t\tu, err := Parse(te.url)\n\n\t\tif te.err == nil && err != nil {\n\t\t\tt.Errorf(\"Error: failed to parse URL %q: %s\", te.url, err)\n\t\t\tcontinue\n\t\t}\n\t\tif te.err != nil && err == nil {\n\t\t\tt.Errorf(\"Error: expected error %q, got none when parsing URL %q\", te.err, te.url)\n\t\t\tcontinue\n\t\t}\n\t\tif te.err != nil && err.Error() != te.err.Error() {\n\t\t\tt.Errorf(\"Error: expected error %q, got %q when parsing URL %q\", te.err, err, te.url)\n\t\t\tcontinue\n\t\t}\n\t\tif u != nil && u.String() != te.expectedURL {\n\t\t\tt.Errorf(\"Error: expected URL: %q, but got %q for URL: %q\", te.expectedURL, u.String(), te.url)\n\t\t}\n\t}\n}", "func NormalizeURL(u string) (*url.URL, error) {\n\tif !isHTTPPrepended(u) {\n\t\tu = prependHTTP(u)\n\t}\n\n\t// We do it like browsers, just remove the trailing slash.\n\t// This will save us from a lot of problems later.\n\tu = strings.TrimSuffix(u, \"/\")\n\n\tparsedURL, err := url.Parse(u)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Unable to parse URL/hostname: %s. %s\", u, err)\n\t}\n\n\treturn parsedURL, nil\n}", "func ParseAddr(addr string) (*url.URL, error) {\n\tu, err := url.Parse(addr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif u.Scheme != SchemaADC && u.Scheme != SchemaADCS {\n\t\treturn u, fmt.Errorf(\"unsupported protocol: %q\", u.Scheme)\n\t}\n\tu.Path = strings.TrimRight(u.Path, \"/\")\n\treturn u, nil\n}", "func processUrl(new_url, domain string) string {\n\tu, err := url.Parse(new_url)\n\tif err != nil {\n\t\tlog.Println(\"URL from page so damaged even url.Parse won't handle it -\", err)\n\t\treturn \"\"\n\t}\n\tbase, err := url.Parse(domain)\n\tif err != nil {\n\t\tlog.Fatal(\"Should never happen, domain validated earlier -\", err)\n\t}\n\tprocessed := base.ResolveReference(u)\n\tprocessed.Fragment = \"\"\n\tprocessed.RawQuery = \"\"\n\treturn processed.String()\n}", "func ParseEndpoint(endpt string) (*url.URL, error) {\n\tvar err error\n\tif endpt == \"\" {\n\t\tendpt, err = util.ParseEnvVar(common.ImporterEndpoint, false)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif endpt == \"\" {\n\t\t\treturn nil, errors.Errorf(\"endpoint %q is missing or blank\", common.ImporterEndpoint)\n\t\t}\n\t}\n\treturn url.Parse(endpt)\n}", "func trimURL(URLStr string) (*url.URL, error) {\n\tparsedURL, err := url.Parse(URLStr)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"parsing %s: %v\", URLStr, err)\n\t}\n\tschemeHost := &url.URL{\n\t\tScheme: parsedURL.Scheme,\n\t\tHost: parsedURL.Host,\n\t}\n\treturn schemeHost, nil\n}", "func _gpurl(link, partname string, strict bool) {\n\tres, err := parseURL(link, partname)\n\thandleError(err)\n\tif res != \"\" {\n\t\tfmt.Println(res)\n\t} else if strict {\n\t\thandleError(fmt.Errorf(\"%s is empty and strict checking is enabled\", partname))\n\t}\n}", "func (r *Request) URL(uri string) *Request {\n\tr.URLStruct, r.Error = url.Parse(uri)\n\treturn r\n}", "func parseURI(uri string) (string, string, error) {\n\tu, err := url.Parse(uri)\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\n\treturn strings.Split(u.Host, \".\")[0], strings.TrimLeft(u.Path, \"/\"), nil\n}", "func (entry *sitemap_url_entry) parseSitemapURL() (sitemap_url *SitemapURL, err *urlParseError) {\n\tsitemap_url = &SitemapURL{}\n\tvar temp_err error = nil\n\tsitemap_url.location, temp_err = ParseURL(entry.URL_raw)\n\tif temp_err != nil && err == nil {\n\t\terr = newUrlParseError(temp_err.Error(), true)\n\t}\n\tsitemap_url.priority, temp_err = parsePriority(entry.Priority)\n\tif temp_err != nil && err == nil {\n\t\terr = newUrlParseError(temp_err.Error(), false)\n\t}\n\tsitemap_url.change_frequency, temp_err = parseChangeFrequency(entry.Change_freq)\n\tif temp_err != nil && err == nil {\n\t\terr = newUrlParseError(temp_err.Error(), false)\n\t}\n\tsitemap_url.last_modified, temp_err = parseLastModified(entry.Last_mod)\n\tif temp_err != nil && err == nil {\n\t\terr = newUrlParseError(temp_err.Error(), false)\n\t}\n\n\treturn sitemap_url, err\n\n}", "func ParseHostURL(host string) (*url.URL, error) {\n\tif !strings.Contains(host, \"://\") {\n\t\thost = \"tcp://\" + host\n\t}\n\tprotoAddrParts := strings.SplitN(host, \"://\", 2)\n\tif len(protoAddrParts) == 1 {\n\t\treturn nil, fmt.Errorf(\"unable to parse agent host `%s`\", host)\n\t}\n\tvar basePath string\n\tproto, addr := protoAddrParts[0], protoAddrParts[1]\n\tif proto == \"tcp\" {\n\t\tparsed, err := url.Parse(\"tcp://\" + addr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\taddr = parsed.Host\n\t\tbasePath = parsed.Path\n\t}\n\treturn &url.URL{\n\t\tScheme: proto,\n\t\tHost: addr,\n\t\tPath: basePath,\n\t}, nil\n}", "func mkurl(s string) (u *url.URL) {\n\tvar err error\n\tu, err = url.Parse(s)\n\tif err != nil {\n\t\tpanic(fmt.Sprint(\"string is not a valid URL:\", s))\n\t}\n\treturn\n}", "func ParseURLs(values ...string) (FixedEndpoints, error) {\n\turls, err := xhttp.ApplyURLParser(url.Parse, values...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn FixedEndpoints(urls), nil\n}", "func ParseRemoteURL(remoteURL string) (*RemoteURL, error) {\n\tremoteURL = strings.TrimSpace(remoteURL)\n\tmatch := sshExp.FindStringSubmatch(remoteURL)\n\tif len(match) > 0 {\n\t\tpath, repoName := getPathAndRepoName(match[3])\n\t\treturn &RemoteURL{\n\t\t\tPath: path,\n\t\t\tHost: match[2],\n\t\t\tRepoName: repoName,\n\t\t\tURL: match[0],\n\t\t\tPort: 22,\n\n\t\t\tFormat: SSHFormat,\n\t\t\tSSHUser: match[1],\n\t\t}, nil\n\t}\n\tmatch = httpsExp.FindStringSubmatch(remoteURL)\n\tif len(match) > 0 {\n\t\tvar port int\n\t\tvar err error\n\t\tif len(match[3]) > 0 {\n\t\t\tport, err = strconv.Atoi(match[3])\n\t\t\tif err != nil {\n\t\t\t\tlog.Panicf(\"git: invalid port: %s\", match[3])\n\t\t\t}\n\t\t} else {\n\t\t\tport = 443\n\t\t}\n\t\tpath, repoName := getPathAndRepoName(match[4])\n\t\treturn &RemoteURL{\n\t\t\tPath: path,\n\t\t\tHost: match[1],\n\t\t\tRepoName: repoName,\n\t\t\tURL: match[0],\n\t\t\tPort: port,\n\n\t\t\tFormat: HTTPSFormat,\n\t\t}, nil\n\t}\n\treturn nil, fmt.Errorf(\"Could not parse %s as a git remote\", remoteURL)\n}", "func ParseParams(urlString string) (Params, error) {\n\tp := NewParams()\n\tif !strings.HasPrefix(urlString, \"http\") {\n\t\turlString = \"https://\" + urlString\n\t}\n\tu, err := url.Parse(urlString)\n\tif err != nil {\n\t\treturn p, err\n\t}\n\tp.Query = u.Query()\n\tif u.Scheme == \"\" {\n\t\tu.Scheme = \"https\"\n\t}\n\tif u.Path != \"\" {\n\t\tp.Prefix = strings.Trim(u.Path, \"/\")\n\t}\n\tu.RawQuery = \"\"\n\tu.Fragment = \"\"\n\tu.Path = \"\"\n\tu.RawPath = \"\"\n\tp.Server = u.String()\n\treturn p, nil\n}", "func parseURI(value string) (map[string]interface{}, error) {\n\tm := make(map[string]interface{})\n\n\tif strings.HasPrefix(value, \"?\") {\n\t\t// remove the query string '?' prefix before parsing\n\t\tv, err := url.ParseQuery(value[1:])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn queryToMap(v, m), nil\n\t}\n\n\tx, err := url.ParseRequestURI(value)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn urlToMap(x, m), nil\n}", "func parseUrlforStu(url string) *gofeed.Feed {\n\tfp := gofeed.NewParser()\n\toutput, _ := fp.ParseURL(url)\n\treturn output\n}", "func TestURLParse(t *testing.T) {\n\ttests := []struct {\n\t\tin string\n\t\twant string\n\t}{\n\t\t{\"ssh://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"ssh://[email protected]/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"ssh://[email protected]:1234/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"ssh://[email protected]/~user/grdl/git-get.git\", \"github.com/user/grdl/git-get\"},\n\t\t{\"git+ssh://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"[email protected]:grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"[email protected]:/~user/grdl/git-get.git\", \"github.com/user/grdl/git-get\"},\n\t\t{\"git://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"git://github.com/~user/grdl/git-get.git\", \"github.com/user/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"http://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/git-get.git\", \"github.com/git-get\"},\n\t\t{\"https://github.com/git-get\", \"github.com/git-get\"},\n\t\t{\"https://github.com/grdl/sub/path/git-get.git\", \"github.com/grdl/sub/path/git-get\"},\n\t\t{\"https://github.com:1234/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get.git/\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get/\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get/////\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get.git/////\", \"github.com/grdl/git-get\"},\n\t\t{\"ftp://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"ftps://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"rsync://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"local/grdl/git-get/\", \"github.com/local/grdl/git-get\"},\n\t\t{\"file://local/grdl/git-get\", \"local/grdl/git-get\"},\n\t}\n\n\tfor _, test := range tests {\n\t\turl, err := ParseURL(test.in, cfg.Defaults[cfg.KeyDefaultHost], cfg.Defaults[cfg.KeyDefaultScheme])\n\t\tassert.NoError(t, err)\n\n\t\tgot := URLToPath(*url, false)\n\t\tassert.Equal(t, test.want, got)\n\t}\n}", "func mustPU(rawurl string) *url.URL {\n\tif u, err := url.Parse(rawurl); err != nil {\n\t\tpanic(err)\n\t} else {\n\t\treturn u\n\t}\n}", "func (c *Config) ParseAddress(address string) (*url.URL, error) {\n\tu, err := url.Parse(address)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc.Address = address\n\n\tif strings.HasPrefix(address, \"unix://\") {\n\t\t// When the address begins with unix://, always change the transport's\n\t\t// DialContext (to match previous behaviour)\n\t\tsocket := strings.TrimPrefix(address, \"unix://\")\n\n\t\tif transport, ok := c.HttpClient.Transport.(*http.Transport); ok {\n\t\t\ttransport.DialContext = func(context.Context, string, string) (net.Conn, error) {\n\t\t\t\treturn net.Dial(\"unix\", socket)\n\t\t\t}\n\n\t\t\t// Since the address points to a unix domain socket, the scheme in the\n\t\t\t// *URL would be set to `unix`. The *URL in the client is expected to\n\t\t\t// be pointing to the protocol used in the application layer and not to\n\t\t\t// the transport layer. Hence, setting the fields accordingly.\n\t\t\tu.Scheme = \"http\"\n\t\t\tu.Host = socket\n\t\t\tu.Path = \"\"\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"attempting to specify unix:// address with non-transport transport\")\n\t\t}\n\t} else if strings.HasPrefix(c.Address, \"unix://\") {\n\t\t// When the address being set does not begin with unix:// but the previous\n\t\t// address in the Config did, change the transport's DialContext back to\n\t\t// use the default configuration that cleanhttp uses.\n\n\t\tif transport, ok := c.HttpClient.Transport.(*http.Transport); ok {\n\t\t\ttransport.DialContext = cleanhttp.DefaultPooledTransport().DialContext\n\t\t}\n\t}\n\n\treturn u, nil\n}", "func parseURLForGithubAPIEndpoint(urlString string) (*url.URL, string) {\n\tu, err := url.Parse(urlString)\n\tif err != nil {\n\t\tlog.Fatalf(\"an error occurred parsing url %s for potential repository: %s\", urlString, err)\n\t}\n\n\t// parse hostname and path for potential github repo api endpoint\n\thostname := u.Hostname()\n\tpath := u.Path\n\trepoURL := github.GetApiEndpoint(hostname, path)\n\treturn u, repoURL\n}", "func NewURL(u *url.URL) URL {\n\treturn URL{\n\t\tURL: u,\n\t}\n}", "func (r *Repository) Parse() error {\n\trepository := r.URL\n\ti := strings.IndexByte(repository, '/')\n\tn := strings.Count(repository, \"/\")\n\tswitch {\n\tcase n == 0:\n\t\tuser := \"library\"\n\t\timage := repository\n\t\t*r = Repository{\n\t\t\tDockerHub: &RepositoryDockerHub{\n\t\t\t\tUser: user,\n\t\t\t\tImage: image,\n\t\t\t},\n\t\t}\n\tcase n == 1:\n\t\tuser := repository[:i]\n\t\timage := repository[i+1:]\n\t\t*r = Repository{\n\t\t\tDockerHub: &RepositoryDockerHub{\n\t\t\t\tUser: user,\n\t\t\t\tImage: image,\n\t\t\t},\n\t\t}\n\tcase n >= 2:\n\t\tregistry, image := repository[:i], repository[i+1:]\n\t\t*r = Repository{\n\t\t\tV2: &RepositoryV2{\n\t\t\t\tRegistry: registry,\n\t\t\t\tImage: image,\n\t\t\t},\n\t\t}\n\tdefault:\n\t\treturn fmt.Errorf(\"cannot determine registry: %s\", repository)\n\t}\n\treturn nil\n}", "func parseListOptions(urlStr string) (*ListOptions, error) {\n\tu, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tpage, err := strconv.Atoi(u.Query().Get(\"page\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif page < 0 {\n\t\tpage = 1\n\t}\n\n\tlimit, err := strconv.Atoi(u.Query().Get(\"limit\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Make sure we stay within page size limits.\n\tif limit > 20 || limit < 0 {\n\t\tlimit = 20\n\t}\n\n\treturn &ListOptions{\n\t\tPage: page,\n\t\tPerPage: limit,\n\t}, nil\n}", "func parseDataURL(url string, du *dataURL) error {\n\n\t// Check prefix\n\tif !isDataURL(url) {\n\t\treturn fmt.Errorf(\"specified string is not a data URL\")\n\t}\n\n\t// Separate header from data\n\tbody := url[len(dataURLprefix):]\n\tparts := strings.Split(body, \",\")\n\tif len(parts) != 2 {\n\t\treturn fmt.Errorf(\"data URI contains more than one ','\")\n\t}\n\tdu.Data = parts[1]\n\n\t// Separate media type from optional encoding\n\tres := strings.Split(parts[0], \";\")\n\tdu.MediaType = res[0]\n\tif len(res) < 2 {\n\t\treturn nil\n\t}\n\tif len(res) >= 2 {\n\t\tdu.Encoding = res[1]\n\t}\n\treturn nil\n}", "func ParseURLs(content string) []string { return parseURLsMax(content, -1) }", "func (b *Backend) ParseAddress(addr string) (err error) {\n\tif b.Addr, err = url.Parse(addr); err != nil {\n\t\treturn err\n\t}\n\n\tif b.Addr.Scheme == \"\" {\n\t\tb.Addr.Scheme = \"http\"\n\t}\n\n\thttps := b.Addr.Scheme == \"https\"\n\tb.Host = b.Addr.Host\n\n\tif b.Addr.Port() == \"\" {\n\t\tif https {\n\t\t\tb.Host += \":443\"\n\t\t} else {\n\t\t\tb.Host += \":80\"\n\t\t}\n\t}\n\n\treturn nil\n}", "func (c *IMAGE) Parse() string {\n\tEmpty := IMAGE{}\n\tif *c != Empty {\n\t\treturn fmt.Sprintf(`\"image\": { \"url\": \"%v\" }`, c.URL)\n\t}\n\treturn \"\"\n}", "func ExtractURL(rawURL string) (urlEntity URLEntity, err error) {\n\tu, err := url.Parse(rawURL)\n\tif err != nil {\n\t\treturn urlEntity, fmt.Errorf(\"URL not in a valid format: %s\", err)\n\t}\n\n\tif (u.Scheme != \"http\" && u.Scheme != \"https\") || u.Host == \"\" {\n\t\treturn urlEntity, fmt.Errorf(\"URL provided is not absolute\")\n\t}\n\n\turlEntity.NetLoc = u.Host\n\turlEntity.Raw = fmt.Sprintf(\"%s://%s%s\", u.Scheme, u.Host, u.Path)\n\tif u.RawQuery != \"\" {\n\t\turlEntity.Raw += \"?\" + u.RawQuery\n\t}\n\n\treturn urlEntity, nil\n}", "func parse(target, protocol, port string) string {\n\tline := target\n\n\t// if no protocol, no port was specified\n\tif len(protocol) == 0 && len(port) == 0 {\n\t\t// scope only http/https\n\t\tline = `http(s)?://` + line\n\n\t\t// if port was specified but no protocol\n\t} else if len(protocol) == 0 && len(port) > 0 {\n\t\t// scope any protocol\n\t\tline = `\\w+://` + line\n\t} else {\n\n\t}\n\n\t// escape '.'\n\tline = strings.Replace(line, \".\", `\\.`, -1)\n\t// escape '/'\n\tline = strings.Replace(line, \"/\", `\\/`, -1)\n\t// replace wildcard\n\tline = strings.Replace(line, \"*\", `[\\S]*`, -1)\n\t// Zap needs this to scope URL params\n\tline = `^` + line + `[\\S]*$`\n\n\treturn line\n}", "func parseCloneURL(repo *sourcegraph.Repo) (*url.URL, error) {\n\tif repo.HTTPCloneURL != \"\" {\n\t\treturn giturls.Parse(repo.HTTPCloneURL)\n\t} else if repo.SSHCloneURL != \"\" {\n\t\treturn giturls.Parse(repo.SSHCloneURL)\n\t} else {\n\t\treturn nil, errors.New(\"Must provide either an HTTP(S) or SSH clone URL\")\n\t}\n}", "func ParseScp(rawurl string) (*url.URL, error) {\n\tmatch := scpSyntax.FindAllStringSubmatch(rawurl, -1)\n\tif len(match) == 0 {\n\t\treturn nil, fmt.Errorf(\"no scp URL found in %q\", rawurl)\n\t}\n\tm := match[0]\n\treturn &url.URL{\n\t\tScheme: \"ssh\",\n\t\tUser: url.User(strings.TrimRight(m[1], \"@\")),\n\t\tHost: m[2],\n\t\tPath: m[3],\n\t}, nil\n}", "func parseUrl(ignore string, urlPath string) map[string]string {\n\turlMap := make(map[string]string)\n\tfor _, s := range strings.Split(urlPath,\"/\") {\n\t\tif s == \"\" || s == ignore {\n\t\t\tcontinue\n\t\t}\n\t\tspl := strings.Split(s, \"=\")\n\t\tif len(spl) != 2 {\n\t\t\treturn nil\n\t\t}\n\t\turlMap[spl[0]] = spl[1]\n\t}\n\treturn urlMap\n}", "func parseQuery(url string) ([]string, error) {\n\turlList := strings.Split(url, \"?\")\n\tif len(urlList) < 2 {\n\t\treturn make([]string, 0), nil\n\t}\n\tquery := make([]string, 0)\n\tfor _, val := range strings.Split(urlList[1], \"&\") {\n\t\tv := strings.Split(val, \"=\")\n\t\tif len(v) < 2 {\n\t\t\treturn make([]string, 0), errors.New(\"query parameter error\")\n\t\t}\n\t\tquery = append(query, fmt.Sprintf(\"%s=%s\", v[0], v[1]))\n\t}\n\treturn query, nil\n}", "func (m *MockRaindropRepository) ParseURL(ctx context.Context, url string) (*entity.ParsedURL, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ParseURL\", ctx, url)\n\tret0, _ := ret[0].(*entity.ParsedURL)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func ParseURL(redisURL string) (*Pool, error) {\n\tpool := Pool{}\n\tu, err := url.Parse(redisURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tpool.Dial, err = dialURL(u)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tq := u.Query()\n\n\tif v, ok := q[\"clock-interval\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.ClockInterval = d\n\t\t}\n\t}\n\n\tif v, ok := q[\"max-idle-time\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.MaxIdleTime = d\n\t\t}\n\t}\n\tif v, ok := q[\"max-connections\"]; ok && len(v) > 0 {\n\t\tif size, _ := strconv.Atoi(v[0]); size > 0 {\n\t\t\tpool.MaxConnections = size\n\t\t}\n\t}\n\tif v, ok := q[\"min-connections\"]; ok && len(v) > 0 {\n\t\tif size, _ := strconv.Atoi(v[0]); size > 0 {\n\t\t\tpool.MinConnections = size\n\t\t}\n\t}\n\n\treturn &pool, nil\n}" ]
[ "0.7380766", "0.72652763", "0.72410816", "0.7142754", "0.6947265", "0.6927803", "0.6865738", "0.68655217", "0.68425167", "0.6833722", "0.68201834", "0.6754445", "0.66690123", "0.6664009", "0.6628707", "0.6611244", "0.6603158", "0.6520449", "0.6483474", "0.64653146", "0.64649415", "0.6422294", "0.6368243", "0.633477", "0.6319429", "0.6293287", "0.62473536", "0.62328625", "0.6197537", "0.61934644", "0.6189007", "0.6159586", "0.61410135", "0.60598534", "0.60237795", "0.59665", "0.59535795", "0.59402275", "0.59227103", "0.5918107", "0.5878664", "0.586704", "0.5864598", "0.58309764", "0.58232546", "0.58119243", "0.580509", "0.580362", "0.57917506", "0.57668763", "0.5755969", "0.5732027", "0.5703928", "0.57014334", "0.5697761", "0.5673327", "0.56559086", "0.56309706", "0.56218106", "0.56194425", "0.559116", "0.5590698", "0.55619955", "0.55569553", "0.55234754", "0.55146503", "0.5509425", "0.5500893", "0.5440841", "0.5432917", "0.5376299", "0.53700644", "0.5369842", "0.53612936", "0.5356708", "0.535393", "0.5351497", "0.5343657", "0.5327721", "0.5317138", "0.5297182", "0.52956605", "0.52899426", "0.5279025", "0.52746445", "0.5248007", "0.5245996", "0.52438176", "0.5230613", "0.5230129", "0.52289367", "0.52246255", "0.52146745", "0.5211997", "0.5206589", "0.5206078", "0.52014476", "0.5201348", "0.519914", "0.5198237" ]
0.8605891
0
/ Math Abs returns the absolute value of operand.
func Abs(operand int) int { if operand < 0 { return operand * -1 } return operand }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Abs(arg float64) float64 {\n\treturn math.Abs(arg)\n}", "func Abs(a int) int {\n\treturn neogointernal.Opcode1(\"ABS\", a).(int)\n}", "func Abs(in Res) Res {\n\tsign := in.Output().Copy()\n\tanyvec.GreaterThan(sign, sign.Creator().MakeNumeric(0))\n\tsign.Scale(sign.Creator().MakeNumeric(2))\n\tsign.AddScalar(sign.Creator().MakeNumeric(-1))\n\treturn Mul(in, NewConst(sign))\n}", "func Abs(x float64) float64 {\n\tif x < 0 {\n\t\tx = -x\n\t}\n\treturn x\n}", "func mathAbs(ctx phpv.Context, args []*phpv.ZVal) (*phpv.ZVal, error) {\n\tvar z *phpv.ZVal\n\t_, err := core.Expand(ctx, args, &z)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tz, err = z.AsNumeric(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tswitch z.GetType() {\n\tcase phpv.ZtInt:\n\t\ti := z.AsInt(ctx)\n\t\tif i < 0 {\n\t\t\treturn (-i).ZVal(), nil\n\t\t} else {\n\t\t\treturn i.ZVal(), nil\n\t\t}\n\tcase phpv.ZtFloat:\n\t\treturn phpv.ZFloat(math.Abs(float64(z.AsFloat(ctx)))).ZVal(), nil\n\tdefault:\n\t\treturn phpv.ZNull{}.ZVal(), nil\n\t}\n}", "func Abs(x float64) float64 {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\tif x == 0 {\n\t\treturn 0 // return correctly abs(-0)\n\t}\n\treturn x\n}", "func Abs(a int) int {\n\tif a > 0 {\n\t\treturn a\n\t}\n\treturn -a\n}", "func Abs(a int) int {\n\tif a > 0 {\n\t\treturn a\n\t}\n\treturn -a\n}", "func Abs(a int) int {\n\tif a > 0 {\n\t\treturn a\n\t}\n\treturn -a\n}", "func (self *State)Abs(a any)any{\n self.IncOperations(self.coeff[\"abs\"]+self.off[\"abs\"])\n return wrap1(a,math.Abs)\n}", "func Abs(a int) int {\n\tif a < 0 {\n\t\treturn -a\n\t}\n\treturn a\n}", "func Abs(number int) int {\n\tif number > 0 {\n\t\treturn number\n\t}\n\treturn -number\n}", "func Abs(num int) int{\n\tif num > 0{\n\t\treturn num\n\t}\n\treturn (-1)*num\n}", "func Abs(b int) int {\n\tif b < 0 {\n\t\treturn -b\n\t}\n\n\treturn b\n}", "func Abs(x int64) int64 {\n\ta := int64(x)\n\tif a < 0 {\n\t\treturn (-a)\n\t}\n\treturn (a)\n}", "func Abs(x int) int {\n\tif x > 0 {\n\t\treturn x\n\t}\n\treturn -x\n}", "func Abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func Abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func Abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func Abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func Abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func Abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func Abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func Abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func Abs(v int) int {\n\tif v > 0 {\n\t\treturn v\n\t}\n\treturn -v\n}", "func (c *Clac) Abs() error {\n\treturn c.applyFloat(1, func(vals []value.Value) (value.Value, error) {\n\t\treturn unary(\"abs\", vals[0])\n\t})\n}", "func Abs(a *big.Float) *big.Float {\n\treturn ZeroBigFloat().Abs(a)\n}", "func (z *Int) Abs(x *Int) *Int {}", "func (d Decimal) Abs() Decimal {\n\tif !d.IsNegative() {\n\t\treturn d\n\t}\n\td.ensureInitialized()\n\td2Value := new(big.Int).Abs(d.value)\n\treturn Decimal{\n\t\tvalue: d2Value,\n\t\texp: d.exp,\n\t}\n}", "func Abs(scope *Scope, x tf.Output) (y tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Abs\",\n\t\tInput: []tf.Input{\n\t\t\tx,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func Abs(value gcv.Value) gcv.Value {\n\tif value.Type() == gcv.Complex {\n\t\treturn gcv.MakeValue(cmplx.Abs(value.Complex()))\n\t}\n\treturn gcv.MakeValue(math.Abs(value.Real()))\n}", "func (z *Big) Abs(x *Big) *Big {\n\tif debug {\n\t\tx.validate()\n\t}\n\tif !z.invalidContext(z.Context) && !z.checkNaNs(x, x, absvalue) {\n\t\tz.Context.round(z.copyAbs(x))\n\t}\n\treturn z\n}", "func Abs[T constraints.Number](x T) T {\n\tif x >= 0 {\n\t\treturn x\n\t}\n\treturn -x\n}", "func (d Decimal) Abs() Decimal {\n\td.ensureInitialized()\n\td2Value := new(big.Int).Abs(d.value)\n\treturn Decimal{\n\t\tvalue: d2Value,\n\t\texp: d.exp,\n\t}\n}", "func Abs(n int) int {\n\tif n < 0 {\n\t\treturn -n\n\t}\n\treturn n\n}", "func Abs(n int) int {\n\tif n < 0 {\n\t\treturn -n\n\t}\n\treturn n\n}", "func Abs(n int) int {\n\tif n < 0 {\n\t\treturn -n\n\t}\n\treturn n\n}", "func Abs(n int) int {\n\tif n < 0 {\n\t\treturn -n\n\t}\n\treturn n\n}", "func Abs(n int) int {\n\tif n < 0 {\n\t\treturn -n\n\t}\n\treturn n\n}", "func Abs(n int) int {\n\tif n >= 0 {\n\t\treturn n\n\t}\n\treturn -n\n}", "func (m *MockFinder) Abs(v []byte) []byte {\n\treturn m.fnd.Abs(v)\n}", "func (s VectOp) Abs() VectOp {\n\treturn fs.Abs(s)\n}", "func abs(num int64) int64 {\n\tif num < 0 {\n\t\treturn -num\n\t}\n\treturn num\n}", "func (z *Float) Abs(x *Float) *Float {}", "func Abs(x int32) int32 {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func ( f MyFloat ) Abs() float64 {\n\tif f < 0 { return float64( -f ) }\n\treturn float64(f) \n}", "func abs(x int64) int64 {\n\tif x >= 0 {\n\t\treturn x\n\t}\n\treturn -x\n}", "func TestAbs(t *testing.T) {\n\tvar tests = []struct {\n\t\tOperand int\n\t\tExpected int\n\t}{\n\t\t{Operand: -10, Expected: 10},\n\t\t{Operand: -1, Expected: 1},\n\t\t{Operand: 0, Expected: 0},\n\t\t{Operand: -1, Expected: 1},\n\t\t{Operand: 10, Expected: 10},\n\t}\n\n\tfor _, test := range tests {\n\t\tresult := Abs(test.Operand)\n\t\tif result != test.Expected {\n\t\t\tt.Errorf(\"Abs result incorrect. Operand: %d, Expected: %d, Received: %d\", test.Operand, test.Expected, result)\n\t\t}\n\t}\n}", "func abs(x int64) int64 {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func (f MyFloat) Abs() float64 {\n\tif f < 0 {\n\t\treturn float64(-f)\n\t}\n\treturn float64(f)\n}", "func Abs(t1 TermT) TermT {\n\treturn TermT(C.yices_abs(C.term_t(t1)))\n}", "func Abs(z, x *big.Int) *big.Int {\n\treturn z.Abs(x)\n}", "func (m MyFloat) Abs() float64 {\n if m < 0 {\n return float64(-m)\n }\n return float64(m)\n}", "func abs(val int) int {\n\tif val < 0 {\n\t\treturn -val\n\t}\n\treturn val\n}", "func (m *Money) Abs() *Money {\n\tif m.M < 0 {\n\t\tm.Neg()\n\t}\n\treturn m\n}", "func (fn *formulaFuncs) ABS(argsList *list.List) formulaArg {\n\tif argsList.Len() != 1 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"ABS requires 1 numeric argument\")\n\t}\n\targ := argsList.Front().Value.(formulaArg).ToNumber()\n\tif arg.Type == ArgError {\n\t\treturn arg\n\t}\n\treturn newNumberFormulaArg(math.Abs(arg.Number))\n}", "func (bi Int) Abs() Int {\n\tif bi.GreaterThanEqual(Zero()) {\n\t\treturn bi.Copy()\n\t}\n\treturn bi.Neg()\n}", "func (v Value) Abs() *Value {\n\treturn newValue(v.native, false, v.num, v.offset)\n}", "func (z *Rat) Abs(x *Rat) *Rat {}", "func Absolute(num cty.Value) (cty.Value, error) {\n\treturn AbsoluteFunc.Call([]cty.Value{num})\n}", "func iAbs(x int) int { if x >= 0 { return x } else { return -x } }", "func Abs(x int) int {\n if x < 0 {\n return -x\n }\n return x\n}", "func getAbs(a int) int {\n if a < 0 {\n return -a\n }\n return a\n}", "func Abs(d Number) Number {\n\tif math.Float64bits(d.Real)&(1<<63) == 0 {\n\t\treturn d\n\t}\n\treturn Scale(-1, d)\n}", "func Absolute(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func (g *Graph) Abs(x Node) Node {\n\treturn g.NewOperator(fn.NewAbs(x), x)\n}", "func abs(x int) int {\r\n\tif x < 0 {\r\n\t\treturn -x\r\n\t}\r\n\treturn x\r\n}", "func (z *Int) Abs() *Int {\n\tif z.Lt(SignedMin) {\n\t\treturn z\n\t}\n\tz.Sub(zero, z)\n\treturn z\n}", "func Abs(n int) int {\n if n < 0 {\n return -n\n }\n return n\n}", "func abs(a int) int {\n\tif a < 0 {\n\t\treturn -a\n\t}\n\treturn a\n}", "func abs(a int) int {\n\tif a < 0 {\n\t\treturn -a\n\t}\n\treturn a\n}", "func abs(v int) int {\n\tif v < 0 {\n\t\treturn -v\n\t}\n\treturn v\n}", "func abs(a int) int {\r\n if a < 0 {\r\n return -a\r\n }\r\n return a\r\n}", "func (i *Number) Absolute() *Number {\n\treturn NewNumber(math.Abs(i.value))\n}", "func Abs(x int32) int32 {\n\t// Patented (!) : return (x ^ (x >> 31)) - (x >> 31)\n\treturn (x + (x >> 31)) ^ (x >> 31)\n}", "func (v Vertex) Abs() float64 {\n\treturn math.Sqrt(v.X*v.X + v.Y*v.Y)\n}", "func abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func abs(x int) int {\n\tif x < 0 {\n\t\treturn -x\n\t}\n\treturn x\n}", "func (f *Float) Abs(x *Float) *Float {\n\tx.doinit()\n\tf.doinit()\n\tC.mpf_abs(&f.i[0], &x.i[0])\n\treturn f\n}", "func abs(x int) int {\n\tif x < 0{\n\t\treturn -x\n\t}\n\treturn x\n}", "func PrintAbs(v Abser) {\n\ta := v.Abs()\n\tfmt.Println(a)\n}", "func IntAbs(x int) int {\n\tif x > 0 {\n\t\treturn x\n\t}\n\n\treturn -x\n}", "func (gdt *Vector3) Abs() Vector3 {\n\targ0 := gdt.getBase()\n\n\tret := C.go_godot_vector3_abs(GDNative.api, arg0)\n\n\treturn Vector3{base: &ret}\n\n}", "func getAbsoluteValue(value int) int {\n\tif value < 0 {\n\t\treturn -value\n\t}\n\treturn value\n}", "func abs(n int) int {\n if n > 0 {\n return n\n }\n\n return -n\n}", "func IntAbs(a int) int {\n\tif a < 0 {\n\t\treturn -a\n\t}\n\treturn a\n}", "func (v *Vertex) Abs() float64 {\n\treturn math.Sqrt(v.X*v.X + v.Y*v.Y)\n}", "func (f Fixed) Abs() Fixed {\n\tif f.IsNaN() {\n\t\treturn NaN\n\t}\n\tif f.Sign() >= 0 {\n\t\treturn f\n\t}\n\tf0 := Fixed{fp: f.fp * -1}\n\treturn f0\n}", "func (c Currency) Abs() Currency {\n\tif c.m < 0 {\n\t\treturn c.Neg()\n\t}\n\treturn c\n}", "func abs(n int) int {\n if n < 0 {\n return -n\n }\n return n\n}", "func (v *Vertex) Abs() float64 {\n\treturn math.Sqrt(v.X*v.X + v.Y*v.Y)\n}", "func (i I) Abs() I {\n\tif i.X < 0 {\n\t\ti.X = -i.X\n\t}\n\tif i.Y < 0 {\n\t\ti.Y = -i.Y\n\t}\n\treturn i\n}", "func abs(n int) int {\n\tif n < 0 {\n\t\treturn -n\n\t}\n\treturn n\n}", "func funcAbs(vals []parser.Value, args parser.Expressions, enh *EvalNodeHelper) Vector {\n\treturn simpleFunc(vals, enh, math.Abs)\n}", "func (cf customFloat) abs() customFloat {\n\tif cf < 0 {\n\t\treturn cf * -1\n\t}\n\treturn cf\n}", "func intAbs(n int64) int64 {\n\treturn int64(math.Abs(float64(n)))\n}", "func abs(n int64) int64 {\n\treturn int64(math.Abs(float64(n)))\n}", "func (p *Point) Abs() float64 {\n return math.Sqrt(p.x*p.x + p.y*p.y)\n}", "func abs(i int) int {\n\tif i < 0 {\n\t\treturn -i\n\t}\n\treturn i\n}" ]
[ "0.82674694", "0.815114", "0.80781394", "0.7916723", "0.78819174", "0.7836325", "0.78362995", "0.78362995", "0.78362995", "0.7770546", "0.7764423", "0.77447665", "0.7727246", "0.7717831", "0.77176994", "0.7654214", "0.7606882", "0.7606882", "0.7606882", "0.7606882", "0.7606882", "0.7606882", "0.7606882", "0.7606882", "0.75868654", "0.7549752", "0.7542997", "0.7529634", "0.7511493", "0.75091416", "0.7502329", "0.74893594", "0.74589455", "0.7457587", "0.7456921", "0.7456921", "0.7456921", "0.7456921", "0.7456921", "0.74521714", "0.73947525", "0.73922175", "0.7389891", "0.7389638", "0.7383013", "0.7381265", "0.73679054", "0.7351312", "0.73399085", "0.7325576", "0.732539", "0.7322928", "0.7320134", "0.72702384", "0.7261474", "0.72420913", "0.72133243", "0.7207558", "0.71847224", "0.717977", "0.7170797", "0.71396697", "0.71395606", "0.7110124", "0.7092093", "0.7080206", "0.7053139", "0.7045929", "0.70454425", "0.7036805", "0.7036805", "0.7032841", "0.702675", "0.70183986", "0.70017886", "0.7001601", "0.70015454", "0.70015454", "0.70015454", "0.6998003", "0.69854945", "0.69756895", "0.69594413", "0.69197613", "0.69189006", "0.6916934", "0.69088703", "0.6908345", "0.6902096", "0.6900945", "0.68667203", "0.68616915", "0.68536127", "0.68455565", "0.6838868", "0.6831629", "0.6824975", "0.68169266", "0.681501", "0.68137246" ]
0.8551374
0
Add a to operand.
func Add(a, operand int) int { return operand + a }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Add() {\n\tMatch('+')\n\tTerm()\n\tEmitLn(\"ADD (SP)+,D0\")\n}", "func Add(a, b Expr) Expr {\n\treturn &addOp{&simpleOperator{a, b, scanner.ADD}}\n}", "func Add(a, b Expr) Expr {\n\treturn &arithmeticOperator{&simpleOperator{a, b, scanner.ADD}}\n}", "func (c *OperandsController) Add(ctx *app.AddOperandsContext) error {\n\tsum := ctx.Left + ctx.Right\n\treturn ctx.OK([]byte(strconv.Itoa(sum)))\n}", "func _cgoexp_e93fccc2f088_add(a *struct {\n\t\tp0 _Ctype_int\n\t\tp1 _Ctype_int\n\t\tr0 _Ctype_int\n\t}) {\n\ta.r0 = add(a.p0, a.p1)\n}", "func ADDPD(mx, x operand.Op) { ctx.ADDPD(mx, x) }", "func Add(a cty.Value, b cty.Value) (cty.Value, error) {\n\treturn AddFunc.Call([]cty.Value{a, b})\n}", "func (proxy CalculatorProxy) Add(a int, b int) int {\n\trequestor := distribution.NewRequestor()\n\tinvocation := distribution.NewInvocation(\n\t\tproxy.client.ObjectID,\n\t\tproxy.client.Hostname,\n\t\tproxy.client.Port,\n\t\t\"add\",\n\t\t[]int{a, b},\n\t)\n\n\trequestor.Invoke(invocation)\n\n\treturn 2 + 2\n}", "func (s *Script) AddOperand(operand []byte) *Script {\n\tdataLen := len(operand)\n\n\tif dataLen < int(OPPUSHDATA1) {\n\t\t*s = append(*s, byte(dataLen))\n\t} else if dataLen <= 0xff {\n\t\t*s = append(*s, byte(OPPUSHDATA1), byte(dataLen))\n\t} else if dataLen <= 0xffff {\n\t\tbuf := make([]byte, 2)\n\t\tbinary.LittleEndian.PutUint16(buf, uint16(dataLen))\n\t\t*s = append(*s, byte(OPPUSHDATA2))\n\t\t*s = append(*s, buf...)\n\t} else {\n\t\tbuf := make([]byte, 4)\n\t\tbinary.LittleEndian.PutUint32(buf, uint32(dataLen))\n\t\t*s = append(*s, byte(OPPUSHDATA4))\n\t\t*s = append(*s, buf...)\n\t}\n\n\t// Append the actual operand\n\t*s = append(*s, operand...)\n\treturn s\n}", "func ADDPS(mx, x operand.Op) { ctx.ADDPS(mx, x) }", "func Add( a *context.Value, b *context.Value ) (*context.Value,error) {\n if a != nil && b != nil {\n switch a.OperationType( b ) {\n case context.VAR_BOOL:\n return context.IntValue( a.Int() + b.Int() ), nil\n case context.VAR_INT:\n return context.IntValue( a.Int() + b.Int() ), nil\n case context.VAR_FLOAT:\n return context.FloatValue( a.Float() + b.Float() ), nil\n case context.VAR_STRING:\n return context.StringValue( a.String() + b.String() ), nil\n case context.VAR_COMPLEX:\n return context.ComplexValue( a.Complex() + b.Complex() ), nil\n }\n }\n\n return nil, errors.New( \"Unsupported type for add\" )\n}", "func (vm *VM) opAdd(instr []uint16) int {\n\ta, b, c := vm.getAbc(instr)\n\n\tvm.registers[a] = (b + c) % 32768\n\treturn 4\n}", "func (m *Manager) PushOperand(operand string) {\n\toperandData := m.getOperandData(operand)\n\toperandName := operandData.Name\n\t// If the operand is an attribute from an instance we set tag in it's ID\n\tif operandData.FromSelf {\n\t\toperandName = fmt.Sprintf(\"self_%d_%s\", operandData.SelfDir, operandData.Name)\n\t}\n\telement := NewElement(operandData.Dir, operandName, operandData.TypeOf, operandData.Class)\n\tm.operands.Push(element)\n}", "func (o *Operator) Add(f manager.Runnable) error {\n\treturn o.mgr.Add(f)\n}", "func add(x, y int) int", "func (gdt *Vector3) OperatorAdd(b Vector3) Vector3 {\n\targ0 := gdt.getBase()\n\targ1 := b.getBase()\n\n\tret := C.go_godot_vector3_operator_add(GDNative.api, arg0, arg1)\n\n\treturn Vector3{base: &ret}\n\n}", "func add(a int, b int) int", "func Add(t1 TermT, t2 TermT) TermT {\n\treturn TermT(C.yices_add(C.term_t(t1), C.term_t(t2)))\n}", "func (g *Gini) Add(m z.Lit) {\n\tg.xo.Add(m)\n}", "func (z *Float64) Plus(y *Float64, a float64) *Float64 {\n\tz.l = y.l + a\n\tz.r = y.r\n\treturn z\n}", "func ADDQ(imr, mr operand.Op) { ctx.ADDQ(imr, mr) }", "func ADDW(imr, amr operand.Op) { ctx.ADDW(imr, amr) }", "func (c *Calculator) Add() {\n\tif opValue, err := c.getOperationValue(); err != nil {\n\t\tc.returnError()\n\t} else {\n\t\tlog.Printf(\"%f + %f = \", value, opValue)\n\t\tvalue += opValue\n\t\tlog.Printf(\"%f\\n\", value)\n\t\tc.returnResult()\n\t}\n}", "func (z *Rat) Add(x, y *Rat) *Rat {}", "func (s VectOp) PlusOp(o fs.IndexedFunc) VectOp {\n\treturn fs.PlusOp(s, o)\n}", "func PADDQ(mx, x operand.Op) { ctx.PADDQ(mx, x) }", "func (m *AbilitypatientrightsMutation) AddOperative(i int) {\n\tif m.add_Operative != nil {\n\t\t*m.add_Operative += i\n\t} else {\n\t\tm.add_Operative = &i\n\t}\n}", "func VPADDQ(ops ...operand.Op) { ctx.VPADDQ(ops...) }", "func PHADDSW(mx, x operand.Op) { ctx.PHADDSW(mx, x) }", "func ADDSD(mx, x operand.Op) { ctx.ADDSD(mx, x) }", "func add(this js.Value, args []js.Value) interface{} {\n\ta := args[0].Int()\n\tb := args[1].Int()\n\n\tthis.Set(\"result\", js.ValueOf(a+b))\n\treturn nil\n}", "func (c *OperandsController) Add(ctx *app.AddOperandsContext) error {\n\ttags := map[string]string{`service-name`: `service-foo`}\n\tvar respErr error\n\n\traven.CapturePanic(func() {\n\t\tsum, err := c.interactor.Add(ctx.Left, ctx.Right)\n\t\tif err != nil {\n\t\t\tlog.Println(`an error occured: `, err)\n\t\t\tc.monitor.CaptureError(err)\n\t\t\trespErr = ctx.InternalServerError()\n\t\t\treturn\n\t\t}\n\n\t\traven.CaptureMessage(`everything is ok`, tags)\n\t\trespErr = ctx.OK([]byte(strconv.Itoa(sum)))\n\t\treturn\n\n\t}, tags)\n\n\treturn respErr\n}", "func Add(x, y Dec) (Dec, error) {\n\treturn x.Add(y)\n}", "func (_ConsortiumManagement *ConsortiumManagementTransactor) AddOperator(opts *bind.TransactOpts, operator common.Address, endpoint string) (*types.Transaction, error) {\n\treturn _ConsortiumManagement.contract.Transact(opts, \"addOperator\", operator, endpoint)\n}", "func (a *Vector3) Plus(b Vector3) {\n\t*a = Vector3{a.X + b.X, a.Y + b.Y, a.Z + b.Z}\n}", "func PADDW(mx, x operand.Op) { ctx.PADDW(mx, x) }", "func XADDQ(r, mr operand.Op) { ctx.XADDQ(r, mr) }", "func AddOpAdd(jl *JSONLogic) {\n\tjl.AddOperation(\"+\", opAdd)\n}", "func PADDSW(mx, x operand.Op) { ctx.PADDSW(mx, x) }", "func (v *VerbalExpression) add(s string) *VerbalExpression {\n\tv.compiled = false //reinit previous regexp compilation\n\tv.expression += s\n\treturn v\n}", "func PADDSB(mx, x operand.Op) { ctx.PADDSB(mx, x) }", "func (t *Arith) Add(ctx *httprpc.Context, args Args, reply *Reply) error {\n\tlog.Printf(\"[%s] Add Args: %v\", ctx.TraceID, args)\n\treply.C = args.A + args.B\n\treturn nil\n}", "func add1(x, y int64) int64", "func (l *logicManager) addTerm(op int64, term ast.Expr) {\n\tl.terms = append(l.terms, term)\n\tl.ops = append(l.ops, op)\n}", "func (g *Graph) Add(x1 Node, x2 Node) Node {\n\tif x1 != nil {\n\t\treturn g.NewOperator(fn.NewAdd(x1, x2), x1, x2)\n\t}\n\tfake := g.NewVariable(nil, false)\n\treturn g.NewOperator(fn.NewAdd(fake, x2), fake, x2)\n}", "func add(x float64,y float64) float64 {\n\treturn x+y\n}", "func ADDSS(mx, x operand.Op) { ctx.ADDSS(mx, x) }", "func (t *infixTokenizer) AddOperator(operator string) error {\n\tif err := t.checkUniqueness(operator); err != nil {\n\t\treturn err\n\t}\n\n\tt.operators[operator] = true\n\n\t_, ok := t.operatorsByLength[len(operator)]\n\tif !ok {\n\t\tt.operatorsByLength[len(operator)] = make(map[string]bool)\n\t}\n\tt.operatorsByLength[len(operator)][operator] = true\n\n\treturn nil\n}", "func (p Point) Add(q Point) Point { return Point{p.X + q.X, p.Y + q.Y} }", "func (p Point) Add(q Point) Point { return Point{p.X + q.X, p.Y + q.Y} }", "func PHADDW(mx, x operand.Op) { ctx.PHADDW(mx, x) }", "func Add(a, b float64) float64 {\n\treturn a + b\n}", "func Add(a, b float64) float64 {\n\treturn a + b\n}", "func (i *Number) Add(v Number) *Number {\n\treturn NewNumber(i.value + v.value)\n}", "func (q query) Add(key string, op Operator, val ...interface{}) {\n\tq.init(key)\n\tq[key][op] = append(q[key][op], val...)\n}", "func (v *Vector) Plus(a *Vector) *Vector {\n\treturn &Vector{X: v.X + a.X, Y: v.Y + a.Y, Z: v.Z + a.Z}\n}", "func XADDW(r, mr operand.Op) { ctx.XADDW(r, mr) }", "func (b ValExprBuilder) Plus(expr interface{}) ValExprBuilder {\n\treturn b.makeBinaryExpr('+', expr)\n}", "func (s *Scalar) Add(x, y *Scalar) *Scalar {\n\ts.s.Add(&x.s, &y.s)\n\treturn s\n}", "func add(a, b float64) float64 {\n\treturn a + b\n}", "func Add(a, b float64) (float64, error) {\n\treturn a + b, nil\n}", "func (s *SQLOperations) Add(op SQLOperation) {\n\t*s = append(*s, op)\n}", "func (m *SplitMutation) AddTo(f float64) {\n\tif m.addto != nil {\n\t\t*m.addto += f\n\t} else {\n\t\tm.addto = &f\n\t}\n}", "func (o *Operators) Add(op *Symbol) {\n\t// Delegate to the root\n\tif o.root != nil {\n\t\to.root.Add(op)\n\t\treturn\n\t}\n\n\t// Scan through symbol name rune by rune\n\tpos := 0\n\tnode := o\n\tfor pos < len(op.Name) {\n\t\t// Grab next rune\n\t\tr, w := utf8.DecodeRuneInString(op.Name[pos:])\n\t\tif r == utf8.RuneError && w == 1 {\n\t\t\tpanic(ErrBadRune)\n\t\t}\n\n\t\t// Advance the text position\n\t\tpos += w\n\n\t\t// Make sure the children map exists\n\t\tif node.children == nil {\n\t\t\tnode.children = map[rune]*Operators{}\n\t\t}\n\n\t\t// See if the node has an entry for that rune\n\t\tif tmp, ok := node.children[r]; ok {\n\t\t\tnode = tmp\n\t\t} else {\n\t\t\t// Construct a new one\n\t\t\ttmp = &Operators{\n\t\t\t\tprefix: op.Name[:pos],\n\t\t\t\troot: o,\n\t\t\t\tparent: node,\n\t\t\t\tchildren: map[rune]*Operators{},\n\t\t\t}\n\t\t\tnode.children[r] = tmp\n\t\t\tnode = tmp\n\t\t}\n\t}\n\n\t// Is the operator already set?\n\tif node.Sym != nil {\n\t\treturn\n\t}\n\n\t// Save the symbol\n\tnode.Sym = op\n}", "func (cal *Calculate) add(value float64) (result float64) {\n\tif len(cal.Arg) == 2 {\n\t\treturn (cal.Arg[0] + cal.Arg[1])\n\t} else if len(cal.Arg) == 1 {\n\t\treturn (value + cal.Arg[0])\n\t}\n\n\tlog.Fatalln(\"Please check the data format of the calculation unit\")\n\treturn\n}", "func (v Vector) Add(o Vector) *Vector {\n\treturn &Vector{v[0] + o[0], v[1] + o[1], v[2] + o[2]}\n}", "func (s *set) Add(t *Term) {\n\ts.insert(t)\n}", "func VADDPS(ops ...operand.Op) { ctx.VADDPS(ops...) }", "func Add(ds datastore.Datastore, a Input) (int, error) {\n\treturn add(ds, a)\n}", "func (v *Vector) Add(rhs *Vector) *Vector {\n\tif v == nil {\n\t\treturn rhs\n\t}\n\tif rhs == nil {\n\t\treturn v\n\t}\n\tl := min(len(v.vec), len(rhs.vec))\n\tvec := make([]float64, l)\n\tcopy(vec, v.vec)\n\tsaxpy(l, 1, rhs.vec, 1, vec, 1)\n\telems := make([]string, len(v.elems)+len(rhs.elems))\n\telems = append(elems, rhs.elems...)\n\telems = append(elems, v.elems...)\n\treturn &Vector{\n\t\tword: v.word + \" + \" + rhs.word,\n\t\tvec: vec,\n\t\telems: elems,\n\t}\n}", "func VPADDW(ops ...operand.Op) { ctx.VPADDW(ops...) }", "func Add(a, b int) int {\n\tklog.Infof(\"a:%d, b:%d\", a, b)\n\treturn a + b\n}", "func add(a, b int, result *int) {\n\t*result = a * b\n}", "func (b *ScriptBuilder) AddOp(opcode byte) *ScriptBuilder {\n\tif b.err != nil {\n\t\treturn b\n\t}\n\n\t// Pushes that would cause the script to exceed the largest allowed\n\t// script size would result in a non-canonical script.\n\tif len(b.script)+1 > MaxScriptSize {\n\t\tstr := fmt.Sprintf(\"adding an opcode would exceed the maximum \"+\n\t\t\t\"allowed canonical script length of %d\", MaxScriptSize)\n\t\tb.err = ErrScriptNotCanonical(str)\n\t\treturn b\n\t}\n\n\tb.script = append(b.script, opcode)\n\treturn b\n}", "func SimpleAdd(exp string) string {\n index := strings.Index(exp, \"+\")\n var arg1 string = exp[:index]\n var arg2 string = exp[index + 1:]\n val1 := NotateToDouble(arg1)\n val2 := NotateToDouble(arg2)\n return NegativeNotate(fmt.Sprintf(\"%f\", val1 + val2))\n}", "func VADDPD(ops ...operand.Op) { ctx.VADDPD(ops...) }", "func Add(scope *Scope, x tf.Output, y tf.Output) (z tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Add\",\n\t\tInput: []tf.Input{\n\t\t\tx, y,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (e *ConstantExpr) Add(other *ConstantExpr) *ConstantExpr {\n\tassert(e.Width == other.Width, \"add: width mismatch: %d != %d\", e.Width, other.Width)\n\treturn NewConstantExpr(e.Value+other.Value, e.Width)\n}", "func calcAdd(rOpd, lOpd formulaArg, opdStack *Stack) error {\n\tlOpdVal := lOpd.ToNumber()\n\tif lOpdVal.Type != ArgNumber {\n\t\treturn errors.New(lOpdVal.Value())\n\t}\n\trOpdVal := rOpd.ToNumber()\n\tif rOpdVal.Type != ArgNumber {\n\t\treturn errors.New(rOpdVal.Value())\n\t}\n\topdStack.Push(newNumberFormulaArg(lOpdVal.Number + rOpdVal.Number))\n\treturn nil\n}", "func Command_Add(script *rex.Script, params []*rex.Value) {\n\tif len(params) != 2 {\n\t\trex.ErrorParamCount(\"float:add\", \"2\")\n\t}\n\n\tscript.RetVal = rex.NewValueFloat64(params[0].Float64() + params[1].Float64())\n\treturn\n}", "func (rc *RuleContext) Add(ea *ExpressionAtom) *ExpressionAtom {\n\tfor _, eat := range rc.ExpressionAtoms {\n\t\tif ea.EqualsTo(eat) {\n\t\t\treturn eat\n\t\t}\n\t}\n\trc.ExpressionAtoms = append(rc.ExpressionAtoms, ea)\n\treturn ea\n}", "func (t *Tx) AddOp(op Op) {\n\tt.Ops = append(t.Ops, op)\n}", "func Add(a, b int64) int64 {\n\treturn a + b\n}", "func (pc *programCode) createAdd(a, b, sum string) {\n\tcode := \"\\n\\t\\t\\t; Add \" + b + \" to \" +\n\t\ta + \" and save sum in \" + sum + \"\\n\"\n\n\tif _, err := strconv.Atoi(a); err == nil {\n\t\tcode += \"\\tmov rax, \" + a + \"\\n\"\n\t} else {\n\t\tcode += \"\\tmov rax, [\" + a + \"]\\n\"\n\t}\n\tif _, err := strconv.Atoi(b); err == nil {\n\t\tcode += \"\\tadd rax, \" + b + \"\\n\"\n\t} else {\n\t\tcode += \"\\tadd rax, [\" + b + \"]\\n\"\n\t}\n\tcode += \"\\tmov [\" + sum + \"], rax\\n\"\n\n\tpc.appendCode(code)\n}", "func (t *Tuple) Add(o *Tuple) *Tuple {\n\treturn &Tuple{\n\t\tt.x + o.x,\n\t\tt.y + o.y,\n\t\tt.z + o.z,\n\t\tt.w + o.w,\n\t}\n}", "func add(x, y int) (answer int, err error) {\n\tanswer = x + y\n\treturn\n}", "func (a Point) Plus(b Point) Point {\n\treturn Point{a[0] + b[0], a[1] + b[1], a[2] + b[2]}\n}", "func VPADDSW(ops ...operand.Op) { ctx.VPADDSW(ops...) }", "func (v *Vector2) Add(b Vector2) {\r\n\tv.x += b.x\r\n\tv.y += b.y\r\n}", "func (m *NodeMutation) AddValue(i int) {\n\tif m.addvalue != nil {\n\t\t*m.addvalue += i\n\t} else {\n\t\tm.addvalue = &i\n\t}\n}", "func ADDSUBPD(mx, x operand.Op) { ctx.ADDSUBPD(mx, x) }", "func Add(x, y Number) Number {\n\treturn Number{\n\t\tReal: x.Real + y.Real,\n\t\tE1mag: x.E1mag + y.E1mag,\n\t\tE2mag: x.E2mag + y.E2mag,\n\t\tE1E2mag: x.E1E2mag + y.E1E2mag,\n\t}\n}", "func (o *FloatObject) Add(r Object) (Object) {\n result := new (FloatObject)\n result.Value = o.Value + r.AsFloat()\n \n return result\n}", "func (v Vec2) Add(x Vec2) Vec2 {\n\treturn Vec2{v[0] + x[0], v[1] + x[1]}\n}", "func TestAdd(t *testing.T) {\n\toperand, a, expected := 42, 3, 45\n\tresult := Add(a, operand)\n\tif result != expected {\n\t\tt.Errorf(\"Add result incorrect. Operand: %d, A: %d, Expected: %d, Received: %d\", operand, a, expected, result)\n\t}\n}", "func (s *ServerCommands) Add(ctx context.Context, in *commands.Message) (*commands.Message, error) {\n\tope1 := GetParam(in.Msg, \"op1\")\n\tope2 := GetParam(in.Msg, \"op2\")\n\tif ope1 == nil || ope2 == nil {\n\t\treturn NewReply(\"wrong_syntax\", -1), nil\n\t}\n\tlog.Printf(\"%d + %d = %d\\n\", ope1.Value, ope2.Value, ope1.Value+ope2.Value)\n\treturn NewReply(\"sum\", ope1.Value+ope2.Value), nil\n}", "func VPHADDSW(mxy, xy, xy1 operand.Op) { ctx.VPHADDSW(mxy, xy, xy1) }", "func (a AmountFigure) Add(b AmountFigure) AmountFigure {\n\treturn AmountFigure(int64(a) + int64(b))\n}", "func (e *GT) Add(a, b *GT) *GT {\n\tif e.p == nil {\n\t\te.p = &gfP12{}\n\t}\n\te.p.Mul(a.p, b.p)\n\treturn e\n}", "func (z *Big) Add(x, y *Big) *Big { return z.Context.Add(z, x, y) }" ]
[ "0.65395206", "0.65035975", "0.641603", "0.6332773", "0.63169646", "0.62367994", "0.6209624", "0.61428565", "0.6119819", "0.6083758", "0.6069107", "0.6053807", "0.60509676", "0.60224986", "0.59966093", "0.5969306", "0.5967391", "0.59664905", "0.5937366", "0.59199166", "0.59162104", "0.5904514", "0.5896647", "0.5887452", "0.58444315", "0.5839", "0.5829261", "0.58005357", "0.58002615", "0.57982254", "0.5786698", "0.5776328", "0.576403", "0.57608324", "0.5756618", "0.5748492", "0.5745734", "0.57352185", "0.5731395", "0.5720034", "0.5715203", "0.57094324", "0.5704483", "0.57039326", "0.5701136", "0.56799877", "0.5669399", "0.56641424", "0.5648578", "0.5648578", "0.5638945", "0.5621744", "0.5621744", "0.5612012", "0.5609422", "0.56003314", "0.5598395", "0.5589811", "0.55833036", "0.55827504", "0.5579156", "0.5577294", "0.55768675", "0.5575299", "0.55695397", "0.55511534", "0.55497587", "0.5546084", "0.5544738", "0.5523049", "0.5522485", "0.55144626", "0.55105746", "0.55090636", "0.5505952", "0.5499473", "0.5498909", "0.5498491", "0.5486077", "0.5478889", "0.547856", "0.5473311", "0.5471421", "0.5469749", "0.54634637", "0.5461401", "0.5460838", "0.5458446", "0.545762", "0.54422647", "0.5439992", "0.54366374", "0.5436396", "0.54269594", "0.5424539", "0.54240805", "0.54221267", "0.54207534", "0.5419694", "0.54176253" ]
0.68668467
0
Subtract a from operand.
func Subtract(a, operand int) int { return operand - a }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Subtract(a cty.Value, b cty.Value) (cty.Value, error) {\n\treturn SubtractFunc.Call([]cty.Value{a, b})\n}", "func Subtract() {\n\tMatch('-')\n\tTerm()\n\tEmitLn(\"SUB (SP)+,D0\")\n\tEmitLn(\"NEG D0\")\n}", "func subtract(x, y int) (res int) {\n\tres = x - y\n\treturn\n}", "func (expr Expression) Subtract() (float64, error) {\n\treturn (expr.FirstOperand.Value - expr.SecondOperand.Value), nil\n}", "func (s *Scalar) Subtract(x, y *Scalar) *Scalar {\n\ts.s.Sub(&x.s, &y.s)\n\treturn s\n}", "func (gdt *Vector3) OperatorSubtract(b Vector3) Vector3 {\n\targ0 := gdt.getBase()\n\targ1 := b.getBase()\n\n\tret := C.go_godot_vector3_operator_subtract(GDNative.api, arg0, arg1)\n\n\treturn Vector3{base: &ret}\n\n}", "func subtract(x, y float64) float64 {\n\treturn x - y\n}", "func Subtract(a, b int64) int64 {\n\treturn a - b\n}", "func Subtract(x, y int) (res int) {\n\treturn x - y\n}", "func Subtract(a, b int) int {\n\treturn a - b\n}", "func Subtract(a, b int) int {\n\treturn a - b\n}", "func Subtract(a, b int) int {\n\treturn a - b\n}", "func Subtract(a float64, b float64) float64 {\n\treturn a - b\n}", "func (s *server) Subtract(ctx context.Context, in *pb.SubtractRequest) (*pb.SubtractReply, error) {\n\treturn &pb.SubtractReply{N1: in.N1 - in.N2}, nil\n}", "func Sub( a *context.Value, b *context.Value ) (*context.Value,error) {\n if a != nil && b != nil {\n switch a.OperationType( b ) {\n case context.VAR_BOOL:\n return context.IntValue( a.Int() - b.Int() ), nil\n case context.VAR_INT:\n return context.IntValue( a.Int() - b.Int() ), nil\n case context.VAR_FLOAT:\n return context.FloatValue( a.Float() - b.Float() ), nil\n case context.VAR_COMPLEX:\n return context.ComplexValue( a.Complex() - b.Complex() ), nil\n default:\n }\n }\n\n return nil, errors.New( \"Unsupported type for sub\" )\n}", "func calcSubtract(rOpd, lOpd formulaArg, opdStack *Stack) error {\n\tlOpdVal := lOpd.ToNumber()\n\tif lOpdVal.Type != ArgNumber {\n\t\treturn errors.New(lOpdVal.Value())\n\t}\n\trOpdVal := rOpd.ToNumber()\n\tif rOpdVal.Type != ArgNumber {\n\t\treturn errors.New(rOpdVal.Value())\n\t}\n\topdStack.Push(newNumberFormulaArg(lOpdVal.Number - rOpdVal.Number))\n\treturn nil\n}", "func Subtract(a, b float64) float64 {\n\treturn a - b\n}", "func Subtract(a, b float64) float64 {\n\treturn a - b\n}", "func Subtract(x, y int) int {\n\treturn x - y\n}", "func (cal *Calculate) sub(value float64) (result float64) {\n\tif len(cal.Arg) == 2 {\n\t\treturn (cal.Arg[0] - cal.Arg[1])\n\t} else if len(cal.Arg) == 1 {\n\t\treturn (value - cal.Arg[0])\n\t}\n\n\tlog.Fatalln(\"Please check the data format of the calculation unit\")\n\treturn\n}", "func Subtract(left, right int) int {\n\treturn left - right\n}", "func (date Nakamura) Subtract(value int, format string) Nakamura {\n\treturn Add(date, -value, format)\n}", "func Subtract(args ...int) int {\n\tif len(args) < 2 {\n\t\treturn 0\n\t}\n\n\tres := args[0]\n\tfor i := 1; i < len(args); i++ {\n\t\tres -= args[i]\n\t}\n\n\treturn res\n}", "func (z *Int) Sub(x, y *Int) *Int {}", "func (c *Calculator) Sub() {\n\tif opValue, err := c.getOperationValue(); err != nil {\n\t\tc.returnError()\n\t} else {\n\t\tlog.Printf(\"%f - %f = \", value, opValue)\n\t\tvalue -= opValue\n\t\tlog.Printf(\"%f\\n\", value)\n\t\tc.returnResult()\n\t}\n}", "func SUBPD(mx, x operand.Op) { ctx.SUBPD(mx, x) }", "func minus(x, y int) int {\n\treturn x - y\n}", "func subtract(this js.Value, i []js.Value) interface{} {\n\tint1, int2 := getInputValues(i)\n\tsetValueById(i[2].String(), int1-int2)\n\treturn nil\n}", "func (v *Vector2) Subtract(b Vector2) {\r\n\tv.x -= b.x\r\n\tv.y -= b.y\r\n}", "func sub(x, y int) (answer int, err error) {\n\tanswer = x - y\n\treturn\n}", "func (i *Number) Subtract(v Number) *Number {\n\treturn NewNumber(i.value - v.value)\n}", "func substract(a, b float64) float64 {\n\treturn a - b\n}", "func Subtract(v1, v2 sqltypes.Value) (sqltypes.Value, error) {\n\tif v1.IsNull() || v2.IsNull() {\n\t\treturn sqltypes.NULL, nil\n\t}\n\te1, err := valueToEval(v1, collationNumeric)\n\tif err != nil {\n\t\treturn sqltypes.NULL, err\n\t}\n\te2, err := valueToEval(v2, collationNumeric)\n\tif err != nil {\n\t\treturn sqltypes.NULL, err\n\t}\n\tr, err := subtractNumericWithError(e1, e2)\n\tif err != nil {\n\t\treturn sqltypes.NULL, err\n\t}\n\treturn evalToSQLValue(r), nil\n}", "func (z *Rat) Sub(x, y *Rat) *Rat {}", "func SUBW(imr, amr operand.Op) { ctx.SUBW(imr, amr) }", "func (n *bigNumber) sub(x *bigNumber, y *bigNumber) *bigNumber {\n\treturn n.subRaw(x, y).bias(2).weakReduce()\n}", "func SUBSD(mx, x operand.Op) { ctx.SUBSD(mx, x) }", "func (a Balance) Sub(b *Balance) Balance {\n\tfor i, v := range b {\n\t\ta[i] -= v\n\t}\n\treturn a\n}", "func (v Vector) Subtract(other Vector) Vector {\n\treturn Vector{\n\t\tX: v.X - other.X,\n\t\tY: v.Y - other.Y,\n\t\tZ: v.Z - other.Z,\n\t}\n}", "func (a *Int64s) Subtract(b *Int64s) *Int64s {\n\tresult := a.Clone()\n\tfor x, _ := range b.backing {\n\t\tresult.Remove(x)\n\t}\n\treturn result\n}", "func (a *AtomicInt64) Sub(delta int64) int64 {\n\treturn atomic.AddInt64((*int64)(a), -delta)\n}", "func (k Keeper) Sub(ctx sdk.Context, address sdk.AccAddress, amount sdk.Int) (sdk.Int, error) {\n\tvalue, err := k.Get(ctx, address)\n\tif err != nil {\n\t\treturn sdk.Int{}, sdkerrors.Wrapf(sdkerrors.ErrInvalidRequest, err.Error())\n\t}\n\tres := value.Sub(amount)\n\t// emit event\n\tctx.EventManager().EmitEvent(\n\t\tsdk.NewEvent(\n\t\t\ttypes.EventType,\n\t\t\tsdk.NewAttribute(sdk.AttributeKeyAction, types.AttributeActionSubtracted),\n\t\t\tsdk.NewAttribute(types.AttributeKeyAddress, address.String()),\n\t\t\tsdk.NewAttribute(types.AttributeKeyAmount, amount.String()),\n\t\t),\n\t)\n\treturn k.set(ctx, address, res)\n}", "func (v Vector2) Subtract(other Vector) Vector {\r\n\totherv := checkVector2(other)\r\n\treturn Vector2{\r\n\t\tv[0] - otherv[0],\r\n\t\tv[1] - otherv[1],\r\n\t}\r\n}", "func SimpleSubtract(exp string) string {\n var index int = strings.Index(exp, \"-\")\n var arg1 string = exp[:index]\n var arg2 string = exp[index + 1:]\n val1 := NotateToDouble(arg1)\n val2 := NotateToDouble(arg2)\n return NegativeNotate(fmt.Sprintf(\"%f\", val1 - val2))\n}", "func (d Duration) Sub(v Value) (Value, error) {\n\tif y, ok := v.GoValue().(time.Duration); ok {\n\t\tx := time.Duration(d)\n\t\treturn Duration(x - y), nil\n\t}\n\treturn nil, ErrOperationNotDefined\n}", "func (b *IBFCell) Subtract(b2 *IBFCell) {\n\tb.IDSum.XOR(&b2.IDSum)\n\tb.HashSum.XOR(&b2.HashSum)\n\tb.Count -= b2.Count\n}", "func (a ImpactAmount) sub(b ImpactAmount) ImpactAmount {\n\tif b >= a {\n\t\treturn 0\n\t}\n\treturn a - b\n}", "func ValueSub(a, b reflect.Value) (reflect.Value, error) {\n\taBkind := GetBaseKind(a)\n\tbBkind := GetBaseKind(b)\n\n\tswitch aBkind {\n\tcase reflect.Int64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(a.Int() - b.Int()), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Int() - int64(b.Uint())), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(float64(a.Int()) - b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do subtraction math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tcase reflect.Uint64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(int64(a.Uint()) - b.Int()), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Uint() - b.Uint()), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(float64(a.Uint()) - b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do subtraction math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tcase reflect.Float64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(a.Float() - float64(b.Int())), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Float() - float64(b.Uint())), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(a.Float() - b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do subtraction math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tdefault:\n\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do subtraction math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t}\n}", "func SUBL(imr, emr operand.Op) { ctx.SUBL(imr, emr) }", "func (op *OpMinus) Eval(x, y float32) float32 {\n\treturn op.LeftChild.Eval(x, y) - op.RightChild.Eval(x, y)\n}", "func Subtract(v, u *Vec) *Vec {\n\treturn &Vec{\n\t\tv.X - u.X,\n\t\tv.Y - u.Y,\n\t}\n}", "func (p Point2) Sub(ps ...Point2) Point2 {\n\tfor _, p2 := range ps {\n\t\tp[0] -= p2[0]\n\t\tp[1] -= p2[1]\n\t}\n\treturn p\n}", "func (vn *VecN) Sub(dst *VecN, addend *VecN) *VecN {\n\tif vn == nil || addend == nil {\n\t\treturn nil\n\t}\n\tsize := intMin(len(vn.vec), len(addend.vec))\n\tdst = dst.Resize(size)\n\n\tfor i := 0; i < size; i++ {\n\t\tdst.vec[i] = vn.vec[i] - addend.vec[i]\n\t}\n\n\treturn dst\n}", "func Sub(x, y int) int {\n\treturn x - y\n}", "func (p *Int64) Sub(q, r *Int64) *Int64 {\n\tx := new(Int64).Set(q)\n\ty := new(Int64).Set(r)\n\tz := NewInt64()\n\tfor n, a := range x.c {\n\t\tif b, ok := y.Coeff(n); ok {\n\t\t\tz.SetCoeff(n, a-b)\n\t\t} else {\n\t\t\tz.SetCoeff(n, a)\n\t\t}\n\t}\n\tfor n, b := range y.c {\n\t\tif _, ok := x.Coeff(n); !ok {\n\t\t\tz.SetCoeff(n, -b)\n\t\t}\n\t}\n\treturn p.Set(z)\n}", "func (z *Int) Sub(x, y *Int) {\n\tvar underflow bool\n\n\tz[0], underflow = u64Sub(x[0], y[0], underflow)\n\tz[1], underflow = u64Sub(x[1], y[1], underflow)\n\tz[2], underflow = u64Sub(x[2], y[2], underflow)\n\tif underflow {\n\t\tz[3] = x[3] - y[3] - 1\n\t} else {\n\t\tz[3] = x[3] - y[3]\n\t}\n}", "func SUBB(imr, amr operand.Op) { ctx.SUBB(imr, amr) }", "func sub(x, y int) int {\n\treturn x - y\n}", "func (x Dec) Sub(y Dec) (Dec, error) {\n\tvar z Dec\n\t_, err := apd.BaseContext.Sub(&z.dec, &x.dec, &y.dec)\n\treturn z, errorsmod.Wrap(err, \"decimal subtraction error\")\n}", "func (t Tuple) Sub(o Tuple) Tuple {\n\tif t.IsVector() && o.IsPoint() {\n\t\tpanic(\"cannot subtract point from vector\")\n\t}\n\treturn Tuple{t.X - o.X, t.Y - o.Y, t.Z - o.Z, t.W - o.W}\n}", "func SUBQ(imr, mr operand.Op) { ctx.SUBQ(imr, mr) }", "func (ai *Arith) Sub(decimal1 *ZnDecimal, others ...*ZnDecimal) *ZnDecimal {\n\tvar result = copyZnDecimal(decimal1)\n\tif len(others) == 0 {\n\t\treturn result\n\t}\n\n\tfor _, item := range others {\n\t\tr1, r2 := rescalePair(result, item)\n\t\tresult.co.Sub(r1.co, r2.co)\n\t\tresult.exp = r1.exp\n\t}\n\treturn result\n}", "func (z *Float64) Sub(x, y *Float64) *Float64 {\n\tz.l = x.l - y.l\n\tz.r = x.r - y.r\n\treturn z\n}", "func Sub(a int, b int) int {\n\treturn a - b\n}", "func Sub(a int, b int) int {\n\treturn a - b\n}", "func Sub(a int, b int) int {\n\treturn a - b\n}", "func Sub(valueA gcv.Value, valueB gcv.Value) gcv.Value {\n\tif valueA.Type() == gcv.Complex || valueB.Type() == gcv.Complex {\n\t\treturn gcv.MakeValue(valueA.Complex() - valueB.Complex())\n\t}\n\treturn gcv.MakeValue(valueA.Real() - valueB.Real())\n}", "func Sub(v1, v2 *Vec) *Vec {\n\tnegV2 := Negate(v2)\n\treturn Add(v1, negV2)\n}", "func Sub(a, b Expr) Expr {\n\treturn &arithmeticOperator{&simpleOperator{a, b, scanner.SUB}}\n}", "func (c MethodsCollection) Subtract() pSubtract {\n\treturn pSubtract{\n\t\tMethod: c.MustGet(\"Subtract\"),\n\t}\n}", "func Substract(a, b NumberArray) (resultingMatrix NumberArray, err error) {\n\treturn binaryOperation(\"Substract\", a, b)\n}", "func (v Vec3i) Sub(other Vec3i) Vec3i {\n\treturn Vec3i{v.X - other.X, v.Y - other.Y, v.Z - other.Z}\n}", "func Sub(a, b int) int {\n\treturn a - b\n}", "func Sub(a, b int) int {\n\treturn a - b\n}", "func Sub(a, b int) int {\n\treturn a - b\n}", "func Subtract(x, y *Matrix) (*Matrix, error) {\n\tminusY, _ := Scale(-1, y)\n\tresult, err := Add(x, minusY)\n\treturn result, err\n}", "func Subtract(t, other Tuplelike) Tuplelike {\n\tresult := []float64{}\n\n\tfor idx, value := range t.Values() {\n\t\tresult = append(result, value-other.At(idx))\n\t}\n\n\treturn Tuple(result)\n}", "func (a Vec2) Sub(b Vec2) Vec2 {\n\treturn Vec2{a.X - b.X, a.Y - b.Y}\n}", "func (v Vec2) Sub(x Vec2) Vec2 {\n\treturn Vec2{v[0] - x[0], v[1] - x[1]}\n}", "func (a *Allocation) Subtract(other *Allocation) *Allocation {\n\tresult := initializeZeroAlloc()\n\tfor t, v := range a.Value {\n\t\tresult.Value[t] = v.Subtract(other.Value[t])\n\t}\n\treturn result\n}", "func SUBPS(mx, x operand.Op) { ctx.SUBPS(mx, x) }", "func (p Point3) Sub(ps ...Point3) Point3 {\n\tfor _, p2 := range ps {\n\t\tp[0] -= p2[0]\n\t\tp[1] -= p2[1]\n\t\tp[2] -= p2[2]\n\t}\n\treturn p\n}", "func (p Vector3) Sub(o Vector3) Vector3 {\n\treturn Vector3{p.X - o.X, p.Y - o.Y, p.Z - o.Z}\n}", "func PHSUBD(mx, x operand.Op) { ctx.PHSUBD(mx, x) }", "func (m *Uint64) Subtract(key interface{}, delta uint64) (new uint64) {\n\treturn m.Value(key).Subtract(delta)\n}", "func (k *Keeper) Subtract(ctx sdk.Context, address sdk.AccAddress, coins sdk.Coins) error {\n\tdeposit, found := k.GetDeposit(ctx, address)\n\tif !found {\n\t\treturn types.ErrorDepositDoesNotExist\n\t}\n\n\tdeposit.Coins, _ = deposit.Coins.SafeSub(coins)\n\tif deposit.Coins.IsAnyNegative() {\n\t\treturn types.ErrorInsufficientDepositFunds\n\t}\n\n\tif err := k.bank.SendCoinsFromModuleToAccount(ctx, types.ModuleName, address, coins); err != nil {\n\t\treturn err\n\t}\n\n\tk.SetDeposit(ctx, deposit)\n\treturn nil\n}", "func SUBSS(mx, x operand.Op) { ctx.SUBSS(mx, x) }", "func (i I) Subtract(i2 I) I {\n\treturn I{i.X - i2.X, i.Y - i2.Y}\n}", "func (v Vector) Sub(o Vector) *Vector {\n\treturn &Vector{v[0] - o[0], v[1] - o[1], v[2] - o[2]}\n}", "func (p *Point) Sub(p2 Point) {\n\tp.X -= p2.X\n\tp.Y -= p2.Y\n\tp.Z -= p2.Z\n}", "func (v Vector3D) Sub(other Vector3D) Vector3D {\n\treturn Vector3D{\n\t\tx: v.x - other.x,\n\t\ty: v.y - other.y,\n\t\tz: v.z - other.z,\n\t}\n}", "func (m *Money) Sub(n *Money) *Money {\n\tr := m.M - n.M\n\tif (r^m.M)&^(r^n.M) < 0 {\n\t\tpanic(ErrMoneyOverflow)\n\t}\n\tm.M = r\n\treturn m\n}", "func (v *V) Sub(x *V) *V {\n\tif !IsVSameShape(x, v) {\n\t\tpanic(ErrShape)\n\t}\n\tfor i, e := range x.Data {\n\t\tv.Data[i] -= e\n\t}\n\treturn v\n}", "func (v1 Vector2) Sub(v2 Vector2) Vector2 {\n\treturn Vector2{v1.X - v2.X, v1.Y - v2.Y}\n}", "func (a *Person) Subtract() {\r\n\ta.Account--\r\n}", "func Sub(x, y Number) Number {\n\treturn Number{\n\t\tReal: x.Real - y.Real,\n\t\tE1mag: x.E1mag - y.E1mag,\n\t\tE2mag: x.E2mag - y.E2mag,\n\t\tE1E2mag: x.E1E2mag - y.E1E2mag,\n\t}\n}", "func Sub(t1 TermT, t2 TermT) TermT {\n\treturn TermT(C.yices_sub(C.term_t(t1), C.term_t(t2)))\n}", "func (v Vector2D) Subtract(v1 Vector) Vector {\n\tv2 := v1.(Vector2D)\n\treturn Vector2D{v.X - v2.X, v.Y - v2.Y}\n}", "func Sub(z, x, y *Elt)", "func (v Vec2) Sub(other Vec2) Vec2 {\n\treturn Vec2{v.X - other.X, v.Y - other.Y}\n}" ]
[ "0.7223267", "0.7035692", "0.6913862", "0.6877539", "0.6848358", "0.68113536", "0.68023574", "0.6774074", "0.6738111", "0.6730106", "0.6730106", "0.6730106", "0.6708217", "0.6704883", "0.67037517", "0.6674617", "0.6651721", "0.6651721", "0.66430914", "0.6631221", "0.6624722", "0.6585441", "0.6574272", "0.6567326", "0.65583825", "0.65243244", "0.65150625", "0.6485563", "0.6472358", "0.64466774", "0.6438324", "0.63855183", "0.6333745", "0.63119996", "0.6276321", "0.62678355", "0.624328", "0.62334675", "0.62216926", "0.6217695", "0.62012565", "0.61943763", "0.6184754", "0.61832523", "0.61787754", "0.6175323", "0.6146909", "0.61363566", "0.6107499", "0.60948765", "0.60887975", "0.608726", "0.6077132", "0.60770535", "0.6070858", "0.6062652", "0.6044442", "0.6036385", "0.6033081", "0.6018071", "0.60157305", "0.6002541", "0.59886426", "0.5987637", "0.5987637", "0.5987637", "0.59697586", "0.5965833", "0.5965015", "0.5940893", "0.59400326", "0.5921759", "0.5913859", "0.5913859", "0.5913859", "0.5909082", "0.5903141", "0.58960056", "0.58918524", "0.5886793", "0.5883857", "0.5877677", "0.5869719", "0.5869333", "0.5864477", "0.5862361", "0.58603704", "0.58592045", "0.58546686", "0.5852079", "0.5845632", "0.58372235", "0.5831086", "0.5825608", "0.5814943", "0.5804343", "0.58031523", "0.57977927", "0.5797696", "0.5772435" ]
0.8263535
0
Divide operand by a.
func Divide(a, operand int) int { return operand / a }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *calculon) Divide(ctx context.Context, arg calculator.Operand) (calculator.Result, error) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\tif arg.Value == 0 {\n\t\treturn calculator.Result{}, calculator.InvalidArgumentError(map[string]string{\"value\": \"cannot be 0\"})\n\t}\n\tc.current /= arg.Value\n\treturn calculator.Result{c.current}, nil\n}", "func Div(a, b Expr) Expr {\n\treturn &arithmeticOperator{&simpleOperator{a, b, scanner.DIV}}\n}", "func Divide(a cty.Value, b cty.Value) (cty.Value, error) {\n\treturn DivideFunc.Call([]cty.Value{a, b})\n}", "func DIVB(mr operand.Op) { ctx.DIVB(mr) }", "func d(a int, b int) int {\n\treturn a/b\n}", "func Divide(a float64, b float64) float64 {\n\treturn a / b\n}", "func Divide() {\n\tMatch('/')\n\tFactor()\n\tEmitLn(\"MOVE (SP)+,D1\")\n\tEmitLn(\"DIVS D1,D0\")\n}", "func ValueDiv(a, b reflect.Value) (reflect.Value, error) {\n\taBkind := GetBaseKind(a)\n\tbBkind := GetBaseKind(b)\n\n\tswitch aBkind {\n\tcase reflect.Int64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(a.Int() / b.Int()), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Int() / int64(b.Uint())), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(float64(a.Int()) / b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do division math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tcase reflect.Uint64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(int64(a.Uint()) / b.Int()), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Uint() / b.Uint()), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(float64(a.Uint()) / b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do division math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tcase reflect.Float64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(a.Float() / float64(b.Int())), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Float() / float64(b.Uint())), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(a.Float() / b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do division math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tdefault:\n\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do division math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t}\n}", "func Div(a, b Expr) Expr {\n\treturn &divOp{&simpleOperator{a, b, scanner.DIV}}\n}", "func Div( a *context.Value, b *context.Value ) (*context.Value,error) {\n if a != nil && b != nil {\n if b.IsZero() {\n return nil, errors.New( \"Division by zero\")\n }\n\n switch a.OperationType( b ) {\n case context.VAR_BOOL:\n return context.IntValue( a.Int() / b.Int() ), nil\n case context.VAR_INT:\n return context.FloatValue( a.Float() / b.Float() ), nil\n case context.VAR_FLOAT:\n return context.FloatValue( a.Float() / b.Float() ), nil\n case context.VAR_COMPLEX:\n return context.ComplexValue( a.Complex() / b.Complex() ), nil\n }\n }\n\n return nil, errors.New( \"Unsupported type for div\" )\n}", "func DIVPD(mx, x operand.Op) { ctx.DIVPD(mx, x) }", "func DIVL(mr operand.Op) { ctx.DIVL(mr) }", "func (gf *GaloisField) Divide(a, b int) int {\n\tif b == 0 {\n\t\tpanic(\"divide by zero\")\n\t} else if a == 0 {\n\t\treturn 0\n\t}\n\treturn gf.ALogTbl[(gf.LogTbl[a]-gf.LogTbl[b])%(gf.Size-1)]\n}", "func Div(a float64, b float64) (float64, error) {\n\tif b == 0 {\n\t\treturn 0.0, errors.New(\"Can't devide by zero\")\n\t}\n\n\treturn a / b, nil\n}", "func DIVQ(mr operand.Op) { ctx.DIVQ(mr) }", "func Command_Div(script *rex.Script, params []*rex.Value) {\n\tif len(params) != 2 {\n\t\trex.ErrorParamCount(\"float:div\", \"2\")\n\t}\n\n\tscript.RetVal = rex.NewValueFloat64(params[0].Float64() / params[1].Float64())\n\treturn\n}", "func (expr Expression) Divide() (float64, error) {\n\tif expr.SecondOperand.Value == 0 {\n\t\treturn 0, errors.New(\"division by zero\")\n\t}\n\n\treturn (expr.FirstOperand.Value / expr.SecondOperand.Value), nil\n}", "func (z *Float64) Divide(y *Float64, a float64) *Float64 {\n\tz.l = y.l / a\n\tz.r = y.r / a\n\treturn z\n}", "func Divide(a, b int) (float64, error) {\n\tif b == 0 {\n\t\treturn 0, errors.New(\"You cannot divide by zero\")\n\t}\n\n\treturn float64(a) / float64(b), nil\n}", "func Divide(a, b int) int {\n\treturn a / b\n}", "func Divide(a, b int) int {\n\treturn a / b\n}", "func (b ValExprBuilder) Div(expr interface{}) ValExprBuilder {\n\treturn b.makeBinaryExpr('/', expr)\n}", "func gfDivide(a, b gfElement) gfElement {\n\tif a == gfZero {\n\t\treturn gfZero\n\t} else if b == gfZero {\n\t\tlog.Panicln(\"Divide by zero\")\n\t}\n\n\treturn gfMultiply(a, gfInverse(b))\n}", "func DIVSD(mx, x operand.Op) { ctx.DIVSD(mx, x) }", "func (v *Vector) DividedBy(a float64) *Vector {\n\treturn &Vector{X: v.X / a, Y: v.Y / a, Z: v.Z / a}\n}", "func div(a, b int32) int32 {\n\tif a >= 0 {\n\t\treturn (a + (b >> 1)) / b\n\t}\n\treturn -((-a + (b >> 1)) / b)\n}", "func (c *Calculator) Div() {\n\tif opValue, err := c.getOperationValue(); err != nil || opValue == 0 {\n\t\tc.returnError()\n\t} else {\n\t\tlog.Printf(\"%f / %f = \", value, opValue)\n\t\tvalue /= opValue\n\t\tlog.Printf(\"%f\\n\", value)\n\t\tc.returnResult()\n\t}\n}", "func DIVW(mr operand.Op) { ctx.DIVW(mr) }", "func DIVPS(mx, x operand.Op) { ctx.DIVPS(mx, x) }", "func Div(a, b int) (int, error) {\n\tif b == 0 {\n\t\treturn 0, errors.New(\"can't divide by 0\")\n\t}\n\n\treturn a / b, nil\n\n}", "func SimpleDivide(exp string) string {\n var index int = strings.Index(exp, \"/\")\n var arg1 string = exp[:index]\n var arg2 string = exp[index + 1:]\n val1 := NotateToDouble(arg1)\n val2 := NotateToDouble(arg2)\n return NegativeNotate(fmt.Sprintf(\"%f\", val1 / val2))\n}", "func (cal *Calculate) div(value float64) (result float64) {\n\tif len(cal.Arg) == 2 {\n\t\treturn (cal.Arg[0] / cal.Arg[1])\n\t} else if len(cal.Arg) == 1 {\n\n\t\treturn (value / cal.Arg[0])\n\t}\n\n\tlog.Fatalln(\"Please check the data format of the calculation unit\")\n\treturn\n}", "func Div(a, b *big.Float) *big.Float {\n\treturn ZeroBigFloat().Quo(a, b)\n}", "func Divide(a, b float64) (result float64, err error) {\n\tif b != 0 {\n\t\treturn a / b, err\n\t}\n\terr = errors.New(\"you cannot divide by 0\")\n\n\treturn 555, err\n}", "func Divide(x, y int) float64 {\n\treturn float64(x) / float64(y)\n}", "func (A Matrix) ScalarDivide(a float64) {\n\tA.ScalarMultiply(1.0 / a)\n}", "func (z *Int) Div(x, y *Int) *Int {}", "func calcDiv(rOpd, lOpd formulaArg, opdStack *Stack) error {\n\tlOpdVal := lOpd.ToNumber()\n\tif lOpdVal.Type != ArgNumber {\n\t\treturn errors.New(lOpdVal.Value())\n\t}\n\trOpdVal := rOpd.ToNumber()\n\tif rOpdVal.Type != ArgNumber {\n\t\treturn errors.New(rOpdVal.Value())\n\t}\n\tif rOpdVal.Number == 0 {\n\t\treturn errors.New(formulaErrorDIV)\n\t}\n\topdStack.Push(newNumberFormulaArg(lOpdVal.Number / rOpdVal.Number))\n\treturn nil\n}", "func Divide(a, b int) (int, error) {\n\tif b == 0 {\n\t\treturn 0, errors.New(\"cannot divide by 0\")\n\t}\n\treturn a / b, nil\n}", "func Div(valueA gcv.Value, valueB gcv.Value) gcv.Value {\n\tif valueA.Type() == gcv.Complex || valueB.Type() == gcv.Complex {\n\t\treturn gcv.MakeValue(valueA.Complex() / valueB.Complex())\n\t}\n\treturn gcv.MakeValue(valueA.Real() / valueB.Real())\n}", "func Divide(a int, b int) (int, error) {\n\t// cannot divide by 0\n\tif b == 0 {\n\t\t// use in-built errors interface to return error message\n\t\treturn 0, errors.New(\"cannot divide by 0\")\n\t}\n\t// otherwise perform calculation\n\treturn (a / b), nil\n}", "func (i *Number) Divide(v Number) *Number {\n\tif v.value == 0 {\n\t\treturn NewNumber(i.value / 1)\n\t}\n\treturn NewNumber(i.value / v.value)\n}", "func (s *server) Divide(ctx context.Context, in *pb.DivideRequest) (*pb.DivideReply, error) {\n\treturn &pb.DivideReply{N1: in.N1 / in.N2}, nil\n}", "func cdiv(a, b int) int { return (a + b - 1) / b }", "func DIVSS(mx, x operand.Op) { ctx.DIVSS(mx, x) }", "func Divide(a, b int64) (int64, int64, error) {\n\tif b == 0 {\n\t\treturn 0, 0, ErrDivideByZero\n\t}\n\treturn a / b, a % b, nil\n}", "func (a *Vec4) Divide(s float32) {\n\ta.X /= s\n\ta.Y /= s\n\ta.Z /= s\n\ta.W /= s\n}", "func Modulo(a, operand int) int { return operand % a }", "func IDIVB(mr operand.Op) { ctx.IDIVB(mr) }", "func (gdt *Vector3) OperatorDivideScalar(b Real) Vector3 {\n\targ0 := gdt.getBase()\n\targ1 := b.getBase()\n\n\tret := C.go_godot_vector3_operator_divide_scalar(GDNative.api, arg0, arg1)\n\n\treturn Vector3{base: &ret}\n\n}", "func (a *Array64) Div(b *Array64) *Array64 {\n\tif a.valRith(b, \"Div\") {\n\t\treturn a\n\t}\n\n\tif b.shape[len(b.shape)-1] == a.shape[len(a.shape)-1] {\n\t\tasm.Div(a.data, b.data)\n\t\treturn a\n\t}\n\n\tst := a.strides[len(a.strides)-1] * a.shape[len(a.shape)-1]\n\tfor i := 0; i < len(b.data); i++ {\n\t\tasm.DivC(b.data[i], a.data[i*st:(i+1)*st])\n\t}\n\treturn a\n}", "func Div(x, y int) int {\n\treturn x / y\n}", "func NewDivExpr(scanner parser.Scanner, a, b Expr) Expr {\n\treturn newArithExpr(scanner, a, b, \"/\", func(a, b float64) float64 { return a / b })\n}", "func (c *Clac) Div() error {\n\treturn c.applyFloat(2, func(vals []value.Value) (value.Value, error) {\n\t\treturn binary(vals[1], \"/\", vals[0])\n\t})\n}", "func (f Fixed8) Div(i int64) Fixed8 {\n\treturn f / Fixed8FromInt64(i)\n}", "func divide(divisor int, dividends ...int) (float64, error) {\n\tif divisor == 0 {\n\t\treturn 0, errors.New(\"Division by zero\")\n\t} else {\n\t\tdividend := 0\n\t\tfor _, d := range dividends {\n\t\t\tdividend += d\n\t\t}\n\t\treturn float64(dividend / divisor), nil\n\t}\n}", "func Divide(a, b int) (int, error) { // Go一般习惯把最后一个返回值设置为返回错误类型\n\tif b == 0 {\n\t\treturn 0, errors.New(\"divide by zero\")\n\t}\n\treturn a / b, nil // 最后一个值返回为nil表示没有错误\n}", "func IDIVL(mr operand.Op) { ctx.IDIVL(mr) }", "func Div(a, b interface{}, opts ...FuncOpt) (retVal Tensor, err error) {\n\tad, adok := a.(*Dense)\n\tbd, bdok := b.(*Dense)\n\n\tswitch {\n\tcase adok && bdok:\n\t\treturn ad.Div(bd, opts...)\n\tcase adok && !bdok:\n\t\treturn ad.ScaleInv(b, opts...)\n\tcase !adok && bdok:\n\t\treturn bd.ScaleInvR(a, opts...)\n\t}\n\tpanic(\"Unreachable\")\n}", "func (a Value) Ratio(b Value) (*Value, error) {\n\tvar err error\n\tnum := &a\n\tden := &b\n\n\tif num.IsNative() {\n\t\tnum, err = num.NonNative()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnum, err = num.Divide(*xrpMultipler)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tif den.IsNative() {\n\t\tden, err = den.NonNative()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tden, err = den.Divide(*xrpMultipler)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tquotient, err := num.Divide(*den)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn quotient, nil\n}", "func (op *OpDiv) Eval(x, y float32) float32 {\n\treturn op.LeftChild.Eval(x, y) / op.RightChild.Eval(x, y)\n}", "func Div(dividend, divisor *big.Int) *big.Int { return I().Div(dividend, divisor) }", "func (a Vec2) Div(b Vec2) Vec2 {\n\treturn Vec2{a.X / b.X, a.Y / b.Y}\n}", "func ScalarDivision(m Multivector, d string) Multivector {\n\tm.E0.SetString(m.E0.Num().String() + \"/\" + d)\n\tm.E1.SetString(m.E1.Num().String() + \"/\" + d)\n\tm.E2.SetString(m.E2.Num().String() + \"/\" + d)\n\tm.E3.SetString(m.E3.Num().String() + \"/\" + d)\n\tm.E12.SetString(m.E12.Num().String() + \"/\" + d)\n\tm.E13.SetString(m.E13.Num().String() + \"/\" + d)\n\tm.E23.SetString(m.E23.Num().String() + \"/\" + d)\n\tm.E123.SetString(m.E123.Num().String() + \"/\" + d)\n\n\treturn m\n}", "func Divide(t Tuplelike, n float64) Tuplelike {\n\treturn Multiply(t, 1/n)\n}", "func divideNumber(a, b int) int {\n\tif b == 0 {\n\t\tfmt.Printf(\"%d cannot divided by \", a)\n\t\treturn b\n\t}\n\treturn a / b\n}", "func (r *ImageRef) Divide(denominator *ImageRef) error {\n\tout, err := vipsDivide(r.image, denominator.image)\n\tif err != nil {\n\t\treturn err\n\t}\n\tr.setImage(out)\n\treturn nil\n}", "func (g *Group) Div(divisor int) Group {\n\tg.GrainDishes /= divisor\n\tg.VegetableDishes /= divisor\n\tg.FishAndMealDishes /= divisor\n\tg.Milk /= divisor\n\tg.Fruit /= divisor\n\treturn *g\n}", "func Divisor(numeroA int, numeroB int) (resultado int) {\n\tresultado = numeroA / numeroB\n\treturn\n}", "func divmod(a, b int) (q int, r int) {\n\tif b == 0 {\n\t\treturn\n\t}\n\tq = a / b\n\tr = a % b\n\treturn\n}", "func divide(S, D uint32) uint32 {\n\treturn 65535 * S / D\n}", "func divSlice(out, a, b []float64)", "func DivMod(a int, b int, div *int, mod *int) {\n\t*div = a / b\n\t*mod = a % b\n}", "func Divide(v1, v2 sqltypes.Value) (sqltypes.Value, error) {\n\tif v1.IsNull() || v2.IsNull() {\n\t\treturn sqltypes.NULL, nil\n\t}\n\te1, err := valueToEval(v1, collationNumeric)\n\tif err != nil {\n\t\treturn sqltypes.NULL, err\n\t}\n\te2, err := valueToEval(v2, collationNumeric)\n\tif err != nil {\n\t\treturn sqltypes.NULL, err\n\t}\n\tr, err := divideNumericWithError(e1, e2, true)\n\tif err != nil {\n\t\treturn sqltypes.NULL, err\n\t}\n\treturn evalToSQLValue(r), nil\n}", "func (i I) Divide(i2 I) I {\n\ti.X /= i2.X\n\ti.Y /= i2.Y\n\treturn i\n}", "func TestPerformArithmeticOperationDivideByZero(t *testing.T) {\n\t// operand stack (also known as data stack)\n\tstack := evaluator.Stack{}\n\n\t// push two values onto the stack\n\tstack.Push(1)\n\tstack.Push(0)\n\n\t// any token that is not token.QUO or token.REM\n\ttok := token.QUO\n\n\t// perform the selected arithmetic operation\n\taddOperation := func(x int, y int) int { return x + y }\n\n\t// perform the selected arithmetic operation\n\terr := evaluator.PerformArithmeticOperation(&stack, addOperation, tok)\n\tassert.Error(t, err)\n}", "func divide(dividend int, divisor int) int {\n\tif dividend == 0 {\n\t\treturn dividend\n\t}\n\tif dividend == -2147483648 && divisor == -1 {\n\t\treturn 2147483647\n\t}\n\tres := 0\n\tcount := 1\n\tsymbol := true\n\n\tif (dividend > 0 && divisor < 0) || (dividend < 0 && divisor > 0) {\n\t\tsymbol = false\n\t}\n\n\tif dividend < 0 {\n\t\tdividend = -dividend\n\t}\n\tif divisor < 0 {\n\t\tdivisor = -divisor\n\t}\n\tmultiDivisor := divisor\n\tfor dividend >= divisor {\n\t\tif dividend > multiDivisor {\n\t\t\tdividend -= multiDivisor\n\t\t\tres += count\n\t\t\tmultiDivisor += multiDivisor\n\t\t\tcount += count\n\t\t} else if dividend == multiDivisor {\n\t\t\tdividend -= multiDivisor\n\t\t\tres += count\n\t\t\tbreak\n\t\t} else {\n\t\t\tif multiDivisor > divisor { //use >>\n\t\t\t\tmultiDivisor = multiDivisor >> 1\n\t\t\t\tcount = count >> 1\n\t\t\t}\n\t\t}\n\t}\n\tif !symbol {\n\t\tres = -res\n\t}\n\treturn res\n}", "func (d *Division) Evaluate(left, right EvalResult) (EvalResult, error) {\n\treturn divideNumericWithError(left, right)\n}", "func (f Fixed) Div(f0 Fixed) Fixed {\n\tif f.IsNaN() || f0.IsNaN() {\n\t\treturn NaN\n\t}\n\treturn NewF(f.Float() / f0.Float())\n}", "func (l Integer) Div(r Number) Number {\n\tif ri, ok := r.(Integer); ok {\n\t\tif ri == 0 {\n\t\t\tpanic(errors.New(ErrDivideByZero))\n\t\t}\n\t\tres := big.NewRat(int64(l), int64(ri))\n\t\treturn maybeWhole(res)\n\t}\n\tpl, pr := purify(l, r)\n\treturn pl.Div(pr)\n}", "func IDIVW(mr operand.Op) { ctx.IDIVW(mr) }", "func IDIVQ(mr operand.Op) { ctx.IDIVQ(mr) }", "func (f Fixed) Div(f0 Fixed) Fixed {\n\tif f.IsNaN() || f0.IsNaN() {\n\t\treturn NaN\n\t}\n\treturn NewFromFloat(f.Float() / f0.Float())\n}", "func division(dividend, divisor int) int {\n\tsign := 1\n\tif (dividend < 0 && divisor > 0) || (dividend > 0 && divisor < 0) {\n\t\tsign = -1\n\t}\n\n\tdividend = abs(dividend)\n\tdivisor = abs(divisor)\n\n\tquotient := 0\n\tfor dividend > divisor {\n\t\tdividend -= divisor\n\t\tquotient++\n\t}\n\n\treturn sign * quotient\n}", "func division(x, y int) (int, error, error) {\n\tif y == 0 {\n\t\treturn 0, nil, errors.New(\"Cannot divide by zero!\")\n\t}\n\tif x%y != 0 {\n\t\tremainder := errors.New(\"There is a remainder!\")\n\t\treturn x / y, remainder, nil\n\t} else {\n\t\treturn x / y, nil, nil\n\t}\n\n}", "func (q Quat) Div(other Quat) Quat {\n\treturn Quat{q.W / other.W, q.X / other.X, q.Y / other.Y, q.Z / other.Z}\n}", "func (p Point2) Div(ps ...Point2) Point2 {\n\tfor _, p2 := range ps {\n\t\tp[0] /= p2[0]\n\t\tp[1] /= p2[1]\n\t}\n\treturn p\n}", "func rcDiv(p *TCompiler, code *TCode) (*value.Value, error) {\n\tv := value.Div(p.regGet(code.B), p.regGet(code.C))\n\tp.regSet(code.A, v)\n\tp.moveNext()\n\treturn v, nil\n}", "func (f F128d16) Div(value F128d16) F128d16 {\n\treturn F128d16{data: f.data.Mul(multiplierF128d16).Div(value.data)}\n}", "func (sig Signature) DivideBy(signatures []Signature) (Signature, error) {\n\tif len(signatures) == 0 {\n\t\treturn sig, nil\n\t}\n\n\t// Get a C pointer to an array of signatures\n\tcSigArrPtr := C.AllocPtrArray(C.size_t(len(signatures)))\n\tdefer C.FreePtrArray(cSigArrPtr)\n\t// Loop thru each sig and add the pointer to it, to the C pointer array at\n\t// the given index.\n\tfor i, sig := range signatures {\n\t\tC.SetPtrArray(cSigArrPtr, unsafe.Pointer(sig.sig), C.int(i))\n\t}\n\n\tvar quo Signature\n\tvar cDidErr C.bool\n\tquo.sig = C.CSignatureDivideBy(sig.sig, cSigArrPtr, C.size_t(len(signatures)), &cDidErr)\n\tif bool(cDidErr) {\n\t\tcErrMsg := C.GetLastErrorMsg()\n\t\terr := errors.New(C.GoString(cErrMsg))\n\t\treturn Signature{}, err\n\t}\n\n\truntime.SetFinalizer(&quo, func(p *Signature) { p.Free() })\n\treturn quo, nil\n}", "func div(x, y int) (answer int, err error) {\n\tif y == 0 {\n\t\terr = fmt.Errorf(\"Cannot Divid by zero\")\n\t} else {\n\t\tanswer = x / y\n\t}\n\treturn\n}", "func (p Point) Div(k int) Point { return Point{p.X / k, p.Y / k} }", "func TestPerformArithmeticOperationDivideByNotZero(t *testing.T) {\n\t// operand stack (also known as data stack)\n\tstack := evaluator.Stack{}\n\n\t// push two values onto the stack\n\tstack.Push(4)\n\tstack.Push(2)\n\n\t// any token that is not token.QUO or token.REM\n\ttok := token.QUO\n\n\t// perform the selected arithmetic operation\n\taddOperation := func(x int, y int) int { return x + y }\n\n\t// perform the selected arithmetic operation\n\terr := evaluator.PerformArithmeticOperation(&stack, addOperation, tok)\n\tassert.NoError(t, err)\n}", "func VDIVPD(ops ...operand.Op) { ctx.VDIVPD(ops...) }", "func div(x byte, y byte) byte {\n\tif x == 0 || y == 0 {\n\t\treturn 0\n\t}\n\treturn expOp[0xff+logOp[x]-logOp[y]]\n}", "func divmod(a, b, mod *big.Int) *big.Int {\n\tb = b.ModInverse(b, mod)\n\tif b == nil {\n\t\treturn nil\n\t}\n\treturn a.Mul(a, b)\n}", "func (s *Server) Divide(ctx context.Context, message *calcpb.DivideRequest) (*calcpb.DivideResponse, error) {\n\tctx = context.WithValue(ctx, goa.MethodKey, \"divide\")\n\tctx = context.WithValue(ctx, goa.ServiceKey, \"calc\")\n\tresp, err := s.DivideH.Handle(ctx, message)\n\tif err != nil {\n\t\tif en, ok := err.(ErrorNamer); ok {\n\t\t\tswitch en.ErrorName() {\n\t\t\tcase \"div_by_zero\":\n\t\t\t\ter := err.(*calc.DivByZero)\n\t\t\t\treturn nil, goagrpc.NewStatusError(codes.InvalidArgument, err, NewDivideDivByZeroError(er))\n\t\t\tcase \"timeout\":\n\t\t\t\treturn nil, goagrpc.NewStatusError(codes.DeadlineExceeded, err, goagrpc.NewErrorResponse(err))\n\t\t\t}\n\t\t}\n\t\treturn nil, goagrpc.EncodeError(err)\n\t}\n\treturn resp.(*calcpb.DivideResponse), nil\n}", "func newDivExpr(op BinaryOp, lhs, rhs Expr) Expr {\n\tassert(op == UDIV || op == SDIV, \"invalid div op: %s\", op)\n\n\tif lhs, ok := lhs.(*ConstantExpr); ok {\n\t\tif rhs, ok := rhs.(*ConstantExpr); ok {\n\t\t\tif op == UDIV {\n\t\t\t\treturn lhs.UDiv(rhs)\n\t\t\t}\n\t\t\treturn lhs.SDiv(rhs)\n\t\t}\n\t}\n\tif ExprWidth(lhs) == WidthBool {\n\t\treturn lhs // rhs must be 1\n\t}\n\treturn &BinaryExpr{Op: op, LHS: lhs, RHS: rhs}\n}", "func DecimalDiv(from1, from2, to *MyDecimal, fracIncr int) error {\n\tfrom1, from2, to = validateArgs(from1, from2, to)\n\tto.resultFrac = mathutil.Min(from1.resultFrac+int8(fracIncr), mysql.MaxDecimalScale)\n\treturn doDivMod(from1, from2, to, nil, fracIncr)\n}", "func (rn *RangedNumber) Div(other *RangedNumber) *RangedNumber {\n\treturn rn.Set(rn.min/other.min, rn.max/other.max)\n}" ]
[ "0.69174546", "0.6843468", "0.6825028", "0.6747531", "0.6714473", "0.67077965", "0.6701392", "0.67010087", "0.6653182", "0.65372056", "0.65251654", "0.65114766", "0.6497556", "0.6494549", "0.64721185", "0.64691997", "0.64665496", "0.6429992", "0.63459116", "0.63432753", "0.63432753", "0.63379186", "0.6302599", "0.6257926", "0.6216312", "0.62111425", "0.62080646", "0.61707777", "0.6166701", "0.6140232", "0.6118848", "0.610607", "0.607825", "0.6060286", "0.6057493", "0.6056412", "0.6056297", "0.60526586", "0.60263824", "0.6007897", "0.5985724", "0.5971579", "0.59663326", "0.5965059", "0.59491295", "0.5939528", "0.5936868", "0.59360033", "0.5928073", "0.5908438", "0.5907627", "0.58309704", "0.5829491", "0.58207446", "0.5790025", "0.57662994", "0.5762752", "0.5744213", "0.5740562", "0.57200694", "0.5676443", "0.56709146", "0.5667499", "0.56590396", "0.5619878", "0.5605265", "0.5602394", "0.5592949", "0.5582361", "0.5581927", "0.55810165", "0.557016", "0.5566253", "0.55655223", "0.5535975", "0.55292267", "0.55034816", "0.54875743", "0.5484622", "0.5475811", "0.5471382", "0.54621416", "0.5461742", "0.54458004", "0.54097533", "0.5407439", "0.5401157", "0.5395769", "0.5395684", "0.53720176", "0.536626", "0.5365374", "0.53621", "0.53580344", "0.5348644", "0.5338557", "0.53376853", "0.5330131", "0.53264636", "0.532582" ]
0.8398136
0
Modulo returns operand modulo a.
func Modulo(a, operand int) int { return operand % a }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Modulo(a cty.Value, b cty.Value) (cty.Value, error) {\n\treturn ModuloFunc.Call([]cty.Value{a, b})\n}", "func mod(a int, n int) int {\n\tif n == 0 {\n\t\treturn 0\n\t}\n\tval := a - (n * int(a/n))\n\tif val < 0 {\n\t\treturn val + n\n\t}\n\treturn val\n}", "func Mod(a, b Expr) Expr {\n\treturn &arithmeticOperator{&simpleOperator{a, b, scanner.MOD}}\n}", "func (this *BigInteger) Mod(a *BigInteger) *BigInteger {\n\tvar r *BigInteger = NewBigInteger()\n\tthis.Abs().DivRemTo(a, nil, r)\n\tif this.S < 0 && r.CompareTo(ZERO) > 0 {\n\t\ta.SubTo(r, r)\n\t}\n\treturn r\n}", "func (z *Int) Mod(x, y *Int) *Int {}", "func mod(v, modulus int) int {\n\treturn (v%modulus + modulus) % modulus\n}", "func mod(val, m int) int {\n\tres := val % m\n\tif res < 0 {\n\t\tres += m\n\t}\n\treturn res\n}", "func Mod(a, b Expr) Expr {\n\treturn &modOp{&simpleOperator{a, b, scanner.MOD}}\n}", "func calMod(b, mod byte) byte {\n\treturn b % mod\n}", "func Mod(dividend, divisor *big.Int) *big.Int { return I().Mod(dividend, divisor) }", "func (l Integer) Mod(r Number) Number {\n\tif ri, ok := r.(Integer); ok {\n\t\tif ri == 0 {\n\t\t\tpanic(errors.New(ErrDivideByZero))\n\t\t}\n\t\treturn l % ri\n\t}\n\tpl, pr := purify(l, r)\n\treturn pl.Mod(pr)\n}", "func (vm *VM) opMod(instr []uint16) int {\n\ta, b, c := vm.getAbc(instr)\n\n\tvm.registers[a] = b % c\n\treturn 4\n}", "func ModInt(a, m int) int {\n\ta = a % m\n\tif a < 0 {\n\t\ta += m\n\t}\n\treturn a\n}", "func (r *Ring) mod(p int) int {\n\treturn p % len(r.buff)\n}", "func DivMod(a int, b int, div *int, mod *int) {\n\t*div = a / b\n\t*mod = a % b\n}", "func (z *Int) DivMod(x, y, m *Int) (*Int, *Int) {}", "func modi(val, m int) int {\n\tres := val % m\n\tif res < 0 {\n\t\tres += m\n\t}\n\treturn res\n}", "func PowMod(a, b, m int) int {\n\ta = a % m\n\tp := 1 % m\n\tfor b > 0 {\n\t\tif b&1 != 0 {\n\t\t\tp = (p * a) % m\n\t\t}\n\t\tb >>= 1\n\t\ta = (a * a) % m\n\t}\n\treturn p\n}", "func PowMod(a, b, m int) int {\n\ta = a % m\n\tp := 1 % m\n\tfor b > 0 {\n\t\tif b&1 != 0 {\n\t\t\tp = (p * a) % m\n\t\t}\n\t\tb >>= 1\n\t\ta = (a * a) % m\n\t}\n\treturn p\n}", "func PowMod(a, b, m int) int {\n\ta = a % m\n\tp := 1 % m\n\tfor b > 0 {\n\t\tif b&1 != 0 {\n\t\t\tp = (p * a) % m\n\t\t}\n\t\tb >>= 1\n\t\ta = (a * a) % m\n\t}\n\treturn p\n}", "func (b ValExprBuilder) Mod(expr interface{}) ValExprBuilder {\n\treturn b.makeBinaryExpr('%', expr)\n}", "func Mod(d, m int) int {\n\tvar res int = d % m\n\tif (res < 0 && m > 0) || (res > 0 && m < 0) {\n\t\treturn res + m\n\t}\n\treturn res\n}", "func Mod(d, m int) int {\n\tvar res int = d % m\n\tif ((res < 0 && m > 0) || (res > 0 && m < 0)) {\n\t return res + m\n\t}\n\treturn res\n }", "func (r Rand) Modulo(n int) int {\n\t// modulo len(groups) with big.Ints (Mod method works on pointers)\n\t//var b big.Int\n\tb := big.NewInt(0)\n\tb.SetBytes(r)\n\tb.Mod(b, big.NewInt(int64(n)))\n\treturn int(b.Int64())\n}", "func (r Rand) Modulo(n int) int {\n\t// modulo len(groups) with big.Ints (Mod method works on pointers)\n\t//var b big.Int\n\tb := big.NewInt(0)\n\tb.SetBytes(r.Bytes())\n\tb.Mod(b, big.NewInt(int64(n)))\n return int(b.Int64())\n}", "func (l *BigInt) Mod(r Number) Number {\n\tif ri, ok := r.(*BigInt); ok {\n\t\tlb := (*big.Int)(l)\n\t\trb := (*big.Int)(ri)\n\t\tif rb.IsInt64() && rb.Int64() == 0 {\n\t\t\tpanic(errors.New(ErrDivideByZero))\n\t\t}\n\t\tres := new(big.Int).Rem(lb, rb)\n\t\treturn maybeInteger(res)\n\t}\n\tlp, rp := purify(l, r)\n\treturn lp.Mod(rp)\n}", "func modll(val, m int64) int64 {\n\tres := val % m\n\tif res < 0 {\n\t\tres += m\n\t}\n\treturn res\n}", "func ModMul(a, b, mod int) int {\n\ta, b = a%mod, b%mod\n\tif b == 0 {\n\t\treturn 0\n\t}\n\tif a*b/b == a {\n\t\treturn a * b % mod\n\t}\n\tpanic(\"overflow\")\n}", "func divmod(a, b int) (q int, r int) {\n\tif b == 0 {\n\t\treturn\n\t}\n\tq = a / b\n\tr = a % b\n\treturn\n}", "func ModInv(a, mod int) int {\n\treturn ModPow(a, mod-2, mod)\n}", "func IntMod(a, b int) int {\n\treturn int(math.Abs(float64(a - b*int(math.Floor(float64(a)/float64(b))))))\n}", "func NewModExpr(scanner parser.Scanner, a, b Expr) Expr {\n\treturn newArithExpr(scanner, a, b, \"%%\", func(a, b float64) float64 {\n\t\treturn math.Mod(a, b)\n\t})\n}", "func IntMod(z *big.Int, x, y *big.Int,) *big.Int", "func divmod(x, m uint64) (quo, rem uint64) {\n\tquo = x / m\n\trem = x % m\n\treturn\n}", "func divmod(a, b, mod *big.Int) *big.Int {\n\tb = b.ModInverse(b, mod)\n\tif b == nil {\n\t\treturn nil\n\t}\n\treturn a.Mul(a, b)\n}", "func Mod(x, y float64) float64 {\n\tres := math.Mod(x, y)\n\tif (res < 0 && y > 0) || (res > 0 && y < 0) {\n\t\treturn res + y\n\t}\n\n\treturn res\n}", "func MulMod(a, b, m int64) int64 {\n\tvar sum int64 = 0\n\ta, b = a%m, b%m\n\tfor b != 0 {\n\t\tif b&1 == 1 {\n\t\t\tsum = (sum + a) % m\n\t\t}\n\t\ta, b = (2*a)%m, b>>1\n\t}\n\treturn sum\n}", "func mod(c uint) uint {\n\tconst MODULUS uint = 0x14d\n\n\t// c2 = (c << 1) ^ ((c & 0x80) ? MODULUS : 0)\n\tif c & 0x80 {\n\t\tv := MODULUS\n\t} else {\n\t\tv := 0\n\t}\n\tc2 := (c << 1) ^ v\n\n\t// c1 = c2 ^ (c >> 1) ^ ((c & 1) ? (MODULUS >> 1) : 0)\n\tif c & 1 {\n\t\tv = MODULUS >> 1\n\t} else {\n\t\tv = 0\n\t}\n\tc1 := c2 ^ (c >> 1) ^ v\n\n\treturn c | (c1 << 8) | (c2 << 16) | (c1 << 24)\n}", "func Mod(x, y *big.Int) *big.Int {\n\treturn new(big.Int).Mod(x, y)\n}", "func IntDivMod(z *big.Int, x, y, m *big.Int,) (*big.Int, *big.Int,)", "func Mod(valueA gcv.Value, valueB gcv.Value) (gcv.Value, error) {\n\tif valueA.Type() == gcv.Complex || valueB.Type() == gcv.Complex {\n\t\treturn nil, errors.New(\"Modulo is not supported for Complex numbers\")\n\t}\n\treturn gcv.MakeValue(math.Mod(valueA.Real(), valueB.Real())), nil\n}", "func powmod(a, k int) int {\n\ta %= 1337\n\tresult := 1\n\tfor i := 0; i < k; i++ {\n\t\tresult = (result * a) % 1337\n\t}\n\treturn result\n}", "func Divide(a, operand int) int { return operand / a }", "func (d Decimal) Mod(d2 Decimal) Decimal {\n\tquo := d.Div(d2).Truncate(0)\n\treturn d.Sub(d2.Mul(quo))\n}", "func (d Decimal) Mod(d2 Decimal) Decimal {\n\tquo := d.Div(d2).Truncate(0)\n\treturn d.Sub(d2.Mul(quo))\n}", "func DivMod(x, y, m *big.Int) (*big.Int, *big.Int) {\n\treturn new(big.Int).DivMod(x, y, m)\n}", "func (z *Int) Mod(x, y *Int) *Int {\n\tif x.IsZero() || y.IsZero() {\n\t\treturn z.Clear()\n\t}\n\tswitch x.Cmp(y) {\n\tcase -1:\n\t\t// x < y\n\t\tcopy(z[:], x[:])\n\t\treturn z\n\tcase 0:\n\t\t// x == y\n\t\treturn z.Clear() // They are equal\n\t}\n\n\t// At this point:\n\t// x != 0\n\t// y != 0\n\t// x > y\n\n\t// Shortcut trivial case\n\tif x.IsUint64() {\n\t\treturn z.SetUint64(x.Uint64() % y.Uint64())\n\t}\n\n\tq := NewInt()\n\tq.Div(x, y)\n\tq.Mul(q, y)\n\tz.Sub(x, q)\n\treturn z\n}", "func (z *polyGF2) Mod(a, b *polyGF2) *polyGF2 {\n\tvar out *polyGF2\n\tif z != a && z != b {\n\t\tout = z\n\t} else {\n\t\tout = &polyGF2{}\n\t}\n\n\t// Compute the remainder using synthetic division.\n\tda, db := a.Degree(), b.Degree()\n\tif db < 0 {\n\t\tpanic(\"divide by zero\")\n\t}\n\tout.coeff.Set(&a.coeff)\n\tvar tmp polyGF2\n\tfor i := da - db; i >= 0; i-- {\n\t\tif out.coeff.Bit(i+db) != 0 {\n\t\t\ttmp.MulX(b, i)\n\t\t\tout.Sub(out, &tmp)\n\t\t}\n\t}\n\n\tif z != out {\n\t\tz.coeff.Set(&out.coeff)\n\t}\n\treturn z\n}", "func floorMod(x, y int64) int64 {\n\tm := x % y\n\tif m == 0 || ((x >= 0 && y > 0) || (x < 0 && y < 0)) {\n\t\treturn m\n\t}\n\treturn m + y\n}", "func (i I) Mod(i2 I) I {\n\ti = I{i.X % i2.X, i.Y % i2.Y}\n\tif i.X < 0 {\n\t\ti.X += i2.X\n\t}\n\tif i.Y < 0 {\n\t\ti.Y += i2.Y\n\t}\n\treturn i\n}", "func walkDivMod(n *ir.BinaryExpr, init *ir.Nodes) ir.Node {\n\tn.X = walkExpr(n.X, init)\n\tn.Y = walkExpr(n.Y, init)\n\n\t// rewrite complex div into function call.\n\tet := n.X.Type().Kind()\n\n\tif types.IsComplex[et] && n.Op() == ir.ODIV {\n\t\tt := n.Type()\n\t\tcall := mkcall(\"complex128div\", types.Types[types.TCOMPLEX128], init, typecheck.Conv(n.X, types.Types[types.TCOMPLEX128]), typecheck.Conv(n.Y, types.Types[types.TCOMPLEX128]))\n\t\treturn typecheck.Conv(call, t)\n\t}\n\n\t// Nothing to do for float divisions.\n\tif types.IsFloat[et] {\n\t\treturn n\n\t}\n\n\t// rewrite 64-bit div and mod on 32-bit architectures.\n\t// TODO: Remove this code once we can introduce\n\t// runtime calls late in SSA processing.\n\tif types.RegSize < 8 && (et == types.TINT64 || et == types.TUINT64) {\n\t\tif n.Y.Op() == ir.OLITERAL {\n\t\t\t// Leave div/mod by constant powers of 2 or small 16-bit constants.\n\t\t\t// The SSA backend will handle those.\n\t\t\tswitch et {\n\t\t\tcase types.TINT64:\n\t\t\t\tc := ir.Int64Val(n.Y)\n\t\t\t\tif c < 0 {\n\t\t\t\t\tc = -c\n\t\t\t\t}\n\t\t\t\tif c != 0 && c&(c-1) == 0 {\n\t\t\t\t\treturn n\n\t\t\t\t}\n\t\t\tcase types.TUINT64:\n\t\t\t\tc := ir.Uint64Val(n.Y)\n\t\t\t\tif c < 1<<16 {\n\t\t\t\t\treturn n\n\t\t\t\t}\n\t\t\t\tif c != 0 && c&(c-1) == 0 {\n\t\t\t\t\treturn n\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tvar fn string\n\t\tif et == types.TINT64 {\n\t\t\tfn = \"int64\"\n\t\t} else {\n\t\t\tfn = \"uint64\"\n\t\t}\n\t\tif n.Op() == ir.ODIV {\n\t\t\tfn += \"div\"\n\t\t} else {\n\t\t\tfn += \"mod\"\n\t\t}\n\t\treturn mkcall(fn, n.Type(), init, typecheck.Conv(n.X, types.Types[et]), typecheck.Conv(n.Y, types.Types[et]))\n\t}\n\treturn n\n}", "func (c *Clac) Mod() error {\n\treturn c.applyFloat(2, func(vals []value.Value) (value.Value, error) {\n\t\treturn binary(vals[1], \"mod\", vals[0])\n\t})\n}", "func DivAndRemainder(a, b int) (int, int, error) {\n\tif b == 0 {\n\t\treturn 0, 0, errors.New(\"Cannot divide by zero\")\n\t}\n\td := int(math.Floor(float64(a) / float64(b)))\n\tr := a % b\n\treturn d, r, nil\n\n}", "func NewSubModExpr(scanner parser.Scanner, a, b Expr) Expr {\n\treturn newArithExpr(scanner, a, b, \"-%\", func(a, b float64) float64 {\n\t\treturn a - math.Mod(a, b)\n\t})\n}", "func (fn *formulaFuncs) MOD(argsList *list.List) formulaArg {\n\tif argsList.Len() != 2 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"MOD requires 2 numeric arguments\")\n\t}\n\tnumber := argsList.Front().Value.(formulaArg).ToNumber()\n\tif number.Type == ArgError {\n\t\treturn number\n\t}\n\tdivisor := argsList.Back().Value.(formulaArg).ToNumber()\n\tif divisor.Type == ArgError {\n\t\treturn divisor\n\t}\n\tif divisor.Number == 0 {\n\t\treturn newErrorFormulaArg(formulaErrorDIV, \"MOD divide by zero\")\n\t}\n\ttrunc, rem := math.Modf(number.Number / divisor.Number)\n\tif rem < 0 {\n\t\ttrunc--\n\t}\n\treturn newNumberFormulaArg(number.Number - divisor.Number*trunc)\n}", "func FMod(arg float64, arg2 float64) float64 {\n\tflooRit := Floor(arg / arg2)\n\treturn arg - (flooRit * arg2)\n}", "func main() {\n\tN := scanInt()\n\tn, e, t := 1, 1, 1\n\tfor i := 0; i < N; i++ {\n\t\t// [+/*]you can divide by mod each add/time ops\n\t\te = (e * 8) % mod\n\t\tn = (n * 9) % mod\n\t\tt = (t * 10) % mod\n\t}\n\tans := t - n - n + e\n\t// [+]consider if above time calc result over mod\n\tans %= mod\n\t// [-]consider if above sub calc result goes negative number\n\tans = (ans + mod) % mod\n\tfmt.Println(ans)\n}", "func opUI64Mod(prgrm *CXProgram) {\n\texpr := prgrm.GetExpr()\n\tfp := prgrm.GetFramePointer()\n\n\toutV0 := ReadUI64(fp, expr.Inputs[0]) % ReadUI64(fp, expr.Inputs[1])\n\tWriteUI64(GetFinalOffset(fp, expr.Outputs[0]), outV0)\n}", "func Modp(z *Elt)", "func (z *Int) ModSqrt(x, p *Int) *Int {}", "func (p Point) Mod(r Rectangle) Point {\n\tw, h := r.Dx(), r.Dy()\n\tp = p.Sub(r.Min)\n\tp.X = p.X % w\n\tif p.X < 0 {\n\t\tp.X += w\n\t}\n\tp.Y = p.Y % h\n\tif p.Y < 0 {\n\t\tp.Y += h\n\t}\n\treturn p.Add(r.Min)\n}", "func ModExp(x, y, N int64) int64 {\n\tif y == 0 {\n\t\treturn 1\n\t}\n\tz := ModExp(x, y/2, N)\n\tt := (z * z) % N\n\tif y%2 == 0 {\n\t\treturn t\n\t}\n\treturn (x * t) % N\n}", "func (n null) Mod(v Val) Val {\n\tpanic(ErrInvalidOpModOnNil)\n}", "func opUI8Mod(inputs []ast.CXValue, outputs []ast.CXValue) {\n\toutV0 := inputs[0].Get_ui8() % inputs[1].Get_ui8()\n\toutputs[0].Set_ui8(outV0)\n}", "func IntModInverse(z *big.Int, g, n *big.Int,) *big.Int", "func rcMod(p *TCompiler, code *TCode) (*value.Value, error) {\n\tv := value.Mod(p.regGet(code.B), p.regGet(code.C))\n\tp.regSet(code.A, v)\n\tp.moveNext()\n\treturn v, nil\n}", "func factmod(n, p int) int {\n\tres := 1\n\tfor n > 1 {\n\t\tf := 1\n\t\tif (n/p)%2 == 1 {\n\t\t\tf = p - 1\n\t\t}\n\t\tres = (res * f) % p\n\t\tfor i := 2; i <= n%p; i++ {\n\t\t\tres = (res * i) % p\n\t\t}\n\t\tn /= p\n\t}\n\treturn res % p\n}", "func bigIntModExp(b, e, m *big.Int) *big.Int {\n\t// For simplicity panic here rather than returning an error.\n\t// Akin to how division by zero is handled.\n\n\tzero := big.NewInt(0)\n\tone := big.NewInt(1)\n\ttwo := big.NewInt(2)\n\n\tif m.Cmp(one) == -1 {\n\t\tlog.Panic(\"modExp: m < 1\")\n\t} else if m.Cmp(one) == 0 {\n\t\treturn zero\n\t}\n\n\tif e.Cmp(zero) == -1 {\n\t\t// TODO Add support for this\n\t\tlog.Panic(\"modExp: e < 0\")\n\t} else if e.Cmp(zero) == 0 {\n\t\treturn one\n\t} else if e.Cmp(one) == 0 {\n\t\treturn new(big.Int).Mod(b, m)\n\t}\n\n\tc := big.NewInt(1)\n\tb2 := new(big.Int).Mod(b, m)\n\te2 := new(big.Int).Set(e)\n\n\t// while e2 > 0\n\tfor e2.Cmp(zero) == 1 {\n\n\t\t// if e2 % 2 == 1\n\t\tif new(big.Int).Mod(e2, two).Cmp(one) == 0 {\n\t\t\t// c = (c * b2) % m\n\t\t\tc.Mul(c, b2).Mod(c, m)\n\t\t}\n\n\t\t// e2 >>= 1\n\t\te2.Rsh(e2, 1)\n\n\t\t// b2 = (b2 * b2) % m\n\t\tb2.Mul(b2, b2).Mod(b2, m)\n\t}\n\n\treturn c\n}", "func DivisorComResto(numeroA int, numberoB int) (resultado int, resto int) {\n\tresultado = numeroA / numberoB\n\tresto = numeroA % numberoB\n\treturn\n}", "func Multiply(a, operand int) int { return operand * a }", "func Div(a, b Expr) Expr {\n\treturn &arithmeticOperator{&simpleOperator{a, b, scanner.DIV}}\n}", "func div(a, b int32) int32 {\n\tif a >= 0 {\n\t\treturn (a + (b >> 1)) / b\n\t}\n\treturn -((-a + (b >> 1)) / b)\n}", "func opI64Mod(expr *CXExpression, fp int) {\n\toutB0 := ReadI64(fp, expr.Inputs[0]) % ReadI64(fp, expr.Inputs[1])\n\tWriteI64(GetOffset_i64(fp, expr.Outputs[0]), outB0)\n}", "func modInverse(a, n *big.Int) (ia *big.Int, ok bool) {\n\tg := new(big.Int)\n\tx := new(big.Int)\n\ty := new(big.Int)\n\tg.GCD(x, y, a, n)\n\tif g.Cmp(bigOne) != 0 {\n\t\treturn\n\t}\n\n\tif x.Cmp(bigOne) < 0 {\n\t\tx.Add(x, n)\n\t}\n\n\treturn x, true\n}", "func Modadd(a, b, m *ED25519.BIG) *ED25519.BIG {\n\tc := a.Plus(b)\n\tc.Mod(m)\n\treturn c\n}", "func Modadd(a, b, m *ED25519.BIG) *ED25519.BIG {\n\tc := a.Plus(b)\n\tc.Mod(m)\n\treturn c\n}", "func Modadd(a, b, m *ED25519.BIG) *ED25519.BIG {\n\tc := a.Plus(b)\n\tc.Mod(m)\n\treturn c\n}", "func divShift(s string, n int) string {\n\tn = n % len(s)\n\tif n == 0 {\n\t\treturn s\n\t}\n\tbuf := make([]byte, len(s))\n\tcopy(buf, s)\n\tswap(buf, 0, n, len(buf))\n\treturn string(buf)\n}", "func DivMod(dvdn, dvsr int) (q, r int) {\n\tr = dvdn\n\tfor r >= dvsr {\n\t\tq += 1\n\t\tr = r - dvsr\n\t}\n\treturn\n}", "func DIVL(mr operand.Op) { ctx.DIVL(mr) }", "func NewMod(first, second Expression) Function {\n\trv := &Mod{\n\t\t*NewBinaryFunctionBase(\"mod\", first, second),\n\t}\n\n\trv.expr = rv\n\treturn rv\n}", "func mod2(n []byte, e int) []byte {\n\n\tres := make([]byte, len(n))\n\tcopy(res, n)\n\n\tmod := sub(pow2(e), []byte{0x01})\n\n\tdiff := len(n) - len(mod)\n\tfor i := len(res) - 1; i >= 0; i-- {\n\n\t\tif i-diff < 0 {\n\t\t\tres[i] = 0x00\n\t\t} else {\n\t\t\tres[i] &= mod[i-diff]\n\t\t}\n\n\t}\n\n\treturn res\n}", "func Rem(x, y reflect.Value) reflect.Value {\n\tmustSameType(x, y)\n\tz := reflect.New(x.Type()).Elem()\n\tswitch x.Type().Kind() {\n\tcase reflect.Int:\n\t\txx := int(x.Int())\n\t\tyy := int(y.Int())\n\t\tzz := int64(xx % yy)\n\t\tz.SetInt(zz)\n\t\treturn z\n\tcase reflect.Int8:\n\t\txx := int8(x.Int())\n\t\tyy := int8(y.Int())\n\t\tzz := int64(xx % yy)\n\t\tz.SetInt(zz)\n\t\treturn z\n\tcase reflect.Int16:\n\t\txx := int16(x.Int())\n\t\tyy := int16(y.Int())\n\t\tzz := int64(xx % yy)\n\t\tz.SetInt(zz)\n\t\treturn z\n\tcase reflect.Int32:\n\t\txx := int32(x.Int())\n\t\tyy := int32(y.Int())\n\t\tzz := int64(xx % yy)\n\t\tz.SetInt(zz)\n\t\treturn z\n\tcase reflect.Int64:\n\t\txx := int64(x.Int())\n\t\tyy := int64(y.Int())\n\t\tzz := int64(xx % yy)\n\t\tz.SetInt(zz)\n\t\treturn z\n\tcase reflect.Uint:\n\t\txx := uint(x.Uint())\n\t\tyy := uint(y.Uint())\n\t\tzz := uint64(xx % yy)\n\t\tz.SetUint(zz)\n\t\treturn z\n\tcase reflect.Uint8:\n\t\txx := uint8(x.Uint())\n\t\tyy := uint8(y.Uint())\n\t\tzz := uint64(xx % yy)\n\t\tz.SetUint(zz)\n\t\treturn z\n\tcase reflect.Uint16:\n\t\txx := uint16(x.Uint())\n\t\tyy := uint16(y.Uint())\n\t\tzz := uint64(xx % yy)\n\t\tz.SetUint(zz)\n\t\treturn z\n\tcase reflect.Uint32:\n\t\txx := uint32(x.Uint())\n\t\tyy := uint32(y.Uint())\n\t\tzz := uint64(xx % yy)\n\t\tz.SetUint(zz)\n\t\treturn z\n\tcase reflect.Uint64:\n\t\txx := uint64(x.Uint())\n\t\tyy := uint64(y.Uint())\n\t\tzz := uint64(xx % yy)\n\t\tz.SetUint(zz)\n\t\treturn z\n\tcase reflect.Uintptr:\n\t\txx := uintptr(x.Uint())\n\t\tyy := uintptr(y.Uint())\n\t\tzz := uint64(xx % yy)\n\t\tz.SetUint(zz)\n\t\treturn z\n\t}\n\tpanic(fmt.Sprintf(\"operator % not defined on %v\", x.Type()))\n}", "func opI8Mod(expr *CXExpression, fp int) {\n\toutV0 := ReadI8(fp, expr.Inputs[0]) % ReadI8(fp, expr.Inputs[1])\n\tWriteI8(GetOffset_i8(fp, expr.Outputs[0]), outV0)\n}", "func ParseModulo(ns string) *big.Int {\n\tbuf, _ := Base64UrlDecode(ns)\n\treturn new(big.Int).SetBytes(buf)\n}", "func InvModPanic(a, n *big.Int) *big.Int {\n\ti, err := mtsn.InvMod(a, n)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn i\n}", "func cdiv(a, b int) int { return (a + b - 1) / b }", "func Modsub(a, b, m *ED25519.BIG) *ED25519.BIG {\n\treturn Modadd(a, ED25519.Modneg(b, m), m)\n}", "func Modsub(a, b, m *ED25519.BIG) *ED25519.BIG {\n\treturn Modadd(a, ED25519.Modneg(b, m), m)\n}", "func Modsub(a, b, m *ED25519.BIG) *ED25519.BIG {\n\treturn Modadd(a, ED25519.Modneg(b, m), m)\n}", "func Div(a, b Expr) Expr {\n\treturn &divOp{&simpleOperator{a, b, scanner.DIV}}\n}", "func IntModSqrt(z *big.Int, x, p *big.Int,) *big.Int", "func sumOfPowersMod(base int64, power int64, mod int64) int64 {\n\tif power&1 == 1 {\n\t\t// (base * sumOfPowersMod(base, power-1, mod) + 1) % mod\n\n\t\tsum := sumOfPowersMod(base, power-1, mod) // partial_sum (see below)\n\n\t\tresult := big.NewInt(sum)\n\t\tresult.Mul(result, big.NewInt(base)) // base * partial_sum\n\t\tresult.Add(result, big.NewInt(int64(1))) // base * partial_sum + 1\n\t\tresult.Mod(result, big.NewInt(mod)) // (base * partial_sum + 1) % mod\n\n\t\tif !result.IsInt64() {\n\t\t\tpanic(\"result should be int64!\")\n\t\t}\n\n\t\treturn result.Int64()\n\n\t} else if power > 0 {\n\t\t// ((base + 1) * sumOfPowersMod(base * base % mod, power // 2, mod)) % mod\n\t\tpb := big.NewInt(mod)\n\n\t\tresult := big.NewInt(base)\n\t\tresult.Mul(result, big.NewInt(base)) // base * base\n\t\tresult.Mod(result, pb) // base * base % mod\n\n\t\tif !result.IsInt64() {\n\t\t\tpanic(\"result should be int64!\")\n\t\t}\n\n\t\thalfN := power / 2\n\t\tsum := sumOfPowersMod(result.Int64(), halfN, mod)\n\n\t\tresult.SetInt64(sum) // partial_sum (see below)\n\t\tresult.Mul(result, big.NewInt(base+1)) // (base + 1) * partial_sum\n\t\tresult.Mod(result, pb) // ((base + 1) * partial_sum) % mod\n\n\t\tif !result.IsInt64() {\n\t\t\tpanic(\"result should be int64!\")\n\t\t}\n\n\t\treturn result.Int64()\n\t}\n\n\treturn 0\n}", "func FloorMod(scope *Scope, x tf.Output, y tf.Output) (z tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"FloorMod\",\n\t\tInput: []tf.Input{\n\t\t\tx, y,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func divmod(dvdn, dvsr int) (q, r int) {\n\tr = dvdn\n\tfor r >= dvsr {\n\t\tq++\n\t\tr = r - dvsr\n\t}\n\treturn\n}", "func h(n int) int {\n var a int = n % 2\n if a == 1 {\n return n * 3 + 1\n }\n return n / 2\n}", "func int256ToBigMod(a *[4]uint64, m *big.Int) big.Int {\n\tvar x, y big.Int\n\tfor i := 3; i >= 0; i -- {\n\t\ty.SetUint64(a[i])\n\t\tx.Lsh(&x, 64).Add(&x, &y)\n\t}\n\tfor x.Cmp(m) >= 0 {\n\t\tx.Sub(&x, m)\n\t}\n\treturn x\n}", "func Divisor(numeroA int, numeroB int) (resultado int) {\n\tresultado = numeroA / numeroB\n\treturn\n}", "func DecimalMod(from1, from2, to *MyDecimal) error {\n\tfrom1, from2, to = validateArgs(from1, from2, to)\n\tto.resultFrac = mathutil.Max(from1.resultFrac, from2.resultFrac)\n\treturn doDivMod(from1, from2, nil, to, 0)\n}", "func wrapPlaneExtension(index, limit int) int { return index % limit }" ]
[ "0.7940327", "0.7586372", "0.7558664", "0.74437463", "0.73613614", "0.7257822", "0.72539693", "0.7203363", "0.6978081", "0.6941233", "0.6902041", "0.6896566", "0.6878835", "0.68742746", "0.6847059", "0.68291336", "0.6719125", "0.6716226", "0.6716226", "0.6716226", "0.6611154", "0.65528214", "0.6543972", "0.6542431", "0.6537188", "0.6520334", "0.6476835", "0.64680046", "0.64459103", "0.6379853", "0.6377088", "0.6359456", "0.63259596", "0.6281688", "0.6278108", "0.62479687", "0.6199337", "0.61975193", "0.6173993", "0.6171825", "0.6138359", "0.6128595", "0.61254984", "0.6085342", "0.6085342", "0.60768414", "0.6067463", "0.60238427", "0.59494907", "0.5912762", "0.5891599", "0.5855484", "0.5768079", "0.5764872", "0.57227665", "0.5593183", "0.55800796", "0.55645466", "0.55467707", "0.5491327", "0.5480641", "0.54474574", "0.54370004", "0.54360676", "0.5430951", "0.5379168", "0.53382325", "0.53347653", "0.5324175", "0.5315091", "0.5313183", "0.52843046", "0.5268682", "0.52616835", "0.5235563", "0.5235563", "0.5235563", "0.52057076", "0.5201699", "0.5161742", "0.5152138", "0.5131914", "0.5116611", "0.5100301", "0.50859547", "0.5084663", "0.50804377", "0.50705457", "0.50705457", "0.50705457", "0.5055105", "0.50502545", "0.5044644", "0.504397", "0.50123596", "0.5006854", "0.5000158", "0.49988383", "0.49725866", "0.49639305" ]
0.9219174
0
Multiply operand and a.
func Multiply(a, operand int) int { return operand * a }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *Arith) Multiply(args *Args, reply *int) error {\n *reply = args.A * args.B\n return nil\n}", "func Mul(a, b Expr) Expr {\n\treturn &arithmeticOperator{&simpleOperator{a, b, scanner.MUL}}\n}", "func (t *Arith) Multiply(args *Args, reply *int) error {\n\t*reply = args.A * args.B\n\treturn nil\n}", "func MULPD(mx, x operand.Op) { ctx.MULPD(mx, x) }", "func MULB(mr operand.Op) { ctx.MULB(mr) }", "func MULPS(mx, x operand.Op) { ctx.MULPS(mx, x) }", "func Multiply(a cty.Value, b cty.Value) (cty.Value, error) {\n\treturn MultiplyFunc.Call([]cty.Value{a, b})\n}", "func MULQ(mr operand.Op) { ctx.MULQ(mr) }", "func IMULQ(ops ...operand.Op) { ctx.IMULQ(ops...) }", "func (i *MyInt) MultiplyBy(operand MyInt) {\n\t*i *= operand\n}", "func Add(a, operand int) int { return operand + a }", "func (c *calculon) Multiply(ctx context.Context, arg calculator.Operand) (calculator.Result, error) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\tc.current *= arg.Value\n\treturn calculator.Result{c.current}, nil\n}", "func (n *bigNumber) mul(x *bigNumber, y *bigNumber) *bigNumber {\n\t//it does not work in place, that why the temporary bigNumber is necessary\n\treturn karatsubaMul(n, x, y)\n}", "func Mul(a, b Expr) Expr {\n\treturn &mulOp{&simpleOperator{a, b, scanner.MUL}}\n}", "func Mult( a *context.Value, b *context.Value ) (*context.Value,error) {\n if a != nil && b != nil {\n switch a.OperationType( b ) {\n case context.VAR_BOOL:\n return context.IntValue( a.Int() * b.Int() ), nil\n case context.VAR_INT:\n return context.IntValue( a.Int() * b.Int() ), nil\n case context.VAR_FLOAT:\n return context.FloatValue( a.Float() * b.Float() ), nil\n case context.VAR_COMPLEX:\n return context.ComplexValue( a.Complex() * b.Complex() ), nil\n }\n }\n return nil, errors.New( \"Unsupported type for mult\" )\n}", "func MULXL(mr, r, r1 operand.Op) { ctx.MULXL(mr, r, r1) }", "func ANDPS(mx, x operand.Op) { ctx.ANDPS(mx, x) }", "func mul(x byte, y byte) byte {\n\tif x == 0 || y == 0 {\n\t\treturn 0\n\t}\n\treturn expOp[logOp[x]+logOp[y]]\n}", "func multiply(a, b float64) float64 {\n\treturn a * b\n}", "func Multiply(a, b int, reply *int) {\n\t*reply = a * b\n}", "func Multiply(a, b int, reply *int) {\n\t*reply = a * b\n}", "func (s *SourceControl) Multiply(args *FactorArgs, reply *int) error {\n\t*reply = args.A * args.B\n\treturn nil\n}", "func MULXQ(mr, r, r1 operand.Op) { ctx.MULXQ(mr, r, r1) }", "func PAND(mx, x operand.Op) { ctx.PAND(mx, x) }", "func calcMultiply(rOpd, lOpd formulaArg, opdStack *Stack) error {\n\tlOpdVal := lOpd.ToNumber()\n\tif lOpdVal.Type != ArgNumber {\n\t\treturn errors.New(lOpdVal.Value())\n\t}\n\trOpdVal := rOpd.ToNumber()\n\tif rOpdVal.Type != ArgNumber {\n\t\treturn errors.New(rOpdVal.Value())\n\t}\n\topdStack.Push(newNumberFormulaArg(lOpdVal.Number * rOpdVal.Number))\n\treturn nil\n}", "func ANDPD(mx, x operand.Op) { ctx.ANDPD(mx, x) }", "func Multiply() {\n\tMatch('*')\n\tFactor()\n\tEmitLn(\"MULS (SP)+,D0\")\n}", "func feMul(out *fieldElement, a *fieldElement, b *fieldElement)", "func Multiply(a, b int64) int64 {\n\treturn a * b\n}", "func add(a, b int, result *int) {\n\t*result = a * b\n}", "func (self *State)Pow(a,b any)any{\n self.IncOperations(self.coeff[\"pow\"]+self.off[\"pow\"])\n return wrap2(a,b,math.Pow)\n}", "func mul(a, b big.Int) big.Int {\n\treturn *big.NewInt(1).Mul(&a, &b)\n}", "func Multiply(a, b interface{}) (interface{}, error) {\n\tswitch at := a.(type) {\n\tcase faces.Multiply:\n\t\treturn at.Multiply(b)\n\tcase int:\n\t\treturn multiply.Int(at, b)\n\tcase float64:\n\t\treturn multiply.Float(at, b)\n\tcase faces.Int:\n\t\treturn multiply.Int(at.Int(), b)\n\tcase faces.Float:\n\t\treturn multiply.Float(at.Float(), b)\n\t}\n\n\treturn nil, multiply.Cant(a, b)\n}", "func Multiply(a int, b int) int {\n\treturn a * b\n}", "func Multiply(a uint32) uint32 {\n\treturn a * 2\n}", "func (ai *Arith) Mul(decimal1 *ZnDecimal, others ...*ZnDecimal) *ZnDecimal {\n\t// init result from decimal1\n\tvar result = copyZnDecimal(decimal1)\n\tif len(others) == 0 {\n\t\treturn result\n\t}\n\n\tfor _, item := range others {\n\t\tresult.co.Mul(result.co, item.co)\n\t\tresult.exp = result.exp + item.exp\n\t}\n\n\treturn result\n}", "func Multiply(a, b int) int {\n\treturn a * b\n}", "func Multiply(a, b int) int {\n\treturn a * b\n}", "func Multiply(a, b int) int {\n\treturn a * b\n}", "func Multiply(a, b float64) float64 {\n\treturn a * b\n}", "func Multiply(a, b float64) float64 {\n\treturn a * b\n}", "func mul(m ast.Mul, isparam map[ast.Variable]bool) (Operation, error) {\n\t// Expect the second operand to always be a variable.\n\tif _, ok := m.Y.(ast.Variable); !ok {\n\t\treturn nil, errutil.AssertionFailure(\"expect second multiply operand to be variable\")\n\t}\n\n\t// Check for a const multiply.\n\tif c, ok := m.X.(ast.Constant); ok {\n\t\treturn ConstMul(c), nil\n\t}\n\n\t// Check for parameter multiply.\n\tif v, ok := m.X.(ast.Variable); ok && isparam[v] {\n\t\treturn ParamMul(v), nil\n\t}\n\n\tif v, ok := m.Y.(ast.Variable); ok && isparam[v] {\n\t\treturn ParamMul(v), nil\n\t}\n\n\t// Fallback to a generic multiply.\n\treturn Mul{}, nil\n}", "func ANDNPS(mx, x operand.Op) { ctx.ANDNPS(mx, x) }", "func Multiply(a float64, b float64) float64 {\n\treturn a * b\n}", "func (vm *VM) opMult(instr []uint16) int {\n\ta, b, c := vm.getAbc(instr)\n\n\tvm.registers[a] = (b * c) % 32768\n\treturn 4\n}", "func mathOp(a, b int, operator string, doOperation func(int, int, string) int) int {\n\tresult := doOperation(a, b, operator)\n\treturn result\n}", "func MULW(mr operand.Op) { ctx.MULW(mr) }", "func MOVAPD(mx, mx1 operand.Op) { ctx.MOVAPD(mx, mx1) }", "func SimpleMultiply(exp string) string {\n var index int = strings.Index(exp, \"*\")\n var arg1 string = exp[:index]\n var arg2 string = exp[index + 1:]\n val1 := NotateToDouble(arg1)\n val2 := NotateToDouble(arg2)\n return NegativeNotate(fmt.Sprintf(\"%f\", val1 * val2))\n}", "func multi(x, y int) (answer int, err error) {\n\tanswer = x * y\n\treturn\n}", "func ValueMul(a, b reflect.Value) (reflect.Value, error) {\n\taBkind := GetBaseKind(a)\n\tbBkind := GetBaseKind(b)\n\n\tswitch aBkind {\n\tcase reflect.Int64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(a.Int() * b.Int()), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Int() * int64(b.Uint())), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(float64(a.Int()) * b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do multiplication math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tcase reflect.Uint64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(int64(a.Uint()) * b.Int()), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Uint() * b.Uint()), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(float64(a.Uint()) * b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do multiplication math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tcase reflect.Float64:\n\t\tswitch bBkind {\n\t\tcase reflect.Int64:\n\t\t\treturn reflect.ValueOf(a.Float() * float64(b.Int())), nil\n\t\tcase reflect.Uint64:\n\t\t\treturn reflect.ValueOf(a.Float() * float64(b.Uint())), nil\n\t\tcase reflect.Float64:\n\t\t\treturn reflect.ValueOf(a.Float() * b.Float()), nil\n\t\tdefault:\n\t\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do multiplication math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t\t}\n\tdefault:\n\t\treturn reflect.ValueOf(nil), errors.Errorf(\"Can not do multiplication math operator between %s and %s\", a.Kind().String(), b.Kind().String())\n\t}\n}", "func ADDPS(mx, x operand.Op) { ctx.ADDPS(mx, x) }", "func PANDN(mx, x operand.Op) { ctx.PANDN(mx, x) }", "func (this *BigInteger) MultiplyTo(a *BigInteger, r *BigInteger) {\n\tx := this.Abs()\n\ty := a.Abs()\n\ti := x.T\n\tr.T = i + y.T\n\tfor i--; i >= 0; i-- {\n\t\tr.V[i] = 0\n\t}\n\tfor i = 0; i < y.T; i++ {\n\t\tr.V[i+x.T] = x.AM(0, y.V[i], r, i, 0, x.T)\n\t}\n\tr.S = 0\n\tr.clamp()\n\tif this.S != a.S {\n\t\tZERO.SubTo(r, r)\n\t}\n}", "func (cal *Calculate) mul(value float64) (result float64) {\n\tif len(cal.Arg) == 2 {\n\t\treturn (cal.Arg[0] * cal.Arg[1])\n\t} else if len(cal.Arg) == 1 {\n\t\treturn (value * cal.Arg[0])\n\t}\n\n\tlog.Fatalln(\"Please check the data format of the calculation unit\")\n\treturn\n}", "func IMULW(ops ...operand.Op) { ctx.IMULW(ops...) }", "func mul(x, y int) int {\n\treturn x * y\n}", "func (self *State)Exp(a any)any{\n self.IncOperations(self.coeff[\"exp\"]+self.off[\"exp\"])\n return wrap1(a,math.Exp)\n}", "func multiply(x, y int) int {\n\treturn x * y\n}", "func (z *Int) Mul(x, y *Int) *Int {}", "func MultiplyAddOnly(a, b int) int {\n\tif a < b {\n\t\tMultiplyAddOnly(b, a)\n\t}\n\n\tresult := 0\n\tfor x := 0; x < Abs(b); x++ {\n\t\tresult += a\n\t}\n\n\tif b < 0 {\n\t\tresult = Min(result)\n\t}\n\treturn result\n}", "func Mod(a, b Expr) Expr {\n\treturn &arithmeticOperator{&simpleOperator{a, b, scanner.MOD}}\n}", "func ADDPD(mx, x operand.Op) { ctx.ADDPD(mx, x) }", "func (a *Mtx) Mult(b *Mtx) *Mtx {\n\tm := Mtx{}\n\tfor i := 0; i < 4; i++ {\n\t\tfor j := 0; j < 4; j++ {\n\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\tm.el[j][i] += a.el[k][i] * b.el[j][k]\n\t\t\t}\n\t\t}\n\t}\n\treturn &m\n}", "func MOVAPS(mx, mx1 operand.Op) { ctx.MOVAPS(mx, mx1) }", "func add(x int, y int) (p,q int) {\n a := x + y\n b := 34 * 34\n return a,b\n}", "func perform_arthmitic(result *int, operation string, left_value int, right_value int) {\n if operation == \"+\" {\n *result = left_value + right_value\n\n } else if operation == \"-\" {\n *result = left_value - right_value\n\n } else if operation == \"*\" {\n *result = left_value * right_value\n\n } else if operation == \".\" {\n *result = left_value / right_value\n }\n}", "func (b ValExprBuilder) Mul(expr interface{}) ValExprBuilder {\n\treturn b.makeBinaryExpr('*', expr)\n}", "func (expr Expression) Multiply() (float64, error) {\n\treturn (expr.FirstOperand.Value * expr.SecondOperand.Value), nil\n}", "func JA(r operand.Op) { ctx.JA(r) }", "func ANDW(imr, amr operand.Op) { ctx.ANDW(imr, amr) }", "func PHADDSW(mx, x operand.Op) { ctx.PHADDSW(mx, x) }", "func VPAND(mxy, xy, xy1 operand.Op) { ctx.VPAND(mxy, xy, xy1) }", "func ANDNPD(mx, x operand.Op) { ctx.ANDNPD(mx, x) }", "func newMulExpr(lhs, rhs Expr) Expr {\n\t// If constant is on right side, swap to left side.\n\tif IsConstantExpr(rhs) && !IsConstantExpr(lhs) {\n\t\tlhs, rhs = rhs, lhs\n\t}\n\n\t// Compute constant if both sides are constant.\n\tif lhs, ok := lhs.(*ConstantExpr); ok {\n\t\tif rhs, ok := rhs.(*ConstantExpr); ok {\n\t\t\treturn lhs.Mul(rhs)\n\t\t}\n\t}\n\n\t// Refactor to XOR for boolean expressions.\n\tif ExprWidth(lhs) == WidthBool {\n\t\treturn NewBinaryExpr(AND, lhs, rhs)\n\t}\n\n\t// Optimize for multiplication with a constant 1 or 0.\n\tif lhs, ok := lhs.(*ConstantExpr); ok {\n\t\tif lhs.Value == 1 {\n\t\t\treturn rhs\n\t\t} else if lhs.Value == 0 {\n\t\t\treturn lhs\n\t\t}\n\t}\n\treturn &BinaryExpr{Op: MUL, LHS: lhs, RHS: rhs}\n}", "func PADDSW(mx, x operand.Op) { ctx.PADDSW(mx, x) }", "func MULSD(mx, x operand.Op) { ctx.MULSD(mx, x) }", "func IMULB(mr operand.Op) { ctx.IMULB(mr) }", "func multiply(x, y int) (r1, r2 int) {\n\tr1 = x * 10\n\tr2 = y * 20\n\treturn r1, r2\n}", "func (n *bigNumber) mulCopy(x *bigNumber, y *bigNumber) *bigNumber {\n\t//it does not work in place, that why the temporary bigNumber is necessary\n\treturn n.set(new(bigNumber).mul(x, y))\n}", "func (f Fraction) Multiply(multiplicand Fraction) (product Fraction) {\n\ttmp := f.normalize()\n\tproduct = Fraction{\n\t\ttmp.numerator * multiplicand.numerator, tmp.denominator * multiplicand.denominator,\n\t}\n\treturn product.normalize()\n}", "func main() {\n // a := 44\n \n var a2 int = 10\n\n a2 = a2 * 10\n fmt.Println(a2) \n}", "func MULSS(mx, x operand.Op) { ctx.MULSS(mx, x) }", "func (s *Scalar) Multiply(x, y *Scalar) *Scalar {\n\ts.s.Mul(&x.s, &y.s)\n\treturn s\n}", "func PADDW(mx, x operand.Op) { ctx.PADDW(mx, x) }", "func Multiply(a float64, b float64) float64 {\n return a*b;\n}", "func KaratsubaMult(a, b int) int {\n\tvar useNeg bool\n\tif (a < 0 && b > 0) || (a > 0 && b < 0) {\n\t\tuseNeg = true\n\t}\n\ta, b = absInt(a), absInt(b)\n\n\tm := getMValue(a, b)\n\n\tx1, x0 := separateInts(a, m)\n\ty1, y0 := separateInts(b, m)\n\n\tz2, z0 := x1*y1, x0*y0\n\tz1 := (x1+x0)*(y1+y0) - z0 - z2\n\n\tresult := (z2 * int(math.Pow10(m*2))) + (z1 * int(math.Pow10(m))) + z0\n\tif useNeg {\n\t\tresult = result * -1\n\t}\n\n\treturn result\n}", "func _a(args ...interface{}) *ast.AssignStmt {\n\tlhs := []ast.Expr(nil)\n\ttok := token.ILLEGAL\n\trhs := []ast.Expr(nil)\n\n\tsetTok := func(t token.Token) {\n\t\tif tok != token.ILLEGAL {\n\t\t\tpanic(\"too many assignment operators\")\n\t\t}\n\t\ttok = t\n\t}\n\n\tfor _, arg := range args {\n\t\tif s, ok := arg.(string); ok {\n\t\t\tswitch s {\n\t\t\tcase \"=\", \":=\", \"+=\", \"-=\", \"*=\", \"/=\", \"%=\",\n\t\t\t\t\"&=\", \"|=\", \"^=\", \"<<=\", \">>=\", \"&^=\":\n\t\t\t\tsetTok(_aop(s))\n\t\t\t\tcontinue\n\t\t\tdefault:\n\t\t\t\targ = _x(s)\n\t\t\t}\n\t\t}\n\t\t// append to lhs or rhs depending on tok.\n\t\tif tok == token.ILLEGAL {\n\t\t\tlhs = append(lhs, arg.(ast.Expr))\n\t\t} else {\n\t\t\trhs = append(rhs, arg.(ast.Expr))\n\t\t}\n\t}\n\n\treturn &ast.AssignStmt{\n\t\tLhs: lhs,\n\t\tTok: tok,\n\t\tRhs: rhs,\n\t}\n}", "func (z *Rat) Mul(x, y *Rat) *Rat {}", "func TestMultiply(t *testing.T) {\n\tassert.Equal(t, multiply(0,0), 0);\n\tassert.Equal(t, multiply(2,0), 0);\n\tassert.Equal(t, multiply(2,3), 6);\n\tassert.Equal(t, multiply(5,-2), -10);\n}", "func (c *Calculator) Mul() {\n\tif opValue, err := c.getOperationValue(); err != nil {\n\t\tc.returnError()\n\t} else {\n\t\tlog.Printf(\"%f * %f = \", value, opValue)\n\t\tvalue *= opValue\n\t\tlog.Printf(\"%f\\n\", value)\n\t\tc.returnResult()\n\t}\n}", "func MultiplyElementwise(a, b NumberArray) (resultingMatrix NumberArray, err error) {\n\treturn binaryOperation(\"MultiplyElementwise\", a, b)\n}", "func PHADDW(mx, x operand.Op) { ctx.PHADDW(mx, x) }", "func (x Rational) Multiply(y Rational) Rational {\n\treturn NewRational(x.numerator*y.numerator, x.denominator*y.denominator)\n}", "func recMultiply(a, b uint) uint {\n\tif b <= a {\n\t\treturn doRecMultiply(a, b)\n\t}\n\treturn doRecMultiply(b, a)\n}", "func PADDQ(mx, x operand.Op) { ctx.PADDQ(mx, x) }", "func (gf *GaloisField) Multiply(a, b int) int {\n\tif a == 0 || b == 0 {\n\t\treturn 0\n\t}\n\treturn gf.ALogTbl[(gf.LogTbl[a]+gf.LogTbl[b])%(gf.Size-1)]\n}", "func Mul(t1 TermT, t2 TermT) TermT {\n\treturn TermT(C.yices_mul(C.term_t(t1), C.term_t(t2)))\n}", "func (z *Int) And(x, y *Int) *Int {}", "func mulr(a, b, c int, r register) register {\n\tr[c] = r[a] * r[b]\n\treturn r\n}" ]
[ "0.64794683", "0.6352899", "0.63007265", "0.6270386", "0.62166786", "0.62012625", "0.6192971", "0.61921877", "0.61724627", "0.6120581", "0.61101425", "0.6107033", "0.60812974", "0.60760665", "0.605561", "0.60107994", "0.5994247", "0.59858465", "0.5966094", "0.59659034", "0.59659034", "0.5965105", "0.59143835", "0.5912405", "0.5892514", "0.58866626", "0.58852", "0.5878717", "0.58446854", "0.5834896", "0.58248633", "0.57982755", "0.57934", "0.5776499", "0.57709855", "0.5757906", "0.57413346", "0.57413346", "0.57413346", "0.5729531", "0.5729531", "0.5726006", "0.572244", "0.5722373", "0.57179177", "0.5713152", "0.5697197", "0.569578", "0.5684562", "0.56772554", "0.5669", "0.5668048", "0.5642373", "0.56419575", "0.564138", "0.5617099", "0.5611323", "0.5609107", "0.5604905", "0.5600799", "0.55955064", "0.55920863", "0.55841506", "0.55826116", "0.5582464", "0.5576209", "0.55706424", "0.5568807", "0.55584973", "0.5553029", "0.5549702", "0.55374116", "0.55267763", "0.55235124", "0.551418", "0.5510604", "0.5502433", "0.54999924", "0.5485678", "0.54838616", "0.5475938", "0.545287", "0.5448652", "0.5448007", "0.5447703", "0.54466283", "0.54387593", "0.5432889", "0.5429119", "0.54211247", "0.5420534", "0.5419141", "0.54188794", "0.541732", "0.54142976", "0.5413547", "0.54010254", "0.5385961", "0.53794277", "0.53787553" ]
0.76100755
0
Min returns the minimum of a and operand.
func Min(a, operand int) int { if a < operand { return a } return operand }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Min(a, b int) int {\n\treturn neogointernal.Opcode2(\"MIN\", a, b).(int)\n}", "func (self *State)Min(a,b any)any{\n self.IncOperations(self.coeff[\"min\"]+self.off[\"min\"])\n return wrap2(a,b,math.Min)\n}", "func Min(a, b float64) float64 {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b float64) float64 {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b float64) float64 {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a int, b int) (res int) {\n\tif a < b {\n\t\tres = a\n\t} else {\n\t\tres = b\n\t}\n\n\treturn\n}", "func Min(a int, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a int, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif b < a {\n\t\treturn b\n\t}\n\treturn a\n}", "func Min(a int, b int) int {\n\tif (a > b) {\n\t\treturn b\n\t}\n\n\treturn a\n}", "func Min(a int, b int) int {\n\tif a <= b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a > b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Min(a, b int) int {\n\tif a > b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Min(a, b int) int {\n\tif a > b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Min(a, b int) int {\r\n\tif a < b {\r\n\t\treturn a\r\n\t}\r\n\treturn b\r\n}", "func min(a int, b int) (int) {\n if a < b { return a }\n return b\n}", "func min(a, b int) (int) {\n\tif a < b {\n\t\treturn a\n\t}\n\n\treturn b\n}", "func Min(a, b uint32) uint32 {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(a, b int) int {\n\tif a-b < 0 {\n\t\treturn a\n\t}\n\n\treturn b\n}", "func Min(a int, b int) int {\n if (b < a) {\n return b;\n }\n\n return a;\n}", "func MIN(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(arg, arg2 float64) float64 {\n\treturn math.Min(arg, arg2)\n}", "func min(a int, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\n\treturn b\n}", "func min(a, b float64) float64 {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(x, y float64) float64 {\n\n\treturn math.Min(x, y)\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(x, min int) int { return x }", "func min(a, b int) int {\n\tif b < a {\n\t\treturn b\n\t}\n\treturn a\n}", "func min(a, b int) int {\n\tif b < a {\n\t\treturn b\n\t}\n\treturn a\n}", "func (g *Graph) Min(x1 Node, x2 Node) Node {\n\treturn g.NewOperator(fn.NewMin(x1, x2), x1, x2)\n}", "func Min(a interface{}, b interface{}) interface{} {\n\tif a == nil {\n\t\treturn b\n\t}\n\tif b == nil {\n\t\treturn a\n\t}\n\tif Cmp(a, b) > 0 {\n\t\treturn b\n\t}\n\treturn a\n}", "func min(a, b int) int {\r\n\tif a < b {\r\n\t\treturn a\r\n\t}\r\n\treturn b\r\n}", "func min(a, b int) int {\r\n\tif a < b {\r\n\t\treturn a\r\n\t}\r\n\treturn b\r\n}", "func min(a, b int32) int32 {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b int) int {\n\tif a > b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Min(numbers ...cty.Value) (cty.Value, error) {\n\treturn MinFunc.Call(numbers)\n}", "func Min[T comdef.XintOrFloat](x, y T) T {\n\tif x < y {\n\t\treturn x\n\t}\n\treturn y\n}", "func min(one int, two int) (rtn int) {\n if one < two {\n rtn = one\n } else {\n rtn = two\n }\n\n return\n}", "func min(a, b int32) int32 {\n\tif a <= b {\n\t\treturn a\n\t}\n\treturn b\n}", "func min(a, b uint32) uint32 {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func (b ValExprBuilder) Min() ValExprBuilder {\n\treturn b.makeFunc(\"MIN\", false)\n}", "func Min(x int, y int) int {\n\tif x < y { return x }\n\treturn y\n}", "func Min(x, y int) int {\n\tif x < y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Min(x, y int) int {\n\tif x < y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Min(x, y int) int {\n\tif x < y {\n\t\treturn x\n\t}\n\treturn y\n}", "func ElemMin(a, b Res) Res {\n\taMask := a.Output().Copy()\n\taMask.Sub(b.Output())\n\tanyvec.LessThan(aMask, aMask.Creator().MakeNumeric(0))\n\tbMask := aMask.Copy()\n\tanyvec.Complement(bMask)\n\treturn Add(Mul(a, NewConst(aMask)), Mul(b, NewConst(bMask)))\n}", "func min(a, b uint64) uint64 {\n\tif a <= b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Min(x, y int64) int64 {\n if x < y {\n return x\n }\n return y\n}", "func mini(x int, y int) int {\n // Return the minimum of two integers.\n if x <= y {\n return x\n } else {\n return y\n }\n}", "func (fn *formulaFuncs) min(mina bool, argsList *list.List) formulaArg {\n\tmin := math.MaxFloat64\n\tfor token := argsList.Front(); token != nil; token = token.Next() {\n\t\targ := token.Value.(formulaArg)\n\t\tswitch arg.Type {\n\t\tcase ArgString:\n\t\t\tif !mina && (arg.Value() == \"TRUE\" || arg.Value() == \"FALSE\") {\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tnum := arg.ToBool()\n\t\t\t\tif num.Type == ArgNumber && num.Number < min {\n\t\t\t\t\tmin = num.Number\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\tnum := arg.ToNumber()\n\t\t\tif num.Type != ArgError && num.Number < min {\n\t\t\t\tmin = num.Number\n\t\t\t}\n\t\tcase ArgNumber:\n\t\t\tif arg.Number < min {\n\t\t\t\tmin = arg.Number\n\t\t\t}\n\t\tcase ArgList, ArgMatrix:\n\t\t\tmin = calcListMatrixMin(mina, min, arg)\n\t\tcase ArgError:\n\t\t\treturn arg\n\t\t}\n\t}\n\tif min == math.MaxFloat64 {\n\t\tmin = 0\n\t}\n\treturn newNumberFormulaArg(min)\n}", "func min(x, y int) int {\n\tif x < y {\n\t\treturn x\n\t} else {\n\t\treturn y\n\t}\n}", "func Min(valueA gcv.Value, valueB gcv.Value) (gcv.Value, error) {\n\tif valueA.Type() == gcv.Complex || valueB.Type() == gcv.Complex {\n\t\treturn nil, errors.New(\"Min is not supported for Complex numbers\")\n\t}\n\treturn gcv.MakeValue(math.Min(valueA.Real(), valueB.Real())), nil\n}", "func (fn *formulaFuncs) MIN(argsList *list.List) formulaArg {\n\tif argsList.Len() == 0 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"MIN requires at least 1 argument\")\n\t}\n\treturn fn.min(false, argsList)\n}", "func min(x, y int64) int64 {\n\tif x > y {\n\t\treturn y\n\t}\n\treturn x\n}", "func minInt(a, b int) int {\n if a < b {\n return a\n }\n return b\n}", "func Min(x, y int32) int32 {\n\treturn y + (((x - y) >> 31) & (x - y))\n}", "func Min(x, y int64) int64 {\n\treturn y + ((x - y) & ((x - y) >> 63))\n}", "func min(x, y int) int {\n\tif x < y {\n\t\treturn x\n\t}\n\treturn y\n}", "func min(x, y int) int {\n\tif x < y {\n\t\treturn x\n\t}\n\treturn y\n}", "func min(x, y int) int {\n\tif x < y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Min(a ...int) (min int) {\n\tfor i, ai := range a {\n\t\tif i == 0 || ai < min {\n\t\t\tmin = ai\n\t\t}\n\t}\n\treturn min\n}", "func Min[\n\tValueT typecons.Ordered,\n](refValue ValueT) OrderedConstraint[ValueT] {\n\treturn Func(\n\t\tfmt.Sprintf(\"min %v\", refValue),\n\t\tGreaterThanOrEqualTo(refValue).IsValid)\n}", "func Min(first Decimal, rest ...Decimal) Decimal {\n\tans := first\n\tfor _, item := range rest {\n\t\tif item.Cmp(ans) < 0 {\n\t\t\tans = item\n\t\t}\n\t}\n\treturn ans\n}", "func Min(first Decimal, rest ...Decimal) Decimal {\n\tans := first\n\tfor _, item := range rest {\n\t\tif item.Cmp(ans) < 0 {\n\t\t\tans = item\n\t\t}\n\t}\n\treturn ans\n}", "func Minimum(scope *Scope, x tf.Output, y tf.Output) (z tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Minimum\",\n\t\tInput: []tf.Input{\n\t\t\tx, y,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func Min(x, y *big.Int) *big.Int {\n\tif x.Cmp(y) == -1 {\n\t\treturn x\n\t}\n\treturn y\n}", "func MinInt(a, b int64) int64 {\n\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}", "func MINPD(mx, x operand.Op) { ctx.MINPD(mx, x) }", "func min(x, y int) int {\n\tif x > y {\n\t\treturn y\n\t}\n\n\treturn x\n}", "func smallest(a, b, fa, fb int) int {\n\tif fa*fa < fb*fb {\n\t\treturn a\n\t} else {\n\t\treturn b\n\t}\n}", "func Min(field string) AggregateFunc {\n\treturn func(start, end string) (string, *dsl.Traversal) {\n\t\tif end == \"\" {\n\t\t\tend = DefaultMinLabel\n\t\t}\n\t\treturn end, __.As(start).Unfold().Values(field).Min().As(end)\n\t}\n}", "func min(x, y float64) float64 {\n\tswitch {\n\tcase math.IsNaN(x) || math.IsNaN(y):\n\t\treturn math.NaN()\n\tcase math.IsInf(x, -1) || math.IsInf(y, -1):\n\t\treturn math.Inf(-1)\n\tcase x == 0 && x == y:\n\t\tif math.Signbit(x) {\n\t\t\treturn x\n\t\t}\n\t\treturn y\n\t}\n\tif x < y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Min(min interface{}) ThresholdRule {\n\treturn ThresholdRule{\n\t\tthreshold: min,\n\t\toperator: greaterEqualThan,\n\t\terr: ErrMinGreaterEqualThanRequired,\n\t}\n\n}", "func (_TellorMesosphere *TellorMesosphereCallerSession) Min(_a *big.Int, _b *big.Int) (*big.Int, error) {\n\treturn _TellorMesosphere.Contract.Min(&_TellorMesosphere.CallOpts, _a, _b)\n}", "func min(x, y int) int {\n\tif x > y {\n\t\treturn y\n\t}\n\treturn x\n}", "func (r *Retry) min(a, b uint64) uint64 {\n\tif a > b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Min(val, min any) bool { return valueCompare(val, min, \">=\") }", "func (c *Clac) Min() error {\n\treturn c.applyFloat(2, func(vals []value.Value) (value.Value, error) {\n\t\treturn binary(vals[1], \"min\", vals[0])\n\t})\n}", "func (_TellorMesosphere *TellorMesosphereSession) Min(_a *big.Int, _b *big.Int) (*big.Int, error) {\n\treturn _TellorMesosphere.Contract.Min(&_TellorMesosphere.CallOpts, _a, _b)\n}", "func MinInt(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\n\treturn b\n}", "func MinInt(a, b int) int {\n\tmin := b\n\tif a < b {\n\t\tmin = a\n\t}\n\treturn min\n}", "func MinInt(a, b int) int {\n\tif a < b {\n\t\treturn a\n\t}\n\treturn b\n}" ]
[ "0.7898636", "0.75427616", "0.7273848", "0.7273848", "0.7273848", "0.7243717", "0.7240031", "0.7240031", "0.71972597", "0.71972597", "0.71972597", "0.71972597", "0.71972597", "0.71972597", "0.71972597", "0.71962607", "0.71934444", "0.7160514", "0.7096722", "0.7096722", "0.7096722", "0.7091117", "0.7086176", "0.70619977", "0.7040246", "0.7038739", "0.7014818", "0.70113075", "0.7006508", "0.70019305", "0.69563746", "0.69321537", "0.69137484", "0.6909084", "0.6909084", "0.6909084", "0.6909084", "0.6909084", "0.6909084", "0.6909084", "0.6909084", "0.6909084", "0.6909084", "0.6909084", "0.6902271", "0.6902113", "0.6902113", "0.6872711", "0.6860385", "0.6847696", "0.6847696", "0.68158084", "0.6785197", "0.6749928", "0.6740962", "0.67077243", "0.670697", "0.6700042", "0.66939396", "0.6685863", "0.66499925", "0.66499925", "0.66499925", "0.6646284", "0.6623924", "0.65791714", "0.65451133", "0.65229", "0.65094763", "0.6475299", "0.6473478", "0.6470654", "0.6466778", "0.64595443", "0.645457", "0.6451043", "0.6451043", "0.6451043", "0.6445966", "0.6444868", "0.64424884", "0.64424884", "0.6437537", "0.643585", "0.6429203", "0.6423891", "0.64139616", "0.6410996", "0.6401231", "0.639825", "0.6385528", "0.6377375", "0.6376195", "0.6347387", "0.63416225", "0.63285697", "0.6324575", "0.6319446", "0.6297565", "0.62923646" ]
0.79488987
0
Max returns the maximum of a and operand.
func Max(a, operand int) int { if a > operand { return a } return operand }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Max(a, b int) int {\n\treturn neogointernal.Opcode2(\"MAX\", a, b).(int)\n}", "func (self *State)Max(a,b any)any{\n self.IncOperations(self.coeff[\"max\"]+self.off[\"max\"])\n return wrap2(a,b,math.Max)\n}", "func Max(a uint64, b uint64) uint64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a int, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a interface{}, b interface{}) interface{} {\n\tif a == nil {\n\t\treturn b\n\t}\n\tif b == nil {\n\t\treturn a\n\t}\n\tif Cmp(a, b) > 0 {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a-b > 0 {\n\t\treturn a\n\t}\n\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a int, b int) int {\n\tif a >= b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\treturn int(math.Max(float64(a), float64(b)))\n}", "func Max(a, b int) int {\r\n\tif a > b {\r\n\t\treturn a\r\n\t}\r\n\treturn b\r\n}", "func Max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Max(a, b uint32) uint32 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int64) int64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int64) int64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a int, b int) int {\n if (b > a) {\n return b;\n }\n\n return a;\n}", "func max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(x, y int64) int64 {\n\treturn x ^ ((x ^ y) & ((x - y) >> 63))\n}", "func max(a int, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func (fn *formulaFuncs) max(maxa bool, argsList *list.List) formulaArg {\n\tmax := -math.MaxFloat64\n\tfor token := argsList.Front(); token != nil; token = token.Next() {\n\t\targ := token.Value.(formulaArg)\n\t\tswitch arg.Type {\n\t\tcase ArgString:\n\t\t\tif !maxa && (arg.Value() == \"TRUE\" || arg.Value() == \"FALSE\") {\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tnum := arg.ToBool()\n\t\t\t\tif num.Type == ArgNumber && num.Number > max {\n\t\t\t\t\tmax = num.Number\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\tnum := arg.ToNumber()\n\t\t\tif num.Type != ArgError && num.Number > max {\n\t\t\t\tmax = num.Number\n\t\t\t}\n\t\tcase ArgNumber:\n\t\t\tif arg.Number > max {\n\t\t\t\tmax = arg.Number\n\t\t\t}\n\t\tcase ArgList, ArgMatrix:\n\t\t\tmax = calcListMatrixMax(maxa, max, arg)\n\t\tcase ArgError:\n\t\t\treturn arg\n\t\t}\n\t}\n\tif max == -math.MaxFloat64 {\n\t\tmax = 0\n\t}\n\treturn newNumberFormulaArg(max)\n}", "func Max(x, y int64) int64 {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Max(arg, arg2 float64) float64 {\n\treturn math.Max(arg, arg2)\n}", "func max(a, b int) int {\n\tif b > a {\n\t\treturn b\n\t}\n\treturn a\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(val, max any) bool { return valueCompare(val, max, \"<=\") }", "func max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Max(x, y float64) float64 {\n\n\treturn math.Max(x, y)\n}", "func Max(x, y *big.Int) *big.Int {\n\tif x.Cmp(y) == 1 {\n\t\treturn x\n\t}\n\treturn y\n}", "func Max[T constraints.Ordered](x T, y T) T {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(a, b int32) int32 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(valueA gcv.Value, valueB gcv.Value) (gcv.Value, error) {\n\tif valueA.Type() == gcv.Complex || valueB.Type() == gcv.Complex {\n\t\treturn nil, errors.New(\"Max is not supported for Complex numbers\")\n\t}\n\treturn gcv.MakeValue(math.Max(valueA.Real(), valueB.Real())), nil\n}", "func Max(x, y int64) int64 {\n if x > y {\n return x\n }\n return y\n}", "func (b ValExprBuilder) Max() ValExprBuilder {\n\treturn b.makeFunc(\"MAX\", false)\n}", "func Max(max interface{}) ThresholdRule {\n\treturn ThresholdRule{\n\t\tthreshold: max,\n\t\toperator: lessEqualThan,\n\t\terr: ErrMaxLessEqualThanRequired,\n\t}\n}", "func max(a, b int32) int32 {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}", "func max(a, b int) int {\nif a < b {\nreturn b\n}\nreturn a\n}", "func Max[T comdef.XintOrFloat](x, y T) T {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(a, b ImpactAmount) ImpactAmount {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(x int, y int) int {\n\tif x > y { return x }\n\treturn y\n}", "func (g *Graph) Max(x1 Node, x2 Node) Node {\n\treturn g.NewOperator(fn.NewMax(x1, x2), x1, x2)\n}", "func Max[\n\tValueT typecons.Ordered,\n](refValue ValueT) OrderedConstraint[ValueT] {\n\treturn Func(\n\t\tfmt.Sprintf(\"max %v\", refValue),\n\t\tLessThanOrEqualTo(refValue).IsValid)\n}", "func max(m, n int) (int, bool) {\n\tif m > n {\n\t\treturn m, true\n\t}\n\treturn n, false\n}", "func max(a, b int32) int32 {\n\tif a >= b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func (m *Arg) GetMax() (val float64, set bool) {\n\tif m.Max == nil {\n\t\treturn\n\t}\n\n\treturn *m.Max, true\n}", "func Max(x, y int) int {\n\tif y > x {\n\t\treturn y\n\t}\n\treturn x\n}", "func MAX(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(numbers ...cty.Value) (cty.Value, error) {\n\treturn MaxFunc.Call(numbers)\n}", "func Max(x, y int) int {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}", "func Max(x, y string) string {\n\tif Compare(x, y) < 0 {\n\t\treturn y\n\t}\n\treturn x\n}", "func max(x, y int64) int64 {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}", "func Max(x, y int32) int32 {\n\treturn x - (((x - y) >> 31) & (x - y))\n}", "func maxI(a int, b int) (res int) {\n\tif a < b {\n\t\tres = b\n\t} else {\n\t\tres = a\n\t}\n\n\treturn\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t} else {\n\t\treturn y\n\t}\n}", "func Max(a ...int) (max int) {\n\tfor i, ai := range a {\n\t\tif i == 0 || ai > max {\n\t\t\tmax = ai\n\t\t}\n\t}\n\treturn max\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func (fn *formulaFuncs) MAX(argsList *list.List) formulaArg {\n\tif argsList.Len() == 0 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"MAX requires at least 1 argument\")\n\t}\n\treturn fn.max(false, argsList)\n}", "func max(x int, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func (_TellorMesosphere *TellorMesosphereCaller) Max(opts *bind.CallOpts, _a *big.Int, _b *big.Int) (*big.Int, error) {\n\tvar out []interface{}\n\terr := _TellorMesosphere.contract.Call(opts, &out, \"max\", _a, _b)\n\n\tif err != nil {\n\t\treturn *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\n\treturn out0, err\n\n}", "func Max(field string) AggregateFunc {\n\treturn func(start, end string) (string, *dsl.Traversal) {\n\t\tif end == \"\" {\n\t\t\tend = DefaultMaxLabel\n\t\t}\n\t\treturn end, __.As(start).Unfold().Values(field).Max().As(end)\n\t}\n}", "func (fn *formulaFuncs) MAXA(argsList *list.List) formulaArg {\n\tif argsList.Len() == 0 {\n\t\treturn newErrorFormulaArg(formulaErrorVALUE, \"MAXA requires at least 1 argument\")\n\t}\n\treturn fn.max(true, argsList)\n}", "func ElemMax(a, b Res) Res {\n\taMask := a.Output().Copy()\n\taMask.Sub(b.Output())\n\tanyvec.GreaterThan(aMask, aMask.Creator().MakeNumeric(0))\n\tbMask := aMask.Copy()\n\tanyvec.Complement(bMask)\n\treturn Add(Mul(a, NewConst(aMask)), Mul(b, NewConst(bMask)))\n}", "func max(x, y float64) float64 {\n\tswitch {\n\tcase math.IsNaN(x) || math.IsNaN(y):\n\t\treturn math.NaN()\n\tcase math.IsInf(x, 1) || math.IsInf(y, 1):\n\t\treturn math.Inf(1)\n\n\tcase x == 0 && x == y:\n\t\tif math.Signbit(x) {\n\t\t\treturn y\n\t\t}\n\t\treturn x\n\t}\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}", "func Max(first Decimal, rest ...Decimal) Decimal {\n\tans := first\n\tfor _, item := range rest {\n\t\tif item.Cmp(ans) > 0 {\n\t\t\tans = item\n\t\t}\n\t}\n\treturn ans\n}", "func Max(first Decimal, rest ...Decimal) Decimal {\n\tans := first\n\tfor _, item := range rest {\n\t\tif item.Cmp(ans) > 0 {\n\t\t\tans = item\n\t\t}\n\t}\n\treturn ans\n}", "func Max(x, y int) int {\n if x < y {\n return y\n }\n return x\n}", "func (a Vec2) Max(b Vec2) Vec2 {\n\tvar r Vec2\n\tif a.X > b.X {\n\t\tr.X = a.X\n\t} else {\n\t\tr.X = b.X\n\t}\n\tif a.Y > b.Y {\n\t\tr.Y = a.Y\n\t} else {\n\t\tr.Y = b.Y\n\t}\n\treturn r\n}", "func Max(i, j int) int {\n\tif i > j {\n\t\treturn i\n\t}\n\treturn j\n}", "func (c *Clac) Max() error {\n\treturn c.applyFloat(2, func(vals []value.Value) (value.Value, error) {\n\t\treturn binary(vals[1], \"max\", vals[0])\n\t})\n}", "func max(num1, num2 int) int {\nresult int\n\n\tif (num1 > num2){\n\t\tresult = num1\n\t} else{\n\t\tresult = num2\n\t}\nreturn result\n}", "func (_TellorMesosphere *TellorMesosphereSession) Max(_a *big.Int, _b *big.Int) (*big.Int, error) {\n\treturn _TellorMesosphere.Contract.Max(&_TellorMesosphere.CallOpts, _a, _b)\n}", "func Max(scope *Scope, input tf.Output, axis tf.Output, optional ...MaxAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Max\",\n\t\tInput: []tf.Input{\n\t\t\tinput, axis,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func Max(vals ...float64) float64 {\n\tmax := vals[0]\n\tfor _, v := range vals {\n\t\tif v > max {\n\t\t\tmax = v\n\t\t}\n\t}\n\n\treturn max\n}", "func Imax(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Imax(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Maximum(scope *Scope, x tf.Output, y tf.Output) (z tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Maximum\",\n\t\tInput: []tf.Input{\n\t\t\tx, y,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func MaxInt(a, b int64) int64 {\n\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}" ]
[ "0.76541007", "0.75704294", "0.7160812", "0.7106996", "0.7106996", "0.7106996", "0.70229346", "0.6992478", "0.69921494", "0.6984913", "0.6984913", "0.6984913", "0.6984913", "0.6984913", "0.6974091", "0.6950687", "0.69445604", "0.69353336", "0.69353336", "0.69049996", "0.6876312", "0.6876312", "0.6789064", "0.67790186", "0.6765124", "0.6734706", "0.67209226", "0.6719933", "0.6712761", "0.6675576", "0.6675335", "0.6675335", "0.6675335", "0.6675335", "0.6675335", "0.6675335", "0.6675335", "0.6675335", "0.6675335", "0.6675335", "0.6675335", "0.6652151", "0.66262907", "0.6624699", "0.6624069", "0.6588682", "0.6575164", "0.65717363", "0.6569868", "0.6565681", "0.6554253", "0.65411764", "0.65409625", "0.65336514", "0.65255654", "0.6521499", "0.65168184", "0.65121025", "0.64980537", "0.6497725", "0.64949507", "0.64949507", "0.6479125", "0.6463722", "0.6426867", "0.6420864", "0.6409916", "0.6342616", "0.63341576", "0.6332803", "0.6326981", "0.62606084", "0.62528914", "0.6249891", "0.6249891", "0.6249891", "0.6249891", "0.6249891", "0.62463987", "0.6230152", "0.62244874", "0.6212205", "0.6201431", "0.61971945", "0.6193698", "0.6173437", "0.61713094", "0.61713094", "0.61690587", "0.616618", "0.6154511", "0.61437804", "0.61072695", "0.6106232", "0.6093482", "0.60763913", "0.60759944", "0.60759944", "0.6063006", "0.6059946" ]
0.78921455
0
InSequence does work in sequence and waits for the previous work item to complete. it should write results on channel in order of input work array.
func InSequence(ch chan Result, work []int) []Result { results := make([]Result, len(work)) go buildInSeq(ch, work, results) return results }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func InParallel(ch chan Result, work []int) []Result {\n\tresults := make([]Result, len(work))\n\tvar mutex = &sync.Mutex{}\n\tfor i, num := range work {\n\t\tgo func(w int, res *Result, ch chan Result) {\n\t\t\t*res = executeWork(w)\n\t\t\tmutex.Lock()\n\t\t\tch <- *res\n\t\t\tmutex.Unlock()\n\t\t}(num, &results[i], ch)\n\t}\n\treturn results\n}", "func (s *intSequence) Items() <-chan string { return s.data }", "func main() {\n\t// EXAMPLE 1\n\tch := make(chan string, 2)\n\tch <- \"naveen\"\n\tch <- \"paul\"\n\tfmt.Println(<-ch)\n\tfmt.Println(<-ch)\n\n\t// EXAMPLE 2\n\tch2 := make(chan int, 2)\n\tgo write(ch2)\n\ttime.Sleep(2 * time.Second)\n\tfor v := range ch2 {\n\t\tfmt.Println(\"read value\", v, \"from ch2\")\n\t\ttime.Sleep(2 * time.Second)\n\n\t}\n\n\t// LENGTH VS CAPACITY\n\t// length: number of elements currently queued in it\n\t// capacity: number of values that the channel can hold\n\tch3 := make(chan string, 3)\n\tch3 <- \"naveen\"\n\tch3 <- \"paul\"\n\tfmt.Println(\"capacity is\", cap(ch3))\n\tfmt.Println(\"length is\", len(ch3))\n\tfmt.Println(\"read value\", <-ch3)\n\tfmt.Println(\"capacity is\", cap(ch3))\n\tfmt.Println(\"new length is\", len(ch3))\n\n\t// WAIT GROUP: used to wait for a collection of Goroutines to finish executing, control is blocked until all Goroutines are are finished\n\tno := 3\n\tvar wg sync.WaitGroup\n\tfor i := 0; i < no; i++ {\n\t\twg.Add(1) // increment counter value\n\t\tgo process(i, &wg) // must be the address, otherwise each goroutine will have its onw copy of the WaitGroup\n\t}\n\twg.Wait() // blocks the Goroutine until the counter becomes zero\n\tfmt.Println(\"All go routines finished executing\")\n\n\t// WORKER POOL\n\tstartTime := time.Now()\n\n\tnoOfJobs := 100\n\tgo allocate(noOfJobs)\n\n\tdone := make(chan bool)\n\tgo result(done)\n\n\tnoOfWorkers := 10\n\tcreateWorkerPool(noOfWorkers)\n\n\t<-done\n\tendTime := time.Now()\n\tdiff := endTime.Sub(startTime)\n\tfmt.Println(\"total time taken\", diff.Seconds(), \"seconds\")\n}", "func (w *Worker) Work() {\n\tfor {\n\t\tselect {\n\t\tcase <-w.done:\n\t\t\treturn\n\t\tdefault:\n\t\t\tevent := <-w.channel\n\t\t\tw.writeToBuffer(event)\n\t\t}\n\t}\n}", "func TestChannelAsOutputRoutine(t *testing.T) {\n\tfor _, array := range bigStruct {\n\t\tch := make(chan int)\n\t\tgo sum(array.bigArray, ch)\n\t\tsum := <-ch // this is the blocking operation.\n\t\tif sum != array.result {\n\t\t\tt.Errorf(\"Expected result %d but got %d\\n\", array.result, sum)\n\t\t} else {\n\t\t\tt.Logf(\"The correct sum was calculated: %d\\n\", sum)\n\t\t}\n\t}\n}", "func doWrite(work <-chan t, done <-chan struct{}) chan t {\n\tout := make(chan t)\n\tgo func() {\n\t\tdefer close(out)\n\t\tfor item := range work {\n\t\t\titem = process(item, \"written\")\n\t\t\tsleep := rand.Int63n(100)\n\t\t\ttime.Sleep(time.Duration(sleep) * time.Millisecond)\n\t\t\tselect {\n\t\t\tcase out <- item:\n\t\t\tcase <-done:\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\treturn out\n}", "func blockingExample() {\n\tintChan := make(chan int)\n\n\tintChan <- 5\n\t// go func() { intChan <- 5 }()\n\t// go func() {\n\t// \tfmt.Println(\"staring goroutine\")\n\t// \tintChan <- 5\n\t// \tfmt.Println(\"finishing goroutine\")\n\t// }()\n\n\tfmt.Println(\"after input\")\n\n\toutput := <-intChan\n\n\tfmt.Println(\"after output1\")\n\n\tfmt.Println(output)\n\n\t// output = <-intChan\n\n\t// fmt.Println(\"after output2\")\n\n\t// fmt.Println(output)\n}", "func (s Sequence) Output(c SeqChan) {s.Do(func(el El){c <- el})}", "func (b *bufferedChan) Run() {\n\tdefer close(b.OutChannel)\n\tfor value := range b.inChannel {\n\t\tselect {\n\t\tcase <-b.ctx.Done():\n\t\t\tfmt.Println(\"Run: Time to return\")\n\t\t\treturn\n\t\tcase b.OutChannel <- value:\n\t\tdefault:\n\t\t\treturn\n\t\t}\n\t}\n\n}", "func (w *DeploymentWrite) Intake() chan msg.Request {\n\treturn w.Input\n}", "func sequential_sort(seq ElementSlice, done chan bool) {\n sort.Sort(seq)\n done <- true\n}", "func (ow *ordered[T, U]) Wait() []U {\n\tow.wg.Wait()\n\treturn ow.results\n}", "func readArchive(work <-chan t, done <-chan struct{}) chan t {\n\tout := make(chan t, 50) // Oo, I learned from Guido that there's a bug in this code\n\tgo func(input <-chan t, output chan<- t, done <-chan struct{}) {\n\t\tdefer close(out)\n\t\tfor item := range input {\n\t\t\titem = process(item, \"archive\") // HL\n\t\t\tselect {\n\t\t\tcase output <- item: // HL\n\t\t\tcase <-done: // HL\n\t\t\t\treturn // HL\n\t\t\t}\n\t\t}\n\t}(work, out, done)\n\treturn out\n}", "func (twrkr *twerk) Work(args ...interface{}) (<-chan []interface{}, error) {\n\n\tif twrkr.stop {\n\t\treturn nil, fmt.Errorf(\"twerker has been stopped. no more work can be done with this twerker\")\n\t}\n\n\targumentValues, err := twrkr.callable.TransformToValues(args...)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar returnToChan chan []interface{}\n\n\tif twrkr.callable.NumberOfReturnValues() > 0 {\n\t\treturnToChan = make(chan []interface{})\n\t}\n\n\tnewJobInstruction := jobInstruction{\n\t\targuments: argumentValues,\n\t\treturnTo: returnToChan,\n\t}\n\n\tgo func() {\n\t\ttwrkr.jobListener <- newJobInstruction\n\t}()\n\n\treturn returnToChan, nil\n}", "func doneChannelExample() {\n\tjobs := make(chan int)\n\t// jobs := make(chan int, 5)\n\tdone := make(chan bool)\n\n\tgo func() {\n\t\tfor {\n\t\t\tj, isOpen := <-jobs\n\t\t\tif isOpen {\n\t\t\t\t//process job\n\t\t\t\tfmt.Println(\"received job\", j)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tfmt.Println(\"no more jobs to recive\")\n\t\t\tdone <- true\n\t\t}\n\t}()\n\n\tfor j := 1; j <= 3; j++ {\n\t\tjobs <- j\n\t\tfmt.Println(\"sent job\", j)\n\t}\n\tclose(jobs)\n\tfmt.Println(\"sent all jobs\")\n\n\t<-done\n}", "func out(ch chan int) {\n\tfor i := 1; i < 100; i++ {\n\t\tch <- i\n\t\ttime.Sleep(time.Second * 1)\n\t}\n\n}", "func (j *Job) Wait() { <-j.isDone }", "func (p *AsyncProducer) Input() chan<- *sarama.ProducerMessage { return p.input }", "func (m Mine) WorkSequential(f Finder, mr Miner, s Smelter) Ingot {\n\tpanic(\"non implementato\")\n}", "func (m *Mare) InChannel(items chan interface{}) *Mare {\n\tm.prepareInput()\n\tgo func() {\n\t\tdefer m.mapInWorkers.Done()\n\t\tfor item := range items {\n\t\t\tm.mapInChan <- item\n\t\t}\n\t}()\n\t// Return m\n\treturn m\n}", "func in(ch chan int) {\n\tfor {\n\t\tmsg := <-ch\n\t\tfmt.Println(msg)\n\t}\n}", "func worker(id int, jobs <-chan int, res chan<- int) {\n\n\tfmt.Printf(\"Worker %d : started\\n\", id)\n\tfor job := range jobs {\n\t\tfmt.Printf(\"worker %d : received job %d\\n\", id, job)\n\t\ttime.Sleep(time.Second)\n\t\tres <- job * job\n\t\tfmt.Printf(\"worker %d : finished job %d\", id, job)\n\t}\n\t//close(results)\n\t//putting close here and having a range call over the results channel below wont work becuase,\n\t//the first worker to finish its jobs closes the results channel, and other workers wont be able to send their results\n\t//sending to a closed channel will cause panic\n\n}", "func (ow *ordered[T, U]) Push(input T) error {\n\tcurrent := ow.count\n\tif current >= len(ow.results) {\n\t\treturn fmt.Errorf(\"MaxCapacity\")\n\t}\n\n\tonComplete := make(chan U)\n\terr := ow.workPool.Run(input, onComplete)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tow.count++\n\tow.wg.Add(1)\n\n\tgo func(index int) {\n\t\tdefer close(onComplete)\n\n\t\tow.results[index] = <-onComplete\n\t\tow.wg.Done()\n\t}(int(current))\n\n\treturn nil\n}", "func (p *Pool) work() {\n\tfor task := range p.queueTasks {\n\t\t//when the work is idle it receive a new task from the channel\n\t\ttask.DoWork()\n\t\tp.wg.Done()\n\t}\n}", "func (b ArrayBucket) Worker(inChan <-chan string, outChan chan<- string, wg *sync.WaitGroup) {\n\twg.Add(1)\n\tgo func() {\n\t\tfor cand := range inChan {\n\t\t\t//Test against the bucket if\n\t\t\t// it the cand can be made from available tokens\n\t\t\tif b.testMatch(cand) {\n\t\t\t\toutChan <- cand\n\t\t\t}\n\t\t}\n\t\twg.Done()\n\t}()\n}", "func (sub *BufferedSubscription[T]) Out() <-chan T {\n\treturn sub.result\n}", "func (async *async) wait() {\n\t<-async.done\n}", "func TestInProgress(t *testing.T) {\n\ttestCase := func(c chan<- fs.Event) {\n\t\tevents.Default.Log(events.ItemStarted, map[string]string{\n\t\t\t\"item\": \"inprogress\",\n\t\t})\n\t\tsleepMs(100)\n\t\tc <- fs.Event{Name: \"inprogress\", Type: fs.NonRemove}\n\t\tsleepMs(1000)\n\t\tevents.Default.Log(events.ItemFinished, map[string]interface{}{\n\t\t\t\"item\": \"inprogress\",\n\t\t})\n\t\tsleepMs(100)\n\t\tc <- fs.Event{Name: \"notinprogress\", Type: fs.NonRemove}\n\t\tsleepMs(800)\n\t}\n\n\texpectedBatches := []expectedBatch{\n\t\t{[][]string{{\"notinprogress\"}}, 2000, 3500},\n\t}\n\n\ttestScenario(t, \"InProgress\", testCase, expectedBatches)\n}", "func (c *QueuedChan) run() {\n\t// Notify close channel coroutine.\n\tdefer close(c.done)\n\n\tfor {\n\t\tvar elem *list.Element\n\t\tvar item interface{}\n\t\tvar popc chan<- interface{}\n\n\t\t// Get front element of the queue.\n\t\tif elem = c.Front(); nil != elem {\n\t\t\tpopc, item = c.popc, elem.Value\n\t\t}\n\n\t\tselect {\n\t\t// Put the new object into the end of queue.\n\t\tcase i := <-c.pushc:\n\t\t\tc.PushBack(i)\n\t\t// Remove the front element from queue if send out success\n\t\tcase popc <- item:\n\t\t\tc.List.Remove(elem)\n\t\t// Control command\n\t\tcase cmd := <-c.ctrlc:\n\t\t\tc.control(cmd)\n\t\t// Channel is closed\n\t\tcase <-c.close:\n\t\t\treturn\n\t\t}\n\t\t// Update channel length\n\t\tatomic.StoreInt32(&c.len, int32(c.List.Len()))\n\t}\n}", "func (p *SingleProcess) In(val interface{}) {\n\tp.in <- val\n}", "func (p *pipe) inputAdvance(count int) {\n\tp.inPos += int32(count)\n\tif p.inPos >= p.size {\n\t\tp.inPos -= p.size\n\t}\n\tatomic.AddInt32(&p.free, -int32(count))\n\n\tselect {\n\tcase p.outWake <- struct{}{}:\n\tdefault:\n\t}\n}", "func JoinWriteCloserChan(out chan<- io.WriteCloser, inp <-chan io.WriteCloser) chan struct{} {\n\tdone := make(chan struct{})\n\tgo func() {\n\t\tdefer close(done)\n\t\tfor i := range inp {\n\t\t\tout <- i\n\t\t}\n\t\tdone <- struct{}{}\n\t}()\n\treturn done\n}", "func main() {\n\tdoWork := func(strings <-chan string ) <-chan interface{} {\n\t\tcompleted := make(chan interface{})\n\t\tgo func() {\n\t\t\tdefer fmt.Println(\"doWork Exited\")\n\t\t\tdefer close(completed)\n\t\t\tfor s := range strings {\n\t\t\t\tfmt.Println(s)\n\t\t\t}\n\t\t}()\n\t\treturn completed\n\t}\n\n\tdoWork2 := func (\n\t\tdone <-chan interface{},\n\t\tstrings <-chan string,\n\t) <-chan interface{} {\n\t\tterminated := make(chan interface{})\n\t\tgo func() {\n\t\t\tdefer fmt.Println(\"doWork exited\")\n\t\t\tdefer close(terminated)\n\t\t\tfor {\n\t\t\t\tselect {\n\t\t\t\tcase s:= <-strings:\n\t\t\t\t\tfmt.Println(s)\n\t\t\t\tcase <-done:\n\t\t\t\t\treturn \n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t\treturn terminated\n\t}\n\n\tnewRandStream := func() <-chan int {\n\t\trandStream:= make(chan int)\n\t\tgo func(){\n\t\t\tdefer fmt.Println(\"newRandStream closure exited\")\n\t\t\tdefer close(randStream)\n\t\t\tfor {\n\t\t\t\trandStream <- rand.Intn(strings.NewReader(), 10)\n\t\t\t}\n\t\t}()\n\t\treturn randStream\n\t}\n\n\trandStream := newRandStream()\n\tfmt.Println(\"3 Random Ints:\")\n\tfor i:= 0; i < 3; i++{\n\t\tfmt.Printf(\"%d: %d\\n\", i, <-randStream)\n\t}\n\n\tdoWork(nil)\n\tdone := make(chan interface{})\n\tstringChan := make(chan string)\n\tterminated := doWork2(done, stringChan)\n\tgo func(){\n\t\ttime.Sleep(1 * time.Second)\n\t\tfmt.Println(\"Cancelling doWork Goroutine\")\n\t\tclose(done)\n\t}()\t\n\tfor i:= 0; i < 10; i++{\n\t\tstringChan<-\"test\"\n\t}\n\t<-terminated\n\tfmt.Println(\"Done.\")\n}", "func input_job(size int, query_chan chan<- []byte, scope int){\n var random_value int = -1\n for i := 1; i <= size; i++ {\n //get random num between 1 and size\n random_value = rand.Intn(scope) + 1\n //convert the int into []byte\n query_chan <- []byte(strconv.Itoa(random_value))\n fmt.Printf(\"输入随机数: %d\\n\", random_value)\n }\n //the last finished signal \"0\"\n query_chan <- []byte(strconv.Itoa(0))\n fmt.Println(\"随机数输入完成\")\n}", "func analysisProduce() {\n\ttotalJob := 10\n\tqueue := make([]int, 0, totalJob)\n\tlog.Println(\n\t\t\"initial queue length: \",\n\t\tlen(queue))\n\n\tgo concurrentProduce(totalJob, &queue)\n\n\tlog.Println(\"main goroutine go to sleep...\")\n\ttime.Sleep(time.Second * 3)\n\n\tlog.Println(\"final queue length:\", len(queue))\n}", "func DoneWriteCloserSlice(inp <-chan io.WriteCloser) chan []io.WriteCloser {\n\tdone := make(chan []io.WriteCloser)\n\tgo func() {\n\t\tdefer close(done)\n\t\tWriteCloserS := []io.WriteCloser{}\n\t\tfor i := range inp {\n\t\t\tWriteCloserS = append(WriteCloserS, i)\n\t\t}\n\t\tdone <- WriteCloserS\n\t}()\n\treturn done\n}", "func (inp ThingFrom) ThingDoneSlice() (done <-chan []Thing) {\n\tsig := make(chan []Thing)\n\tgo inp.doneThingSlice(sig)\n\treturn sig\n}", "func processMessagesInOrder(in chan kafka.Message, timeWindow time.Duration) {\n\tw := window.New(100, timeWindow)\n\n\tgo func() {\n\t\tvar lastMessage kafka.Message\n\t\tfor batch := range w.Out {\n\t\t\tfor _, m := range batch {\n\t\t\t\tprintMessage(&lastMessage, m.Value.(kafka.Message))\n\t\t\t\tlastMessage = m.Value.(kafka.Message)\n\t\t\t}\n\t\t}\n\t}()\n\n\tfor m := range in {\n\t\tw.Insert(&window.Element{\n\t\t\tTimestamp: m.Timestamp,\n\t\t\tValue: m,\n\t\t})\n\n\t}\n\n\tw.Close()\n}", "func Put(queue chan<- byte, items []byte) {\n\tfor _, i := range items {\n\t\tqueue <- i\n\t}\n}", "func (c *ConnectionMock) WriteLoop(out <-chan []byte) {\n\tc.Called(out)\n}", "func (ig *Instagram) producer(results chan *InstagramResponse, work string) {\n\t//Todo: Error handling\n\tr, err := ig.getDecode(work)\n\tif err != nil {\n\t\tr = &InstagramResponse{}\n\t}\n\n\tresults <- r\n\treturn\n}", "func routineBuffer(){\n\tchannel := make(chan int,4)//make channel and add buffer\n\twg.Add(2)\n\t//receive from channel\n\tgo func(channel <- chan int){\n\t\t//loop throug the channel\n\t\tfor i := range channel{\n\t\t\tfmt.Println(i)\n\t\t}\n\t\twg.Done()\n\t}(channel)\n\t//sending to channel\n\tgo func(channel chan <- int){\n\t\tchannel <- 100\n\t\tchannel <- 200\n\t\tclose(channel) //after sending close the channel \n\t\twg.Done()\n\t}(channel)\n\twg.Wait()\n}", "func (suite *KeeperTestSuite) TestSetSequence() {\n\tpath := ibctesting.NewPath(suite.chainA, suite.chainB)\n\tsuite.coordinator.Setup(path)\n\n\tctxA := suite.chainA.GetContext()\n\tone := uint64(1)\n\n\t// initialized channel has next send seq of 1\n\tseq, found := suite.chainA.App.GetIBCKeeper().ChannelKeeper.GetNextSequenceSend(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID)\n\tsuite.True(found)\n\tsuite.Equal(one, seq)\n\n\t// initialized channel has next seq recv of 1\n\tseq, found = suite.chainA.App.GetIBCKeeper().ChannelKeeper.GetNextSequenceRecv(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID)\n\tsuite.True(found)\n\tsuite.Equal(one, seq)\n\n\t// initialized channel has next seq ack of\n\tseq, found = suite.chainA.App.GetIBCKeeper().ChannelKeeper.GetNextSequenceAck(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID)\n\tsuite.True(found)\n\tsuite.Equal(one, seq)\n\n\tnextSeqSend, nextSeqRecv, nextSeqAck := uint64(10), uint64(10), uint64(10)\n\tsuite.chainA.App.GetIBCKeeper().ChannelKeeper.SetNextSequenceSend(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID, nextSeqSend)\n\tsuite.chainA.App.GetIBCKeeper().ChannelKeeper.SetNextSequenceRecv(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID, nextSeqRecv)\n\tsuite.chainA.App.GetIBCKeeper().ChannelKeeper.SetNextSequenceAck(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID, nextSeqAck)\n\n\tstoredNextSeqSend, found := suite.chainA.App.GetIBCKeeper().ChannelKeeper.GetNextSequenceSend(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID)\n\tsuite.True(found)\n\tsuite.Equal(nextSeqSend, storedNextSeqSend)\n\n\tstoredNextSeqRecv, found := suite.chainA.App.GetIBCKeeper().ChannelKeeper.GetNextSequenceSend(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID)\n\tsuite.True(found)\n\tsuite.Equal(nextSeqRecv, storedNextSeqRecv)\n\n\tstoredNextSeqAck, found := suite.chainA.App.GetIBCKeeper().ChannelKeeper.GetNextSequenceAck(ctxA, path.EndpointA.ChannelConfig.PortID, path.EndpointA.ChannelID)\n\tsuite.True(found)\n\tsuite.Equal(nextSeqAck, storedNextSeqAck)\n}", "func drainWorkQueue(work chan *v1.Node) {\n\tfor {\n\t\t_, ok := <-work\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t}\n}", "func (s *SyncTask) Start(input io.Reader, synced, failed io.Writer) error {\n\n\tstart := time.Now()\n\n\tkeysIn := make(chan s3.Key, s.SyncPara*BufferFactor)\n\tkeysOk := make(chan s3.Key, s.SyncPara*BufferFactor)\n\tkeysFail := make(chan s3.Key, s.SyncPara*BufferFactor)\n\n\tdecoders := make(chan []byte, s.DecodePara*BufferFactor)\n\n\t// start JSON decoders\n\tlogrus.WithFields(logrus.Fields{\n\t\t\"key_decoders\": s.DecodePara,\n\t\t\"buffer_size\": cap(decoders),\n\t}).Info(\"starting key decoders\")\n\n\tdecGroup := sync.WaitGroup{}\n\tfor i := 0; i < s.DecodePara; i++ {\n\t\tdecGroup.Add(1)\n\t\tgo s.decode(&decGroup, decoders, keysIn)\n\t}\n\n\t// start S3 sync workers\n\tlogrus.WithFields(logrus.Fields{\n\t\t\"sync_workers\": s.SyncPara,\n\t\t\"buffer_size\": cap(keysIn),\n\t}).Info(\"starting key sync workers\")\n\tsyncGroup := sync.WaitGroup{}\n\tfor i := 0; i < s.SyncPara; i++ {\n\t\tsyncGroup.Add(1)\n\t\tgo s.syncKey(&syncGroup, s.src, s.dst, keysIn, keysOk, keysFail)\n\t}\n\n\t// track keys that have been sync'd, and those that we failed to sync.\n\tlogrus.Info(\"starting to write progress\")\n\tencGroup := sync.WaitGroup{}\n\tencGroup.Add(2)\n\tgo s.encode(&encGroup, synced, keysOk)\n\tgo s.encode(&encGroup, failed, keysFail)\n\n\t// feed the pipeline by reading the listing file\n\tlogrus.Info(\"starting to read key listing file\")\n\terr := s.readLines(input, decoders)\n\n\t// when done reading the source file, wait until the decoders\n\t// are done.\n\tlogrus.WithFields(logrus.Fields{\n\t\t\"since_start\": time.Since(start),\n\t\t\"line_count\": metrics.fileLines.String(),\n\t}).Info(\"done reading lines from sync list\")\n\tclose(decoders)\n\tdecGroup.Wait()\n\n\t// when the decoders are all done, wait for the sync workers to finish\n\n\tlogrus.WithFields(logrus.Fields{\n\t\t\"since_start\": time.Since(start),\n\t\t\"line_count\": metrics.decodedKeys.String(),\n\t}).Info(\"done decoding keys from sync list\")\n\n\tclose(keysIn)\n\tsyncGroup.Wait()\n\n\tclose(keysOk)\n\tclose(keysFail)\n\n\tencGroup.Wait()\n\n\t// the source file is read, all keys were decoded and sync'd. we're done.\n\tlogrus.WithFields(logrus.Fields{\n\t\t\"since_start\": time.Since(start),\n\t\t\"sync_ok\": metrics.syncOk.String(),\n\t\t\"sync_fail\": metrics.syncAbandoned.String(),\n\t}).Info(\"done syncing keys\")\n\n\treturn err\n}", "func (o *outQueue) shift() {\n\tif len(o.data) > 0 {\n\t\tselect {\n\t\tcase o.out <- o.data[0]:\n\t\t\to.data = o.data[1:]\n\t\tdefault:\n\t\t}\n\t}\n}", "func queueWorker() {\n\tcrossRefClient := crossref.NewCrossRefClient(&http.Client{})\n\tbibClient := bibtex.NewDXDOIClient(&http.Client{})\n\tfor {\n\t\tx := <-enhancementChan\n\t\tfor counter.Get() > 40 {\n\t\t\ttime.Sleep(time.Millisecond * 10)\n\t\t}\n\t\tcounter.Inc()\n\t\tgo gatherAdditionalInfo(crossRefClient, bibClient, x)\n\n\t}\n}", "func continuePipelineFunctionA(numbers <-chan int, squared chan<- string) {\n\tfor {\n\t\tres, open := <- numbers\n\t\tif !open {\n\t\t\tbreak\n\t\t}\n\t\tsquared <- strconv.Itoa(res) + \" is \" + strconv.Itoa(res * res) // send 'numbers' to receiving 'squared' when pipeline is un-sync'd\n\t}\n\tclose(squared)\n}", "func (r *Reflex) batch(out chan<- string, in <-chan string) {\n\tfor name := range in {\n\t\tr.backlog.Add(name)\n\t\ttimer := time.NewTimer(300 * time.Millisecond)\n\touter:\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase name := <-in:\n\t\t\t\tr.backlog.Add(name)\n\t\t\tcase <-timer.C:\n\t\t\t\tfor {\n\t\t\t\t\tselect {\n\t\t\t\t\tcase name := <-in:\n\t\t\t\t\t\tr.backlog.Add(name)\n\t\t\t\t\tcase out <- r.backlog.Next():\n\t\t\t\t\t\tif r.backlog.RemoveOne() {\n\t\t\t\t\t\t\tbreak outer\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func main() {\n\tin, out := make(chan *Work), make(chan *Work)\n\tfor i := 0; i < NumWorkers; i++ {\n\t\tgo worker(in, out)\n\t}\n\tgo sendLotsOfWork(in)\n\treceiveLotsOfResults(out)\n}", "func (s *Chan) In() chan []byte {\n\treturn s.inMsgChan\n}", "func (b ArrayBucket) BlockWorker(inChan <-chan []string, outChan chan<- string, wg *sync.WaitGroup) {\n\n\tfor candA := range inChan {\n\t\tfor _, cand := range candA {\n\t\t\tif b.testMatch(cand) {\n\t\t\t\toutChan <- cand\n\t\t\t}\n\t\t}\n\t}\n\twg.Done()\n}", "func (inp ThingFrom) ThingFiniSlice() func(inp ThingFrom) (done <-chan []Thing) {\n\n\treturn func(inp ThingFrom) (done <-chan []Thing) {\n\t\treturn inp.ThingDoneSlice()\n\t}\n}", "func (p *Pool) worker() {\n\tdefer p.wg.Done()\n\tfor job := range p.In {\n\t\tatomic.AddInt64(&p.Stats.Pending, -1)\n\t\tatomic.AddInt64(&p.Stats.Running, 1)\n\t\tjob.Result, job.Error = job.F(job.Args...)\n\t\tatomic.AddInt64(&p.Stats.Running, -1)\n\t\tatomic.AddInt64(&p.Stats.Completed, 1)\n\t\tp.Out <- job\n\t}\n}", "func wait_and_process_results(results <-chan stat, done <-chan doneStatus,\n num_workers int) []stat {\n\n output := make([]stat, 0)\n\n for w := 0; w < num_workers; {\n select { // Blocking\n case result := <-results:\n output = append(output, result)\n case <-done:\n num_workers--\n }\n }\n\nDONE:\n\n // process any remaining results\n for {\n select {\n case result := <-results:\n output = append(output, result)\n default:\n break DONE\n }\n }\n\n return output\n}", "func threadWait(ch chan []string, callerSignal chan bool, threadReturn chan []interface{}, size int) {\n\treturnContainer := make([]interface{}, size)\n\titemCounter := 0\n\n\tfor itemCounter < (size - 1) {\n\t\tret := <-ch\n\t\t//fmt.Println(\"Thread finished\")\n\t\treturnContainer[itemCounter] = ret\n\t\tcallerSignal <- true // signal to caller when we see a thread return\n\t\titemCounter++\n\t}\n\n\tthreadReturn <- returnContainer\n}", "func worker(id int, jobs <-chan int, results chan<- int) {\n\tfor j := range jobs {\n\t\tfmt.Println(\"worker\", id, \"started job\", j)\n\t\ttime.Sleep(time.Second)\n\t\tfmt.Println(\"worker\", id, \"finished job\", j)\n\t\tresults <- j * 2\n\t}\n}", "func (c *ConnectionMock) Out() chan []byte {\n\targs := c.Called()\n\treturn args.Get(0).(chan []byte)\n}", "func (c *ConnectionMock) In() (chan []byte, chan string) {\n\targs := c.Called()\n\treturn args.Get(0).(chan []byte), args.Get(1).(chan string)\n}", "func Work() {\n\tgo func() {\n\t\tpublisher()\n\t\tsubscriber()\n\t}()\n}", "func seq(id string, vals ...int64) chan Ordered {\n\tch := make(chan Ordered)\n\n\tgo func() {\n\t\tfor _, v := range vals {\n\t\t\tch <- item{id, v}\n\t\t}\n\t\tclose(ch)\n\t}()\n\n\treturn ch\n}", "func SimpleSequence(ctx context.Context, t *testing.T, client *entroq.EntroQ, qPrefix string) {\n\tnow := time.Now()\n\n\tqueue := path.Join(qPrefix, \"simple_sequence\")\n\n\t// Claim from empty queue.\n\ttask, err := client.TryClaim(ctx, entroq.From(queue), entroq.ClaimFor(100*time.Millisecond))\n\tif err != nil {\n\t\tt.Fatalf(\"Got unexpected error claiming from empty queue: %v\", err)\n\t}\n\tif task != nil {\n\t\tt.Fatalf(\"Got unexpected non-nil claim response from empty queue:\\n%s\", task)\n\t}\n\n\tconst futureTaskDuration = 2 * time.Second\n\tfutureTime := now.Add(futureTaskDuration)\n\n\tinsWant := []*entroq.Task{\n\t\t{\n\t\t\tQueue: queue,\n\t\t\tAt: now,\n\t\t\tValue: []byte(\"hello\"),\n\t\t\tClaimant: client.ID(),\n\t\t},\n\t\t{\n\t\t\tQueue: queue,\n\t\t\tAt: futureTime,\n\t\t\tValue: []byte(\"there\"),\n\t\t\tClaimant: client.ID(),\n\t\t},\n\t}\n\tvar insData []*entroq.TaskData\n\tfor _, task := range insWant {\n\t\tinsData = append(insData, task.Data())\n\t}\n\n\tinserted, changed, err := client.Modify(ctx, entroq.Inserting(insData...))\n\tif err != nil {\n\t\tt.Fatalf(\"Got unexpected error inserting two tasks: %+v\", err)\n\t}\n\tif changed != nil {\n\t\tt.Fatalf(\"Got unexpected changes during insertion: %+v\", err)\n\t}\n\tif diff := EqualAllTasksOrderedSkipIDAndTime(insWant, inserted); diff != \"\" {\n\t\tt.Fatalf(\"Modify tasks unexpected result, ignoring ID and time fields (-want +got):\\n%v\", diff)\n\t}\n\t// Also check that their arrival times are 100 ms apart as expected:\n\tif diff := inserted[1].At.Sub(inserted[0].At); diff != futureTaskDuration {\n\t\tt.Fatalf(\"Wanted At difference to be %v, got %v\", futureTaskDuration, diff)\n\t}\n\n\t// Get queues.\n\tqueuesWant := map[string]int{queue: 2}\n\tqueuesGot, err := client.Queues(ctx, entroq.MatchPrefix(qPrefix))\n\tif err != nil {\n\t\tt.Fatalf(\"Getting queues failed: %v\", err)\n\t}\n\tif diff := cmp.Diff(queuesWant, queuesGot); diff != \"\" {\n\t\tt.Fatalf(\"Queues (-want +got):\\n%v\", diff)\n\t}\n\n\t// Get all tasks.\n\ttasksGot, err := client.Tasks(ctx, queue)\n\tif err != nil {\n\t\tt.Fatalf(\"Tasks call failed after insertions: %v\", err)\n\t}\n\tif diff := EqualAllTasksUnorderedSkipTimesAndCounters(inserted, tasksGot); diff != \"\" {\n\t\tt.Fatalf(\"Tasks unexpected return, ignoring ID and time fields (-want +got):\\n%+v\", diff)\n\t}\n\n\t// Claim ready task.\n\tclaimCtx, _ := context.WithTimeout(ctx, 5*time.Second)\n\tclaimed, err := client.Claim(claimCtx, entroq.From(queue), entroq.ClaimFor(10*time.Second))\n\n\tif err != nil {\n\t\tt.Fatalf(\"Got unexpected error for claiming from a queue with one ready task: %+v\", err)\n\t}\n\tif claimed == nil {\n\t\tt.Fatalf(\"Unexpected nil result from blocking Claim\")\n\t}\n\tif diff := EqualTasksVersionIncr(inserted[0], claimed, 1); diff != \"\" {\n\t\tt.Fatalf(\"Claim tasks differ, ignoring ID and times:\\n%v\", diff)\n\t}\n\tif claimed.Claims != 1 {\n\t\tt.Fatalf(\"Expected claim to increment task claims to %d, got %d\", 1, claimed.Claims)\n\t}\n\n\t// TryClaim not ready task.\n\ttryclaimed, err := client.TryClaim(ctx, entroq.From(queue), entroq.ClaimFor(10*time.Second))\n\tif err != nil {\n\t\tt.Fatalf(\"Got unexpected error for claiming from a queue with no ready tasks: %v\", err)\n\t}\n\tif tryclaimed != nil {\n\t\tt.Fatalf(\"Got unexpected non-nil claim response from a queue with no ready tasks:\\n%s\", tryclaimed)\n\t}\n\n\t// Make sure the next claim will work.\n\tclaimCtx, cancel := context.WithTimeout(ctx, 2*futureTaskDuration)\n\tdefer cancel()\n\tclaimed, err = client.Claim(claimCtx,\n\t\tentroq.From(queue),\n\t\tentroq.ClaimFor(5*time.Second),\n\t\tentroq.ClaimPollTime(time.Second))\n\tif err != nil {\n\t\tt.Fatalf(\"Got unexpected error for claiming from a queue with one ready task: %v\", err)\n\t}\n\tif diff := EqualTasksVersionIncr(inserted[1], claimed, 1); diff != \"\" {\n\t\tt.Fatalf(\"Claim got unexpected task, ignoring ID and time fields (-want +got):\\n%v\", diff)\n\t}\n\tlog.Printf(\"Now: %v\", now)\n\tif got := claimed.At; got.Before(futureTime) {\n\t\tt.Fatalf(\"Claimed arrival time %v came earlier than expedcted time %v\", got, futureTime)\n\t}\n\tif claimed.Claims != 1 {\n\t\tt.Fatalf(\"Expected claim to increment task claims to %d, got %d\", 1, claimed.Claims)\n\t}\n}", "func TestConcurrentActivityExecution(t *testing.T) {\n\tr := task.NewTaskRegistry()\n\tr.AddOrchestratorN(\"ActivityFanOut\", func(ctx *task.OrchestrationContext) (any, error) {\n\t\ttasks := []task.Task{}\n\t\tfor i := 0; i < 10; i++ {\n\t\t\ttasks = append(tasks, ctx.CallActivity(\"ToString\", task.WithActivityInput(i)))\n\t\t}\n\t\tresults := []string{}\n\t\tfor _, t := range tasks {\n\t\t\tvar result string\n\t\t\tif err := t.Await(&result); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tresults = append(results, result)\n\t\t}\n\t\tsort.Sort(sort.Reverse(sort.StringSlice(results)))\n\t\treturn results, nil\n\t})\n\tr.AddActivityN(\"ToString\", func(ctx task.ActivityContext) (any, error) {\n\t\tvar input int\n\t\tif err := ctx.GetInput(&input); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\t// Sleep for 1 second to ensure that the test passes only if all activities execute in parallel.\n\t\ttime.Sleep(1 * time.Second)\n\t\treturn fmt.Sprintf(\"%d\", input), nil\n\t})\n\n\tctx := context.Background()\n\tclient, engine := startEngine(ctx, t, r)\n\tfor _, opt := range GetTestOptions() {\n\t\tt.Run(opt(engine), func(t *testing.T) {\n\t\t\tid, err := client.ScheduleNewOrchestration(ctx, \"ActivityFanOut\")\n\t\t\tif assert.NoError(t, err) {\n\t\t\t\tmetadata, err := client.WaitForOrchestrationCompletion(ctx, id)\n\t\t\t\tif assert.NoError(t, err) {\n\t\t\t\t\tassert.True(t, metadata.IsComplete())\n\t\t\t\t\tassert.Equal(t, `[\"9\",\"8\",\"7\",\"6\",\"5\",\"4\",\"3\",\"2\",\"1\",\"0\"]`, metadata.SerializedOutput)\n\n\t\t\t\t\t// Because all the activities run in parallel, they should complete very quickly\n\t\t\t\t\tassert.Less(t, metadata.LastUpdatedAt.Sub(metadata.CreatedAt), 3*time.Second)\n\t\t\t\t}\n\t\t\t}\n\t\t})\n\t}\n}", "func (s *SetFilter) Proceed(ctx context.Context, input chan structs.Message, output chan structs.Message) (err error) {\n\ts.log.Printf(\"Set filter started. Total items: %d\", len(s.Updates))\n\n\tfor msg := range input {\n\t\tpayload := msg.Payload\n\n\t\tfor key, value := range s.Updates {\n\t\t\tpayload[key] = value\n\t\t}\n\t\tmsg.Payload = payload\n\n\t\toutput <- msg\n\t}\n\n\ts.log.Printf(\"Channel processing finished. Exiting\")\n\n\treturn\n}", "func (p *pipe) outputAdvance(count int) {\n\tp.outPos += int32(count)\n\tif p.outPos >= p.size {\n\t\tp.outPos -= p.size\n\t}\n\tatomic.AddInt32(&p.free, int32(count))\n\n\tselect {\n\tcase p.inWake <- struct{}{}:\n\tdefault:\n\t}\n}", "func TestBufferedStorageClientWorker(t *testing.T) {\n\t// Disable the logger\n\tlog.SetOutput(ioutil.Discard)\n\n\t// Define the job Queue and the Buffered Storage Client\n\tjobQueue = make(chan *Job, 10)\n\tstorageClient = &BufferedStorageClient{}\n\n\t// Make testing.T and the response global\n\tT = t\n\tsResp = nil\n\n\t// Create a worker\n\tt.Log(\"Creating a single worker\")\n\tpool := make(chan chan *Job, 2)\n\tworker := NewWorker(1, 10, pool)\n\tworker.Start()\n\n\t// Stop the worker on the end\n\tvar wg sync.WaitGroup\n\twg.Add(1)\n\tdefer worker.Stop(&wg)\n\n\t// Start the test\n\tjobChannel := <-pool\n\tvar job Job\n\n\tt.Log(\"Creating 9 job and send it to the worker\")\n\tpartStr := \"\"\n\tfor i := 0; i < 9; i++ {\n\t\tjob = Job{GetTestEvent(uint32(56746535 + i)), 1}\n\t\tpart, _ := dialects.ConvertJSON(job.Event)\n\t\tpartStr += part.String()\n\t\tjobChannel <- &job\n\t\tjobChannel = <-pool\n\n\t\tif exnr := i + 1; len(worker.BufferedEvents) != exnr {\n\t\t\tt.Errorf(\"Worker's buffered events count should be %d but it was %d instead\", exnr, len(worker.BufferedEvents))\n\t\t}\n\t}\n\n\tt.Log(\"Creating the 10th job and send it to the worker that will proceed the buffer\")\n\tjob = Job{GetTestEvent(1), 1}\n\tpart, _ := dialects.ConvertJSON(job.Event)\n\tpartStr += part.String()\n\texp = bytes.NewBuffer([]byte(partStr))\n\tjobChannel <- &job\n\tjobChannel = <-pool\n\n\tif exnr := 0; len(worker.BufferedEvents) != exnr {\n\t\tt.Errorf(\"Worker's buffered events count should be %d but it was %d instead\", exnr, len(worker.BufferedEvents))\n\t}\n\tif expen := float32(1.0); worker.Penalty != expen {\n\t\tt.Errorf(\"Expected worker's penalty was %d but it was %d instead\", expen, worker.Penalty)\n\t}\n\tif exnr := 10; worker.GetBufferSize() != exnr {\n\t\tt.Errorf(\"Expected worker's buffer size after the error was %d but it was %d instead\", exnr, worker.GetBufferSize())\n\t}\n\n\tt.Log(\"Creating 14 job and send it to the worker, during the process it'll fail after the 10th\")\n\tsResp = fmt.Errorf(\"Error was intialized for testing\")\n\tpartStr = \"\"\n\tfor i := 0; i < 14; i++ {\n\t\tjob = Job{GetTestEvent(uint32(213432 + i)), 1}\n\t\tpart, _ := dialects.ConvertJSON(job.Event)\n\t\tpartStr += part.String()\n\t\tjobChannel <- &job\n\t\tjobChannel = <-pool\n\n\t\tif exnr := i + 1; len(worker.BufferedEvents) != exnr {\n\t\t\tt.Errorf(\"Worker's buffered events count should be %d but it was %d instead\", exnr, len(worker.BufferedEvents))\n\t\t}\n\t}\n\n\tif expen := float32(1.5); worker.Penalty != expen {\n\t\tt.Errorf(\"Expected worker's penalty was %d but it was %d instead\", expen, worker.Penalty)\n\t}\n\tif exnr := 15; worker.GetBufferSize() != exnr {\n\t\tt.Errorf(\"Expected worker's buffer size after the error was %d but it was %d instead\", exnr, worker.GetBufferSize())\n\t}\n\n\tsResp = nil\n\tt.Log(\"Creating the 15th job and send it to the worker that will proceed the buffer\")\n\tjob = Job{GetTestEvent(1), 1}\n\tpart, _ = dialects.ConvertJSON(job.Event)\n\tpartStr += part.String()\n\texp = bytes.NewBuffer([]byte(partStr))\n\tjobChannel <- &job\n\tjobChannel = <-pool\n\n\tif exnr := 0; len(worker.BufferedEvents) != exnr {\n\t\tt.Errorf(\"Worker's buffered events count should be %d but it was %d instead\", exnr, len(worker.BufferedEvents))\n\t}\n\tif expen := float32(1.0); worker.Penalty != expen {\n\t\tt.Errorf(\"Expected worker's penalty was %d but it was %d instead\", expen, worker.Penalty)\n\t}\n\tif exnr := 10; worker.GetBufferSize() != exnr {\n\t\tt.Errorf(\"Expected worker's buffer size after the error was %d but it was %d instead\", exnr, worker.GetBufferSize())\n\t}\n}", "func Synchronising() {\n\tdone := make(chan bool, 1)\n\n\tgo worker(done)\n\n\t<-done\n}", "func main() {\n\tsize := 3\n\n\tvar buffChan = make(chan int, size)\n\n\t// reader\n\tgo func(){\n\t\tfor {\n\t\t\t_ = <- buffChan\n\t\t\ttime.Sleep(time.Second * 3)\n\t\t}\n\t}()\n\n\t// writer\n\twriter := func() {\n\t\tfor i := 1 ; i <= 10; i++ {\n\t\t\tbuffChan <- i\n\t\t\tprintln(i)\n\t\t}\n\t}\n\n\twriter()\n}", "func (q *priorityLocalQueue) Results(ctx context.Context) <-chan amboy.Job {\n\toutput := make(chan amboy.Job)\n\n\tgo func() {\n\t\tdefer close(output)\n\t\tfor job := range q.storage.Contents() {\n\t\t\tif ctx.Err() != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif job.Status().Completed {\n\t\t\t\toutput <- job\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn output\n}", "func (p *asyncProducerMock) acknowledge(n int, ch chan *sarama.ProducerMessage) {\n\tfor n > 0 {\n\t\tvar outstanding []*sarama.ProducerMessage\n\t\tp.mu.Lock()\n\t\toutstanding = append(outstanding, p.mu.outstanding...)\n\t\tp.mu.outstanding = p.mu.outstanding[:0]\n\t\tp.mu.Unlock()\n\n\t\tfor _, m := range outstanding {\n\t\t\tch <- m\n\t\t}\n\t\tn -= len(outstanding)\n\t}\n}", "func (s FakeSerial) RunSerialTx(writeChan <-chan []byte) error {\n\t// Iterate over the channel looking for new stuff to shoot out over serial\n\tfor v := range writeChan {\n\t\tfmt.Println(\"Serial Write: \", v)\n\t}\n\treturn nil\n}", "func work(messageChan chan response, messageWg *sync.WaitGroup, work chan int, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\tfor {\n\t\tselect {\n\t\tcase _ = <-work:\n\t\t\tmessageWg.Add(1)\n\t\t\tmakeRequest(messageChan)\n\t\tdefault:\n\t\t\treturn\n\t\t}\n\t}\n}", "func main() {\r\n\r\n\t//Creating a Buffered channel of size 3, holding any type of values\r\n\tc := make(chan interface{}, 3)\r\n\tfmt.Println(c)\r\n\r\n\t//Creating a Structure\r\n\ttype stud struct{\r\n\r\n\t\tName string\r\n\t}\r\n\t//Sender Go Routine(Anonymous) to send an integer, two structure objects and a string\r\n\t//If the size exceeds 3, for every item after 3rd, once an item is sniffed, nnext item will be sent\r\n\tgo func(){\r\n\r\n\t\tc <- 1\r\n\t\tc <- stud{\"Kailash\"}\r\n\t\tc <- stud{\"SK7\"}\r\n\t\tc <- \"End of Channel items\"\r\n\r\n\t\t//\"close\" is used to indicate end of items in channel when iterating over items in buffered channel\r\n\t\t//If not used, leads to deadlock, as iterator will forever be expecting items in channel\r\n\t\tclose(c)\r\n\t}()\r\n\r\n\t//When using 2 or more Go Routines, give a sleep time of 1 second after every Go Routine to compensate for concurrency\r\n\ttime.Sleep(time.Second * 1)\r\n\r\n\t//Sniff Go Routine(Anonymous) to receive the channel items\r\n\tgo func(){\r\n\r\n\t\tfor i := range c{\r\n\r\n\t\t\tfmt.Println(i)\r\n\t\t}\r\n\t}()\r\n\r\n\ttime.Sleep(time.Second * 1)\r\n}", "func main() {\n\tgo produce()\n\tgo consume()\n\t<-done\n}", "func f(tester *exec.Cmd, testee *exec.Cmd, giveInput chan io.Writer) {\n var(\n testerBuf *bytes.Buffer\n testeeBuf *bytes.Buffer\n testerArr = make([]byte, 1024)\n testeeArr = make([]byte, 1024)\n )\n\n testerIn, err := tester.StdinPipe()\n testeeIn, err := testee.StdinPipe()\n testerOut, err := tester.StdoutPipe()\n testeeOut, err := testee.StdoutPipe()\n\n if err != nil {\n return nil\n }\n\n stdin := io.MultiWriter(testerIn, testeeIn)\n giveInput <- stdin\n\n for {\n a, err := io.ReadFull(testerOut, testerArr)\n b, err := io.ReadFull(testeeOut, testeeArr)\n\n switch(err) {\n case io.EOF:\n //do something\n case nil:\n //do nothing\n default:\n //??\n }\n\n if(flag && !bytes.Equal(testerArr, testeeArr)) {\n // error - wrong answer, send signals to other threads\n flag = false\n }\n }\n\n return\n}", "func writeToArchive(work <-chan t, done <-chan struct{}) chan t {\n\tout := make(chan t)\n\tgo func() {\n\t\tdefer close(out)\n\t\tfanout := make([]<-chan t, *concurrency) // HL\n\t\tfor j := 0; j < *concurrency; j++ {\n\t\t\tfanout[j] = doWrite(work, done) // HL\n\t\t}\n\n\t\tfor merged := range merge(fanout, done) { // HL\n\t\t\tselect {\n\t\t\tcase out <- merged:\n\t\t\tcase <-done:\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t}()\n\treturn out\n}", "func queue(initialTasks []*task, in <-chan *task, doneTask <-chan *task, out chan<- *task) {\n\taddedKeys := make(map[string]struct{})\n\tvar q linkedQueue\n\tfor _, t := range initialTasks {\n\t\tif _, ok := addedKeys[t.key]; ok {\n\t\t\t// already added this key, skip it\n\t\t\tcontinue\n\t\t}\n\t\tq.pushRight(t)\n\t}\n\tincompleteTasks := len(initialTasks)\nLoop:\n\tfor incompleteTasks > 0 {\n\t\tvar sendChan chan<- *task\n\t\tcurrentTask := q.popLeft()\n\t\tif currentTask != nil {\n\t\t\tsendChan = out\n\t\t}\n\t\tselect {\n\t\tcase t, ok := <-in:\n\t\t\tif currentTask != nil {\n\t\t\t\t// need to restore the task for next iteration.\n\t\t\t\tq.pushLeft(currentTask)\n\t\t\t}\n\t\t\tif !ok {\n\t\t\t\tin = nil\n\t\t\t\tcontinue Loop\n\t\t\t}\n\t\t\tif _, ok := addedKeys[t.key]; ok {\n\t\t\t\t// already added this key, skip it\n\t\t\t\tcontinue Loop\n\t\t\t}\n\t\t\taddedKeys[t.key] = struct{}{}\n\t\t\tq.pushRight(t)\n\t\t\tincompleteTasks++\n\t\tcase sendChan <- currentTask:\n\t\t\t// successfully sent\n\t\tcase _, ok := <-doneTask:\n\t\t\tif currentTask != nil {\n\t\t\t\t// need to restore the task for next iteration.\n\t\t\t\tq.pushLeft(currentTask)\n\t\t\t}\n\t\t\tif ok {\n\t\t\t\tincompleteTasks--\n\t\t\t}\n\t\t}\n\t}\n}", "func (ch *RingChannel) In() chan<- interface{} {\n\treturn ch.input\n}", "func TestArray(t *testing.T) {\n\tt.Skip(\"Bug(sickyoon): array as an argument returns nil: https://github.com/gocelery/gocelery/issues/74\")\n\ttestCases := []struct {\n\t\tname string\n\t\tbroker CeleryBroker\n\t\tbackend CeleryBackend\n\t\ttaskName string\n\t\ttaskFunc interface{}\n\t\tinA []string\n\t\tinB []string\n\t\texpected []string\n\t}{\n\t\t{\n\t\t\tname: \"array addition with redis broker/backend\",\n\t\t\tbroker: redisBroker,\n\t\t\tbackend: redisBackend,\n\t\t\ttaskName: uuid.Must(uuid.NewV4(), nil).String(),\n\t\t\ttaskFunc: addArr,\n\t\t\tinA: []string{\"a\", \"b\", \"c\", \"d\"},\n\t\t\tinB: []string{\"e\", \"f\", \"g\", \"h\"},\n\t\t\texpected: []string{\"a\", \"b\", \"c\", \"d\", \"e\", \"f\", \"g\", \"h\"},\n\t\t},\n\t}\n\tfor _, tc := range testCases {\n\t\tcli, _ := NewCeleryClient(tc.broker, tc.backend, 1)\n\t\tcli.Register(tc.taskName, tc.taskFunc)\n\t\tcli.StartWorker()\n\t\tasyncResult, err := cli.Delay(tc.taskName, tc.inA, tc.inB)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"test '%s': failed to get result for task %s: %+v\", tc.name, tc.taskName, err)\n\t\t\tcli.StopWorker()\n\t\t\tcontinue\n\t\t}\n\t\tres, err := asyncResult.Get(TIMEOUT)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"test '%s': failed to get result for task %s: %+v\", tc.name, tc.taskName, err)\n\t\t\tcli.StopWorker()\n\t\t\tcontinue\n\t\t}\n\t\tif !reflect.DeepEqual(tc.expected, res.([]string)) {\n\t\t\tt.Errorf(\"test '%s': returned result %+v is different from expected result %+v\", tc.name, res, tc.expected)\n\t\t}\n\t\tcli.StopWorker()\n\t}\n}", "func (a *Actor) Issue(order func()) {\n a.orders <- func() {\n a.loop = false\n order()\n }\n}", "func (r *InstanceRead) Intake() chan msg.Request {\n\treturn r.Input\n}", "func sumWorker(nums chan int, out chan int) {\n\t// TODO: implement me\n\t// HINT: use for loop over `nums`\n\tvar sum int\n\tdefer wg.Done()\n\tcounterA:= len(nums)\n\tfor i := 0; i<counterA; i++ {\n\t\telement := <- nums\n\t\tsum+=element\n\t}\n\tout <- sum\n}", "func main() {\n\tresultChan := make(chan *result)\n\n\t//Async\n\tgo doWorkAsync(resultChan)\n\t//Sync\n\tresult := doWork()\n\n\t//Wait for Async to catch up\n\tresultAsync := <-resultChan\n\t//Merge results\n\tfmt.Println(result.success && resultAsync.success)\n}", "func in(channel chan <- string, msg string) {\n\tchannel <- msg\n}", "func queryWorker(\n workerIndex int,\n in <-chan string,\n out chan<- result,\n group *sync.WaitGroup,\n client *api.BingClient) {\n\n defer group.Done()\n\n for queryString := range in {\n if queryString == \"\" { continue }\n log.Printf(\"[worker:%d] sending search string: %s\", workerIndex, queryString)\n\n currOffset := 0\n running := true\n var err error\n for running {\n params := api.CreateQuery(queryString, currOffset)\n paramsString := params.AsQueryParameters()\n log.Printf(\"[worker:%d] running query with params: %s\", workerIndex, paramsString)\n images := client.RequestImages(params)\n if images.Values == nil {\n err = fmt.Errorf(\"[worker:%d] failed to pull query: %s/%s\",\n workerIndex, client.Endpoint, paramsString)\n running = false\n } else {\n running = images.NextOffset != currOffset\n currOffset = images.NextOffset\n }\n out <- result{images, err}\n }\n }\n\n log.Printf(\"[worker:%d] terminated\", workerIndex)\n}", "func (ig *Instagram) consumer(results chan *InstagramResponse, done chan []InstagramData, simultaneous int, keepCriterion func(InstagramData) bool) {\n\tvar igData []InstagramData\n\n\ttimeout := make(chan bool)\n\tgo func() {\n\t\t//TODO: Make timeout a config'able setting\n\t\ttime.Sleep(10 * time.Second)\n\t\ttimeout <- true\n\t}()\n\n\ti := 0\n\t//Pull down results forever until we've hit some satisfaction criterion\n\tfor {\n\t\ti++\n\t\tselect {\n\n\t\tcase <-timeout:\n\t\t\tfmt.Printf(\"We timed out after %d attempts. Unlocking.\\n\", (i - 1))\n\n\t\t\t//Drain the channel\n\t\t\tgo drain(results, simultaneous+1) //+1 for timeout case\n\t\t\tdone <- igData //Semaphore{}\n\t\t\tfmt.Printf(\"Unlocked\\n\")\n\t\t\treturn\n\n\t\tcase res, ok := <-results:\n\t\t\tif !ok {\n\t\t\t\tfmt.Printf(\"I is %d but the channel is closed.\", i)\n\t\t\t\tdone <- igData //Semaphore{}\n\t\t\t\tfmt.Printf(\"Unlocked\\n\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfmt.Printf(\"%s is overall #%d\\n\", res, i)\n\n\t\t\t//Append\n\t\t\tfor _, datum := range res.Data {\n\t\t\t\tif keepCriterion(datum) {\n\t\t\t\t\tigData = append(igData, datum)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif !ig.satisfied(igData, i) {\n\t\t\t\t//If it's not happy after this result, the consumer\n\t\t\t\t// instructs a producer to start on something new\n\t\t\t\t//job := URL(random(300))\n\t\t\t\tfmt.Printf(\"Consumer is not satisfied after job #%d. Fetching next: %d\\n\", i, res.Pagination.NextURL)\n\t\t\t\tif res.Pagination.NextURL == \"\" {\n\t\t\t\t\tfmt.Printf(\"Consumer is not satisfied after job #%d but no next page was provided.\", i)\n\t\t\t\t\t//1 fewer goroutine is running at the same time\n\t\t\t\t\tsimultaneous = simultaneous - 1\n\t\t\t\t\tif simultaneous == 0 {\n\t\t\t\t\t\tgo drain(results, simultaneous)\n\t\t\t\t\t\t//Nothing worked.\n\t\t\t\t\t\tdone <- igData\n\t\t\t\t\t\tfmt.Printf(\"Unlocked\\n\")\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tgo ig.producer(results, res.Pagination.NextURL)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tfmt.Printf(\"Consumer is satisfied after job #%d. Unlocking.\\n\", i)\n\n\t\t\t\t//Drain the channel\n\t\t\t\tgo drain(results, simultaneous)\n\t\t\t\tdone <- igData //Semaphore{}\n\t\t\t\tfmt.Printf(\"Unlocked\\n\")\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n}", "func Transform(ctx context.Context, parallelism int, bufferSize int, in chan OutResult,\n\ttransformer func(interface{}) (interface{}, error), errhandler func(error),\n) chan InOutResult {\n\t// TODO: can we have a channel factory to do this?\n\toutChan := make(chan InOutResult, bufferSize)\n\tvar wg sync.WaitGroup\n\tif parallelism < 1 {\n\t\tparallelism = 1\n\t}\n\twg.Add(parallelism)\n\ti := func() {\n\t\tdefer wg.Done()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\tout := simpleInOut{\n\t\t\t\t\tsimpleOut: simpleOut{err: ctx.Err()},\n\t\t\t\t}\n\t\t\t\tselect {\n\t\t\t\tcase outChan <- out:\n\t\t\t\tdefault:\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\tcase sr, ok := <-in:\n\t\t\t\t// do stuff, write to out maybe\n\t\t\t\tif !ok {\n\t\t\t\t\t// channel is closed, time to exit\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif sr.Err() != nil {\n\t\t\t\t\tif errhandler != nil {\n\t\t\t\t\t\terrhandler(sr.Err())\n\t\t\t\t\t}\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tres, err := transformer(sr.Output())\n\t\t\t\tif err == ErrSkip {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tout := simpleInOut{\n\t\t\t\t\tsimpleOut: simpleOut{err: err, out: res},\n\t\t\t\t\tin: sr.Output(),\n\t\t\t\t}\n\t\t\t\t// TODO: this section will never cancel if this write blocks. Problem?\n\t\t\t\toutChan <- out\n\t\t\t}\n\t\t}\n\t}\n\tfor x := 0; x < parallelism; x++ {\n\t\tgo i()\n\t}\n\tgo func() {\n\t\twg.Wait()\n\t\tclose(outChan)\n\t}()\n\treturn outChan\n}", "func (d *disp) process() {\n\tfor {\n\t\tselect {\n\t\tcase pipeline := <-d.PipelineChan:\n\n\t\t\tPipelineChan := <-d.PipelineQueue\n\n\t\t\tPipelineChan <- pipeline\n\n\t\tcase job := <-d.JobChan: // listen to any submitted job on the WorkChan\n\t\t\t// wait for a worker2 to submit JobChan to JobQueue\n\t\t\t// note that this JobQueue is shared among all workers.\n\t\t\t// Whenever there is an available JobChan on JobQueue pull it\n\t\t\tJobChan := <-d.Queue\n\n\t\t\t// Once a jobChan is available, send the submitted Job on this JobChan\n\t\t\tJobChan <- job\n\t\t}\n\t}\n}", "func (j *jobsheet) work() {\n\tfor {\n\t\tselect {\n\t\tcase worksheet := <-j.pendingWorkload:\n\t\t\tworksheet.Work()\n\t\tcase <-j.stopWork:\n\t\t\treturn\n\t\t}\n\t}\n}", "func (op Operator) transfer(path, format, target string, lim int64, c chan string) {\n\tinsc := make(chan string)\n\tvideos := op.search(target, lim)\n\tprintSR(videos)\n\n\tfor id, item := range videos {\n\t\tgo dWorker(path+item.Snippet.Title, id, format, insc)\n\t}\n\n\tfor i := 0; i < len(videos); i++ {\n\t\tfmt.Println(<-insc)\n\t}\n\tc <- \"done in \" + time.Now().String()\n}", "func TestSequential(t *testing.T) {\n\tt.Parallel()\n\tctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)\n\tt.Cleanup(cancel)\n\n\tconfig := Config{MaxConcurrency: 4}\n\tcountdown := NewCountdown(config.MaxConcurrency)\n\tprocess := NewMockEventsProcess(ctx, t, config, func(ctx context.Context, event types.Event) error {\n\t\tdefer countdown.Decrement()\n\t\ttime.Sleep(time.Second)\n\t\treturn trace.Wrap(ctx.Err())\n\t})\n\n\ttimeBefore := time.Now()\n\tfor i := 0; i < config.MaxConcurrency; i++ {\n\t\tresource, err := types.NewAccessRequest(\"REQ-SAME\", \"foo\", \"admin\")\n\t\trequire.NoError(t, err)\n\t\tprocess.Events.Fire(types.Event{Type: types.OpPut, Resource: resource})\n\t}\n\trequire.NoError(t, countdown.Wait(ctx))\n\n\ttimeAfter := time.Now()\n\tassert.InDelta(t, 4*time.Second, timeAfter.Sub(timeBefore), float64(750*time.Millisecond))\n}", "func reader(threads int, parallel bool, queue Queue, wg *sync.WaitGroup, shots []Shot, m map[string]int) {\n\n\t// Add jobs from the input to the queue\n\tscanner := bufio.NewScanner(os.Stdin)\n\tfor scanner.Scan(){\n\t\ttempJob := scanner.Text()\n\t\ttempJobP := unsafe.Pointer(&tempJob)\n\t\tqueue.enq(&tempJobP)\n\n\t}\n\n\t// Boolean that signals when to stop reading from queue\n\tvar done bool\n\n\t// Variable that checks if we have gone through all items from the queue\n\tcounter := int32(0)\n\n\t// Value to reach before signaling done to be true\n\ttail := queue.end\n\n\t// Dequeue until done is true\n\tfor !done{\n\n\t\t// If there are still tasks to read\n\t\tif counter != tail{\n\n\t\t\t// Grab a job\n\t\t\ttemp := queue.deq()\n\n\t\t\t// Visualization Job\n\t\t\tif strings.Contains(temp, \"visualize\"){\n\n\t\t\t\t// Run Parallel Version\n\t\t\t\tif parallel{\n\t\t\t\t\twg.Add(1)\n\t\t\t\t\tgo visualizationWorker(temp, threads, wg, parallel, shots, m)\n\n\t\t\t\t\t// Run Sequential Version\n\t\t\t\t} else{\n\t\t\t\t\tvisualizationWorker(temp, threads, wg, parallel, shots, m)\n\t\t\t\t}\n\t\t\t// ML Model Job\n\t\t\t} else if strings.Contains(temp, \"ml\"){\n\t\t\t\t\n\t\t\t\t// Run Parallel Version\n\t\t\t\tif parallel{\n\t\t\t\t\twg.Add(1)\n\t\t\t\t\tgo mlWorker(temp, threads, wg, parallel)\n\n\t\t\t\t// Run Sequential Version\n\t\t\t\t}else{\n\t\t\t\t\tmlWorker(temp, threads, wg, parallel)\n\n\t\t\t\t}\n\n\t\t\t}\n\t\t\t// Increment Counter\n\t\t\tcounter++\n\t\t\n\t\t// We have read everything\n\t\t// break out of for loop\n\t\t} else {\n\n\t\t\tdone = true\n\t\t}\n\n\t}\n\n}", "func printNumsBuffered(consumeChannel chan int, endChannel chan bool) {\r\n\t// Continually recieve and print values from the provided channel until it is closed\r\n\tfor number := range consumeChannel {\r\n\t\tfmt.Println(\"Got number: \", number)\r\n\r\n\t\ttime.Sleep(time.Second)\r\n\t}\r\n\r\n\t// Indicate we are finished to the provided channel\r\n\tendChannel <- true\r\n}", "func ExecuteInParallel(q *Queue, fn func(interface{})) {\n\tif q == nil {\n\t\treturn\n\t}\n\n\tq.lock.Lock()\n\ttodo, done := int64(len(q.items)), int64(-1)\n\tif todo == 0 {\n\t\treturn\n\t}\n\n\tnumCPU := 1\n\tif runtime.NumCPU() > 1 {\n\t\tnumCPU = runtime.NumCPU() - 1\n\t}\n\n\tvar wg sync.WaitGroup\n\twg.Add(numCPU)\n\titems := q.items\n\n\tfor i := 0; i < numCPU; i++ {\n\t\tgo func() {\n\t\t\tfor {\n\t\t\t\tindex := atomic.AddInt64(&done, 1)\n\t\t\t\tif index >= todo {\n\t\t\t\t\twg.Done()\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tfn(items[index])\n\t\t\t\titems[index] = 0\n\t\t\t}\n\t\t}()\n\t}\n\n\twg.Wait()\n\tq.lock.Unlock()\n\tq.Dispose()\n}", "func AnyDoneSlice(inp <-chan Any) (done <-chan []Any) {\n\tsig := make(chan []Any)\n\tgo doneAnySlice(sig, inp)\n\treturn sig\n}", "func (r *CheckConfigurationRead) Intake() chan msg.Request {\n\treturn r.Input\n}", "func (ep *defaultEventProcessor) waitUntilInactive() {\n\tm := syncEventsMessage{replyCh: make(chan struct{})}\n\tep.inputCh <- m\n\t<-m.replyCh // Now we know that all events prior to this call have been processed\n}", "func main() {\n\tprintln(\"start - async\")\n\tgo process(nil)\n\tprintln(\"end - async\")\n\n\ttime.Sleep(time.Second)\n\tprintln(\"---------\")\n\n\tsig := make(chan int)\n\tprintln(\"start - sync\")\n\tgo process(sig)\n\t<-sig\n\tprintln(\"end - sync\")\n}", "func (client *Client) writeJob(job *Job) {\n client.out <- job\n}", "func worker(done chan bool) {\r\n\tfmt.Print(\"working\")\r\n\tseconds := 10\r\n\tfor i:= 1; i <= seconds; i++ {\r\n\t\tfmt.Print(\".\")\r\n\t\ttime.Sleep(time.Second)\r\n\t}\r\n\t//after delay send notofication to channel\r\n\tfmt.Println(\"done the work\")\r\n\tdone <- true \r\n}" ]
[ "0.5639781", "0.5458441", "0.51682293", "0.51466763", "0.50523716", "0.50321937", "0.49681675", "0.4961015", "0.4910961", "0.48957968", "0.48762244", "0.48419768", "0.4810957", "0.47938824", "0.4775782", "0.47745758", "0.47491193", "0.47487912", "0.4732034", "0.47135037", "0.47122923", "0.4710638", "0.47091413", "0.47047526", "0.47008064", "0.4695484", "0.4690679", "0.46783498", "0.46765015", "0.46745548", "0.46690097", "0.46688053", "0.46555644", "0.4655261", "0.46534088", "0.46517292", "0.4634863", "0.46321297", "0.46284282", "0.46204185", "0.46160248", "0.46105814", "0.45861468", "0.45707723", "0.4564629", "0.45603603", "0.45564088", "0.45556295", "0.4554878", "0.45536003", "0.45501184", "0.45478562", "0.45462134", "0.45444426", "0.4543071", "0.4531207", "0.45262295", "0.4519472", "0.45189175", "0.45114785", "0.45101812", "0.4508767", "0.4499405", "0.44983783", "0.44932353", "0.44879842", "0.4483157", "0.44823158", "0.44811445", "0.44781205", "0.4476046", "0.4465586", "0.44645947", "0.4462125", "0.44614595", "0.4455205", "0.44483075", "0.4446695", "0.44442064", "0.44441888", "0.44380367", "0.44380304", "0.44361785", "0.44279915", "0.44277146", "0.44272768", "0.44220695", "0.44138753", "0.4411125", "0.44104698", "0.44086537", "0.44062215", "0.44041446", "0.44033426", "0.44011846", "0.43993917", "0.4398251", "0.43978494", "0.43928966", "0.43919653" ]
0.7534587
0
InParallel does work in parallel and should write results on channels as work item complete
func InParallel(ch chan Result, work []int) []Result { results := make([]Result, len(work)) var mutex = &sync.Mutex{} for i, num := range work { go func(w int, res *Result, ch chan Result) { *res = executeWork(w) mutex.Lock() ch <- *res mutex.Unlock() }(num, &results[i], ch) } return results }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *Pool) worker() {\n\tdefer p.wg.Done()\n\tfor job := range p.In {\n\t\tatomic.AddInt64(&p.Stats.Pending, -1)\n\t\tatomic.AddInt64(&p.Stats.Running, 1)\n\t\tjob.Result, job.Error = job.F(job.Args...)\n\t\tatomic.AddInt64(&p.Stats.Running, -1)\n\t\tatomic.AddInt64(&p.Stats.Completed, 1)\n\t\tp.Out <- job\n\t}\n}", "func (h *connHandler) parallel(m *reqData, args []interface{}) {\n\th.Lock()\n\theap.Push(&h.pQueue, m)\n\th.Unlock()\n\n\tm.resp = h.srv.pool.Cmd(m.cmd, args...)\n\n\th.Lock()\n\tfor h.pQueue.Len() > 0 {\n\t\titem := heap.Pop(&h.pQueue).(*reqData)\n\t\tif item.resp == nil {\n\t\t\theap.Push(&h.pQueue, item)\n\t\t\tbreak\n\t\t}\n\t\tif item.answerCh != nil {\n\t\t\titem.answerCh <- item.resp\n\t\t}\n\t}\n\th.Unlock()\n}", "func writingWorker(\n workerIndex int,\n outputFolder string,\n in <-chan result,\n group *sync.WaitGroup,\n exportFunc io.Exporter) {\n\n defer group.Done()\n\n for result := range in {\n outputFile := path.Join(outputFolder, utils.SimpleRandomString(20))\n log.Printf(\"[worker:%d] saving file %s\", workerIndex, outputFile)\n if result.err != nil {\n log.Printf(result.err.Error())\n } else {\n query := result.collection.Query\n log.Printf(\n \"[worker:%d] exporting query results for '%s' into file '%s\",\n workerIndex, query, outputFile)\n if err := exportFunc(result.collection, outputFile); err != nil {\n log.Printf(err.Error())\n }\n }\n }\n\n log.Printf(\"[worker:%d] terminated\", workerIndex)\n}", "func worker(id int, jobs <-chan int, res chan<- int) {\n\n\tfmt.Printf(\"Worker %d : started\\n\", id)\n\tfor job := range jobs {\n\t\tfmt.Printf(\"worker %d : received job %d\\n\", id, job)\n\t\ttime.Sleep(time.Second)\n\t\tres <- job * job\n\t\tfmt.Printf(\"worker %d : finished job %d\", id, job)\n\t}\n\t//close(results)\n\t//putting close here and having a range call over the results channel below wont work becuase,\n\t//the first worker to finish its jobs closes the results channel, and other workers wont be able to send their results\n\t//sending to a closed channel will cause panic\n\n}", "func main() {\n\tin, out := make(chan *Work), make(chan *Work)\n\tfor i := 0; i < NumWorkers; i++ {\n\t\tgo worker(in, out)\n\t}\n\tgo sendLotsOfWork(in)\n\treceiveLotsOfResults(out)\n}", "func (t *T) Parallel()", "func ExecuteInParallel(q *Queue, fn func(interface{})) {\n\tif q == nil {\n\t\treturn\n\t}\n\n\tq.lock.Lock()\n\ttodo, done := int64(len(q.items)), int64(-1)\n\tif todo == 0 {\n\t\treturn\n\t}\n\n\tnumCPU := 1\n\tif runtime.NumCPU() > 1 {\n\t\tnumCPU = runtime.NumCPU() - 1\n\t}\n\n\tvar wg sync.WaitGroup\n\twg.Add(numCPU)\n\titems := q.items\n\n\tfor i := 0; i < numCPU; i++ {\n\t\tgo func() {\n\t\t\tfor {\n\t\t\t\tindex := atomic.AddInt64(&done, 1)\n\t\t\t\tif index >= todo {\n\t\t\t\t\twg.Done()\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tfn(items[index])\n\t\t\t\titems[index] = 0\n\t\t\t}\n\t\t}()\n\t}\n\n\twg.Wait()\n\tq.lock.Unlock()\n\tq.Dispose()\n}", "func main() {\n\t// EXAMPLE 1\n\tch := make(chan string, 2)\n\tch <- \"naveen\"\n\tch <- \"paul\"\n\tfmt.Println(<-ch)\n\tfmt.Println(<-ch)\n\n\t// EXAMPLE 2\n\tch2 := make(chan int, 2)\n\tgo write(ch2)\n\ttime.Sleep(2 * time.Second)\n\tfor v := range ch2 {\n\t\tfmt.Println(\"read value\", v, \"from ch2\")\n\t\ttime.Sleep(2 * time.Second)\n\n\t}\n\n\t// LENGTH VS CAPACITY\n\t// length: number of elements currently queued in it\n\t// capacity: number of values that the channel can hold\n\tch3 := make(chan string, 3)\n\tch3 <- \"naveen\"\n\tch3 <- \"paul\"\n\tfmt.Println(\"capacity is\", cap(ch3))\n\tfmt.Println(\"length is\", len(ch3))\n\tfmt.Println(\"read value\", <-ch3)\n\tfmt.Println(\"capacity is\", cap(ch3))\n\tfmt.Println(\"new length is\", len(ch3))\n\n\t// WAIT GROUP: used to wait for a collection of Goroutines to finish executing, control is blocked until all Goroutines are are finished\n\tno := 3\n\tvar wg sync.WaitGroup\n\tfor i := 0; i < no; i++ {\n\t\twg.Add(1) // increment counter value\n\t\tgo process(i, &wg) // must be the address, otherwise each goroutine will have its onw copy of the WaitGroup\n\t}\n\twg.Wait() // blocks the Goroutine until the counter becomes zero\n\tfmt.Println(\"All go routines finished executing\")\n\n\t// WORKER POOL\n\tstartTime := time.Now()\n\n\tnoOfJobs := 100\n\tgo allocate(noOfJobs)\n\n\tdone := make(chan bool)\n\tgo result(done)\n\n\tnoOfWorkers := 10\n\tcreateWorkerPool(noOfWorkers)\n\n\t<-done\n\tendTime := time.Now()\n\tdiff := endTime.Sub(startTime)\n\tfmt.Println(\"total time taken\", diff.Seconds(), \"seconds\")\n}", "func queryWorker(\n workerIndex int,\n in <-chan string,\n out chan<- result,\n group *sync.WaitGroup,\n client *api.BingClient) {\n\n defer group.Done()\n\n for queryString := range in {\n if queryString == \"\" { continue }\n log.Printf(\"[worker:%d] sending search string: %s\", workerIndex, queryString)\n\n currOffset := 0\n running := true\n var err error\n for running {\n params := api.CreateQuery(queryString, currOffset)\n paramsString := params.AsQueryParameters()\n log.Printf(\"[worker:%d] running query with params: %s\", workerIndex, paramsString)\n images := client.RequestImages(params)\n if images.Values == nil {\n err = fmt.Errorf(\"[worker:%d] failed to pull query: %s/%s\",\n workerIndex, client.Endpoint, paramsString)\n running = false\n } else {\n running = images.NextOffset != currOffset\n currOffset = images.NextOffset\n }\n out <- result{images, err}\n }\n }\n\n log.Printf(\"[worker:%d] terminated\", workerIndex)\n}", "func workers(done *int, jobs <-chan int, arrayA, arrayB, arrayF *arr) {\r\n\t// Inicia uma trhead para cada linha da matriz.\r\n\tfor job := range jobs {\r\n\t\tgo partial(job, done, arrayA, arrayB, arrayF)\r\n\t}\r\n}", "func worker(id int, jobs <-chan int, results chan<- int) {\n\tfor j := range jobs {\n\t\tfmt.Println(\"worker\", id, \"started job\", j)\n\t\ttime.Sleep(time.Second)\n\t\tfmt.Println(\"worker\", id, \"finished job\", j)\n\t\tresults <- j * 2\n\t}\n}", "func wait_and_process_results(results <-chan stat, done <-chan doneStatus,\n num_workers int) []stat {\n\n output := make([]stat, 0)\n\n for w := 0; w < num_workers; {\n select { // Blocking\n case result := <-results:\n output = append(output, result)\n case <-done:\n num_workers--\n }\n }\n\nDONE:\n\n // process any remaining results\n for {\n select {\n case result := <-results:\n output = append(output, result)\n default:\n break DONE\n }\n }\n\n return output\n}", "func worker(id int, jobs <-chan int, results chan<- int) {\n\tfor j := range jobs {\n\t\tfmt.Println(\"worker\", id, \"processing job\", j)\n\t\t//simulate an expensive work\n\t\ttime.Sleep(time.Second)\n\t\tresults <- j * 10\n\t}\n}", "func workerpool() {\n\tworkers := 3\n\tworkchan := make(chan int)\n\tfor i := 0; i < workers; i++ {\n\t\tgo func() {\n\t\t\tfor i := range workchan {\n\t\t\t\ttime.Sleep(3 * time.Second)\n\t\t\t\tfmt.Println(\"Workerpool worked on \", i)\n\t\t\t}\n\t\t}()\n\t}\n\tamountOfWork := 10\n\tfor i := 0; i < amountOfWork; i++ {\n\t\tworkchan <- i\n\t}\n\tfmt.Println(\"Finished workerpool work\")\n\t//Give some time for goroutines to finish. To avoid using WaitGroup and loosing focus.\n\ttime.Sleep(5 * time.Second)\n}", "func (pip Pipeline) worker(done <-chan bool, fn func(X) X) Pipeline {\n\tout := make(chan X)\n\tgo func() {\n\t\tdefer close(out)\n\t\tfor val := range pip {\n\t\t\tselect {\n\t\t\tcase out <- fn(val):\n\t\t\tcase <-done:\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\treturn out\n}", "func (c *MultiClusterController) worker() {\n\tfor c.processNextWorkItem() {\n\t}\n}", "func testParallelismWithBeverages(bm *BeverageMachine, beverageNames []string) {\n\n\tfmt.Printf(\"\\nstarting test: testParallelismWithBeverages\\n\\n\")\n\n\twg := sync.WaitGroup{}\n\tfor i, beverageName := range beverageNames {\n\t\twg.Add(1)\n\t\tgo func(i int, beverageName string) {\n\t\t\tdefer wg.Done()\n\n\t\t\tfmt.Printf(\"thread %d-> start\\n\", i+1)\n\n\t\t\t//1. get an idle dispenser\n\t\t\tdispenser, err := bm.GetIdleDispenser()\n\t\t\tfor err != nil {\n\t\t\t\tfmt.Printf(\"thread %d-> %s, retrying in 2 seconds...\\n\", i+1, err.Error())\n\t\t\t\ttime.Sleep(2 * time.Second)\n\t\t\t\tdispenser, err = bm.GetIdleDispenser()\n\t\t\t}\n\t\t\tfmt.Printf(\"thread %d-> acquired dispenser %d\\n\", i+1, dispenser.GetId())\n\n\t\t\tfmt.Printf(\"thread %d-> starting to prepare %s on dispenser %d...\\n\", i+1, beverageName, dispenser.GetId())\n\n\t\t\t//2. request the beverage from the dispenser\n\t\t\tbeverage, err := bm.RequestBeverage(dispenser, beverageName)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"thread %d-> dispenser %d says: %s\\n\", i+1, dispenser.GetId(), err.Error())\n\t\t\t\tfmt.Printf(\"thread %d-> end\\n\", i+1)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tfmt.Printf(\"thread %d-> successfully served %s on dispenser %d\\n\", i+1, beverage.GetName(), dispenser.GetId())\n\t\t\tfmt.Printf(\"thread %d-> end\\n\", i+1)\n\t\t}(i, beverageName)\n\t}\n\twg.Wait()\n\n\tfmt.Println(\"\\ncompleted test: testParallelismWithBeverages\\n\")\n}", "func multiWork(nums []float64) {\n\tch := make(chan float64)\n\tvar wg sync.WaitGroup\n\twg.Add(len(nums))\n\n\tfor i := 0; i < runtime.NumCPU(); i++ {\n\t\tgo poolWorker(ch, &wg)\n\t}\n\n\tfor _, i := range nums {\n\t\tch <- i\n\t}\n\n\twg.Wait()\n\tclose(ch)\n}", "func (p *Pool) work() {\n\tfor task := range p.queueTasks {\n\t\t//when the work is idle it receive a new task from the channel\n\t\ttask.DoWork()\n\t\tp.wg.Done()\n\t}\n}", "func (b ArrayBucket) Worker(inChan <-chan string, outChan chan<- string, wg *sync.WaitGroup) {\n\twg.Add(1)\n\tgo func() {\n\t\tfor cand := range inChan {\n\t\t\t//Test against the bucket if\n\t\t\t// it the cand can be made from available tokens\n\t\t\tif b.testMatch(cand) {\n\t\t\t\toutChan <- cand\n\t\t\t}\n\t\t}\n\t\twg.Done()\n\t}()\n}", "func main(){\n\tjobs := make(chan int, 100) //buffered chanel of ints\n\tresult := make(chan int, 100)\n\tvar wg sync.WaitGroup //the mutex\n\twg.Add(1)\n\n\tfibo := 50 //how many numbers we want to calculate\n\n\tgo worker(jobs, result) //parallel workers\n\tgo worker(jobs, result) //parallel workers\n\tgo worker(jobs, result) //parallel workers\n\tgo worker(jobs, result) //parallel workers\n\n\tfor i:=1; i<=fibo; i++ {\n\t\tjobs <- i\n\t}\n\tclose(jobs)\n\n\tgo func() {\n\t\tfor i:=1; i<=fibo; i++ {\n\t\t\tfmt.Printf(\"Number is %v and the result: %+v\\n\", i, <-result)\n\t\t}\n\t\twg.Done()\n\t}()\n\n\twg.Wait()\n\treturn\n}", "func worker(queue chan []string, out chan []byte, opts options, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\tfor batch := range queue {\n\tLoop:\n\t\tfor _, s := range batch {\n\t\t\tvar err error\n\t\t\tis := finc.IntermediateSchema{}\n\t\t\terr = json.Unmarshal([]byte(s), &is)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\n\t\t\t// Skip things, e.g. blacklisted DOIs.\n\t\t\tfor _, f := range opts.filters {\n\t\t\t\tif !f.Apply(is) {\n\t\t\t\t\tcontinue Loop\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Get export format.\n\t\t\tschema := opts.exportSchemaFunc()\n\t\t\terr = schema.Convert(is)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\n\t\t\t// Get list of ISILs to attach.\n\t\t\tschema.Attach(opts.tagger.Tags(is))\n\n\t\t\t// TODO(miku): maybe move marshalling into Exporter, if we have\n\t\t\t// anything else than JSON - function could be somethings like\n\t\t\t// func Marshal() ([]byte, error)\n\t\t\tb, err := json.Marshal(schema)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\t\t\tout <- b\n\t\t}\n\t}\n}", "func Test_Multi_Write_Iteration_Concurrency(t *testing.T) {\n\tc := Config{\n\t\tBucketLimits: []float64{1},\n\t}\n\thv, err := New(c)\n\tif err != nil {\n\t\tt.Fatalf(\"expected nil, got %v\", err)\n\t}\n\n\tvar done = make(chan bool)\n\tvar errors = make(chan error)\n\tvar writeWaitGroup sync.WaitGroup\n\tfor i := 0; i < 1000; i++ {\n\t\twriteWaitGroup.Add(1)\n\n\t\tgo func(i int) {\n\t\t\tdefer writeWaitGroup.Done()\n\n\t\t\terr := hv.Add(fmt.Sprintf(\"%v\", i), float64(i))\n\t\t\tif err != nil {\n\t\t\t\terrors <- err\n\t\t\t}\n\t\t}(i)\n\t}\n\n\tvar iterateWaitGroup sync.WaitGroup\n\tfor i := 0; i < 1000; i++ {\n\t\titerateWaitGroup.Add(1)\n\n\t\tgo func() {\n\t\t\tdefer iterateWaitGroup.Done()\n\n\t\t\tfor range hv.Histograms() {\n\t\t\t}\n\t\t}()\n\t}\n\n\tgo func() {\n\t\twriteWaitGroup.Wait()\n\t\tclose(done)\n\t}()\n\n\tselect {\n\tcase <-done:\n\t\tbreak\n\tcase err := <-errors:\n\t\tt.Fatalf(\"goroutine error : %v\", err)\n\t}\n\n\titerateWaitGroup.Wait()\n}", "func (fc *FederatedController) worker() {\n\tfor fc.processNextWorkItem() {\n\t}\n}", "func (ig *Instagram) consumer(results chan *InstagramResponse, done chan []InstagramData, simultaneous int, keepCriterion func(InstagramData) bool) {\n\tvar igData []InstagramData\n\n\ttimeout := make(chan bool)\n\tgo func() {\n\t\t//TODO: Make timeout a config'able setting\n\t\ttime.Sleep(10 * time.Second)\n\t\ttimeout <- true\n\t}()\n\n\ti := 0\n\t//Pull down results forever until we've hit some satisfaction criterion\n\tfor {\n\t\ti++\n\t\tselect {\n\n\t\tcase <-timeout:\n\t\t\tfmt.Printf(\"We timed out after %d attempts. Unlocking.\\n\", (i - 1))\n\n\t\t\t//Drain the channel\n\t\t\tgo drain(results, simultaneous+1) //+1 for timeout case\n\t\t\tdone <- igData //Semaphore{}\n\t\t\tfmt.Printf(\"Unlocked\\n\")\n\t\t\treturn\n\n\t\tcase res, ok := <-results:\n\t\t\tif !ok {\n\t\t\t\tfmt.Printf(\"I is %d but the channel is closed.\", i)\n\t\t\t\tdone <- igData //Semaphore{}\n\t\t\t\tfmt.Printf(\"Unlocked\\n\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfmt.Printf(\"%s is overall #%d\\n\", res, i)\n\n\t\t\t//Append\n\t\t\tfor _, datum := range res.Data {\n\t\t\t\tif keepCriterion(datum) {\n\t\t\t\t\tigData = append(igData, datum)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif !ig.satisfied(igData, i) {\n\t\t\t\t//If it's not happy after this result, the consumer\n\t\t\t\t// instructs a producer to start on something new\n\t\t\t\t//job := URL(random(300))\n\t\t\t\tfmt.Printf(\"Consumer is not satisfied after job #%d. Fetching next: %d\\n\", i, res.Pagination.NextURL)\n\t\t\t\tif res.Pagination.NextURL == \"\" {\n\t\t\t\t\tfmt.Printf(\"Consumer is not satisfied after job #%d but no next page was provided.\", i)\n\t\t\t\t\t//1 fewer goroutine is running at the same time\n\t\t\t\t\tsimultaneous = simultaneous - 1\n\t\t\t\t\tif simultaneous == 0 {\n\t\t\t\t\t\tgo drain(results, simultaneous)\n\t\t\t\t\t\t//Nothing worked.\n\t\t\t\t\t\tdone <- igData\n\t\t\t\t\t\tfmt.Printf(\"Unlocked\\n\")\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tgo ig.producer(results, res.Pagination.NextURL)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tfmt.Printf(\"Consumer is satisfied after job #%d. Unlocking.\\n\", i)\n\n\t\t\t\t//Drain the channel\n\t\t\t\tgo drain(results, simultaneous)\n\t\t\t\tdone <- igData //Semaphore{}\n\t\t\t\tfmt.Printf(\"Unlocked\\n\")\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n}", "func Synchronising() {\n\tdone := make(chan bool, 1)\n\n\tgo worker(done)\n\n\t<-done\n}", "func Worker(mapf func(string, string) []KeyValue,\n\t\t\treducef func(string, []string) string) {\n\n\t// Your worker implementation here.\n\n\t// TODO: maybe use a channel for in-process comm?\n\t// determine task state to know which master RPC to call\n\t//reply := CallRegisterIdle()\n\tvar reply *RegisterIdleReply\n\n\t//for workerInfo.State == IDLE || workerInfo.State == COMPLETED {\n\tfor {\n\n\t\tif workerInfo.State == IDLE {\n\t\t\treply = CallRegisterIdle()\n\t\t\tif reply == nil {\n\t\t\t\tworker_logger.Error(\"Got Error!!!!!!\")\n\t\t\t}\n\t\t} else if workerInfo.State == COMPLETED {\n\t\t\treply = CallCompletedTask() // override reply\n\t\t\t//if reply != nil {\n\t\t\t//\tresetWorkerInfo()\n\t\t\t//\tworkerInfo.State = IDLE\n\t\t\t//}\n\t\t\tif reply == nil {\n\t\t\t\tworker_logger.Error(\"Got errror!!!!!!!!\")\n\t\t\t}\n\t\t} else {\n\t\t\tworker_logger.Error(\"Shouldn't be in IN_PROGRESS state here...\")\n\t\t}\n\n\t\t// TODO: maybe don't need a mutex?\n\t\tif reply.MasterCommand == ASSIGN_TASK {\n\n\t\t\tworkerInfo.State = IN_PROGRESS\n\t\t\tworkerInfo.Id = reply.WorkerId\n\t\t\tworkerInfo.TaskType = reply.TaskType\n\t\t\tworkerInfo.TaskId = reply.TaskId\n\t\t\tworkerInfo.InputFileLoc = reply.InputFileLoc\n\t\t\tworkerInfo.NReduce = reply.NReduce\n\t\t\t//workerInfo.Progress = 0.0\n\n\t\t\t// TODO: replace this with broadcaster/observer design\n\t\t\tprogress_ch := make(chan float32)\n\t\t\tdone := make(chan struct{})\n\t\t\theartbeatStoped := make(chan struct {})\n\n\n\t\t\t// Actual computing job goroutine\n\t\t\tgo func() {\n\t\t\t\tif workerInfo.TaskType == MAP {\n\t\t\t\t\tdoMapTask(&workerInfo, mapf, progress_ch)\n\t\t\t\t} else if workerInfo.TaskType == REDUCE {\n\t\t\t\t\tdoReduceTask(&workerInfo, reducef, progress_ch)\n\t\t\t\t}/* else { // None task\n\t\t\t\t\tclose(progress_ch)\n\t\t\t\t}*/\n\n\t\t\t}()\n\n\t\t\t// Heartbeat gorountine\n\t\t\tgo func() {\n\t\t\t\tfor {\n\t\t\t\t\tselect {\n\t\t\t\t\t\tcase <-done:\n\t\t\t\t\t\t\tworker_logger.Debug(\"heartbeat job received done signal, stopping!\")\n\t\t\t\t\t\t\theartbeatStoped <- struct{}{}\n\t\t\t\t\t\t\tclose(heartbeatStoped)\n\t\t\t\t\t\t\treturn\n\t\t\t\t\t\tdefault:\n\t\t\t\t\t\t\tCallSendHeartbeat()\n\t\t\t\t\t\t\ttime.Sleep(1*time.Second)\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t}()\n\n\n\t\t\tfor progress := range progress_ch {\n\t\t\t\tworker_logger.Debug(fmt.Sprintf(\"Task(%s) progress: %f\", workerInfo.TaskId, progress))\n\t\t\t}\n\t\t\tdone <- struct{}{}\n\t\t\tclose(done)\n\t\t\t<- heartbeatStoped\n\n\t\t\t// Set result location & worker state\n\t\t\tworkerInfo.State = COMPLETED\n\n\t\t} else if reply.MasterCommand == STAND_BY {\n\t\t\tworker_logger.Debug(fmt.Sprintf(\"Got masterCommand: %s\", reply.MasterCommand))\n\t\t\ttime.Sleep(500*time.Millisecond)\n\t\t} else if reply.MasterCommand == PLEASE_EXIT {\n\t\t\tworker_logger.Info(fmt.Sprintf(\"Got masterCommand: %s\", reply.MasterCommand))\n\t\t\treturn\n\t\t}\n\n\t}\n\n\n\t// uncomment to send the Example RPC to the master.\n\t// CallExample()\n\n}", "func parallel(count int, procs int, fn func(int, int)) {\n\tvar blockSize int\n\tif count > 0 && procs > 0 {\n\t\tblockSize = count / procs\n\t} else {\n\t\tblockSize = 1\n\t}\n\tif blockSize <= 0 {\n\t\tblockSize = 1\n\t}\n\n\tidx := count\n\tvar wg sync.WaitGroup\n\tfor idx > 0 {\n\t\tstart := idx - blockSize\n\t\tend := idx\n\t\tif start < 0 {\n\t\t\tstart = 0\n\t\t}\n\t\tidx -= blockSize\n\n\t\twg.Add(1)\n\t\tgo func() {\n\t\t\tfn(start, end)\n\t\t\twg.Done()\n\t\t}()\n\t}\n\twg.Wait()\n}", "func TestChannelAsOutputRoutine(t *testing.T) {\n\tfor _, array := range bigStruct {\n\t\tch := make(chan int)\n\t\tgo sum(array.bigArray, ch)\n\t\tsum := <-ch // this is the blocking operation.\n\t\tif sum != array.result {\n\t\t\tt.Errorf(\"Expected result %d but got %d\\n\", array.result, sum)\n\t\t} else {\n\t\t\tt.Logf(\"The correct sum was calculated: %d\\n\", sum)\n\t\t}\n\t}\n}", "func process(w io.Writer, concurrency int, limit int, timeout int, sitemapURL string, headers []parameter, query []parameter) bool {\n\twritesToStdout := w == os.Stdout\n\n\tif writesToStdout {\n\t\tuiprogress.Start()\n\t}\n\n\t// Create two channels for our pipeline\n\ttasks := make(chan URL)\n\tresults := make(chan URL)\n\t// Create pre-configured client\n\tclient := newClient()\n\t// Define timeout for workers' pool\n\tworkerTimeout := time.Duration(1000000 * timeout)\n\n\tsitemap, err := requestSitemap(client, sitemapURL, headers)\n\tif err != nil {\n\t\tlog.Fatalf(\"Error: Failed to download the sitemap: %v\", err)\n\t}\n\n\tif len(sitemap.URLS) == 0 {\n\t\tlog.Fatalf(\"Error: The sitemap is empty\")\n\t}\n\n\tvar entiesNum int\n\tif len(sitemap.URLS) > limit && limit > 0 {\n\t\tentiesNum = len(sitemap.URLS[:limit])\n\t} else {\n\t\tentiesNum = len(sitemap.URLS)\n\t}\n\n\tbar := makeProgressBar(entiesNum)\n\n\t// Spawn workers\n\tfor w := 1; w <= concurrency; w++ {\n\t\tworker := newWorker(workerTimeout, tasks, results)\n\t\tgo worker.Perform(func(url URL) URL {\n\t\t\tstatusCode, err := requestPage(client, appendQuery(url.Loc, query), headers)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Error: %v\", err)\n\t\t\t}\n\n\t\t\turl.StatusCode = statusCode\n\t\t\treturn url\n\t\t})\n\t}\n\n\t// Spawn tasks producer\n\tproducer := newProducer(tasks)\n\tgo producer.Perform(sitemap.URLS[:entiesNum])\n\n\t// Create a consumer and join results\n\tconsumer := newConsumer(results)\n\treport := consumer.Perform(entiesNum, func() {\n\t\tif writesToStdout {\n\t\t\tbar.Incr()\n\t\t}\n\t})\n\n\t// Stop the progressbar\n\tif writesToStdout {\n\t\tuiprogress.Stop()\n\t}\n\n\tvar failed []URL\n\n\t// // Write a report\n\t// drawTable(w, report)\n\n\tfor _, url := range report {\n\t\tif url.StatusCode != 200 {\n\t\t\tfailed = append(failed, url)\n\t\t}\n\t}\n\n\tif len(failed) > 0 {\n\t\tdrawTable(w, failed)\n\t} else {\n\t\tfmt.Println(\"+-------------------+\\n| NO PROBLEMS FOUND |\\n+-------------------+\")\n\t}\n\n\treturn len(failed) == 0\n}", "func worker(jobs <-chan int, results chan<- int) {\n\tfor n := range jobs {\n\t\tresults <- fib(n)\n\t}\n}", "func worker(jobs <-chan int, results chan<- int) {\n\tfor n := range jobs {\n\t\tresults <- fib(n)\n\t}\n}", "func worker(threads int, doneWorker chan<- bool, imageTasks <-chan *imagetask.ImageTask, imageResults chan<- *imagetask.ImageTask) {\n\n\t// Initial placing image chunks in to a channel for filter to consume.\n\tchunkStreamGenerator := func(done <- chan interface{}, imageChunks []*imagetask.ImageTask) chan *imagetask.ImageTask {\n\t\tchunkStream := make(chan *imagetask.ImageTask)\n\t\tgo func() {\n\t\t\tdefer close(chunkStream)\n\t\t\tfor _, chunk := range imageChunks {\n\t\t\t\tselect {\n\t\t\t\tcase <-done:\n\t\t\t\t\treturn\n\t\t\t\tcase chunkStream <- chunk:\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t\treturn chunkStream\n\t}\n\n\t// Filter applies a filter in a pipeline fashion. \n\t// A goroutine is spawned for each chunk that needs to be filtered (which is numOfThreads chunks for each filter effect)\n\tfilter := func(threads int, effect string, effectNum int, done <- chan interface{}, chunkStream chan *imagetask.ImageTask) chan *imagetask.ImageTask {\n\t\tfilterStream := make(chan *imagetask.ImageTask, threads) // Only numOfThreads image chunks should be in the local filter channel.\n\t\tdonefilterChunk := make(chan bool)\n\t\tfor chunk := range chunkStream { // For each image chunk ...\n\t\t\tif effectNum > 0 {\n\t\t\t\tchunk.Img.UpdateInImg() // Replace inImg with outImg if not the first effect to compund effects.\n\t\t\t}\t\n\t\t\tgo func(chunk *imagetask.ImageTask) { // Spawn a goroutine for each chunk, which is equal to the numOfThreads. Each goroutine works on a portion of the image.\n\t\t\t\tselect {\n\t\t\t\tcase <-done:\n\t\t\t\t\tdonefilterChunk <- true\n\t\t\t\t\treturn\n\t\t\t\tcase filterStream <- chunk:\n\t\t\t\t\tif effect != \"G\" {\n\t\t\t\t\t\tchunk.Img.ApplyConvolution(effect) // Can wait to apply effect until after chunk is in the channel because has to wait for all goroutines to finish before it can move on to the next filter for a given image.\n\t\t\t\t\t} else {\n\t\t\t\t\t\tchunk.Img.Grayscale()\n\t\t\t\t\t}\n\t\t\t\t\tdonefilterChunk <- true // Indicate that the filtering is done for the given chunk.\n\t\t\t\t}\n\t\t\t}(chunk)\n\t\t}\n\t\tfor i := 0; i < threads; i ++ { // Wait for all portions to be put through one filter because of image dependencies with convolution.\n\t\t\t<-donefilterChunk\n\t\t}\n\t\treturn filterStream\n\t}\n\n\t// While there are more image tasks to grab ...\n\tfor true {\n\t\t// Grab image task from image task channel.\t\n\t\timgTask, more := <-imageTasks\n\n\t\t// If you get an image task, split up the image in to even chunks by y-pixels.\n\t\tif more {\n\t\t\timageChunks := imgTask.SplitImage(threads)\n\t\t\t\n\t\t\t// Iterate through filters on image chunks.\n\t\t\t// Will spawn a goroutine for each chunk in each filter (n goroutines per filter)\n\t\t\tdone := make(chan interface{})\n\t\t\tdefer close(done)\n\t\t\tchunkStream := chunkStreamGenerator(done, imageChunks)\n\t\t\tfor i := 0; i < len(imgTask.Effects); i++ {\n\t\t\t\teffect := imgTask.Effects[i]\n\t\t\t\tchunkStream = filter(threads, effect, i, done, chunkStream)\n\t\t\t\tclose(chunkStream)\n\t\t\t}\n\n\t\t\t// Put the image back together.\n\t\t\treconstructedImage, _ := imgTask.Img.NewImage()\n\t\t\tfor imgChunk := range chunkStream {\n\t\t\t\treconstructedImage.ReAddChunk(imgChunk.Img, imgChunk.YPixelStart, imgChunk.ChunkPart)\n\t\t\t}\n\t\t\timgTask.Img = reconstructedImage\n\t\t\timageResults <- imgTask // Send image to results channel to be saved.\n\n\t\t} else { // Otherwise, if there are no more image tasks, then goroutine worker exits.\n\t\t\tdoneWorker <- true // Indicate that the worker is done.\n\t\t\treturn\n\t\t}\n\t}\n}", "func worker(wg *sync.WaitGroup) {\n\tvar b bytes.Buffer\n\t// a defer function to recover is a good idea here because any panics would otherwise crash the entire program\n\tdefer wg.Done()\n\n\t// pull urls from queue until it's done/closed so no need to check if channel is closed with ok variable\n\tfor url := range urls {\n\t\tfmt.Println(\"concurrently read value\", url, \"from channel\")\n\t\timg, err := loadImage(url)\n\t\tif err != nil {\n\t\t\tcheckError(&b, err)\n\t\t\tcontinue\n\t\t}\n\t\tgetThreePrevalentColours(img, url)\n\t}\n}", "func worker(wg *sync.WaitGroup) {\n\tfor job := range jobs {\n\t\toutput := Result{job, digits(job.randomno)}\n\t\tresults <- output\n\t}\n\twg.Done()\n}", "func worker(cfg Config, out chan<- result) {\n\tzoom := 1.0\n\tfor {\n\t\tout <- result{\n\t\t\toutput: julia(cfg, zoom),\n\t\t\tzoom: zoom,\n\t\t}\n\t\tzoom += .01\n\t}\n}", "func wait_and_process_results(results <-chan column, done <-chan doneStatus,\n num_workers int) column {\n\n var output column\n num_cols := 0\n\n for w := 0; w < num_workers; {\n select { // Blocking\n case result := <-results:\n output = process_column(result, output)\n case d := <-done:\n num_cols += d.files_processed\n num_workers--\n }\n }\n\nDONE:\n\n // process any remaining results\n for {\n select {\n case result := <-results:\n output = process_column(result, output)\n default:\n break DONE\n }\n }\n\n num_cols_f := float64(num_cols)\n for i, v := range output {\n output[i] = v / num_cols_f\n }\n\n return output\n}", "func (a *ami) processResults(resChan chan *workResult, doneChan chan *workResults) {\n\n\tvar ebb, mbb bytes.Buffer\n\n\tfor r := range resChan {\n\t\tif r.err != nil {\n\t\t\tebb.WriteString(fmt.Sprintf(\"%v\", r.err))\n\t\t}\n\t\tif r.message != \"\" {\n\t\t\tmbb.WriteString(r.message)\n\t\t}\n\t}\n\t// tell the world we are done\n\tdoneChan <- &workResults{\n\t\tmessages: mbb,\n\t\terrs: ebb,\n\t}\n}", "func (a App) Run() error {\n\ta.log.Printf(\"config %+v\", a.params)\n\twg := &sync.WaitGroup{}\n\tqueue := make(chan string)\n\tresults := make(chan result)\n\n\twg.Add(1)\n\tgo func() {\n\t\tdefer close(queue)\n\t\ta.log.Printf(\"queue sender started\")\n\t\tfor _, url := range a.params.URLs {\n\t\t\ta.log.Printf(\"send to queue: %s\", url)\n\t\t\tqueue <- url\n\t\t}\n\t\twg.Done()\n\t}()\n\n\tfor i := 0; i < a.params.Parallel; i++ {\n\t\ti := i\n\t\twg.Add(1)\n\t\tgo func(queue <-chan string, results chan<- result, wg *sync.WaitGroup) {\n\t\t\ta.log.Printf(\"worker %d started\", i)\n\t\t\tfor job := range queue {\n\t\t\t\tif requestedURL, body, err := download(a.client, job); err != nil {\n\t\t\t\t\ta.log.Printf(\"downloaded with error: %s\", err)\n\t\t\t\t\tresults <- result{\n\t\t\t\t\t\terr: err,\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\ta.log.Printf(\"%s downloaded successfully\", requestedURL)\n\t\t\t\t\tresults <- result{\n\t\t\t\t\t\tbody: fmt.Sprintf(\"%x\", md5.Sum(body)),\n\t\t\t\t\t\turl: requestedURL,\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\twg.Done()\n\t\t\ta.log.Printf(\"worker done: %d\", i)\n\t\t}(queue, results, wg)\n\t}\n\n\tgo func() {\n\t\twg.Wait()\n\t\ta.log.Printf(\"close results\")\n\t\tclose(results)\n\t}()\n\n\tfor r := range results {\n\t\tif r.err != nil {\n\t\t\ta.log.Printf(\"error: %s\", r.err)\n\t\t} else {\n\t\t\tif _, err := fmt.Fprintf(a.w, \"%s %s\\n\", r.url, r.body); err != nil {\n\t\t\t\treturn fmt.Errorf(\"error writing results: %s\", err)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func worker(links chan linkPair, results chan listPair) {\n\tfor link := range links {\n\t\tresults <- cardScrape(link)\n\t}\n}", "func worker(w *sync.WaitGroup, goroutineID int) {\n\tfor j := range jobs {\n\t\tdata <- Data{goroutineID, j, double(j.Value)}\n\t}\n\n\t// When no more jobs are on the jobs channel\n\t// resolve waitgroup.\n\tw.Done()\n}", "func (gq *Dispatch) next() {\n for true {\n // Attempt to start processing the file.\n gq.pLock.Lock()\n if gq.processing >= gq.MaxGo {\n gq.waitingToRun = true\n gq.nextWait.Add(1)\n gq.pLock.Unlock()\n gq.nextWait.Wait()\n continue\n }\n // Keep the books and reset wait time before unlocking.\n gq.processing++\n gq.pLock.Unlock()\n\n // Get an element from the queue.\n gq.qLock.Lock()\n var wrapper = gq.queue.Dequeue().(queues.RegisteredTask)\n gq.qLock.Unlock()\n\n // Begin processing and asyncronously return.\n //var task = taskelm.Value.(dispatchTaskWrapper)\n var task = wrapper.Func()\n go task(wrapper.Id())\n return\n }\n}", "func sumWorker(nums chan int, out chan int) {\n\t// TODO: implement me\n\t// HINT: use for loop over `nums`\n\tvar sum int\n\tdefer wg.Done()\n\tcounterA:= len(nums)\n\tfor i := 0; i<counterA; i++ {\n\t\telement := <- nums\n\t\tsum+=element\n\t}\n\tout <- sum\n}", "func queueWorker() {\n\tcrossRefClient := crossref.NewCrossRefClient(&http.Client{})\n\tbibClient := bibtex.NewDXDOIClient(&http.Client{})\n\tfor {\n\t\tx := <-enhancementChan\n\t\tfor counter.Get() > 40 {\n\t\t\ttime.Sleep(time.Millisecond * 10)\n\t\t}\n\t\tcounter.Inc()\n\t\tgo gatherAdditionalInfo(crossRefClient, bibClient, x)\n\n\t}\n}", "func processCountParallel(chunks []*bytes.Buffer, position uint64, condition string, cpus int) (result int64) {\n\n\tvar wg sync.WaitGroup\n\tchunkChan := make(chan chunkInput)\n\tresultChan := make(chan chunkOutput)\n\n\trows := bytes.Count(chunks[0].Bytes(), []byte{0x0a}) * 2\n\n\t// Start one go routine per CPU\n\tfor i := 0; i < cpus; i++ {\n\t\twg.Add(1)\n\t\tgo func() {\n\t\t\tdefer wg.Done()\n\t\t\tprocessWorker(position, condition, rows, chunkChan, resultChan)\n\t\t}()\n\t}\n\n\t// Push chunks onto input channel\n\tgo func() {\n\t\tfor _, buf := range chunks {\n\n\t\t\tchunkChan <- chunkInput{buf: buf}\n\t\t}\n\n\t\t// Close input channel\n\t\tclose(chunkChan)\n\t}()\n\n\t// Wait for workers to complete\n\tgo func() {\n\t\twg.Wait()\n\t\tclose(resultChan) // Close output channel\n\t}()\n\n\tfor r := range resultChan {\n\t\tresult += int64(r.result)\n\t}\n\n\treturn\n}", "func Worker(mapf func(string, string) []KeyValue,\n\treducef func(string, []string) string) {\n\n\t// init\n\ttaskId = 9999\n\n\t//\n\tfor {\n\t\ttime.Sleep(time.Second)\n\n\t\treply := CallAssign()\n\n\t\t// fmt.Println(reply)\n\n\t\tif reply.TaskId < 0 {\n\t\t\t// fmt.Println(\"Waiting for assigning a work...\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// modify taskId and later will tell master who i am\n\t\ttaskId = reply.TaskId\n\n\t\tif reply.TaskType == \"map\" {\n\t\t\tfile, err := os.Open(reply.FileName)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatalf(\"cannot open %v\", reply.FileName)\n\t\t\t}\n\t\t\tcontent, err := ioutil.ReadAll(file)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatalf(\"cannot read %v\", reply.FileName)\n\t\t\t}\n\t\t\tfile.Close()\n\t\t\tkva := mapf(reply.FileName, string(content))\n\n\t\t\t// sort\n\t\t\t// sort.Sort(ByKey(kva))\n\n\t\t\t// store intermediate kvs in tempFile\n\t\t\ttempFileName := \"tmp-\" + reply.TaskType + \"-\" + strconv.Itoa(reply.TaskId)\n\n\t\t\tfile, err = os.Create(tempFileName)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(\"cannot create %v\", tempFileName)\n\t\t\t}\n\n\t\t\t// transform k,v into json\n\t\t\tenc := json.NewEncoder(file)\n\t\t\tfor _, kv := range kva {\n\t\t\t\terr := enc.Encode(&kv)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Fatal(err)\n\t\t\t\t}\n\t\t\t}\n\t\t\t//\n\t\t\tfile.Close()\n\n\t\t\t// try to delay sometime\n\t\t\t// ran := rand.Intn(4)\n\t\t\t// fmt.Printf(\"Sleep %v s\\n\", ran)\n\t\t\t// d := time.Second * time.Duration(ran)\n\t\t\t// time.Sleep(d)\n\n\t\t\t// tell the master the mapwork has done\n\t\t\tCallDoneTask(reply, tempFileName)\n\n\t\t} else if reply.TaskType == \"reduce\" {\n\t\t\t// fmt.Println(reply.TaskType)\n\n\t\t\tkva := []KeyValue{}\n\n\t\t\tfile, err := os.Open(reply.FileName)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\n\t\t\tdec := json.NewDecoder(file)\n\t\t\tfor {\n\t\t\t\tvar kv KeyValue\n\t\t\t\tif err := dec.Decode(&kv); err != nil {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tkva = append(kva, kv)\n\t\t\t}\n\n\t\t\toutputFileName := \"mr-out-\" + strconv.Itoa(reply.TaskIndex)\n\t\t\tofile, _ := os.Create(outputFileName)\n\n\t\t\t// sort\n\t\t\t// sort.Sort(ByKey(kva))\n\n\t\t\ti := 0\n\t\t\tfor i < len(kva) {\n\t\t\t\tj := i + 1\n\t\t\t\tfor j < len(kva) && kva[j].Key == kva[i].Key {\n\t\t\t\t\tj++\n\t\t\t\t}\n\t\t\t\tvalues := []string{}\n\t\t\t\tfor k := i; k < j; k++ {\n\t\t\t\t\tvalues = append(values, kva[k].Value)\n\t\t\t\t}\n\t\t\t\toutput := reducef(kva[i].Key, values)\n\n\t\t\t\t// fmt.Println(output)\n\n\t\t\t\tfmt.Fprintf(ofile, \"%v %v\\n\", kva[i].Key, output)\n\n\t\t\t\ti = j\n\t\t\t}\n\n\t\t\tofile.Close()\n\n\t\t\t// fmt.Printf(\"Reduce task %v has finished.\\n\", reply.TaskIndex)\n\n\t\t\t// ran := rand.Intn(4)\n\t\t\t// fmt.Printf(\"Sleep %v s\\n\", ran)\n\t\t\t// d := time.Second * time.Duration(ran)\n\t\t\t// time.Sleep(d)\n\n\t\t\tCallDoneTask(reply, outputFileName)\n\t\t} else if reply.TaskType == \"close\" {\n\t\t\t// fmt.Println(\"MapReduce has done. Exiting...\")\n\t\t\tbreak\n\t\t} else {\n\t\t\tfmt.Println(\"UnExcepted TaskType\")\n\t\t}\n\n\t}\n\n}", "func TestOrderedWorkerPool(t *testing.T) {\n\ttaskCount := 100000\n\tresults := make([]int, taskCount)\n\n\ttasksCh, resultCh := New(runtime.NumCPU() * 2)\n\tgo dispatchTasks(tasksCh, taskCount)\n\n\t// Cache the results in order they are received\n\ti := 0\n\tfor result := range resultCh {\n\t\tresults[i] = result.Value.(int)\n\t\ti++\n\t}\n\n\t// Verify that results arrived in correct order.\n\tfor i := 0; i < taskCount; i++ {\n\t\tif results[i] != i {\n\t\t\tt.Error(\"Results out of order\", results)\n\t\t}\n\t}\n}", "func worker(subHeight, imageWidth int, sendByte chan byte, out chan byte){\n\n\tworkPlace := make([][]byte,subHeight + 2)\n\tfor i := range workPlace{\n\t\tworkPlace[i] = make([]byte, imageWidth)\n\t}\n\n\n\tfor y := 0; y < subHeight + 2; y++{\n\t\tfor x := 0; x < imageWidth; x++{\n\t\t\tcurrentByte := <- sendByte\n\t\t\tworkPlace[y][x] = currentByte\n\t\t}\n\t}\n\tnextWorldPart := schrodinger(workPlace)\n\n\tfor y := 1; y < subHeight + 1; y++{\n\t\tfor x := 0; x < imageWidth; x++{\n\t\t\tout <- nextWorldPart[y][x]\n\t\t}\n\t}\n}", "func main() {\n\tresultChan := make(chan *result)\n\n\t//Async\n\tgo doWorkAsync(resultChan)\n\t//Sync\n\tresult := doWork()\n\n\t//Wait for Async to catch up\n\tresultAsync := <-resultChan\n\t//Merge results\n\tfmt.Println(result.success && resultAsync.success)\n}", "func ForN(low, high, step, parallelism int, worker interface{}) error {\n\tif low > high {\n\t\treturn fmt.Errorf(\"low (%d) > high (%d)\", low, high)\n\t}\n\tif step <= 0 {\n\t\treturn fmt.Errorf(\"step (%d) must be positive\", step)\n\t}\n\n\ttyp := reflect.TypeOf(worker)\n\tif typ.Kind() != reflect.Func {\n\t\treturn errors.New(\"parallel.ForN worker must be a function.\")\n\t}\n\tif typ.NumIn() != 1 {\n\t\treturn errors.New(\"parallel.ForN worker must have 1 parameter\")\n\t}\n\tif typ.In(0).Kind() != reflect.Int {\n\t\treturn errors.New(\"parallel.ForN worker must have a int param\")\n\t}\n\tif typ.NumOut() > 1 {\n\t\treturn errors.New(\"parallel.ForN worker must return nothing or error\")\n\t}\n\tif typ.NumOut() == 1 && typ.Out(0).Name() != \"error\" {\n\t\treturn errors.New(\"parallel.ForN worker's return type must be error\")\n\t}\n\n\tchin := make(chan int)\n\tchout := make(chan error)\n\tvar wg sync.WaitGroup\n\twg.Add(parallelism)\n\tvar errs string\n\n\tgo func() {\n\t\tfor i := low; i < high; i += step {\n\t\t\tchin <- i\n\t\t}\n\t\tclose(chin)\n\t}()\n\n\tval := reflect.ValueOf(worker)\n\tfor i := 0; i < parallelism; i++ {\n\t\tgo func(val reflect.Value) {\n\t\t\tdefer wg.Done()\n\t\t\tfor input := range chin {\n\t\t\t\tif typ.NumOut() == 0 { // worker returns nothing\n\t\t\t\t\tval.Call([]reflect.Value{reflect.ValueOf(input)})\n\t\t\t\t} else { // worker returns an error\n\t\t\t\t\tr := val.Call([]reflect.Value{reflect.ValueOf(input)})\n\t\t\t\t\tif r[0].Interface() != nil {\n\t\t\t\t\t\tchout <- r[0].Interface().(error)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}(val)\n\t}\n\n\tgo func() {\n\t\twg.Wait()\n\t\tclose(chout)\n\t}()\n\n\tfor e := range chout {\n\t\terrs += e.Error() + \"\\n\"\n\t}\n\tif len(errs) > 0 {\n\t\treturn errors.New(errs)\n\t}\n\treturn nil\n}", "func (mr *Master) schedule(phase jobPhase) {\n\tvar ntasks int\n\tvar nios int // number of inputs (for reduce) or outputs (for map)\n\tswitch phase {\n\tcase mapPhase:\n\t\tntasks = len(mr.files)\n\t\tnios = mr.nReduce\n\tcase reducePhase:\n\t\tntasks = mr.nReduce\n\t\tnios = len(mr.files)\n\t}\n\n\tfmt.Printf(\"Schedule: %v %v tasks (%d I/Os)\\n\", ntasks, phase, nios)\n\n\t// them have been completed successfully should the function return.\n\t// Remember that workers may fail, and that any given worker may finish\n\t// multiple tasks.\t// All ntasks tasks have to be scheduled on workers, and only once all of\n\n\t//\n\t// TODO TODO TODO TODO TODO TODO TODO TODO TODO TODO TODO TODO TODO\n\t// mycode\n\tvar wg sync.WaitGroup\n\twg.Add(ntasks)\n\ti := int32(-1)\n\n\tif phase == mapPhase {\n\t\tfor {\n\t\t\tvar worker string\n\t\t\tworker = <-mr.registerChannel\n\t\t\tif worker == \"done\" {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tgo func() {\n\t\t\t\tfor {\n\t\t\t\t\tt := atomic.AddInt32(&i, 1)\n\t\t\t\t\tfmt.Printf(\"map worker = %s, t = %d\\n\", worker, t)\n\t\t\t\t\tif t > int32(ntasks) {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t\tif t == int32(ntasks) {\n\t\t\t\t\t\tmr.registerChannel <- \"done\"\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t\tfile := mr.files[t]\n\t\t\t\t\targs := DoTaskArgs{\"test\", file, phase, int(t), nios}\n\t\t\t\t\t//rpc worker possible have error\n\t\t\t\t\tif !call(worker, \"Worker.DoTask\", args, new(struct{})) {\n\t\t\t\t\t\tfmt.Printf(\"mapworkfailed %s\\n\", worker)\n\t\t\t\t\t\tt = atomic.AddInt32(&i, -1)\n\t\t\t\t\t\t//recall after one seconds\n\t\t\t\t\t\tbreak\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfmt.Printf(\"success t = %d\\n\", t)\n\t\t\t\t\t\twg.Done()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}()\n\t\t}\n\t\twg.Wait()\n\t}\n\n\tif phase == reducePhase {\n\t\tfmt.Printf(\"workers num=%d\\n\", len(mr.workers))\n\t\tfmt.Println(\"reduce begin\")\n\t\t//new worker\n\t\tgo func() {\n\t\t\tfor {\n\t\t\t\tworker := <-mr.registerChannel\n\t\t\t\tfmt.Printf(\"new worker arrived name = %s\\n\", worker)\n\t\t\t\tif worker == \"done\" {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tgo func() {\n\t\t\t\t\tfor {\n\t\t\t\t\t\tt := atomic.AddInt32(&i, 1)\n\t\t\t\t\t\tfmt.Printf(\"reduce new worker = %s, t = %d\\n\", worker, t)\n\t\t\t\t\t\tif t >= int32(ntasks) {\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\t\t\t\t\t\t// if t == int32(ntasks) {\n\t\t\t\t\t\t// \tmr.registerChannel <- \"done\"\n\t\t\t\t\t\t// \tbreak\n\t\t\t\t\t\t// }\n\t\t\t\t\t\tfile := mr.files[t]\n\t\t\t\t\t\targs := DoTaskArgs{\"test\", file, phase, int(t), nios}\n\t\t\t\t\t\t//rpc worker possible have error\n\t\t\t\t\t\tif !call(worker, \"Worker.DoTask\", args, new(struct{})) {\n\t\t\t\t\t\t\tt = atomic.AddInt32(&i, -1)\n\t\t\t\t\t\t\t//recall after one seconds\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\twg.Done()\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}()\n\t\t\t}\n\t\t}()\n\t\t//use previous worker\n\t\tfor _, val := range mr.workers {\n\t\t\t// fmt.Printf(\"reduce worker=%s\\n\", val)\n\t\t\tgo func(val string) {\n\t\t\t\tfor {\n\t\t\t\t\t// fmt.Printf(\"reduce4 worker=%s\\n\", val)\n\t\t\t\t\tt := atomic.AddInt32(&i, 1)\n\t\t\t\t\tif t >= int32(ntasks) {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t\tfile := mr.files[t]\n\t\t\t\t\targs := DoTaskArgs{\"test\", file, phase, int(t), nios}\n\t\t\t\t\t//rpc worker possible have error\n\t\t\t\t\tif !call(val, \"Worker.DoTask\", args, new(struct{})) {\n\t\t\t\t\t\tt = atomic.AddInt32(&i, -1)\n\t\t\t\t\t\tbreak\n\t\t\t\t\t} else {\n\t\t\t\t\t\twg.Done()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}(val)\n\t\t}\n\n\t\twg.Wait()\n\t\tmr.registerChannel <- \"done\"\n\t}\n\n\tfmt.Printf(\"Schedule: %v phase done\\n\", phase)\n}", "func main(){\n\tlockUsingChannel := make(chan bool, 1)\n\tfor i:=0; i <= 10; i++{\n\t\tgo performJob(i, lockUsingChannel)\n\t}\n\t// Make sure all go routines complete. You can also do runtime.gosched or implement wait groups\n\ttime.Sleep(time.Second * 20)\n}", "func (cc *CertificateController) worker(ctx context.Context) {\n\tfor cc.processNextWorkItem(ctx) {\n\t}\n}", "func (w *Worker) Work() {\n\tfor {\n\t\tselect {\n\t\tcase <-w.done:\n\t\t\treturn\n\t\tdefault:\n\t\t\tevent := <-w.channel\n\t\t\tw.writeToBuffer(event)\n\t\t}\n\t}\n}", "func TestConcurrency(t *testing.T) {\n\tc := New(0.000001)\n\n\tfn := func() ([]model.Torrent, int, error) {\n\t\treturn []model.Torrent{{}, {}, {}}, 10, nil\n\t}\n\n\tvar wg sync.WaitGroup\n\twg.Add(300)\n\tfor i := 0; i < 3; i++ {\n\t\tgo func() {\n\t\t\tfor j := 0; j < 100; j++ {\n\t\t\t\tgo func(j int) {\n\t\t\t\t\tdefer wg.Done()\n\t\t\t\t\tk := common.SearchParam{\n\t\t\t\t\t\tPage: j,\n\t\t\t\t\t}\n\t\t\t\t\tif _, _, err := c.Get(k, fn); err != nil {\n\t\t\t\t\t\tt.Fatal(err)\n\t\t\t\t\t}\n\t\t\t\t}(j)\n\t\t\t}\n\t\t}()\n\t}\n\twg.Wait()\n}", "func worker(jobs <-chan int, result chan<- int){\n\tfor n := range jobs{\n\t\tresult <- fib(n)\n\t}\n}", "func (c *Controller) worker() {\n\tfor c.processNextWorkItem() {\n\t}\n}", "func (c *Controller) worker() {\n\tfor c.processNextWorkItem() {\n\t}\n}", "func (c *Controller) worker() {\n\tfor c.processNextWorkItem() {\n\t}\n}", "func add_jobs(fileNames []string, colID int, wantMedian bool, jobs chan<- job,\n result chan<- stat) {\n for _, name := range fileNames {\n jobs <- job{name, colID, wantMedian, result}\n }\n close(jobs)\n}", "func (i producerConsumerJobImpl[T]) runMultiThreaded(ctx context.Context, cfg jobConfig) {\n\tj := i.j\n\tch := make(chan taskWithLabels[T]) //unbuffered!\n\tvar wg sync.WaitGroup\n\n\t//one goroutine produces tasks\n\twg.Add(1)\n\tgo func(ch chan<- taskWithLabels[T]) {\n\t\tdefer wg.Done()\n\t\tfor ctx.Err() == nil { //while ctx has not expired\n\t\t\ttask, labels, err := j.produceOne(ctx, cfg, true)\n\t\t\tif err == nil {\n\t\t\t\tch <- taskWithLabels[T]{task, labels}\n\t\t\t} else {\n\t\t\t\tlogAndSlowDownOnError(err)\n\t\t\t}\n\t\t}\n\n\t\t//`ctx` has expired -> tell workers to shutdown\n\t\tclose(ch)\n\t}(ch)\n\n\t//multiple goroutines consume tasks\n\t//\n\t//We use `numGoroutines-1` here since we already have spawned one goroutine\n\t//for the polling above.\n\twg.Add(int(cfg.NumGoroutines - 1))\n\tfor i := uint(0); i < cfg.NumGoroutines-1; i++ {\n\t\tgo func(ch <-chan taskWithLabels[T]) {\n\t\t\tdefer wg.Done()\n\t\t\tfor item := range ch {\n\t\t\t\terr := j.consumeOne(ctx, cfg, item.Task, item.Labels, true)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogg.Error(err.Error())\n\t\t\t\t}\n\t\t\t}\n\t\t}(ch)\n\t}\n\n\t//block until they are all done\n\twg.Wait()\n}", "func (b *B) RunParallel(body func(*PB)) {}", "func parMerge(a[] int, fst int, snd int, lng int, c chan bool, done chan bool) {\n <- c; <- c\n merge(a, fst, snd, lng)\n done <- true\n}", "func (a *adapterBase) worker(workerNum int, ctx interface{}) {\n\ta.Trace(\"xfer: adapter %q worker %d starting\", a.Name(), workerNum)\n\twaitForAuth := workerNum > 0\n\tsignalAuthOnResponse := workerNum == 0\n\n\t// First worker is the only one allowed to start immediately\n\t// The rest wait until successful response from 1st worker to\n\t// make sure only 1 login prompt is presented if necessary\n\t// Deliberately outside jobChan processing so we know worker 0 will process 1st item\n\tif waitForAuth {\n\t\ta.Trace(\"xfer: adapter %q worker %d waiting for Auth\", a.Name(), workerNum)\n\t\ta.authWait.Wait()\n\t\ta.Trace(\"xfer: adapter %q worker %d auth signal received\", a.Name(), workerNum)\n\t}\n\n\tfor job := range a.jobChan {\n\t\tt := job.T\n\n\t\tvar authCallback func()\n\t\tif signalAuthOnResponse {\n\t\t\tauthCallback = func() {\n\t\t\t\ta.authWait.Done()\n\t\t\t\tsignalAuthOnResponse = false\n\t\t\t}\n\t\t}\n\t\ta.Trace(\"xfer: adapter %q worker %d processing job for %q\", a.Name(), workerNum, t.Oid)\n\n\t\t// Actual transfer happens here\n\t\tvar err error\n\t\tif t.Size < 0 {\n\t\t\terr = fmt.Errorf(\"object %q has invalid size (got: %d)\", t.Oid, t.Size)\n\t\t} else {\n\t\t\terr = a.transferImpl.DoTransfer(ctx, t, a.cb, authCallback)\n\t\t}\n\n\t\t// Mark the job as completed, and alter all listeners\n\t\tjob.Done(err)\n\n\t\ta.Trace(\"xfer: adapter %q worker %d finished job for %q\", a.Name(), workerNum, t.Oid)\n\t}\n\t// This will only happen if no jobs were submitted; just wake up all workers to finish\n\tif signalAuthOnResponse {\n\t\ta.authWait.Done()\n\t}\n\ta.Trace(\"xfer: adapter %q worker %d stopping\", a.Name(), workerNum)\n\ta.transferImpl.WorkerEnding(workerNum, ctx)\n\ta.workerWait.Done()\n}", "func (w *mongoAutoCommitWorker) work() {\n\tdefer func() {\n\t\tw.p.workerWg.Done()\n\t\tclose(w.flushAckC)\n\t\tclose(w.flushC)\n\t}()\n\n\tvar stop bool\n\tfor !stop {\n\t\tselect {\n\t\tcase req, open := <-w.p.docsInsert:\n\t\t\tif open {\n\t\t\t\t// Received a new request\n\t\t\t\tw.insert(req)\n\t\t\t\tif w.commitRequired() {\n\t\t\t\t\tw.commit()\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// Channel closed: Stop.\n\t\t\t\tstop = true\n\t\t\t\tw.commit()\n\t\t\t}\n\t\tcase req, open := <-w.p.docsUpdate:\n\t\t\tif open {\n\t\t\t\t// Received a new request\n\t\t\t\tw.update(req)\n\t\t\t\tif w.commitRequired() {\n\t\t\t\t\tw.commit()\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// Channel closed: Stop.\n\t\t\t\tstop = true\n\t\t\t\tw.commit()\n\t\t\t}\n\t\tcase req, open := <-w.p.docsUpsert:\n\t\t\tif open {\n\t\t\t\t// Received a new request\n\t\t\t\tw.upsert(req)\n\t\t\t\tif w.commitRequired() {\n\t\t\t\t\tw.commit()\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// Channel closed: Stop.\n\t\t\t\tstop = true\n\t\t\t\tw.commit()\n\t\t\t}\n\n\t\tcase <-w.flushC:\n\t\t\t// Commit outstanding requests\n\t\t\tif w.capacity() > 0 {\n\t\t\t\tw.commit()\n\t\t\t}\n\t\t\tw.flushAckC <- struct{}{}\n\t\t}\n\t}\n}", "func (ssc *StatefulSetController) worker() {\n\tfor ssc.processNextWorkItem() {\n\t}\n}", "func workIpInfoLookup(workCh chan string, resultCh chan ipInfoResult) {\n\tfor ip := range workCh {\n\t\tobj := callRemoteService(ip)\n\t\tresultCh <- obj\n\t}\n}", "func (pm *PipelineManager) runWorker() {\n\tfor pm.processNextWorkItem() {\n\t}\n}", "func worker(cache *utxoCache, jobs <-chan []byte, results chan<- *result) {\n\tvar (\n\t\terr error\n\t\tm = bchec.NewMultiset(bchec.S256())\n\t\tentry *UtxoEntry\n\t\toutpoint *wire.OutPoint\n\t\tstate = &BestState{Hash: chainhash.Hash{}}\n\t)\n\tfor serializedUtxo := range jobs {\n\t\tm.Add(serializedUtxo)\n\n\t\toutpoint, entry, err = deserializeUtxoCommitmentFormat(serializedUtxo)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deserializing UTXO: %s\", err.Error())\n\t\t\tresults <- &result{err: err}\n\t\t\treturn\n\t\t}\n\n\t\tif err = cache.AddEntry(*outpoint, entry, true); err != nil {\n\t\t\tresults <- &result{err: err}\n\t\t\treturn\n\t\t}\n\n\t\tif err = cache.Flush(FlushIfNeeded, state); err != nil {\n\t\t\tresults <- &result{err: err}\n\t\t\treturn\n\t\t}\n\t}\n\tresults <- &result{m: m}\n}", "func MapReduce(input int, reduce func(results []int) int, tasks ...Task) (int, error) {\n\t// 1. tasks => taskCh (chan Task)\n\t// 2. Limited number of gophers work on tasks\n\t// - output => output channel\n\t// 3. goroutine pull data from output channel\n\t// - get a slice of output\n\t// 4. feed reduce func with the slice\n\t// * if any error occured, abort & clean goroutines\n\t// - taskCh = drain\n\t// - outCh = close & drain\n\ttype taskResult struct {\n\t\to int\n\t\te error\n\t}\n\t// for abort\n\tabort := make(chan struct{})\n\n\tabortSwicher := func() bool {\n\t\tselect {\n\t\tcase <-abort:\n\t\t\treturn true\n\t\tdefault:\n\t\t\treturn false\n\t\t}\n\t}\n\n\t// step 1\n\ttaskCh := make(chan Task)\n\tgo func() {\n\t\tfor _, t := range tasks {\n\t\t\ttaskCh <- t\n\t\t}\n\t\tclose(taskCh)\n\t}()\n\t// step 2\n\t// in order to close channel(outCh), use WaitGroup\n\tvar wg sync.WaitGroup\n\tworkerNum := 4\n\toutCh := make(chan taskResult)\n\terrCh := make(chan taskResult, 10) // blocked without a buffer\n\tfor i := 0; i < workerNum; i++ {\n\t\tgo func() {\n\t\t\tfor t := range taskCh {\n\t\t\t\tif abortSwicher() {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\twg.Add(1)\n\t\t\t\to, e := t.Execute(input)\n\t\t\t\tif e != nil && !abortSwicher() {\n\t\t\t\t\terrCh <- taskResult{o, e}\n\t\t\t\t\tclose(abort)\n\t\t\t\t} else {\n\t\t\t\t\toutCh <- taskResult{o, e}\n\t\t\t\t}\n\t\t\t\twg.Done()\n\t\t\t}\n\t\t}()\n\t}\n\n\tgo func() {\n\t\twg.Wait()\n\t\tclose(outCh)\n\t\tclose(errCh)\n\t}()\n\n\t// step 3\n\tres := []int{}\n\tdone := make(chan struct{})\n\tgo func() {\n\t\tfor o := range outCh {\n\t\t\tif abortSwicher() {\n\t\t\t\tfor range outCh {\n\t\t\t\t} // drain outCh\n\t\t\t\tfor range taskCh {\n\t\t\t\t} // drain taskCh\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tres = append(res, o.o)\n\t\t}\n\t\tclose(done)\n\t}()\n\t// step 4\n\t<-done\n\n\tif abortSwicher() {\n\t\to := <-errCh\n\t\tfor range errCh {\n\t\t} // drain errCh\n\t\treturn o.o, o.e\n\t} else {\n\t\treturn reduce(res), nil\n\t}\n}", "func doFetch(cfg config.View, be pb.BackendClient, matches chan *pb.Match) {\n\tstartTime := time.Now()\n\tmprofiles := profiles.Generate(cfg)\n\tfor {\n\t\tvar wg sync.WaitGroup\n\t\tfor _, p := range mprofiles {\n\t\t\twg.Add(1)\n\t\t\tp := p\n\t\t\tgo func(wg *sync.WaitGroup) {\n\t\t\t\tdefer wg.Done()\n\t\t\t\tfetch(be, p, matches)\n\t\t\t}(&wg)\n\t\t}\n\n\t\t// Wait for all FetchMatches calls to complete before proceeding.\n\t\twg.Wait()\n\t\tlogger.Infof(\"FetchedMatches:%v, AssignedTickets:%v, DeletedTickets:%v in time %v\", atomic.LoadUint64(&matchCount), atomic.LoadUint64(&assigned), atomic.LoadUint64(&deleted), time.Since(startTime))\n\t}\n}", "func (f *Fs) multithread(ctx context.Context, fn func(context.Context, *upstream) error) error {\n\tg, gCtx := errgroup.WithContext(ctx)\n\tfor _, u := range f.upstreams {\n\t\tu := u\n\t\tg.Go(func() (err error) {\n\t\t\treturn fn(gCtx, u)\n\t\t})\n\t}\n\treturn g.Wait()\n}", "func (b *B) RunParallel(body func(*PB))", "func BenchmarkAPIRequestPublishParallel(b *testing.B) {\n\tapp := NewTestMemoryNode()\n\tjsonData := getPublishJSON(\"channel\")\n\tb.ResetTimer()\n\tb.RunParallel(func(pb *testing.PB) {\n\t\tfor pb.Next() {\n\t\t\t_, err := apiv1.ProcessAPIData(app, jsonData)\n\t\t\tif err != nil {\n\t\t\t\tb.Error(err)\n\t\t\t}\n\t\t}\n\t})\n}", "func doWrite(work <-chan t, done <-chan struct{}) chan t {\n\tout := make(chan t)\n\tgo func() {\n\t\tdefer close(out)\n\t\tfor item := range work {\n\t\t\titem = process(item, \"written\")\n\t\t\tsleep := rand.Int63n(100)\n\t\t\ttime.Sleep(time.Duration(sleep) * time.Millisecond)\n\t\t\tselect {\n\t\t\tcase out <- item:\n\t\t\tcase <-done:\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\treturn out\n}", "func JoinWriteCloserChan(out chan<- io.WriteCloser, inp <-chan io.WriteCloser) chan struct{} {\n\tdone := make(chan struct{})\n\tgo func() {\n\t\tdefer close(done)\n\t\tfor i := range inp {\n\t\t\tout <- i\n\t\t}\n\t\tdone <- struct{}{}\n\t}()\n\treturn done\n}", "func (ru *resumableUpload) work() {\n\tdefer close(ru.done)\n\n\tsendBuf := new(bytes.Buffer)\n\tsendBuf.Grow(ru.maxChunkGroup * rblockSize)\n\tvar chunkGroupSize int\n\n\t// same as ru.blocks, but `.last` is set properly, no matter\n\t// what the size is\n\tannotatedBlocks := make(chan *rblock, ru.maxChunkGroup)\n\tgo func() {\n\t\tvar lastBlock *rblock\n\t\tdefer close(annotatedBlocks)\n\n\tannotate:\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase b := <-ru.blocks:\n\t\t\t\tif b == nil {\n\t\t\t\t\t// ru.blocks closed!\n\t\t\t\t\tbreak annotate\n\t\t\t\t}\n\n\t\t\t\t// queue block\n\t\t\t\tif lastBlock != nil {\n\t\t\t\t\tannotatedBlocks <- lastBlock\n\t\t\t\t}\n\t\t\t\tlastBlock = b\n\t\t\tcase <-ru.pushedErr:\n\t\t\t\t// stop everything\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tif lastBlock != nil {\n\t\t\tlastBlock.last = true\n\t\t\tannotatedBlocks <- lastBlock\n\t\t}\n\t}()\n\naggregate:\n\tfor {\n\t\tsendBuf.Reset()\n\t\tchunkGroupSize = 0\n\n\t\t{\n\t\t\t// do a block receive for the first vlock\n\t\t\tselect {\n\t\t\tcase <-ru.pushedErr:\n\t\t\t\t// nevermind, stop everything\n\t\t\t\treturn\n\t\t\tcase block := <-annotatedBlocks:\n\t\t\t\tif block == nil {\n\t\t\t\t\t// done receiving blocks!\n\t\t\t\t\tbreak aggregate\n\t\t\t\t}\n\n\t\t\t\t_, err := sendBuf.Write(block.data)\n\t\t\t\tif err != nil {\n\t\t\t\t\tru.pushError(errors.WithStack(err))\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tchunkGroupSize++\n\n\t\t\t\tif block.last {\n\t\t\t\t\t// done receiving blocks\n\t\t\t\t\tbreak aggregate\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t// see if we can't gather any more blocks\n\tmaximize:\n\t\tfor chunkGroupSize < ru.maxChunkGroup {\n\t\t\tselect {\n\t\t\tcase <-ru.pushedErr:\n\t\t\t\t// nevermind, stop everything\n\t\t\t\treturn\n\t\t\tcase block := <-annotatedBlocks:\n\t\t\t\tif block == nil {\n\t\t\t\t\t// done receiving blocks!\n\t\t\t\t\tbreak aggregate\n\t\t\t\t}\n\n\t\t\t\t_, err := sendBuf.Write(block.data)\n\t\t\t\tif err != nil {\n\t\t\t\t\tru.pushError(errors.WithStack(err))\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tchunkGroupSize++\n\n\t\t\t\tif block.last {\n\t\t\t\t\t// done receiving blocks\n\t\t\t\t\tbreak aggregate\n\t\t\t\t}\n\t\t\tcase <-time.After(100 * time.Millisecond):\n\t\t\t\t// no more blocks available right now, that's ok\n\t\t\t\t// let's just send what we got\n\t\t\t\tbreak maximize\n\t\t\t}\n\t\t}\n\n\t\t// send what we have so far\n\t\tru.debugf(\"Uploading %d chunks\", chunkGroupSize)\n\t\terr := ru.chunkUploader.put(sendBuf.Bytes(), false)\n\t\tif err != nil {\n\t\t\tru.pushError(errors.WithStack(err))\n\t\t\treturn\n\t\t}\n\t}\n\n\t// send the last block\n\tru.debugf(\"Uploading last %d chunks\", chunkGroupSize)\n\terr := ru.chunkUploader.put(sendBuf.Bytes(), true)\n\tif err != nil {\n\t\tru.pushError(errors.WithStack(err))\n\t\treturn\n\t}\n}", "func Worker(mapf func(string, string) []Pair, reducef func(string, []string) string) {\n\tclient := MakeRpcClient()\n\tdefer client.Close()\n\tfor {\n\t\t// 对端的 server 如果退出了,下面这个会有什么反应\n\t\ttask := Task{TaskKind: ReduceTaskFlag, TaskId: \"10\"}\n\n\t\t// fmt.Println(\"request task\")\n\t\tstatus := client.Call(\"Coordinator.RequestTask\", struct{}{}, &task)\n\t\t// fmt.Println(\"Get response\", task)\n\t\tif status == false {\n\t\t\tbreak\n\t\t}\n\n\t\tswitch task.TaskKind {\n\t\tcase MapTaskFlag:\n\t\t\t// fmt.Println(\"get map task \", task.TaskId)\n\t\t\tintermediate := mapf(task.File, readFileToString(task.File))\n\t\t\t// fmt.Println(\"map task done\")\n\t\t\tsort.Sort(ByKey(intermediate))\n\t\t\tr := MapResult{TaskId: task.TaskId, Items: divideIntoItems(intermediate)}\n\t\t\tclient.Call(\"Coordinator.UploadMapResult\", r, nil)\n\t\t\t// fmt.Println(\"map result upload\")\n\n\t\tcase ReduceTaskFlag:\n\t\t\tLog(\"get reduce task \", task.TaskId)\n\t\t\tfilename := fmt.Sprint(\"mr-out-\", task.TaskId)\n\t\t\tf, _ := os.Create(filename)\n\t\t\tdefer f.Close()\n\t\t\targFile, _ := os.Open(task.File)\n\t\t\treader := bufio.NewReader(argFile)\n\n\t\t\tfor {\n\t\t\t\tend, k, vs := readFrom(reader)\n\t\t\t\tif end {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tLog(\"reduce func call\", k)\n\t\t\t\t// fmt.Println(\"key: \", k, \"values: \", vs)\n\n\t\t\t\tv := reducef(k, vs)\n\t\t\t\tfmt.Fprintf(f, \"%v %v\\n\", k, v)\n\t\t\t}\n\t\t\tLog(\"reduce task \", task.TaskId, \"done\")\n\n\t\t\tresult := ReduceResult{TaskId: task.TaskId, Filename: filename}\n\t\t\tclient.Call(\"Coordinator.UploadReduceResult\", result, nil)\n\t\t\tLog(\"reduce task\", task.TaskId, \"result upload\")\n\n\t\tcase ShutdownFlag:\n\t\t\tfallthrough\n\t\tdefault:\n\t\t\treturn\n\t\t}\n\t}\n}", "func (mr *MapReduce) RunMaster() *list.List {\n\t// Your code here\n\tsendList := list.New()\t\t// list of jobs that need to be dispatched\n\tjobList := list.New()\t\t// list of jobs that are waiting to finish\n\tdoneChan := make(chan string)\t// dispatcher thread signals on this channel when worker finishes job successfully\n\tfailChan := make(chan struct {jobNumber int; worker string})\t// dispatched thread signals here when worker fails to process request\n\t\n\t\n\t// Add all map jobs to lists\n\tfor i := 0; i < mr.nMap; i++ {\n\t\tsendList.PushBack(i)\n\t\tjobList.PushBack(i)\n\t}\n\t\n\t// Dispatch all map jobs and wait for them to finish\n\te := sendList.Front()\n\tfor jobList.Len() > 0 {\n\t\t// dispatch jobs if any are waiting\n\t\tif e != nil {\n\t\t\tif (mr.SendJob(mr.file, Map, e.Value.(int), mr.nReduce, doneChan, failChan) == true) {\n\t\t\t\tp := e\n\t\t\t\te = e.Next()\t\t// move to next job \n\t\t\t\tsendList.Remove(p)\t// and remove current job from list only if current job successfully sent\n\t\t\t}\n\t\t}\t\n\t\t\n\t\tselect {\n\t\tcase worker := <- mr.registerChannel:\t// process new worker registrations\n\t\t\tmr.Workers[worker] = &WorkerInfo{worker, -1, false, false}\n\t\t\tfmt.Printf(\"Registered worker %v\\n\", mr.Workers[worker])\n\t\t\t\n\t\tcase worker := <- doneChan:\t\t\t\t// take finished jobs off the jobList and mark the worker as free\n\t\t\tmr.Workers[worker].busy = false\n\t\t\tjobList.Remove(FindListElement(jobList, mr.Workers[worker].currentJobNumber))\n\t\t\tmr.Workers[worker].currentJobNumber = -1\n\t\t\t\n\t\tcase failure := <- failChan:\t\t\t// if any job fails, re-add the job to the sendList and mark the worker as failed \n\t\t\tsendList.PushBack(failure.jobNumber)\n\t\t\tmr.Workers[failure.worker].failed = true\n\t\t\t\n\t\t}\n\t\t\n\t}\n\t\n\tsendList.Init()\t// clear the lists\n\tjobList.Init()\n\t\n\t// Add all reduce jobs to the lists\n\tfor i := 0; i < mr.nReduce; i++ {\n\t\tsendList.PushBack(i)\n\t\tjobList.PushBack(i)\n\t}\n\t\n\t// Dispatch all reduce jobs and wait for them to finish\n\te = sendList.Front()\n\tfor jobList.Len() > 0 {\n\t\t// dispatch jobs if any are waiting\n\t\tif e != nil {\n\t\t\tif (mr.SendJob(mr.file, Reduce, e.Value.(int), mr.nMap, doneChan, failChan) == true) {\n\t\t\t\tp := e\n\t\t\t\te = e.Next()\t\t// move to next job \n\t\t\t\tsendList.Remove(p)\t// and remove current job from list only if current job successfully sent\n\t\t\t}\n\t\t}\t\n\t\t\n\t\tselect {\n\t\tcase worker := <- mr.registerChannel:\t// process new worker registrations\n\t\t\tmr.Workers[worker] = &WorkerInfo{worker, -1, false, false}\n\t\t\tfmt.Printf(\"Registered worker %v\\n\", mr.Workers[worker])\n\t\t\t\n\t\tcase worker := <- doneChan:\t\t\t\t// take finished jobs off the jobList and mark the worker as free\n\t\t\tmr.Workers[worker].busy = false\n\t\t\tjobList.Remove(FindListElement(jobList, mr.Workers[worker].currentJobNumber))\n\t\t\tmr.Workers[worker].currentJobNumber = -1\n\t\t\t\n\t\tcase failure := <- failChan:\t\t\t// if any job fails, re-add the job to the sendList and mark the worker as failed \n\t\t\tsendList.PushBack(failure.jobNumber)\n\t\t\tmr.Workers[failure.worker].failed = true\n\t\t}\n\t\t\n\t}\n\t\n\treturn mr.KillWorkers()\t\t// kill the workers and return\n}", "func writeToArchive(work <-chan t, done <-chan struct{}) chan t {\n\tout := make(chan t)\n\tgo func() {\n\t\tdefer close(out)\n\t\tfanout := make([]<-chan t, *concurrency) // HL\n\t\tfor j := 0; j < *concurrency; j++ {\n\t\t\tfanout[j] = doWrite(work, done) // HL\n\t\t}\n\n\t\tfor merged := range merge(fanout, done) { // HL\n\t\t\tselect {\n\t\t\tcase out <- merged:\n\t\t\tcase <-done:\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t}()\n\treturn out\n}", "func (pool *Pool) executor() {\n\tfor {\n\t\tif task, ok := <- pool.taskChannel; ok {\n\t\t\ttask.Run()\n\t\t\tpool.waitGroup.Done()\n\t\t} else {\n\t\t\treturn\n\t\t}\n\t}\n}", "func Worker(mapf func(string, string) []KeyValue,\n\treducef func(string, []string) string) {\n\n\treply := GetTask()\n\tid := reply.Id\n\tfilename := reply.Filename\n\n\tfor {\n\t\tif filename == \"error\" {\n\t\t\t//fmt.Printf(\"Error getting filename from master\\n\")\n\t\t\t//return\n\t\t\ttime.Sleep(1000 * time.Millisecond)\n\t\t\treply = GetTask()\n\t\t\tid = reply.Id\n\t\t\tfilename = reply.Filename\n\t\t}\n\t\t// fmt.Printf(\"Worker received filename: %s\\n\", filename)\n\n\t\tvar intermediate []KeyValue\n\t\t//intermediate := []KeyValue{}\n\n\t\tif reply.Type == \"map\" {\n\t\t\tfile, err := os.Open(filename)\n\t\t\tif err != nil {\n\t\t\t\t//log.Fatalf(\"cannot open %v\", filename)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tcontent, err := ioutil.ReadAll(file)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatalf(\"cannot read %v\", filename)\n\t\t\t}\n\t\t\tfile.Close()\n\n\t\t\tkva := mapf(filename, string(content))\n\t\t\tintermediate = append(intermediate, kva...)\n\t\t\tWriteIntermediate(intermediate, id, reply.NReduce)\n\t\t\tCompleteMapTask(id)\n\t\t} else if reply.Type == \"reduce\" {\n\t\t\tfor _, reduce_filename := range reply.FileList {\n\t\t\t\tfile, err := os.Open(reduce_filename)\n\t\t\t\tif err != nil {\n\t\t\t\t\t//log.Fatalf(\"cannot open %v\", reduce_filename)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tdec := json.NewDecoder(file)\n\t\t\t\tfor {\n\t\t\t\t\tvar kv KeyValue\n\t\t\t\t\tif err := dec.Decode(&kv); err != nil {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t\tintermediate = append(intermediate, kv)\n\t\t\t\t}\n\t\t\t\tfile.Close()\n\t\t\t\tdefer os.Remove(reduce_filename)\n\t\t\t}\n\t\t\tsort.Sort(ByKey(intermediate))\n\t\t\t// fmt.Println(intermediate)\n\t\t\ts := []string{\"mr-out\", \"-\", strconv.Itoa(reply.Id)}\n\t\t\toname := strings.Join(s, \"\")\n\t\t\t// oname := \"mr-out-0\"\n\t\t\tofile, _ := os.Create(oname)\n\n\t\t\t//\n\t\t\t// call Reduce on each distinct key in intermediate[],\n\t\t\t// and print the result to mr-out-0.\n\t\t\t//\n\t\t\ti := 0\n\t\t\tfor i < len(intermediate) {\n\t\t\t\tj := i + 1\n\t\t\t\tfor j < len(intermediate) && intermediate[j].Key == intermediate[i].Key {\n\t\t\t\t\tj++\n\t\t\t\t}\n\t\t\t\tvalues := []string{}\n\t\t\t\tfor k := i; k < j; k++ {\n\t\t\t\t\tvalues = append(values, intermediate[k].Value)\n\t\t\t\t}\n\t\t\t\toutput := reducef(intermediate[i].Key, values)\n\n\t\t\t\t// this is the correct format for each line of Reduce output.\n\t\t\t\tfmt.Fprintf(ofile, \"%v %v\\n\", intermediate[i].Key, output)\n\n\t\t\t\ti = j\n\t\t\t}\n\t\t\tCompleteReduceTask(id)\n\t\t} else if reply.Type == \"exit\" {\n\t\t\tbreak\n\t\t}\n\n\t\ttime.Sleep(1000 * time.Millisecond)\n\t\treply = GetTask()\n\t\tid = reply.Id\n\t\tfilename = reply.Filename\n\t\tintermediate = []KeyValue{}\n\t}\n\n}", "func continuePipelineFunctionA(numbers <-chan int, squared chan<- string) {\n\tfor {\n\t\tres, open := <- numbers\n\t\tif !open {\n\t\t\tbreak\n\t\t}\n\t\tsquared <- strconv.Itoa(res) + \" is \" + strconv.Itoa(res * res) // send 'numbers' to receiving 'squared' when pipeline is un-sync'd\n\t}\n\tclose(squared)\n}", "func BenchmarkAPIRequestPublishManyParallel(b *testing.B) {\n\tapp := NewTestMemoryNode()\n\tjsonData := getNPublishJSON(\"channel\", 1000)\n\tb.ResetTimer()\n\tb.RunParallel(func(pb *testing.PB) {\n\t\tfor pb.Next() {\n\t\t\t_, err := apiv1.ProcessAPIData(app, jsonData)\n\t\t\tif err != nil {\n\t\t\t\tb.Error(err)\n\t\t\t}\n\t\t}\n\t})\n}", "func fanOutData() (output chan []mapreduce.KeyValue, done chan bool) {\n\tvar (\n\t\terr error\n\t\tfile *os.File\n\t\tfileEncoder *json.Encoder\n\t\treduceCounter int\n\t)\n\n\toutput = make(chan []mapreduce.KeyValue, REDUCE_BUFFER_SIZE)\n\tdone = make(chan bool)\n\n\tgo func() {\n\t\tfor v := range output {\n\t\t\tlog.Println(\"Fanning out file\", resultFileName(reduceCounter))\n\t\t\tif file, err = os.Create(resultFileName(reduceCounter)); err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\n\t\t\tfileEncoder = json.NewEncoder(file)\n\n\t\t\tfor _, value := range v {\n\t\t\t\tfileEncoder.Encode(value)\n\t\t\t}\n\n\t\t\tfile.Close()\n\t\t\treduceCounter++\n\t\t}\n\n\t\tdone <- true\n\t}()\n\n\treturn output, done\n}", "func (r *Resolver) worker(instances []string, lastIndex uint64) {\n\tvar err error\n\tvar newInstances []string\n\n\tfor {\n\t\ttime.Sleep(5 * time.Second)\n\t\tselect {\n\t\tcase <-r.done:\n\t\t\treturn\n\t\tdefault:\n\t\t\tnewInstances, lastIndex, err = r.getInstances(lastIndex, true)\n\t\t\tif err != nil {\n\t\t\t\t// log\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tupdatedInstances := r.updateInstances(instances, newInstances)\n\t\t\tif len(updatedInstances) > 0 {\n\t\t\t\tr.updateCh <- updatedInstances\n\t\t\t}\n\t\t\tinstances = newInstances\n\t\t}\n\t}\n}", "func BenchmarkIntSyncMapAppendParallel(b *testing.B) {\n\tsyncMap := sync.Map{}\n\tresultChan := make(chan bool)\n\n\tgorountinesCount := 8\n\n\t//match real \"hashing\" complexity\n\tnow := time.Now().Unix()\n\n\tfor g := 0; g < gorountinesCount; g++ {\n\t\tgo func() {\n\t\t\tfor n := 0; n < (b.N / gorountinesCount); n++ {\n\t\t\t\tsyncMap.Store(now, n)\n\t\t\t}\n\n\t\t\tresultChan <- true\n\t\t}()\n\t}\n\n\tb.StartTimer()\n\t<-resultChan\n\t<-resultChan\n\t<-resultChan\n\t<-resultChan\n\t<-resultChan\n\t<-resultChan\n\t<-resultChan\n\t<-resultChan\n\tb.StopTimer()\n\n}", "func TestFlowWithJoinAndSplit(t *testing.T) {\n\trequire := require.New(t)\n\n\t// TODO remove this tracing from workfow package\n\tgo func() {\n\t\tfor s := range main.TracingChan {\n\t\t\tfmt.Println(s)\n\t\t}\n\t}()\n\n\tt.Run(`Given two parallel jobs that joins to a third job,\n\twhen the workflow runs,\n\tthen the workflow outputs the result of the third job`, func(t *testing.T) {\n\n\t\t// Create workers\n\t\tjob1 := func(ctx context.Context, postman main.Postman) {\n\t\t\tv, err := postman.Receive(ctx)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif v != nil {\n\t\t\t\tfor !postman.Send(ctx, *v) {\n\t\t\t\t\t//main.Infof(\"retry send job1 %#v\", v.Value)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tw1In := make(chan main.Param)\n\t\tw1 := main.NewWorker(\n\t\t\tnil,\n\t\t\tjob1,\n\t\t\tmain.NameWOpt(\"w1\"),\n\t\t\tmain.AddOrphanInWOpt(w1In),\n\t\t\tmain.AddInfoFuncWOpt(main.Infof),\n\t\t\tmain.ErrHndFuncWOpt(errHndFunc),\n\t\t)\n\n\t\tjob2 := func(ctx context.Context, postman main.Postman) {\n\t\t\tv, err := postman.Receive(ctx)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif v != nil {\n\t\t\t\tmain.Infof(\"job2 to send %#v\", v.Value)\n\t\t\t\tfor !postman.Send(ctx, *v) {\n\t\t\t\t\t//main.Infof(\"retry send job2 %#v\", v.Value)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tw2In := make(chan main.Param)\n\t\tw2 := main.NewWorker(\n\t\t\tnil,\n\t\t\tjob2,\n\t\t\tmain.NameWOpt(\"w2\"),\n\t\t\tmain.AddOrphanInWOpt(w2In),\n\t\t\tmain.AddInfoFuncWOpt(main.Infof),\n\t\t\tmain.ErrHndFuncWOpt(errHndFunc),\n\t\t)\n\n\t\tvar job3Max = math.MinInt64\n\t\toutJob3 := make(main.JobOut)\n\t\tjob3 := func(ctx context.Context, postman main.Postman) {\n\t\t\tv, err := postman.Receive(ctx)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif v != nil {\n\t\t\t\tn := v.Value.(int)\n\t\t\t\tmain.Infof(\"job3 received %d\", n)\n\t\t\t\tif n > job3Max {\n\t\t\t\t\tjob3Max = n\n\t\t\t\t}\n\t\t\t\toutJob3 <- main.Param{Value: job3Max}\n\t\t\t}\n\t\t}\n\t\tw3 := main.NewWorker(\n\t\t\t[]*main.Worker{&w1, &w2},\n\t\t\tjob3,\n\t\t\tmain.NameWOpt(\"w3\"),\n\t\t\tmain.AddInfoFuncWOpt(main.Infof),\n\t\t\tmain.ErrHndFuncWOpt(errHndFunc),\n\t\t)\n\n\t\tvar job4Min = math.MaxInt64\n\t\toutJob4 := make(main.JobOut)\n\t\tjob4 := func(ctx context.Context, postman main.Postman) {\n\t\t\tv, err := postman.Receive(ctx)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif v != nil {\n\t\t\t\tn := v.Value.(int)\n\t\t\t\tmain.Infof(\"job4 received %d\", n)\n\t\t\t\tif n < job4Min {\n\t\t\t\t\tjob4Min = n\n\t\t\t\t}\n\t\t\t\toutJob4 <- main.Param{Value: job4Min}\n\t\t\t}\n\t\t}\n\t\tw4 := main.NewWorker(\n\t\t\t[]*main.Worker{&w2},\n\t\t\tjob4,\n\t\t\tmain.NameWOpt(\"w4\"),\n\t\t\tmain.AddInfoFuncWOpt(main.Infof),\n\t\t\tmain.ErrHndFuncWOpt(errHndFunc),\n\t\t)\n\n\t\t// create the flow\n\t\tctx, cancelFunc := context.WithCancel(context.Background())\n\t\tflow := main.NewFlow(ctx, main.AddInfoFuncFOpt(main.Infof))\n\n\t\t// Add workers to the workflow\n\t\tflow.AddWorker(w1)\n\t\tflow.AddWorker(w2)\n\t\tflow.AddWorker(w3)\n\t\tflow.AddWorker(w4)\n\n\t\t// Start the workflow\n\t\tflow.Run()\n\n\t\t// Fixtures to feed the workflow\n\t\tvalues := []main.Param{\n\t\t\t{Value: 10},\n\t\t\t{Value: 32},\n\t\t\t{Value: math.MaxInt64},\n\t\t\t{Value: math.MinInt64},\n\t\t\t{Value: 999999},\n\t\t\t{Value: -45},\n\t\t\t{Value: 100},\n\t\t}\n\n\t\t// Start the workflow output MAX consumer\n\t\tvar (\n\t\t\ti int\n\t\t\tmax int\n\t\t\twg = sync.WaitGroup{}\n\t\t)\n\t\twg.Add(1)\n\t\tgo func() {\n\t\t\tdefer wg.Done()\n\t\t\tfor {\n\t\t\t\t// finishing the consumer when all values have been received\n\t\t\t\t// this step is only necessary if we do know when finish the flow\n\t\t\t\tif i == len(values) {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\treturn\n\t\t\t\tcase p, ok := <-outJob3:\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t\tmax = p.Value.(int)\n\t\t\t\t\ti++\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\n\t\t// Start the workflow output consumer\n\t\tvar (\n\t\t\tmin int\n\t\t\todds = countOdds(len(values))\n\t\t\ti2 = 0\n\t\t)\n\t\twg.Add(1)\n\t\tgo func() {\n\t\t\tdefer wg.Done()\n\t\t\tfor {\n\t\t\t\t// finishing the consumer when all values have been received\n\t\t\t\t// this step is only necessary if we do know when finish the flow\n\t\t\t\tif i2 == odds {\n\t\t\t\t\t// all odd values positions have been readed\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\treturn\n\t\t\t\tcase p, ok := <-outJob4:\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t\tmin = p.Value.(int)\n\t\t\t\t\ti2++\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\n\t\t// Feed the workflow\n\t\tfor i := range values {\n\t\t\tif i%2 == 0 {\n\t\t\t\tw1In <- values[i]\n\t\t\t} else {\n\t\t\t\tw2In <- values[i]\n\t\t\t}\n\t\t}\n\n\t\t// Wait for all fixture data has been read\n\t\twg.Wait()\n\n\t\t// Finish the workflow\n\t\tcancelFunc()\n\t\tflow.Kill()\n\n\t\trequire.Equal(math.MaxInt64, max)\n\t\trequire.Equal(math.MinInt64, min)\n\t})\n}", "func ConcurrentRetry(tasks []func() (string, error), concurrent int, retry int) <-chan Result {\n\n\tvar wg sync.WaitGroup\n\n\tfmt.Println(\"Creating job channel\")\n\tjobsChannel := make(chan func() (string, int, error), len(tasks))\n\tfmt.Println(\"Creating result channel\")\n\tresultChannel := make(chan Result, len(tasks))\n\n\tfmt.Println(\"Builing\", concurrent, \"microwaves for the Food Heating Factory\")\n\n\t// Start microwaves\n\tfor i := 0; i < concurrent; i++ {\n\t\tgo microwave(i, jobsChannel, resultChannel, retry, &wg)\n\t}\n\n\tfmt.Println(\"Gattering jobs...\")\n\tfmt.Println(\"Filling job queue\")\n\tfmt.Println(\"Passing job queue to jobsChannel\")\n\n\tfor i := 0; i < len(tasks); i++ {\n\n\t\twg.Add(1)\n\t\tfun := tasks[i]\n\t\tind := i\n\t\ttask := func() (string, int, error) {\n\t\t\tstr, err := fun()\n\t\t\treturn str, ind, err\n\t\t}\n\n\t\tjobsChannel <- task\n\t}\n\n\tfmt.Println(\"Closing job queue channel\")\n\n\tclose(jobsChannel)\n\n\t// When all tasks are done close the result channel\n\tgo func() {\n\t\twg.Wait()\n\t\tclose(resultChannel)\n\t}()\n\n\tfmt.Println(\"Returning the output channel\")\n\n\treturn resultChannel\n}", "func (g *getter) worker() {\n for c := range g.get_ch { // HL\n g.retryGetChunk(c)\n }\n\n}", "func (localOptimizer) finish(operation chan<- Task, result <-chan Task) {\n\t// Guarantee that result is closed before operation is closed.\n\tfor range result {\n\t}\n}", "func (c *Operator) worker() {\n\tfor {\n\t\tp, ok := c.queue.pop()\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tif err := c.reconcile(p); err != nil {\n\t\t\tutilruntime.HandleError(logger.LogError(\"reconciliation failed: %v\", err))\n\t\t}\n\t}\n}", "func (pc *PolicyController) worker() {\n\tfor pc.processNextWorkItem() {\n\t}\n}", "func (d *disp) process() {\n\tfor {\n\t\tselect {\n\t\tcase pipeline := <-d.PipelineChan:\n\n\t\t\tPipelineChan := <-d.PipelineQueue\n\n\t\t\tPipelineChan <- pipeline\n\n\t\tcase job := <-d.JobChan: // listen to any submitted job on the WorkChan\n\t\t\t// wait for a worker2 to submit JobChan to JobQueue\n\t\t\t// note that this JobQueue is shared among all workers.\n\t\t\t// Whenever there is an available JobChan on JobQueue pull it\n\t\t\tJobChan := <-d.Queue\n\n\t\t\t// Once a jobChan is available, send the submitted Job on this JobChan\n\t\t\tJobChan <- job\n\t\t}\n\t}\n}", "func sumWorker(nums chan int, out chan int) {\n\t// TODO: implement me\n\t// HINT: use for loop over `nums`\n sum := 0\n for num := range nums {\n sum += num\n }\n out <- sum\n close(out)\n}", "func worker() {\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase job := <-jobs:\n\t\t\t\tupdate(job)\n\t\t\t}\n\t\t}\n\t}()\n}", "func TestConcurrencyAndRaceConditions(t *testing.T) {\n\t//\tdefer profile.Start(profile.MemProfile).Stop()\n\tfmt.Println(\"Testing for concurrency and race conditions\")\n\tq := New()\n\tq.Sched()\n\tnotifyChannel := make(chan bool)\n\tfor i := 0; i < 1000; i++ {\n\t\tgo func(j int) {\n\t\t\tfor j := 0; j < 2000; j++ {\n\t\t\t\tjd := jobData{\n\t\t\t\t\tstatusChannel: notifyChannel,\n\t\t\t\t\tdata: int32(j),\n\t\t\t\t}\n\t\t\t\tjob := Job{\n\t\t\t\t\tJobData: jd,\n\t\t\t\t\tTaskCreator: taskHandler{},\n\t\t\t\t}\n\t\t\t\tq.PushChannel() <- job\n\t\t\t}\n\t\t}(i)\n\t}\n\tcounter := 0\n\tdone := false\n\tfor {\n\t\tselect {\n\t\tcase <-notifyChannel:\n\t\t\tcounter++\n\t\t\tif counter == 2000000 {\n\t\t\t\tdone = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif done {\n\t\t\tbreak\n\t\t}\n\t}\n\tassert.Equal(t, 1999000000, int(sum))\n\tassert.Equal(t, 2000000, counter)\n}", "func (m *Manager) run() {\n\tfor i := 0; i < m.workerPool.MaxWorker; i++ {\n\t\twID := i + 1\n\t\t//log.Printf(\"[workerPool] worker %d spawned\", wID)\n\t\tgo func(workerID int) {\n\t\t\tfor task := range m.workerPool.queuedTaskC {\n\t\t\t\tlog.Printf(\"[workerPool] worker %d is processing task\", wID)\n\t\t\t\ttask()\n\t\t\t\tlog.Printf(\"[workerPool] worker %d has finished processing task\", wID)\n\t\t\t}\n\t\t}(wID)\n\t}\n}", "func (w *WaitTask) startInner(taskContext *TaskContext) {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\n\tklog.V(3).Infof(\"wait task progress: %d/%d\", 0, len(w.Ids))\n\n\tpending := object.ObjMetadataSet{}\n\tfor _, id := range w.Ids {\n\t\tswitch {\n\t\tcase w.skipped(taskContext, id):\n\t\t\terr := taskContext.InventoryManager().SetSkippedReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as skipped reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.sendEvent(taskContext, id, event.ReconcileSkipped)\n\t\tcase w.changedUID(taskContext, id):\n\t\t\t// replaced\n\t\t\tw.handleChangedUID(taskContext, id)\n\t\tcase w.reconciledByID(taskContext, id):\n\t\t\terr := taskContext.InventoryManager().SetSuccessfulReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as successful reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.sendEvent(taskContext, id, event.ReconcileSuccessful)\n\t\tdefault:\n\t\t\terr := taskContext.InventoryManager().SetPendingReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as pending reconcile: %v\", err)\n\t\t\t}\n\t\t\tpending = append(pending, id)\n\t\t\tw.sendEvent(taskContext, id, event.ReconcilePending)\n\t\t}\n\t}\n\tw.pending = pending\n\n\tklog.V(3).Infof(\"wait task progress: %d/%d\", len(w.Ids)-len(w.pending), len(w.Ids))\n\n\tif len(pending) == 0 {\n\t\t// all reconciled - clear pending and exit\n\t\tklog.V(3).Infof(\"all objects reconciled or skipped (name: %q)\", w.TaskName)\n\t\tw.cancelFunc()\n\t}\n}", "func (c *EgressController) worker() {\n\tfor c.processNextWorkItem() {\n\t}\n}" ]
[ "0.64908785", "0.625165", "0.6146995", "0.59722465", "0.5944965", "0.59253687", "0.5891594", "0.57865125", "0.5781637", "0.5769501", "0.5758638", "0.5753309", "0.5699973", "0.5681928", "0.5672774", "0.56393874", "0.5605364", "0.56021065", "0.5571599", "0.5538197", "0.55216354", "0.55019146", "0.54696214", "0.54567325", "0.5434515", "0.54329306", "0.5432889", "0.5424204", "0.5422709", "0.54206115", "0.54200053", "0.54200053", "0.541862", "0.5417252", "0.5415719", "0.54133886", "0.5403733", "0.5375567", "0.53734034", "0.53686893", "0.53667796", "0.5359984", "0.53582966", "0.5357149", "0.5345506", "0.5342561", "0.533416", "0.5329784", "0.5327439", "0.53269017", "0.5302806", "0.52940077", "0.5286193", "0.5271574", "0.5270691", "0.5267579", "0.5259068", "0.5259068", "0.5259068", "0.5254727", "0.52542764", "0.5242307", "0.5241437", "0.524055", "0.5234735", "0.5228225", "0.5222815", "0.5220473", "0.5213985", "0.5205999", "0.52037376", "0.5203331", "0.5198832", "0.5198058", "0.5196122", "0.5193239", "0.5187261", "0.5185635", "0.5183636", "0.5179039", "0.517752", "0.51768494", "0.51767397", "0.51678973", "0.51673585", "0.51602197", "0.5159613", "0.51592934", "0.5154375", "0.5153196", "0.5144991", "0.5143371", "0.5138021", "0.5136859", "0.5136726", "0.5126407", "0.51236284", "0.5118296", "0.51177907", "0.5116575" ]
0.6966479
0
build sleep for time mentioned in num and returns the same number
func executeWork(num int) Result { time.Sleep(time.Duration(num) * SleepUnit) return Result(num) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func sleep(sec int) {\n <-time.After(time.Second * time.Duration(sec))\n}", "func FindStuff(num int, c chan int) { \n pause := NextPause();\n\n for i := 0; i < 10; i++ { \n time.Sleep(pause())\n c <- num\n } \n}", "func sleep(n time.Duration) {\n\ttime.Sleep(n * time.Second)\n}", "func longFunc(num int) int {\n\ttime.Sleep(time.Second * 2)\n\treturn num * num\n}", "func send_msg_per_time(number int, duration int) {\n\tsleep := time.Duration(int64(float64(duration) / float64(number) * 1000)) * time.Millisecond\n\n\tvar count = 0\n\n\tfor count < number {\n\t\tcount++\n\t\tfmt.Printf(\"Message number %s of %s (%s)\\n\", strconv.Itoa(count), strconv.Itoa(number), time.Now().Format(\"15:04:05.00000\"))\n\t\ttime.Sleep(sleep)\n\t}\n}", "func (t *Time) Sleep(s time.Duration) {\n\tt.current = t.current.Add(s)\n}", "func WaitSecond() {time.Sleep(1*1000 * time.Millisecond)}", "func randSleep(n, m int) {\n\tr := m\n\tif m-n > 0 {\n\t\tr = rand.Intn(m-n) + n\n\t}\n\ttime.Sleep(time.Duration(r) * time.Second)\n}", "func printSquare (idx int, sleeptime int) {\n if idx % 2 == 0 {\n time.Sleep(time.Millisecond * time.Duration(sleeptime))\n }\n fmt.Println(\"square of \", idx, \"is\", idx*idx)\n}", "func getSleepTime() time.Duration {\n\treturn time.Duration(150+rand.Int31n(200)) * time.Millisecond\n}", "func exponentialBackoffSleep(\n\tbase time.Duration,\n\tcap time.Duration,\n\tattempt int,\n) time.Duration {\n\tvar pow time.Duration\n\tif attempt < 63 {\n\t\tpow = 1 << attempt\n\t} else {\n\t\tpow = math.MaxInt64\n\t}\n\n\tsleep := base * pow\n\tif sleep > cap || sleep/pow != base {\n\t\tsleep = cap\n\t}\n\n\texponentialBackoffMutex.Lock()\n\tsleep = time.Duration(exponentialBackoffRand.Int63n(int64(sleep)))\n\texponentialBackoffMutex.Unlock()\n\n\treturn sleep\n}", "func (jit *Jitter) Sleep() {\n\ttime.Sleep(jit.Increase())\n}", "func (s *SpySleeper) Sleep(){\n\ts.Calls++\n}", "func Sleep(instructionData reflect.Value, finished chan bool) int {\n\tduration, err := variable.GetValue(instructionData, \"DurationVarName\", \"DurationIsVar\", \"Duration\")\n\tif err != nil {\n\t\tfinished <- true\n\t\treturn -1\n\t}\n\n\tfmt.Println(\"FIBER INFO: Sleeping for: \", duration.(float64), \"s\")\n\ttime.Sleep(time.Duration(duration.(float64)) * time.Second)\n\tfinished <- true\n\treturn -1\n}", "func (s *SpySleeper) Sleep() {\n s.Calls++\n}", "func MilliSleep(instructionData reflect.Value, finished chan bool) int {\n\n\tduration, err := variable.GetValue(instructionData, \"DurationVarName\", \"DurationIsVar\", \"Duration\")\n\tif err != nil {\n\t\tfinished <- true\n\t\treturn -1\n\t}\n\n\tfmt.Println(\"FIBER INFO: Sleeping for: \", duration.(float64), \"ms\")\n\ttime.Sleep(time.Duration(duration.(float64)) * time.Millisecond)\n\tfinished <- true\n\treturn -1\n}", "func sleepMillisec(millisec time.Duration) {\n\ttime.Sleep(millisec * time.Millisecond)\n}", "func (t *Time) Sleep(distance Distance) {\n\ttime.Sleep(time.Duration(distance))\n}", "func sleep(length time.Duration) {\n\n\ttime.Sleep(time.Second * length)\n\n}", "func backoff(i float64) time.Duration {\n\t// Use crypt random to ensure source is good (not needed for security,\n\t// but better than seeding on time, which wouldn't avoid the clash we're using the backoff for...)\n\trandMilliSecs, _ := rand.Int(rand.Reader, big.NewInt(1000))\n\t// waitTime in milliseconds for precision\n\t// Add some jitter to avoid clashes with other backed-off attempts.\n\twaitTime := (time.Duration(math.Pow(2, i)) * time.Second) +\n\t\t(time.Duration(randMilliSecs.Int64()) * time.Millisecond)\n\ttime.Sleep(waitTime)\n\treturn waitTime\n}", "func main () {\n\tfor timer := 20; timer >= 0; timer-- {\n\t\tif timer == 0 {\n\t\t\tfmt.Println(\"Bboooomm!!!!\")\n\t\t\tbreak\n\t\t}\n\tfmt.Println(timer)\n\ttime.Sleep(1 * time.Second)\n\t}\n}", "func (gen *Gen) sleep() {\n\tnext := gen.epochns + gen.bucket*BucketLen\n\tsleep := time.Duration(next - gen.now().UTC().UnixNano())\n\ttime.Sleep(sleep)\n}", "func futexsleep(addr *uint32, val uint32, ns int64)", "func sleepSec(sec time.Duration) {\n\ttime.Sleep(sec * time.Second)\n}", "func Sleep(d time.Duration) {\n\tif d <= 0 {\n\t\treturn\n\t}\n\n\tdefaultTimerWheel.Sleep(d)\n}", "func sleep() string {\n var reply string;\n t := time.Now()\n switch {\n case t.Hour() > 23:\n reply = \"You should be asleep.\"\n case t.Hour() > 9:\n reply = \"You should be at work\"\n case t.Hour() > 17:\n reply = \"You should go home.\"\n }\n return reply\n}", "func compute(value int) {\n\tfor i := 0; i < value; i++ {\n\t\ttime.Sleep(time.Second)\n\t\tfmt.Println(i)\n\t}\n}", "func sleep(wg *sync.WaitGroup, duration time.Duration) {\n\twg.Add(1)\n\ttime.Sleep(duration)\n\twg.Done()\n}", "func SleepUntil(t int64) {\n\tsleepUntil(t)\n}", "func delay(v int) {\n\tfor i := 0; i < 684000; i++ {\n\t\tfor j := 0; j < v; j++ {\n\t\t}\n\t}\n}", "func Sleep(duration int, unit time.Duration) {\n\ttime.Sleep(time.Duration(duration) * unit)\n}", "func gosched() { time.Sleep(1 * time.Millisecond) }", "func (es *ExponentialSleeper) Sleep() {\n\ttime.Sleep(es.delay)\n\n\tnextDelay := int64(float64(es.delay) * factor)\n\tif nextDelay > es.maxDelay.Nanoseconds() {\n\t\tnextDelay = es.maxDelay.Nanoseconds()\n\t}\n\tnextDelay = int64(rand.NormFloat64()*jitter*1000000000) + nextDelay\n\tes.delay = time.Duration(nextDelay)\n}", "func busySleep(dur time.Duration) {\n\tstart := time.Now()\n\tfor time.Now().Sub(start) < dur {\n\t}\n}", "func sleep(delay uint) {\n\tfmt.Println(\"Time Now is:\",time.Now());\n\t//Time.After returns current time\n\tdelayTime := <-time.After(time.Second * time.Duration(delay))\n\tfmt.Println(\"Timestamp after delay:\",delay,\" by time.After is:\", delayTime);\n}", "func Sleep(d time.Duration) {\n\tWork.Sleep(d)\n}", "func boring1(msg string) {\n\tfor i := 0; ; i++ {\n\t\tfmt.Println(msg, i)\n\t\t//time.Sleep(time.Second)\n\t\ttime.Sleep(time.Duration(rand.Intn(1e3)) * time.Millisecond)\n\t}\n}", "func Sleep(milliseconds int64) {\n\t//fmt.Println(\"sleep: \" + strconv.FormatInt(milliseconds, 10))\n\ttime.Sleep(time.Duration(milliseconds * int64(time.Millisecond)))\n}", "func wait() {\n\t// 1000 - 2000 millisecond\n\tn := 1000 + CryptoRandNonNegInt(1000)\n\ttime.Sleep(time.Millisecond * time.Duration(n))\n}", "func Sleep(mean time.Duration, stdDev time.Duration) {\n\tfMean := float64(mean)\n\tfStdDev := float64(stdDev)\n\tdelay := time.Duration(math.Max(1, rand.NormFloat64()*fStdDev+fMean))\n\ttime.Sleep(delay)\n}", "func Backoff(try int) {\n\tnf := math.Pow(2, float64(try))\n\tnf = math.Max(1, nf)\n\tnf = math.Min(nf, 16)\n\tr := rand.Int31n(int32(nf))\n\td := time.Duration(r) * time.Second\n\ttime.Sleep(d)\n}", "func (r *MethodCallRetrier) sleepAndIncrementRetries() {\n\ttime.Sleep(time.Duration(r.waitTime) * time.Second)\n\n\tr.waitTime *= r.exponent\n\n\tr.currentRetries++\n}", "func run_timeNano() int64", "func longTask(txt string) {\n for i := 0; i < 3; i++ {\n fmt.Println(txt, \":\", i)\n time.Sleep(1 * time.Second)\n }\n}", "func lSleep(L *lua.LState) int {\n\ttc := lContext(L)\n\tmilliseconds := L.ToInt(2)\n\tduration := time.Millisecond * time.Duration(milliseconds)\n\ttc.Sleep(duration)\n\n\treturn 0\n}", "func Delay(in chan int,\n\tout chan int,\n\tseconds int64) {\n\tdelay := time.Second * time.Duration(seconds)\n\tvar v int = 0\n\tfor {\n\t\tv = <-in\n\t\ttime.Sleep(delay) //should work in go1.2\n\t\tout <- v\n\t}\n}", "func numbers(done chan<- bool) {\n\tfor i := 0; i < 5; i++ {\n\t\ttime.Sleep(2 * time.Second)\n\t\tfmt.Println(i)\n\t}\n\n\tdone <- true\n}", "func new_time() time.Duration {\n\treturn time.Duration((rand.Intn(300) + 150)) * time.Millisecond\n}", "func waitForSplitSecond() {\n\tcurNs := time.Now().Nanosecond()\n\tfor curNs > 500000000 {\n\t\ttime.Sleep(100 * time.Millisecond)\n\t\tcurNs = time.Now().Nanosecond()\n\t}\n}", "func sleepTicks(d timeUnit) {\n\tfor d != 0 {\n\t\tticks() // update timestamp\n\t\tticks := uint32(d) // current scaling only supports 100 usec to 6553 msec\n\t\ttimerSleep(ticks)\n\t\td -= timeUnit(ticks)\n\t}\n}", "func sleepTicks(d timeUnit) {\n\ttimerSleep(uint32(d))\n}", "func timerSleep(ticks uint32) {\n\ttimerWakeup.Set(0)\n\n\t// CK_INT = APB1 x2 = 84mhz\n\t// prescale counter down from 84mhz to 10khz aka 0.1 ms frequency.\n\tstm32.TIM3.PSC.Set(84000000/10000 - 1) // 8399\n\n\t// set duty aka duration\n\tarr := (ticks / 100) - 1 // convert from microseconds to 0.1 ms\n\tif arr == 0 {\n\t\tarr = 1 // avoid blocking\n\t}\n\tstm32.TIM3.ARR.Set(arr)\n\n\t// Enable the hardware interrupt.\n\tstm32.TIM3.DIER.SetBits(stm32.TIM_DIER_UIE)\n\n\t// Enable the timer.\n\tstm32.TIM3.CR1.SetBits(stm32.TIM_CR1_CEN)\n\n\t// wait till timer wakes up\n\tfor timerWakeup.Get() == 0 {\n\t\tarm.Asm(\"wfi\")\n\t}\n}", "func boring(msg string) {\n\tfor i := 0; ; i++ {\n\t\ttime.Sleep(time.Duration(rand.Intn(2e3)) * time.Millisecond)\n\t\tfmt.Println(msg, i)\n\t}\n}", "func timerSleep(ticks uint32) {\n\ttimerWakeup.Set(0)\n\n\t// STM32 timer update event period is calculated as follows:\n\t//\n\t// \t\t\tUpdate_event = TIM_CLK/((PSC + 1)*(ARR + 1)*(RCR + 1))\n\t//\n\t// Where:\n\t//\n\t//\t\t\tTIM_CLK = timer clock input\n\t// \t\t\tPSC = 16-bit prescaler register\n\t// \t\t\tARR = 16/32-bit Autoreload register\n\t// \t\t\tRCR = 16-bit repetition counter\n\t//\n\t// Example:\n\t//\n\t//\t\t\tTIM_CLK = 72 MHz\n\t// \t\t\tPrescaler = 1\n\t// \t\t\tAuto reload = 65535\n\t// \t\t\tNo repetition counter RCR = 0\n\t// \t\t\tUpdate_event = 72*(10^6)/((1 + 1)*(65535 + 1)*(1))\n\t// \t\t\tUpdate_event = 549.3 Hz\n\t//\n\t// Set the timer prescaler/autoreload timing registers.\n\n\t// TODO: support smaller or larger scales (autoscaling) based\n\t// on the length of sleep time requested.\n\t// The current scaling only supports a range of 200 usec to 6553 msec.\n\n\t// prescale counter down from 72mhz to 10khz aka 0.1 ms frequency.\n\tstm32.TIM3.PSC.Set(machine.CPUFrequency()/10000 - 1) // 7199\n\n\t// Set duty aka duration.\n\t// STM32 dividers use n-1, i.e. n counts from 0 to n-1.\n\t// As a result, with these prescaler settings,\n\t// the minimum allowed duration is 200 microseconds.\n\tif ticks < 200 {\n\t\tticks = 200\n\t}\n\tstm32.TIM3.ARR.Set(ticks/100 - 1) // convert from microseconds to 0.1 ms\n\n\t// Enable the hardware interrupt.\n\tstm32.TIM3.DIER.SetBits(stm32.TIM_DIER_UIE)\n\n\t// Enable the timer.\n\tstm32.TIM3.CR1.SetBits(stm32.TIM_CR1_CEN)\n\n\t// wait till timer wakes up\n\tfor timerWakeup.Get() == 0 {\n\t\tarm.Asm(\"wfi\")\n\t}\n}", "func (b *Backoff) Sleep() {\n\ttime.Sleep(time.Duration(b.nextSleepMsec) * time.Millisecond)\n\tb.attempt++\n\tb.nextSleepMsec *= 2\n\tif b.nextSleepMsec > b.maxSleepMsec {\n\t\tb.nextSleepMsec = b.maxSleepMsec\n\t}\n}", "func (m *Mock) Sleep(d time.Duration) {\n\t<-m.After(d)\n}", "func Sleep(d time.Duration) {\n\tCurrentClock().Sleep(d)\n}", "func (r *RetryPolicy) Sleep() {\n\tdefer r.retryCount.Inc()\n\n\tif r.retryCount.Load() == 0 {\n\t\tduration := r.retryCount.Load() * uint32(r.rangeMillSeconds[0]) / 2\n\t\ttime.Sleep(time.Duration(duration) * time.Millisecond)\n\t\treturn\n\t}\n\n\tif r.retryCount.Load() <= r.immuneCount {\n\t\tduration := r.retryCount.Load() * uint32(r.rangeMillSeconds[0])\n\t\ttime.Sleep(time.Duration(duration) * time.Millisecond)\n\t\treturn\n\t}\n\n\t// no matter retry how many times, sleep a const time and with an extra rand time.\n\trd := rand.New(rand.NewSource(time.Now().UnixNano()))\n\trandTime := rd.Intn(int(r.rangeMillSeconds[1])-int(r.rangeMillSeconds[0])) + int(r.rangeMillSeconds[0])\n\tduration := r.rangeMillSeconds[0] + uint(randTime)\n\ttime.Sleep(time.Duration(duration) * time.Millisecond)\n\treturn\n}", "func TIMER(n int) *Periph { return timer(n) }", "func sleep() {\n\ttime.Sleep(100 * time.Millisecond)\n}", "func backoff(last time.Duration, base time.Duration, max time.Duration) time.Duration {\n\tvar sleep time.Duration\n\tsleep = base + time.Duration(rand.Int63n(int64(last)*3-int64(base)))\n\tif sleep > max {\n\t\treturn max\n\t}\n\treturn sleep\n}", "func main() {\n\tsleeper := &ConfigurableSleeper{1 * time.Second, time.Sleep}\n\tCountdown(os.Stdout, sleeper)\n}", "func Try(a func() error, max time.Duration, extra ...interface{}) {\n\tx, y := 0, 1\n\tfor {\n\t\terr := actual(a, extra...)\n\t\tif err == nil {\n\t\t\treturn\n\t\t}\n\t\tt := time.Duration(x) * time.Second\n\t\tif t < max {\n\t\t\tx, y = y, x+y\n\t\t}\n\t\ttime.Sleep(t)\n\t}\n}", "func waitForSystemTime() {\n\ttime.Sleep(150 * time.Millisecond)\n}", "func SleepTicks(d TimeUnit) {\n\ttimerSleep(uint64(d))\n}", "func main() {\n\tvar wg sync.WaitGroup\n\n\tincrementer := 0\n\tgs := 100\n\twg.Add(gs)\n\tvar m sync.Mutex\n\n\tfor i := 0; i < gs; i++ {\n\t\tgo func() {\n\t\t\tm.Lock()\n\t\t\tv := incrementer\n\t\t\truntime.Gosched()\n\t\t\tv++\n\t\t\tincrementer = v\n\t\t\tfmt.Println(incrementer)\n\t\t\tm.Unlock()\n\t\t\twg.Done()\n\t\t}()\n\t}\n\twg.Wait()\n\tfmt.Println(\"end value:\", incrementer)\n}", "func SleepDelay(delay float64, jitter float64) {\n\tminSleep := delay - (delay * jitter)\n\tmaxSleep := delay + (delay * jitter)\n\tif minSleep < 3 {\n\t\tminSleep = 2\n\t\tmaxSleep += 2\n\t}\n\ttime.Sleep(time.Duration(r.Intn(int((maxSleep-minSleep))+int(minSleep))) * time.Second)\n}", "func sleep() {\n\ttime.Sleep(time.Millisecond * SLEEP_MILLISECONDS)\n}", "func randSleep(d, jitter time.Duration) {\n\tdelta := time.Duration(rand.Int63n(int64(jitter))) - (jitter / 2)\n\ttime.Sleep(d + delta)\n}", "func randomSleep() {\n\trand.Seed(time.Now().UnixNano())\n\tn := rand.Intn(10)\n\ttime.Sleep(time.Duration(n) * time.Millisecond)\n}", "func randomSleep(maxSleep int) (napTime time.Duration, napFunc func()) {\n\tnapTime = time.Duration(rand.Intn(maxSleep)) * time.Millisecond\n\tnapFunc = func() {\n\t\ttime.Sleep(napTime)\n\t}\n\treturn\n}", "func sleepEstimatorTask1(data map[int]guard) int {\n\tmaxTotalSleep := 0\n\tmostAsleepMinuteVal := -1\n\tmaxTotalSleepGuardID := -1\n\n\tfor _, guard := range data {\n\t\te := guard.events\n\t\tsleep := make([]int, 60)\n\t\tfor i := 0; i < len(e); i++ {\n\t\t\tif e[i].etype == fallsAsleep {\n\t\t\t\tif e[i+1].etype != endShift && e[i+1].etype != wakesUp {\n\t\t\t\t\tpanic(\"unexpected etype\" + string(e[i+1].etype))\n\t\t\t\t}\n\t\t\t\tupdateSlepTime(e[i].date, e[i+1].date, sleep)\n\t\t\t}\n\t\t}\n\t\ttmpMaxTotalSleep := sum(sleep)\n\t\tif tmpMaxTotalSleep > maxTotalSleep {\n\t\t\tmaxTotalSleep = tmpMaxTotalSleep\n\t\t\tmostAsleepMinuteVal = mostAsleepMinute(sleep)\n\t\t\tmaxTotalSleepGuardID = guard.id\n\t\t}\n\t}\n\treturn mostAsleepMinuteVal * maxTotalSleepGuardID\n}", "func (t *Tracker) sleepDuration(attempts int) time.Duration {\n\tsleepTime := (t.rand.Float64() + 1) + math.Pow(2, float64(attempts-0))\n\tdurationStr := fmt.Sprintf(\"%ss\", strconv.FormatFloat(sleepTime, 'f', 2, 64))\n\tsleepDuration, _ := time.ParseDuration(durationStr)\n\treturn sleepDuration\n}", "func randomTimeout(minVal, maxVal time.Duration) <-chan time.Time {\n\textra := time.Duration(rand.Int()) % maxVal\n\treturn time.After((minVal + extra) % maxVal)\n}", "func exponentialBackoff(i int) {\n\ttime.Sleep(time.Duration(Round(math.Pow(3.0, float64(i))-1.0)*100.0/2.0) * time.Millisecond)\n}", "func (t thing) run(me int, big int, two_chan chan string, cntl_chan chan string) {\n waituntil := <-cntl_chan\n t.validate_start(waituntil)\n for n:=0; n<big; n++ {\n newt := thing(math.Sqrt(float64(t * t + thing(me))))\n if n % (big / 4) == 0 {\n two_chan <- fmt.Sprintf(\"%s %d %d\\n\", PROGRESS, me, n)\n fmt.Printf(\"%s %d %d\\n\", PROGRESS, me, n)\n }\n t = newt\n }\n fmt.Printf(\"%d ended with final result %f\\n\", me, float64(t))\n two_chan <- STOP\n}", "func (s DefaultSleeper) Sleep() {\n\ttime.Sleep(1 * time.Second)\n}", "func addDelay(d time.Duration) time.Duration {\n\trand.Seed(time.Now().UnixNano())\n\n\tsec := int(math.Max(float64(d/time.Second), 1))\n\tsec = int(math.Min(float64(sec+rand.Intn(9))+1, 60)) // #nosec G404\n\n\treturn time.Duration(sec) * time.Second\n}", "func addDelay(d time.Duration) time.Duration {\n\trand.Seed(time.Now().UnixNano())\n\n\tsec := int(math.Max(float64(d/time.Second), 1))\n\tsec = int(math.Min(float64(sec+rand.Intn(9))+1, 60)) // #nosec G404\n\n\treturn time.Duration(sec) * time.Second\n}", "func HelloServer(w http.ResponseWriter, req *http.Request) {\n timestamp := time.Now().Unix()\n // 毫秒 \n time.Sleep(10 * time.Millisecond)\n io.WriteString(w, strconv.FormatInt(timestamp, 10))\n}", "func timer(testT int, consumerChan chan bool) {\n\ttime.Sleep(time.Second * time.Duration(testT))\n\tconsumerChan <- true\n}", "func Until(t Time) Duration {}", "func tickspersecond() int64", "func loop() {\n\n\tdelta := 0 * time.Nanosecond\n\n\tlast := time.Now()\n\n\tfor true {\n\t\tcur := time.Now()\n\t\tdelta += cur.Sub(last)\n\t\tlast = cur\n\n\t\tfor delta >= 15*time.Millisecond {\n\t\t\tdelta -= time.Millisecond\n\t\t\t//fmt.Println(\"Up\")\n\t\t}\n\t\t//fmt.Println(\"Re\")\n\t}\n\n}", "func main() {\n\tgodur, _ := time.ParseDuration(\"10ms\")\n\t// we are telling our application it can only use 2 processes\n\truntime.GOMAXPROCS(2)\n\n\t// anonamous function\n\tgo func() {\n\t\tfor i := 0; i < 100; i++ {\n\t\t\tprintln(\"Hello\")\n\t\t\ttime.Sleep(godur)\n\t\t}\n\t}()\n\n\t// anonamous function\n\tgo func() {\n\t\tfor i := 0; i < 100; i++ {\n\t\t\tprintln(\"World\")\n\t\t\ttime.Sleep(godur)\n\t\t}\n\t}()\n\n\t// without a sleep the function will close out before the go routines has a chance to finish\n\tdur, _ := time.ParseDuration(\"1s\")\n\ttime.Sleep(dur)\n}", "func doAnything(anything string, t time.Duration) {\r\n\tfor i := 0; i < 5; i++ {\r\n\t\tfmt.Println(anything)\r\n\t\ttime.Sleep(time.Millisecond * t)\r\n\t}\r\n}", "func doSomething() {\n\tfor j := 0; j < 10; j++ {\n\t\ttime.Sleep(50 * time.Millisecond)\n\t\tj--\n\t}\n}", "func doSomething() {\n\tfor j := 0; j < 10; j++ {\n\t\ttime.Sleep(50 * time.Millisecond)\n\t\tj--\n\t}\n}", "func NextPause() func() (pause time.Duration) {\n rando := rand.New(rand.NewSource(time.Now().UnixNano()))\n return func() (pause time.Duration) {\n pause = time.Duration((rando.Intn(500)) * int(time.Millisecond))\n return\n }\n}", "func Countdown(out io.Writer, num int, sleeper Sleeper) {\n\tfor i := num; i > 0; i-- {\n\t\tsleeper.Sleep()\n\t\tfmt.Fprintln(out, i)\n\t}\n\n\tsleeper.Sleep()\n\tfmt.Fprint(out, finalWord)\n}", "func AnounanceTime(){\n\tfor {\n\t\tfmt.Println(time.Now())\n\t\ttime.Sleep(5*time.Second)\n\t}\n}", "func sleep(ctx context.Context, d time.Duration) error {\n\tt := time.NewTimer(d)\n\tdefer t.Stop()\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tcase <-t.C:\n\t\treturn nil\n\t}\n}", "func Delay(n *NetOp, statusCode int, pass int, duration time.Duration) time.Duration {\n\tm := fmt.Sprintf(\"Delay: HTTP error %v on %v. Sleeping %v seconds, pass %d of %d.\",\n\t\tstatusCode, n.Endpoint, duration.Seconds(), pass, MaxWaitIterations)\n\tlog.Println(m)\n\tn.Println(m)\n\ttime.Sleep(duration)\n\tduration = duration * Multiplier\n\treturn duration\n}", "func main() {\n\t//var mynum int\n\t//mynum = 10\n\t//fmt.Println(mynum)\n\n\tc1 := make(chan string, 1)\n\tgo func() {\n\t\ttime.Sleep(2 * time.Second)\n\t\tc1 <- \"result 1\"\n\t}()\n\n\tselect {\n\tcase res := <-c1:\n\t\tfmt.Println(res)\n\tcase <-time.After(1 * time.Second):\n\t\tfmt.Println(\"timeout 1\")\n\t}\n\n\t//var i int\n\t//for i=0; i<5; i++{\n\t//\ttime.Sleep(1)\n\t//\tfmt.Println(i)\n\t//}\n\n}", "func sleepEstimatorTask2(data map[int]guard) int {\n\tmaxTotalSleep := -1\n\tmaxTotalSleepIndx := -1\n\tmaxTotalSleepGuardID := -1\n\n\tfor _, guard := range data {\n\t\te := guard.events\n\t\tsleep := make([]int, 60)\n\t\tfor i := 0; i < len(e); i++ {\n\t\t\tif e[i].etype == fallsAsleep {\n\t\t\t\tif e[i+1].etype != endShift && e[i+1].etype != wakesUp {\n\t\t\t\t\tpanic(\"unexpected etype\" + string(e[i+1].etype))\n\t\t\t\t}\n\t\t\t\tupdateSlepTime(e[i].date, e[i+1].date, sleep)\n\t\t\t}\n\t\t}\n\t\ttmpMaxTotalSleep, tmpIndx := max(sleep)\n\t\tif tmpMaxTotalSleep > maxTotalSleep {\n\t\t\tmaxTotalSleep = tmpMaxTotalSleep\n\t\t\tmaxTotalSleepIndx = tmpIndx\n\t\t\tmaxTotalSleepGuardID = guard.id\n\t\t}\n\t}\n\treturn maxTotalSleepIndx * maxTotalSleepGuardID\n}", "func deliveryTime(min int, max int) int {\n\trand.Seed(time.Now().UTC().UnixNano())\n\treturn min + rand.Intn(max-min)\n}", "func sayRepeat(txt string) {\n\tfor i := 0; i < 10; i++ {\n\t\t// it sleeps for 1 seconds everytime start a goroutine and return back to this position start\n\t\ttime.Sleep(100 * time.Microsecond)\n\t\tfmt.Println(txt)\n\t}\n}", "func lesson56(){\n\ttick := time.Tick(100 * time.Microsecond)\n\tboom := time.After(500 * time.Microsecond)\n\t\n\tfor {\n\t\tselect {\n\t\t\tcase <- tick:\n\t\t\t\tfmt.Println(\"tick.\")\n\t\t\tcase <- boom:\n\t\t\t\tfmt.Println(\"BOOM!\")\n\t\t\t\t//break ここでbreakしてもforからは抜けない\n\t\t\t\treturn\n\t\t\tdefault:\n\t\t\t\tfmt.Println(\" .\")\n\t\t\t\ttime.Sleep(50 * time.Microsecond)\n\t\t}\n\t}\n}", "func (fc *FakeClock) Sleep(d time.Duration) {\n\t<-fc.After(d)\n}", "func worker(done chan bool) {\r\n\tfmt.Print(\"working\")\r\n\tseconds := 10\r\n\tfor i:= 1; i <= seconds; i++ {\r\n\t\tfmt.Print(\".\")\r\n\t\ttime.Sleep(time.Second)\r\n\t}\r\n\t//after delay send notofication to channel\r\n\tfmt.Println(\"done the work\")\r\n\tdone <- true \r\n}" ]
[ "0.6189984", "0.6142917", "0.6070678", "0.60201865", "0.5921345", "0.5911421", "0.58785754", "0.5866416", "0.5822231", "0.5820751", "0.58051276", "0.57652694", "0.5728804", "0.56591916", "0.5655849", "0.5601837", "0.5596157", "0.55948293", "0.5592226", "0.557727", "0.5573602", "0.5570889", "0.55522364", "0.55270696", "0.5520957", "0.5507423", "0.55063254", "0.5502785", "0.54547274", "0.54474974", "0.54308516", "0.5415803", "0.54156137", "0.5412803", "0.54076517", "0.5396702", "0.53838557", "0.5379741", "0.5362207", "0.53308195", "0.53237104", "0.53174955", "0.53118676", "0.53056026", "0.53044266", "0.52956825", "0.52945095", "0.52934486", "0.5257944", "0.5231788", "0.52314985", "0.5211266", "0.5172805", "0.51713216", "0.51597834", "0.5148242", "0.51429427", "0.51357114", "0.51254255", "0.51188767", "0.5115333", "0.5109674", "0.5104213", "0.50896364", "0.50709414", "0.5066072", "0.5056496", "0.504372", "0.50354385", "0.50354356", "0.50341374", "0.5029704", "0.502721", "0.50076234", "0.50047827", "0.49995333", "0.49982518", "0.49940518", "0.49940518", "0.4981032", "0.49730757", "0.4959043", "0.49588248", "0.495681", "0.49495298", "0.4949123", "0.4936109", "0.4936109", "0.49304053", "0.49057287", "0.49021116", "0.48990992", "0.48985633", "0.4883791", "0.4882102", "0.48800635", "0.48793215", "0.48764247", "0.48681557", "0.48603645" ]
0.5885448
6
SpittingWebToNest returns the Nest that this Spitter is creating a Web to spit at, thus connecting them. Nil if not spitting. Value can be returned as a nil pointer.
func (spitterImpl *SpitterImpl) SpittingWebToNest() spiders.Nest { return spitterImpl.spittingWebToNestImpl }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (spitterImpl *SpitterImpl) Spit(nest spiders.Nest) bool {\n\treturn spitterImpl.RunOnServer(\"spit\", map[string]interface{}{\n\t\t\"nest\": nest,\n\t}).(bool)\n}", "func (connection *SenseConnection) Sense() *SenseUplink {\n\treturn connection.SenseUplink\n}", "func Spit(content string, spitType string, exp uint64, multipart bool) (string, error) {\n\tif multipart {\n\t\treturn ShortenMultipart(content, spitType, exp)\n\t} else {\n\t\treturn ShortenURLEnc(content, spitType, exp)\n\t}\n}", "func SearchWeb(ctx context.Context, prodName string, w http.ResponseWriter, r *http.Request) ([]string, error) {\n\tvar (\n\t\tebayInfo webUtil = &ebayUtil{\n\t\t\tName: \"Ebay\",\n\t\t\tNumPerPage: 50,\n\t\t\tOnHTML: \"div[class='s-item__wrapper clearfix']\",\n\t\t\tParallel: 13,\n\t\t\tUserAgent: \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.75 Safari/537.36\",\n\t\t}\n\t\twatsonInfo webUtil = &watsonsUtil{\n\t\t\tName: \"Watsons\",\n\t\t\tNumPerPage: 64,\n\t\t\tOnHTML: \"e2-product-list\",\n\t\t\tParallel: 3,\n\t\t\tUserAgent: \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.75 Safari/537.36\",\n\t\t}\n\n\t\twebsites = []webUtil{\n\t\t\tebayInfo,\n\t\t\twatsonInfo,\n\t\t}\n\n\t\tresultJSON []string\n\t\tmu sync.Mutex\n\t\tErr error\n\t\tch = make(chan error, 2)\n\t\tcounter = 0\n\t)\n\n\tfor _, website := range websites {\n\t\tgo func(web webUtil) {\n\t\t\tcrawlWebsite(ctx, ch, &mu, web, prodName, &resultJSON, w)\n\t\t}(website)\n\t}\n\n\tfor err := range ch {\n\t\tcounter++\n\t\tif err != nil {\n\t\t\tif Err == nil {\n\t\t\t\tErr = err\n\t\t\t} else {\n\t\t\t\tErr = fmt.Errorf(\"%v;%v\", Err, err)\n\t\t\t}\n\t\t}\n\t\tif counter == len(websites) {\n\t\t\tbreak\n\t\t}\n\t}\n\tfmt.Println(\"Done err waiting\")\n\treturn resultJSON, Err\n}", "func (k *KeyRing) buildTrustWeb() {\n\tvar queue []*Key\n\tvisited := make(map[string]bool)\n\n\t// Populate initial trusted peers.\n\t// The queue only contains peers whose signatures can be trusted.\n\tfor _, key := range k.keys {\n\t\tif key.trust >= TrustThreshold {\n\t\t\tqueue = append(queue, key)\n\t\t\tvisited[key.identity] = true\n\t\t}\n\n\t\tkey.effectiveTrust = key.trust\n\t\tkey.signedBy = nil\n\t}\n\n\t// While there are some vertexes to be processed\n\tvar current *Key\n\tfor len(queue) > 0 {\n\t\tcurrent, queue = queue[0], queue[1:]\n\n\t\t// For each signatures\n\t\tfor signee, signature := range current.Signatures {\n\n\t\t\t// The signature is valid, add its value (if exists)\n\t\t\tsigneeKey := k.keys[signee]\n\t\t\tif signeeKey == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// EffectiveTrust calculation takes into account previously\n\t\t\t// accumulated trust wrt signer's trust.\n\t\t\tsigneeKey.effectiveTrust = signeeKey.effectiveTrust.Add(\n\t\t\t\tsignature.Trust.Min(current.effectiveTrust),\n\t\t\t)\n\t\t\tsigneeKey.signedBy = append(signeeKey.signedBy, current)\n\n\t\t\t// Is it the first time we can trust the signee?\n\t\t\tif signeeKey.effectiveTrust >= TrustThreshold {\n\t\t\t\tif !visited[signee] {\n\t\t\t\t\tqueue = append(queue, signeeKey)\n\t\t\t\t\tvisited[signee] = true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tk.stale = false\n}", "func StartTweeting(twitter *twitter.Client, storageBackend storage.Pool, tweetTime time.Duration) {\n\n\t// Setup tweet scheduling\n\tts := &TweetSearch{\n\t\tChannel: make(chan *Tweet),\n\t\tTrending: trendingwrap.NewClient(),\n\t\tStorage: storageBackend,\n\t\tURLLength: twitter.Configuration.ShortUrlLengthHttps,\n\t}\n\tSetupRegularTweetSearchProcess(ts, tweetTime)\n\tlog.Println(\"Setup complete. Lets wait for the first trending project...\")\n\n\t// Waiting for tweets ...\n\tfor tweet := range ts.Channel {\n\t\t// Sometimes it happens that we won`t get a project.\n\t\t// In this situation we try to avoid empty tweets like ...\n\t\t//\t* https://twitter.com/TrendingGithub/status/628714326564696064\n\t\t//\t* https://twitter.com/TrendingGithub/status/628530032361795584\n\t\t//\t* https://twitter.com/TrendingGithub/status/628348405790711808\n\t\t// we will return here\n\t\t// We do this check here and not in tweets.go, because otherwise\n\t\t// a new tweet won`t be scheduled\n\t\tif len(tweet.ProjectName) <= 0 {\n\t\t\tlog.Println(\"No project found. No tweet sent.\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// In debug mode the twitter variable is not available, so we won`t tweet the tweet.\n\t\t// We will just output them.\n\t\t// This is a good development feature ;)\n\t\tif twitter.API == nil {\n\t\t\tlog.Printf(\"Tweet: %s (length: %d)\", tweet.Tweet, len(tweet.Tweet))\n\n\t\t} else {\n\t\t\tpostedTweet, err := twitter.Tweet(tweet.Tweet)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Tweet publishing: ❌ (%s)\\n\", err)\n\t\t\t} else {\n\t\t\t\tlog.Printf(\"Tweet publishing: ✅ (https://twitter.com/TrendingGithub/status/%s)\\n\", postedTweet.IdStr)\n\t\t\t}\n\t\t}\n\t\tts.MarkTweetAsAlreadyTweeted(tweet.ProjectName)\n\t}\n}", "func (n *Node) join() (*Node, error) {\n\tif !n.IsAllowed() {\n\t\terr := errors.New(fmt.Sprintln(n.Nodestr, \"is not allowd\"))\n\t\treturn nil, err\n\t}\n\tres, err := n.Talk(\"/join/\"+n.Myself.toxstring(), true, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Println(n.Nodestr, \"response of join:\", res)\n\tswitch len(res) {\n\tcase 0:\n\t\treturn nil, errors.New(\"illegal response\")\n\tcase 1:\n\t\tif res[0] != \"WELCOME\" {\n\t\t\treturn nil, errors.New(\"not welcomed\")\n\t\t}\n\t\treturn nil, nil\n\t}\n\tnn, err := newNode(res[1])\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil, err\n\t}\n\tif res[0] != \"WELCOME\" {\n\t\terr = errors.New(\"not welcomed\")\n\t}\n\treturn nn, err\n}", "func (ut *sportPayload) Publicize() *SportPayload {\n\tvar pub SportPayload\n\tif ut.Active != nil {\n\t\tpub.Active = *ut.Active\n\t}\n\tif ut.EventTerm != nil {\n\t\tpub.EventTerm = *ut.EventTerm\n\t}\n\tif ut.GameTerm != nil {\n\t\tpub.GameTerm = *ut.GameTerm\n\t}\n\tif ut.ID != nil {\n\t\tpub.ID = ut.ID\n\t}\n\tif ut.MaxPreSplitPrice != nil {\n\t\tpub.MaxPreSplitPrice = *ut.MaxPreSplitPrice\n\t}\n\tif ut.Name != nil {\n\t\tpub.Name = *ut.Name\n\t}\n\treturn &pub\n}", "func StartTweeting(twitter *Twitter, storageBackend storage.Pool) {\n\n\t// Setup tweet scheduling\n\tts := &TweetSearch{\n\t\tChannel: make(chan *Tweet),\n\t\tTrending: NewTrendingClient(),\n\t\tStorage: storageBackend,\n\t\tURLLength: twitter.Configuration.ShortUrlLengthHttps,\n\t}\n\tSetupRegularTweetSearchProcess(ts)\n\n\t// Waiting for tweets ...\n\tfor tweet := range ts.Channel {\n\t\t// Sometimes it happens that we won`t get a project.\n\t\t// In this situation we try to avoid empty tweets like ...\n\t\t//\t* https://twitter.com/TrendingGithub/status/628714326564696064\n\t\t//\t* https://twitter.com/TrendingGithub/status/628530032361795584\n\t\t//\t* https://twitter.com/TrendingGithub/status/628348405790711808\n\t\t// we will return here\n\t\t// We do this check here and not in tweets.go, because otherwise\n\t\t// a new tweet won`t be scheduled\n\t\tif len(tweet.ProjectName) <= 0 {\n\t\t\tlog.Print(\"No project found. No tweet sent.\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// In debug mode the twitter variable is not available, so we won`t tweet the tweet.\n\t\t// We will just output them.\n\t\t// This is a good development feature ;)\n\t\tif twitter.API == nil {\n\t\t\tlog.Printf(\"Tweet: %s (length: %d)\", tweet.Tweet, len(tweet.Tweet))\n\n\t\t} else {\n\t\t\tpostedTweet, err := twitter.Tweet(tweet.Tweet)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(err)\n\t\t\t} else {\n\t\t\t\tlog.Printf(\"Tweet %s posted\", postedTweet.IdStr)\n\t\t\t}\n\t\t}\n\t\tts.MarkTweetAsAlreadyTweeted(tweet.ProjectName)\n\t}\n}", "func New(url, token string, mock bool, l *logrus.Logger) Nest {\n\n\tinitLog(l)\n\n\tlogDebug(funcName(), \"New nest structure\", url)\n\n\t// Read mock file\n\tif mock {\n\t\tlogWarn(funcName(), \"Mock activated !!!\")\n\t\tmockFileByte = readFile(mockFile)\n\t}\n\n\trest = http.New(log)\n\n\treturn &nest{url: url, token: token, mock: mock}\n\n}", "func (n *NetNs) Get() *NetNs {\n\treturn n\n}", "func (m *EducationAssignment) GetWebUrl()(*string) {\n return m.webUrl\n}", "func WebAddress() string {\n\tif IsSukebei() {\n\t\treturn Get().WebAddress.Sukebei\n\t} else {\n\t\treturn Get().WebAddress.Nyaa\n\t}\n}", "func (o *ClientConfiguration) GetIsWebScrapingEnabled() bool {\n\tif o == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\n\treturn o.IsWebScrapingEnabled\n}", "func (o *TenantWithOfferWeb) GetSiret() string {\n\tif o == nil || o.Siret == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Siret\n}", "func LaunchWeb(conf Conf, rp string, hub *hub.Hub, q *event.Queue, addrChan chan string, webDev bool) {\n\tif rp == \"\" {\n\t\trp = rootPath\n\t}\n\tr := httprouter.New()\n\tr.HandleMethodNotAllowed = false\n\tr.NotFound = notFoundHandler{}\n\tr.PanicHandler = panicHandler\n\n\th := handler{hub: hub}\n\n\t// --- authentication ---\n\tr.POST(rp+\"/login\", h.mw(loginHandler, false))\n\tr.POST(rp+\"/logout\", h.mw(logoutHandler, true))\n\n\t// --- api ---\n\tr.GET(rp+\"/flows\", h.mw(hndAllFlows, true)) // list all the flows configs\n\tr.GET(rp+\"/flows/:id\", h.mw(hndFlow, true)) // return highest version of the flow config and run summaries from the cluster\n\tr.GET(rp+\"/flows/:id/runs/:rid\", h.mw(hndRun, true)) // returns the identified run detail (may be on another host)\n\n\t// --- push endpoints ---\n\th.setupPushes(rp+\"/push/\", r, hub)\n\n\t// --- p2p api ---\n\tr.POST(rp+\"/p2p/flows/exec\", h.mw(hndP2PExecFlow, true)) // internal api to pass a pending todo to activate it on this host\n\tr.GET(rp+\"/p2p/flows/:id/runs\", h.mw(hndP2PRuns, true)) // all summary runs from this host for this flow id\n\tr.GET(rp+\"/p2p/flows/:id/runs/:rid\", h.mw(hndP2PRun, true)) // detailed run info from this host for this flow id and run id\n\tr.GET(rp+\"/p2p/config\", h.mw(confHandler, true)) // return host config and what it knows about other hosts\n\n\t// --- static files for the spa ---\n\tif webDev { // local development mode\n\t\tserveFiles(r, \"/static/*filepath\", http.Dir(\"webapp\"))\n\t\tr.GET(\"/app/*filepath\", zipper(singleFile(\"webapp/index.html\")))\n\t} else { // release mode\n\t\tserveFiles(r, \"/static/*filepath\", assetFS())\n\t\tr.GET(\"/app/*filepath\", zipper(assetFile(\"webapp/index.html\")))\n\t}\n\n\t// serveFiles(r, \"/static/img/*filepath\", http.Dir(\"webapp/img\"))\n\t// serveFiles(r, \"/static/js/*filepath\", http.Dir(\"webapp/js\"))\n\t// serveFiles(r, \"/static/font/*filepath\", http.Dir(\"webapp/font\"))\n\n\t// ws endpoint\n\twsh := newWsHub()\n\tq.Register(wsh)\n\tr.GET(\"/ws\", wsh.getWsHandler(&h))\n\n\t// --- CORS ---\n\tr.OPTIONS(rp+\"/*all\", h.mw(nil, false)) // catch all options\n\n\t/*\n\t\tr.GET(rp+\"/flows/:flid\", h.mw(floeHandler, true))\n\t\tr.POST(rp+\"/flows/:flid/exec\", h.mw(execHandler, true))\n\t\tr.POST(rp+\"/flows/:flid/stop\", h.mw(stopHandler, true))\n\t\tr.GET(rp+\"/flows/:flid/run/:agentid/:runid\", h.mw(runHandler, true)) // get the current progress of a run for an agent and run\n\n\t\t// --- web socket connection ---\n\t\tr.GET(rp+\"/msg\", wsHandler)\n\n\n\n\t\t// --- the web page stuff ---\n\t\tr.GET(\"/build/\", indexHandler)\n\t\tserveFiles(r, \"/build/css/*filepath\", http.Dir(\"public/build/css\"))\n\t\tserveFiles(r, \"/build/fonts/*filepath\", http.Dir(\"public/build/fonts\"))\n\t\tserveFiles(r, \"/build/img/*filepath\", http.Dir(\"public/build/img\"))\n\t\tserveFiles(r, \"/build/js/*filepath\", http.Dir(\"public/build/js\"))\n\n\t*/\n\n\t// start the private server if one is configured differently to the public server\n\tif conf.PrvBind != conf.PubBind && conf.PrvBind != \"\" {\n\t\tlog.Debug(\"private server listen on:\", conf.PrvBind)\n\t\tgo launch(conf.PrvBind, conf.PrvCert, conf.PrvKey, r, nil)\n\t}\n\n\t// start the public server\n\tlog.Debug(\"pub server listen on:\", conf.PubBind)\n\tlaunch(conf.PubBind, conf.PubCert, conf.PubKey, r, addrChan)\n}", "func (st *Storage) Website(r scope.WebsiteIDer) (*Website, error) {\n\tw, err := st.website(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn NewWebsite(w).SetGroupsStores(st.groups, st.stores), nil\n}", "func Eden() Scraper {\n\treturn &edenScraper{}\n}", "func (client Client) Web(parameters Parameters) (WebResultContainer, error) {\n\tvar result webResultWrapper\n\n\tif err := client.search(parameters.GetURI(SearchTypeWeb), &result); err != nil {\n\t\treturn WebResultContainer{}, err\n\t}\n\n\treturn result.Data, nil\n}", "func (s *Site) Site() interface{} {\n\treturn s\n}", "func (m *ChatMessage) GetWebUrl()(*string) {\n return m.webUrl\n}", "func (m *BookingBusiness) GetWebSiteUrl()(*string) {\n val, err := m.GetBackingStore().Get(\"webSiteUrl\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (q VariadicQuery) NestThis() Query {\n\tq.Nested = true\n\treturn q\n}", "func (st *Storage) Websites() (WebsiteSlice, error) {\n\twebsites := make(WebsiteSlice, len(st.websites), len(st.websites))\n\tfor i, w := range st.websites {\n\t\twebsites[i] = NewWebsite(w).SetGroupsStores(st.groups, st.stores)\n\t}\n\treturn websites, nil\n}", "func (m *Application) GetWeb()(WebApplicationable) {\n return m.web\n}", "func (o *MicrosoftGraphListItem) GetWebUrl() string {\n\tif o == nil || o.WebUrl == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.WebUrl\n}", "func Web() *WebConfig {\n\treturn &AppConfig.Web\n}", "func (n *NetIter) Net() *net.IPNet {\n\tn.net.IP = n.ips.IP()\n\treturn n.net\n}", "func (m *GraphBaseServiceClient) Sites()(*if5372351befdb652f617b1ee71fbf092fa8dd2a161ba9c021bc265628b6ea82b.SitesRequestBuilder) {\n return if5372351befdb652f617b1ee71fbf092fa8dd2a161ba9c021bc265628b6ea82b.NewSitesRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func (m *GraphBaseServiceClient) Sites()(*if5372351befdb652f617b1ee71fbf092fa8dd2a161ba9c021bc265628b6ea82b.SitesRequestBuilder) {\n return if5372351befdb652f617b1ee71fbf092fa8dd2a161ba9c021bc265628b6ea82b.NewSitesRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func Getstein(c *connection) steinmuster {\n\tvar muster steinmuster\n\talleverwendet := true\n\t\n\tfmt.Println(\"anzahl der steine = \", len(steine))\n\tdieses := rand.Intn(len(steine) - 1)\n\tfmt.Println(\"zufällige zahl =\", dieses)\n\tif steine[dieses].verwendet == true {\n\t\tfor dieser := range steine {\n\t\t\tif steine[dieser].verwendet != true {\n\t\t\t\tfmt.Println(dieser, \"war der nächste noch nicht verwendete Stein\")\n\t\t\t\tdieses = dieser\n\t\t\t\talleverwendet = false\n\t\t\t\tbreak\n\t\t\t} \n\t\t\t\n\t\t\t// debug\n\t\t\tfmt.Println(dieser, \"verwendet ?\", steine[dieser].verwendet)\n\t\t}\n\t} else {\n\t\talleverwendet = false\n\t}\n\t\n\tif alleverwendet == false {\n\t\tmuster = steine[dieses]\n\t\tsteine[dieses].verwendet = true\n\t\tsteine[dieses].verbindung = c\n\t} else {\n\t\t// keine Steine mehr zum Vergeben; deswegen Dummy-Stein\n\t\tmuster = steinmuster{0, 0, 0, false, c}\n\t}\n\t\n\t// debug\n\t// fmt.Println(steine[dieses])\n\n\treturn muster\n}", "func (o *BaseItem) GetWebUrl() string {\n\tif o == nil || o.WebUrl == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.WebUrl\n}", "func (m *SharepointIds) GetSiteUrl()(*string) {\n val, err := m.GetBackingStore().Get(\"siteUrl\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (r *reviewResolver) Wine() *wineResolver {\n\tfor _,k := range data.WineList {\n\t\tif k.Id == r.r.WineId {\n\t\t\treturn &wineResolver{k}\n\t\t}\n\t}\n\treturn nil\n}", "func (*SponsoredWebPage) TypeName() string {\n\treturn \"sponsoredWebPage\"\n}", "func (k *Kite) Kite() *protocol.Kite {\n\treturn &protocol.Kite{\n\t\tUsername: k.Config.Username,\n\t\tEnvironment: k.Config.Environment,\n\t\tName: k.name,\n\t\tVersion: k.version,\n\t\tRegion: k.Config.Region,\n\t\tHostname: hostname,\n\t\tID: k.Id,\n\t}\n}", "func (muo *ModelUpdateOne) SetWeb(s string) *ModelUpdateOne {\n\tmuo.mutation.SetWeb(s)\n\treturn muo\n}", "func (o EndpointsResponsePtrOutput) Web() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *EndpointsResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Web\n\t}).(pulumi.StringPtrOutput)\n}", "func (m *SharepointIds) GetWebId()(*string) {\n val, err := m.GetBackingStore().Get(\"webId\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (o *Transaction) GetWebsite() string {\n\tif o == nil || o.Website.Get() == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Website.Get()\n}", "func Nest(err ...error) Exception {\n\tvar ex *Ex\n\tvar last *Ex\n\tvar didSet bool\n\n\tfor _, e := range err {\n\t\tif e != nil {\n\t\t\tvar wrappedEx *Ex\n\t\t\tif typedEx, isTyped := e.(*Ex); !isTyped {\n\t\t\t\twrappedEx = &Ex{\n\t\t\t\t\tclass: e.Error(),\n\t\t\t\t\tstack: callers(),\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\twrappedEx = typedEx\n\t\t\t}\n\n\t\t\tif wrappedEx != ex {\n\t\t\t\tif ex == nil {\n\t\t\t\t\tex = wrappedEx\n\t\t\t\t\tlast = wrappedEx\n\t\t\t\t} else {\n\t\t\t\t\tlast.inner = wrappedEx\n\t\t\t\t\tlast = wrappedEx\n\t\t\t\t}\n\t\t\t\tdidSet = true\n\t\t\t}\n\t\t}\n\t}\n\tif didSet {\n\t\treturn ex\n\t}\n\treturn nil\n}", "func (this *KeyspaceTerm) IsIndexJoinNest() bool {\n\treturn (this.property & TERM_INDEX_JOIN_NEST) != 0\n}", "func (st *Storage) website(r scope.WebsiteIDer) (*TableWebsite, error) {\n\tif r == nil {\n\t\treturn nil, ErrWebsiteNotFound\n\t}\n\tif c, ok := r.(scope.WebsiteCoder); ok && c.WebsiteCode() != \"\" {\n\t\treturn st.websites.FindByCode(c.WebsiteCode())\n\t}\n\treturn st.websites.FindByWebsiteID(r.WebsiteID())\n}", "func (l *WebListener) Web() net.Listener {\n\treturn l.webListener\n}", "func NewTraceWeb(ts *TraceServer, tt *TestTracker) *TraceWeb {\n\treturn &TraceWeb{\n\t\tts: ts,\n\t\ttt: tt,\n\t}\n}", "func Nest(err ...error) error {\n\tvar ex *Ex\n\tvar last *Ex\n\tvar didSet bool\n\n\tfor _, e := range err {\n\t\tif e != nil {\n\t\t\tvar wrappedEx *Ex\n\t\t\tif typedEx, isTyped := e.(*Ex); !isTyped {\n\t\t\t\twrappedEx = &Ex{\n\t\t\t\t\tclass: e.Error(),\n\t\t\t\t\tstack: callers(),\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\twrappedEx = typedEx\n\t\t\t}\n\n\t\t\tif wrappedEx != ex {\n\t\t\t\tif ex == nil {\n\t\t\t\t\tex = wrappedEx\n\t\t\t\t\tlast = wrappedEx\n\t\t\t\t} else {\n\t\t\t\t\tlast.inner = wrappedEx\n\t\t\t\t\tlast = wrappedEx\n\t\t\t\t}\n\t\t\t\tdidSet = true\n\t\t\t}\n\t\t}\n\t}\n\tif didSet {\n\t\treturn ex\n\t}\n\treturn nil\n}", "func (c *ChromeTarget) Network() *ChromeNetwork {\n\tif c.network == nil {\n\t\tc.network = newChromeNetwork(c)\n\t}\n\treturn c.network\n}", "func (n *NetworkBuilder) Build() workloads.Network {\n\treturn n.Network\n}", "func (nsr *NatsServiceRepository) Transport() natsevents.Transport {\n\treturn nsr.transport\n}", "func (i *treapIter) span() *mspan {\n\treturn i.t.span\n}", "func (r *Resolver) Wht() WhtResolver { return &whtResolver{r} }", "func (o *TransferOriginatorDiligence) GetWebsite() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Website\n}", "func (o *GovChainMetadata) GetWebsite() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Website\n}", "func ExampleScrape(count string, cors bool, tow int) (string, bool) {\n\t// Request the HTML page.\n\tvar c int\n\tvar err error\n\tvar findthis string\n\tc, err = strconv.Atoi(count)\n\tvar url string\n\turl = \"https://t.me/s/V2List\"\n\tif cors {\n\t\turl = strings.Join([]string{\"https://cors.izumana.ml\", url}, \"/?url=\")\n\t}\n\tres, err := http.Get(url)\n\tif err != nil {\n\t\treturn \"bad\", false\n\t}\n\tdefer res.Body.Close()\n\tif res.StatusCode != 200 {\n\t\t// log.Fatalf(\"status code error: %d %s\", res.StatusCode, res.Status)\n\t\treturn res.Status, false\n\t}\n\n\t// Load the HTML document\n\tdoc, err := goquery.NewDocumentFromReader(res.Body)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\troot := doc.Find(\"body.widget_frame_base > main.tgme_main > div.tgme_container > section.tgme_channel_history > div.tgme_widget_message_wrap\")\n\tlength := root.Length()\n\tif tow <= 0 {\n\t\tfindthis = root.Eq(length - c).Find(\"div.tgme_widget_message_text\").Text()\n\t} else {\n\t\tfor ; tow >= 0; tow-- {\n\t\t\t// fmt.Println(c + tow)\n\t\t\tfindthis += root.Eq(length - (c + tow)).Find(\"div.tgme_widget_message_text\").Text()\n\t\t\t// fmt.Println(findthis)\n\t\t}\n\t}\n\t// fmt.Println(findthis)\n\treturn findthis, true\n}", "func FindProspectNetwork(prospectID int) (*Network, bool) {\n\tlog.Printf(\"debug: finding network for %d\", prospectID)\n\tvar network Network\n\terr := connection.QueryRow(\n\t\t`SELECT id, ec2_instance_id, network, mask,\n\t prospect_id, exam_container_ip, proctor_container_ip, gateway_ip, ssh_port,\n\t created, modified\n\tFROM prospect_network\n\tWHERE prospect_id = $1`, prospectID,\n\t).Scan(\n\t\t&network.ID,\n\t\t&network.EC2InstanceID,\n\t\t&network.Mask,\n\t\t&network.ProspectID,\n\t\t&network.ExamContainerIP,\n\t\t&network.ProctorContainerIP,\n\t\t&network.GatewayIP,\n\t\t&network.SSHPort,\n\t\t&network.Created,\n\t\t&network.Modified,\n\t)\n\tif err != nil {\n\t\tlog.Printf(\"error: failed to find prospect network for %d: %s\", prospectID, err)\n\t\treturn nil, false\n\t}\n\treturn &network, true\n}", "func (o *ClientConfiguration) GetIsWebScrapingEnabledOk() (*bool, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.IsWebScrapingEnabled, true\n}", "func JoinTournament(ctx AppContext) error {\n\tform := new(Join)\n\tif err := ctx.Bind(form); err != nil {\n\t\treturn ctx.JSON(http.StatusBadRequest, Error(err))\n\t}\n\n\tservice := services.NewTournamentService(\n\t\tdb.NewTournamentRepo(ctx.Session),\n\t\tdb.NewPlayerRepo(ctx.Session),\n\t)\n\n\tif err := service.Join(form.TournamentID, form.PlayerID, form.Backers...); err != nil {\n\t\treturn ctx.JSON(http.StatusBadRequest, Error(err))\n\t}\n\n\treturn ctx.JSON(http.StatusOK, form)\n}", "func SitedbUrl() string {\n\turl := bareSitedbUrl()\n\tif strings.HasPrefix(url, \"http\") {\n\t\treturn url\n\t}\n\treturn utils.AdjustUrl(fmt.Sprintf(\"%s/%s\", FrontendURL, url))\n}", "func (se *SharedElement) URL() string {\n\treturn \"#\" + se.ID()\n}", "func (se *SharedElement) URL() string {\n\treturn \"#\" + se.ID()\n}", "func (socket *MockSocket) Tethering() *socket.TetheringProtocol {\n\treturn socket.tethering\n}", "func ExampleWebAppsClient_GetNetworkTracesV2() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclient, err := armappservice.NewWebAppsClient(\"34adfa4f-cedf-4dc0-ba29-b6d1a69ab345\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := client.GetNetworkTracesV2(ctx,\n\t\t\"testrg123\",\n\t\t\"SampleApp\",\n\t\t\"c291433b-53ad-4c49-8cae-0a293eae1c6d\",\n\t\tnil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// TODO: use response item\n\t_ = res\n}", "func ExampleWebAppsClient_GetNetworkTracesSlot() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclient, err := armappservice.NewWebAppsClient(\"34adfa4f-cedf-4dc0-ba29-b6d1a69ab345\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := client.GetNetworkTracesSlot(ctx,\n\t\t\"testrg123\",\n\t\t\"SampleApp\",\n\t\t\"c291433b-53ad-4c49-8cae-0a293eae1c6d\",\n\t\t\"Production\",\n\t\tnil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// TODO: use response item\n\t_ = res\n}", "func (c *generalConfig) P2PNetworkingStack() (n ocrnetworking.NetworkingStack) {\n\tstr := c.P2PNetworkingStackRaw()\n\terr := n.UnmarshalText([]byte(str))\n\tif err != nil {\n\t\tc.lggr.Panicf(\"P2PNetworkingStack failed to unmarshal '%s': %s\", str, err)\n\t}\n\treturn n\n}", "func SearchEggWeb(c *bm.Context) {\n\tvar (\n\t\terr error\n\t\teggs []eggModel.SearchEggWeb\n\t)\n\tres := map[string]interface{}{}\n\tif eggs, err = eggSvc.SearchEggWeb(); err != nil {\n\t\tres[\"message\"] = \"Web搜索查询失败:\" + err.Error()\n\t\tc.JSONMap(res, ecode.RequestErr)\n\t\treturn\n\t}\n\tc.JSON(eggs, nil)\n}", "func (e EdgePair) To() Node {\n\tif e[0] != nil {\n\t\treturn e[0].To()\n\t} else if e[1] != nil {\n\t\treturn e[1].To()\n\t}\n\treturn nil\n}", "func spreadingDesease(networkPointer *bigNet, epochs int, epochsResultsPointer *[simulationEpochs][5]int, muskPointer *muskMeasure, socialDistancePointer *socialDistancingMeasure, ssnPointer *nationalHealthcareSystem, trialsResultsPointer *[][3]int, ssnEpochResults *[simulationEpochs][2]int, trial int) error {\n\tfor epoch := 0; epoch < epochs; epoch++ {\n\n\t\tif epoch == 0 {\n\t\t\t// pick a random infect over the graph\n\t\t\tcase0 := rand.Intn(nNodes)\n\t\t\t(*networkPointer)[case0].Infective = true\n\t\t\tlog.Println(\"CASE 0:\", case0)\n\t\t\tinfectiveDaysLen := len((*networkPointer)[case0].InfectiveDays)\n\n\t\t\tfor day := 0; day < infectiveDaysLen; day++ {\n\t\t\t\tif (*networkPointer)[case0].InfectiveDays[day] == 0 {\n\n\t\t\t\t\tisInfected, infected := middlewareContainmentMeasure(&(*networkPointer)[case0], muskPointer, socialDistancePointer, epoch)\n\n\t\t\t\t\tif isInfected {\n\n\t\t\t\t\t\tif (*networkPointer)[infected].InfectiveEpochs > 0 {\n\t\t\t\t\t\t\t// Check if Healthcare is neeeded\n\t\t\t\t\t\t\trequireHealthcare, typeHealthcare := bernoulliHealthcare(pIntensiveCare, pSubIntensiveCare)\n\t\t\t\t\t\t\tif requireHealthcare {\n\t\t\t\t\t\t\t\t// if Healthcare needed, check if there are bed available\n\t\t\t\t\t\t\t\taddedToSSN := addToSSN(ssnPointer, uint32(infected), typeHealthcare)\n\t\t\t\t\t\t\t\tif addedToSSN {\n\t\t\t\t\t\t\t\t\t// if added to SSN can still infect others\n\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].Infective = true\n\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].InfectiveEpochs += hospitalDays\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\t// if not possible to add to SSN the patient is dead\n\t\t\t\t\t\t\t\t\t//log.Println(\"NO BED AVAILABLE\")\n\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].InfectiveEpochs = 0\n\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].Dead = true\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t(*networkPointer)[infected].Infective = true\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\n\t\t\t\t\t// I set to -1 in order to not consider it anymore\n\t\t\t\t\t(*networkPointer)[case0].InfectiveDays[day] = -1\n\n\t\t\t\t} else if (*networkPointer)[case0].InfectiveDays[day] > 0 {\n\t\t\t\t\t(*networkPointer)[case0].InfectiveDays[day]--\n\t\t\t\t}\n\t\t\t}\n\t\t\t// make time pass and reduce the remaining infective days\n\t\t\t_ = reduceInfectiveEpochs(&(*networkPointer)[case0], ssnPointer, uint32(case0))\n\n\t\t} else {\n\t\t\tinfected := getInfected(networkPointer)\n\n\t\t\tfor _, infectedID := range infected {\n\n\t\t\t\tinfectiveDaysLen := len((*networkPointer)[infectedID].InfectiveDays)\n\n\t\t\t\tfor day := 0; day < infectiveDaysLen; day++ {\n\t\t\t\t\tif (*networkPointer)[infectedID].InfectiveDays[day] == 0 {\n\n\t\t\t\t\t\tisInfected, infected := middlewareContainmentMeasure(&(*networkPointer)[infectedID], muskPointer, socialDistancePointer, epoch)\n\n\t\t\t\t\t\tif isInfected {\n\t\t\t\t\t\t\tif (*networkPointer)[infected].Infective == false &&\n\t\t\t\t\t\t\t\t(*networkPointer)[infected].Dead == false &&\n\t\t\t\t\t\t\t\t(*networkPointer)[infected].Survived == false &&\n\t\t\t\t\t\t\t\t(*networkPointer)[infected].InfectiveEpochs > 0 {\n\n\t\t\t\t\t\t\t\t// Check if Healthcare is neeeded\n\t\t\t\t\t\t\t\trequireHealthcare, typeHealthcare := bernoulliHealthcare(pIntensiveCare, pSubIntensiveCare)\n\t\t\t\t\t\t\t\tif requireHealthcare {\n\t\t\t\t\t\t\t\t\t// if Healthcare needed, check if there are bed available\n\t\t\t\t\t\t\t\t\taddedToSSN := addToSSN(ssnPointer, uint32(infected), typeHealthcare)\n\t\t\t\t\t\t\t\t\tif addedToSSN {\n\t\t\t\t\t\t\t\t\t\t// if added to SSN can still infect others\n\t\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].Infective = true\n\t\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].InfectiveEpochs += hospitalDays\n\t\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\t\t//log.Println(\"NO BED AVAILABLE\")\n\t\t\t\t\t\t\t\t\t\t// if not possible to add to SSN the patient is dead\n\t\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].InfectiveEpochs = 0\n\t\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].Dead = true\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\t(*networkPointer)[infected].Infective = true\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\t// I set to -1 in order to not consider it anymore\n\t\t\t\t\t\t(*networkPointer)[infectedID].InfectiveDays[day] = -1\n\t\t\t\t\t} else if (*networkPointer)[infectedID].InfectiveDays[day] > 0 {\n\t\t\t\t\t\t(*networkPointer)[infectedID].InfectiveDays[day]--\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// make time pass and reduce the remaining infective days\n\t\t\t\t_ = reduceInfectiveEpochs(&(*networkPointer)[infectedID], ssnPointer, uint32(infectedID))\n\t\t\t}\n\t\t}\n\n\t\tinfectNumber := countInfected(networkPointer, true, false, false)\n\t\tlog.Println(\"EPOCH\\t\", epoch,\n\t\t\t\"\\tACTIVE:\\t\", infectNumber,\n\t\t\t\"\\t\\tINT.CARE:\\t\", len((*ssnPointer).intensiveCareHospitalization),\n\t\t\t\"\\tSUB.INT.CARE:\", len((*ssnPointer).subIntensiveCareHospitalization))\n\n\t\t// number of infected today\n\t\t(*epochsResultsPointer)[epoch][0] = infectNumber\n\t\t// new number of infected today regards yesterday\n\t\tif epoch != 0 {\n\t\t\tlastInfected := (*epochsResultsPointer)[epoch-1][0]\n\t\t\t(*epochsResultsPointer)[epoch][1] = infectNumber - int(lastInfected)\n\t\t} else {\n\t\t\t(*epochsResultsPointer)[epoch][1] = infectNumber\n\t\t}\n\n\t\t// number of total infected\n\t\t(*epochsResultsPointer)[epoch][2] = countTotalInfected(networkPointer)\n\t\t// number of total recovered\n\t\t(*epochsResultsPointer)[epoch][3] = countInfected(networkPointer, false, true, false)\n\t\t// number of total deaths\n\t\t(*epochsResultsPointer)[epoch][4] = countInfected(networkPointer, false, false, true)\n\n\t\t// number of intensive care\n\t\t(*ssnEpochResults)[epoch][0] = len((*ssnPointer).intensiveCareHospitalization)\n\t\t// number of sub intensive care\n\t\t(*ssnEpochResults)[epoch][1] = len((*ssnPointer).subIntensiveCareHospitalization)\n\n\t\truntime.GC()\n\t}\n\n\t// assign number of total infected to col 0 of trial\n\t(*trialsResultsPointer)[trial][0] = countTotalInfected(networkPointer)\n\t// assign number of total recovered to col 1 of trial\n\t(*trialsResultsPointer)[trial][1] = countInfected(networkPointer, false, true, false)\n\t// assign number of total deaths to col 2 of trial\n\t(*trialsResultsPointer)[trial][2] = countInfected(networkPointer, false, false, true)\n\n\treturn nil\n}", "func GetSriovnetProvider() SriovnetProvider {\n\treturn sriovnetProvider\n}", "func GetSriovnetProvider() SriovnetProvider {\n\treturn sriovnetProvider\n}", "func (s *SponsoredWebPage) GetURL() (value string) {\n\tif s == nil {\n\t\treturn\n\t}\n\treturn s.URL\n}", "func (mu *ModelUpdate) SetWeb(s string) *ModelUpdate {\n\tmu.mutation.SetWeb(s)\n\treturn mu\n}", "func GetNetwork() (network build.Network, err error) {\n\tenv, err := agoraenv.FromEnvVariable()\n\tif err != nil {\n\t\treturn build.Network{}, err\n\t}\n\n\tswitch env {\n\tcase agoraenv.AgoraEnvironmentProd:\n\t\treturn prodNetwork, nil\n\tdefault:\n\t\treturn testNetwork, nil\n\t}\n}", "func (r *Repeater) Hub() *Hub {\n\treturn r.h\n}", "func (m *Group) GetSites()([]Siteable) {\n return m.sites\n}", "func GetWebPort() int {\n\treturn c.serverPort\n}", "func grabThePage() *html.Node {\n\t// grab url from args\n\tvar link = os.Args[1]\n\n\t// add https in front if necessary\n\tif !strings.HasPrefix(link,\"http\") {\n\t\tlink = \"https://\" + link\n\t}\n\t// make request\n\tget, err := http.Get(link)\n\tif err != nil {\n\t\tlog.Fatalf(\"can't load page: %s\\n\", err)\n\t}\n\t// grab the response body\n\tdefer get.Body.Close()\n\tnode, err := html.Parse(get.Body)\n\n\tif err != nil {\n\t\tlog.Fatalf(\"Cant parse html: %s\", err)\n\t}\n\treturn node\n}", "func (rel *Relationship) GetTo() *Node {\n\tnode := &en.ENode{ID: rel.SecondNodeID}\n\tif engine.GetObject(node) {\n\t\treturn &Node{node}\n\t}\n\tlogger.Error.Print(\"Node doesn't exists\")\n\treturn nil\n}", "func (s WashingtonPostScraper) CreateNewWashingtonPostScraper() *WashingtonPostScraper {\n\tc := colly.NewCollector()\n\t// c := colly.NewCollector(colly.Debugger(&debug.LogDebugger{}))\n\tc.UserAgent = s.UserAgent()\n\tc.IgnoreRobotsTxt = false\n\n\t// Adding this wait so AJAX can load, might need to look at https://github.com/chromedp/chromedp in the future\n\tc.Limit(&colly.LimitRule{\n\t\tDelay: 5 * time.Second,\n\t})\n\n\tscraper := WashingtonPostScraper{\n\t\tcollector: c,\n\t}\n\treturn &scraper\n}", "func (m *OnlineMeetingInfo) GetJoinUrl()(*string) {\n val, err := m.GetBackingStore().Get(\"joinUrl\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (k *Kraken) GetWebsocket() (*wshandler.Websocket, error) {\n\treturn k.Websocket, nil\n}", "func (bh *BlankHost) Network() inet.Network {\n\treturn bh.n\n}", "func (i *InputInlineQueryResultSticker) GetStickerURL() (value string) {\n\tif i == nil {\n\t\treturn\n\t}\n\treturn i.StickerURL\n}", "func ExampleWebAppsClient_BeginStartWebSiteNetworkTraceOperation() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclient, err := armappservice.NewWebAppsClient(\"34adfa4f-cedf-4dc0-ba29-b6d1a69ab345\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := client.BeginStartWebSiteNetworkTraceOperation(ctx,\n\t\t\"testrg123\",\n\t\t\"SampleApp\",\n\t\t&armappservice.WebAppsClientBeginStartWebSiteNetworkTraceOperationOptions{DurationInSeconds: to.Ptr[int32](60),\n\t\t\tMaxFrameLength: nil,\n\t\t\tSasURL: nil,\n\t\t})\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\tres, err := poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n\t// TODO: use response item\n\t_ = res\n}", "func connected(url1 string, url2 string) bool {\n\n\turl2 = fixUrl(url2)\n\n\tmasterList := hashset.New() // The list set ensuring we don't revisit web pages\n\tmasterStack := arraystack.New() // The list of elements we need to visit\n\n\t//listified := arraystack.New()\n\t//listified.Push(url1)\n\t//urlTemp, _ := listified.Pop()\n\n\tmasterStack.Push(url1) // Adding the first element to the stack\n\tmasterList.Add(url1)\n\n\tnumTimes := 5\n\n\tfor numTimes > 0 { // each numTimes iteration is another depth level\n\t\tnumTimes--\n\n\t\tretSet := hashset.New() // Form a temporary set to hold elements from our current search\n\t\tfor masterStack.Size() > 0 { // Used to visit all of the elements in the stack\n\t\t\tcurrentUrl, err := masterStack.Pop() // Get the top element of the stack\n\t\t\tif err == false { // If the stack is empty break; redundent\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttempSet := getPageWords(currentUrl.(string)) // Get the urls from a given site\n\t\t\ttoStore := tempSet.Values() // Get the string values of the urls\n\t\t\tfor _, value := range toStore {\n\t\t\t\tretSet.Add(value) // add the string values of our current search\n\t\t\t}\n\t\t}\n\n\t\t// Get the values our the search results from every web page we visited on a particular level\n\t\tlinkNames := retSet.Values()\n\t\tfor _, value := range linkNames {\n\t\t\tif masterList.Contains(value) { // If we have visited the page before don't revisit\n\t\t\t\tcontinue\n\t\t\t} else {\n\n\t\t\t\t// If we have not visited the page add it as a page to visit and one not to visit again\n\t\t\t\tmasterList.Add(value)\n\t\t\t\tmasterStack.Push(value)\n\t\t\t\t//fmt.Println(value)\n\t\t\t}\n\t\t}\n\n\t\tif masterList.Contains(url2) {\n\t\t\treturn true\n\t\t}\n\t}\n\n\tvals := masterList.Values()\n\n\t/*for _, value := range vals {\n\t\tfmt.Println(value)\n\t}*/\n\n\tfor _, value := range vals {\n\t\tif value == url2 {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n\n}", "func (m *TelecomExpenseManagementPartner) GetUrl()(*string) {\n val, err := m.GetBackingStore().Get(\"url\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func WebPort() string {\n\treturn port\n}", "func ExampleWebAppsClient_BeginStartWebSiteNetworkTraceOperationSlot() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclient, err := armappservice.NewWebAppsClient(\"34adfa4f-cedf-4dc0-ba29-b6d1a69ab345\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := client.BeginStartWebSiteNetworkTraceOperationSlot(ctx,\n\t\t\"testrg123\",\n\t\t\"SampleApp\",\n\t\t\"Production\",\n\t\t&armappservice.WebAppsClientBeginStartWebSiteNetworkTraceOperationSlotOptions{DurationInSeconds: to.Ptr[int32](60),\n\t\t\tMaxFrameLength: nil,\n\t\t\tSasURL: nil,\n\t\t})\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\tres, err := poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n\t// TODO: use response item\n\t_ = res\n}", "func (n *Node) join(nodeIp string) {\n\n\t/* Alone, set successor and predecessor to myself */\n\tif nodeIp == \"\" {\n\t\tn.setSuccessor(n.info)\n\t\tn.setPrev(n.info)\n\t\treturn\n\t}\n\tnode := shared.Search {\n\t\tIp: n.info.Ip,\n\t\tId: n.info.Id.Bytes(),\n\t\tRpcPort: n.info.RpcPort,\n\t\tHttpPort: n.info.HttpPort }\n\n\tr := &shared.Reply{}\n\n\taddr := strings.Split(nodeIp, \":\")\n\terr := shared.SingleCall(\"Node.FindSuccessor\", addr[0], addr[1], node, r)\n\tif err != nil {\n\t\tn.logger.Error(err.Error())\n\t\treturn\n\t}\n\n\tn.setPrev(r.Prev)\n\tn.setSuccessor(r.Next)\n}", "func (o EndpointsResponseOutput) Web() pulumi.StringOutput {\n\treturn o.ApplyT(func(v EndpointsResponse) string { return v.Web }).(pulumi.StringOutput)\n}", "func (o *TenantWithOfferWeb) GetOfferSettings() map[string]SettingsSection {\n\tif o == nil || o.OfferSettings == nil {\n\t\tvar ret map[string]SettingsSection\n\t\treturn ret\n\t}\n\treturn *o.OfferSettings\n}", "func (subnet *VirtualNetworksSubnet) Hub() {}", "func WebPageResolver(article *Article) ArticleImage {\n\tvar ret ArticleImage\n\tdoc := article.Doc\n\timgs := doc.Find(\"img\")\n\tvar candidates []candidate\n\tsignificantSurface := 320 * 200\n\tsignificantSurfaceCount := 0\n\tsrc := \"\"\n\timgs.Each(func(i int, tag *goquery.Selection) {\n\t\tsurface := 0\n\t\tsrc, _ = tag.Attr(\"src\")\n\t\tif src == \"\" {\n\t\t\tsrc, _ = tag.Attr(\"data-src\")\n\t\t}\n\t\tif src == \"\" {\n\t\t\tsrc, _ = tag.Attr(\"data-lazy-src\")\n\t\t}\n\t\tif src == \"\" {\n\t\t\treturn\n\t\t}\n\n\t\twidth, _ := tag.Attr(\"width\")\n\t\theight, _ := tag.Attr(\"height\")\n\t\talt, _ := tag.Attr(\"alt\")\n\t\tif width != \"\" {\n\t\t\tw, _ := strconv.Atoi(width)\n\t\t\tif height != \"\" {\n\t\t\t\th, _ := strconv.Atoi(height)\n\t\t\t\tsurface = w * h\n\t\t\t} else {\n\t\t\t\tsurface = w\n\t\t\t}\n\t\t} else {\n\t\t\tif height != \"\" {\n\t\t\t\tsurface, _ = strconv.Atoi(height)\n\t\t\t} else {\n\t\t\t\tsurface = 0\n\t\t\t}\n\t\t}\n\n\t\tif surface > significantSurface {\n\t\t\tsignificantSurfaceCount++\n\t\t}\n\n\t\ttagscore := score(tag)\n\t\tif tagscore >= 0 {\n\t\t\tc := candidate{\n\t\t\t\turl: src,\n\t\t\t\twidth: width,\n\t\t\t\theight: height,\n\t\t\t\tcaption: alt,\n\t\t\t\tsurface: surface,\n\t\t\t\tscore: score(tag),\n\t\t\t}\n\t\t\tcandidates = append(candidates, c)\n\t\t}\n\t})\n\n\tif len(candidates) == 0 {\n\t\treturn ret\n\t}\n\n\tif significantSurfaceCount > 0 {\n\t\tbestCandidate := findBestCandidateFromSurface(candidates)\n\t\tret.URL = bestCandidate.url\n\t\tret.Width, _ = strconv.Atoi(bestCandidate.width)\n\t\tret.Height, _ = strconv.Atoi(bestCandidate.height)\n\t\tret.Caption = bestCandidate.caption\n\t} else {\n\t\tbestCandidate := findBestCandidateFromScore(candidates)\n\t\tret.URL = bestCandidate.url\n\t\tret.Width, _ = strconv.Atoi(bestCandidate.width)\n\t\tret.Height, _ = strconv.Atoi(bestCandidate.height)\n\t\tret.Caption = bestCandidate.caption\n\t}\n\n\ta, err := url.Parse(ret.URL)\n\tif err != nil {\n\t\treturn ret\n\t}\n\tfinalURL, err := url.Parse(article.FinalURL)\n\tif err != nil {\n\t\treturn ret\n\t}\n\tb := finalURL.ResolveReference(a)\n\tret.URL = b.String()\n\n\treturn ret\n}", "func (c *Curl) SqueezeTrytes(tritsCount int) (Trytes, error) {\n\ttrits, err := c.Squeeze(tritsCount)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn MustTritsToTrytes(trits), nil\n}", "func (o *TenantWithOfferWeb) HasSiret() bool {\n\tif o != nil && o.Siret != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (ts *TeeSpan) Tracer() opentracing.Tracer {\n\treturn ts.tracer\n}", "func (t Tunes) StreamURL() string {\n\treturn tunesStreamURL\n}", "func LeninaCrowne(wg *sync.WaitGroup, updateSite bool, knownScenes []string, out chan<- models.ScrapedScene) error {\n\treturn SexLikeReal(wg, updateSite, knownScenes, out, \"leninacrowne\", \"LeninaCrowne\", \"Terrible\")\n}", "func (ut *teamPayload) Publicize() *TeamPayload {\n\tvar pub TeamPayload\n\tif ut.HomeTownID != nil {\n\t\tpub.HomeTownID = *ut.HomeTownID\n\t}\n\tif ut.ID != nil {\n\t\tpub.ID = ut.ID\n\t}\n\tif ut.Name != nil {\n\t\tpub.Name = *ut.Name\n\t}\n\tif ut.ShortName != nil {\n\t\tpub.ShortName = *ut.ShortName\n\t}\n\tif ut.SportID != nil {\n\t\tpub.SportID = *ut.SportID\n\t}\n\treturn &pub\n}", "func JoinNetwork(n *Node, id string) error {\n\terr := n.registerNode()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnodes, err := netutils.GetNodeIPs(n.nameServer)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer func() {\n\t\tif n.graph {\n\t\t\tn.addNodeToGraph()\n\t\t\tgo n.pushState()\n\t\t}\n\t}()\n\n\tif len(nodes) == 1 && nodes[0] == n.IP {\n\t\tn.setSuccessor(n.Rnode)\n\t\tn.setPredecessor(n.Rnode)\n\t\tn.initFTable(true)\n\t\tgo n.periodicRun()\n\t\treturn nil\n\t}\n\n\tn.initFTable(false)\n\n\tnode := n.getRandomNode(nodes)\n\n\trnode := &comm.Rnode{IP: node}\n\tsucc, err := n.remote.FindSuccessor(*rnode, n.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tn.setSuccessor(succ)\n\tn.setPredecessor(n.Rnode)\n\tgo n.periodicRun()\n\treturn nil\n}", "func NewStentor(bufSize int) *Stentor {\n\treturn &Stentor{\n\t\tbufSize: bufSize,\n\t\tsubscribers: make(map[<-chan G]chan G),\n\t}\n}" ]
[ "0.46678373", "0.43025213", "0.41266426", "0.40999755", "0.4031369", "0.40144655", "0.40144596", "0.39685214", "0.39212373", "0.39158916", "0.38860258", "0.38709098", "0.3857416", "0.38451752", "0.383847", "0.3829763", "0.38195068", "0.3810067", "0.38047096", "0.37613794", "0.37592193", "0.37577853", "0.3746965", "0.3719668", "0.3712487", "0.37016645", "0.37012848", "0.3697794", "0.3662746", "0.3662746", "0.3660167", "0.3655175", "0.36546078", "0.36535957", "0.3605385", "0.35999718", "0.35987833", "0.3587019", "0.35841757", "0.35711712", "0.35657182", "0.3565328", "0.35644296", "0.35632256", "0.35625142", "0.35609588", "0.3537943", "0.35345635", "0.3528131", "0.3517552", "0.35156086", "0.35119042", "0.35032618", "0.35005653", "0.34973195", "0.3492214", "0.3491862", "0.348617", "0.34855658", "0.34855658", "0.34808758", "0.34743166", "0.34689555", "0.34579727", "0.34510624", "0.3445542", "0.3444711", "0.34429878", "0.34429878", "0.34427288", "0.34402654", "0.34333313", "0.343329", "0.34321064", "0.34282556", "0.34244788", "0.3424405", "0.34243485", "0.3410434", "0.34095454", "0.3407835", "0.3405729", "0.3401886", "0.34005597", "0.33991805", "0.3397149", "0.3390375", "0.339032", "0.33901873", "0.33899713", "0.3387499", "0.33874488", "0.33839685", "0.33837414", "0.33805576", "0.33799836", "0.33775765", "0.33734217", "0.3373245", "0.3367357" ]
0.8139762
0
Spit runs logic that creates and spits a new Web from the Nest the Spitter is on to another Nest, connecting them.
func (spitterImpl *SpitterImpl) Spit(nest spiders.Nest) bool { return spitterImpl.RunOnServer("spit", map[string]interface{}{ "nest": nest, }).(bool) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func LaunchWeb(conf Conf, rp string, hub *hub.Hub, q *event.Queue, addrChan chan string, webDev bool) {\n\tif rp == \"\" {\n\t\trp = rootPath\n\t}\n\tr := httprouter.New()\n\tr.HandleMethodNotAllowed = false\n\tr.NotFound = notFoundHandler{}\n\tr.PanicHandler = panicHandler\n\n\th := handler{hub: hub}\n\n\t// --- authentication ---\n\tr.POST(rp+\"/login\", h.mw(loginHandler, false))\n\tr.POST(rp+\"/logout\", h.mw(logoutHandler, true))\n\n\t// --- api ---\n\tr.GET(rp+\"/flows\", h.mw(hndAllFlows, true)) // list all the flows configs\n\tr.GET(rp+\"/flows/:id\", h.mw(hndFlow, true)) // return highest version of the flow config and run summaries from the cluster\n\tr.GET(rp+\"/flows/:id/runs/:rid\", h.mw(hndRun, true)) // returns the identified run detail (may be on another host)\n\n\t// --- push endpoints ---\n\th.setupPushes(rp+\"/push/\", r, hub)\n\n\t// --- p2p api ---\n\tr.POST(rp+\"/p2p/flows/exec\", h.mw(hndP2PExecFlow, true)) // internal api to pass a pending todo to activate it on this host\n\tr.GET(rp+\"/p2p/flows/:id/runs\", h.mw(hndP2PRuns, true)) // all summary runs from this host for this flow id\n\tr.GET(rp+\"/p2p/flows/:id/runs/:rid\", h.mw(hndP2PRun, true)) // detailed run info from this host for this flow id and run id\n\tr.GET(rp+\"/p2p/config\", h.mw(confHandler, true)) // return host config and what it knows about other hosts\n\n\t// --- static files for the spa ---\n\tif webDev { // local development mode\n\t\tserveFiles(r, \"/static/*filepath\", http.Dir(\"webapp\"))\n\t\tr.GET(\"/app/*filepath\", zipper(singleFile(\"webapp/index.html\")))\n\t} else { // release mode\n\t\tserveFiles(r, \"/static/*filepath\", assetFS())\n\t\tr.GET(\"/app/*filepath\", zipper(assetFile(\"webapp/index.html\")))\n\t}\n\n\t// serveFiles(r, \"/static/img/*filepath\", http.Dir(\"webapp/img\"))\n\t// serveFiles(r, \"/static/js/*filepath\", http.Dir(\"webapp/js\"))\n\t// serveFiles(r, \"/static/font/*filepath\", http.Dir(\"webapp/font\"))\n\n\t// ws endpoint\n\twsh := newWsHub()\n\tq.Register(wsh)\n\tr.GET(\"/ws\", wsh.getWsHandler(&h))\n\n\t// --- CORS ---\n\tr.OPTIONS(rp+\"/*all\", h.mw(nil, false)) // catch all options\n\n\t/*\n\t\tr.GET(rp+\"/flows/:flid\", h.mw(floeHandler, true))\n\t\tr.POST(rp+\"/flows/:flid/exec\", h.mw(execHandler, true))\n\t\tr.POST(rp+\"/flows/:flid/stop\", h.mw(stopHandler, true))\n\t\tr.GET(rp+\"/flows/:flid/run/:agentid/:runid\", h.mw(runHandler, true)) // get the current progress of a run for an agent and run\n\n\t\t// --- web socket connection ---\n\t\tr.GET(rp+\"/msg\", wsHandler)\n\n\n\n\t\t// --- the web page stuff ---\n\t\tr.GET(\"/build/\", indexHandler)\n\t\tserveFiles(r, \"/build/css/*filepath\", http.Dir(\"public/build/css\"))\n\t\tserveFiles(r, \"/build/fonts/*filepath\", http.Dir(\"public/build/fonts\"))\n\t\tserveFiles(r, \"/build/img/*filepath\", http.Dir(\"public/build/img\"))\n\t\tserveFiles(r, \"/build/js/*filepath\", http.Dir(\"public/build/js\"))\n\n\t*/\n\n\t// start the private server if one is configured differently to the public server\n\tif conf.PrvBind != conf.PubBind && conf.PrvBind != \"\" {\n\t\tlog.Debug(\"private server listen on:\", conf.PrvBind)\n\t\tgo launch(conf.PrvBind, conf.PrvCert, conf.PrvKey, r, nil)\n\t}\n\n\t// start the public server\n\tlog.Debug(\"pub server listen on:\", conf.PubBind)\n\tlaunch(conf.PubBind, conf.PubCert, conf.PubKey, r, addrChan)\n}", "func (spitterImpl *SpitterImpl) SpittingWebToNest() spiders.Nest {\n\treturn spitterImpl.spittingWebToNestImpl\n}", "func (q *Q) Web(port string) {\n\tlog.WithFields(log.Fields{\"port\": port}).Info(\"Crush is starting ...\")\n\tr := mux.NewRouter()\n\n\t// setup our routes\n\tr.HandleFunc(\"/{topic}/{id}\", q.WebTopicID)\n\tr.HandleFunc(\"/{topic}\", q.WebNewMessage)\n\n\t// set some defaults\n\tsrv := &http.Server{\n\t\tHandler: r,\n\t\tAddr: \"0.0.0.0:\" + port,\n\t\tWriteTimeout: 5 * time.Second,\n\t\tReadTimeout: 5 * time.Second,\n\t}\n\n\t// start serving\n\tlog.Fatal(srv.ListenAndServe())\n}", "func main () {\n // Load app configuration\n config.ReadConfig()\n // Initialize database\n database.InitDB()\n // Set up webpage stuff\n webserver.InitTemplates()\n // Generate about, 404, etc. static pages\n webserver.GenerateStaticPages()\n\n /*fmt.Print(\"Debug getting Stonk's clips...\\n\")\n cr := twitchapi.TopClipsWeek(\"stonk\")\n fmt.Printf(\"%v clips this week\\n\", len(cr.Clips))\n\n for _, clip := range cr.Clips {\n fmt.Printf(\"Clip ID: %v\\n\", clip.TrackingId)\n fmt.Printf(\"Clip Url: %v\\n\", clip.Url)\n fmt.Printf(\"Clip Created_At: %v\\n\", clip.Created_At_Time)\n fmt.Printf(\"Clip Thumb Url: %v\\n\", clip.Thumbnails.Small)\n database.AddClipToDB(clip.TrackingId, 25276, clip.Created_At_Time,\n clip.Thumbnails.Small, clip.Url)\n }*/\n\n /*fmt.Print(\"Debug regenerating filters pages...\")\n RegenerateFilterPages()\n fmt.Print(\"Done\\n\")*/\n\n /*fmt.Print(\"Debug Update()...\")\n Update()\n fmt.Print(\"Done\\n\")*/\n\n // Start web server to handle HTTP requets\n go webserver.Serve()\n\n // Start tracking streams\n TrackStreams()\n}", "func StartTweeting(twitter *twitter.Client, storageBackend storage.Pool, tweetTime time.Duration) {\n\n\t// Setup tweet scheduling\n\tts := &TweetSearch{\n\t\tChannel: make(chan *Tweet),\n\t\tTrending: trendingwrap.NewClient(),\n\t\tStorage: storageBackend,\n\t\tURLLength: twitter.Configuration.ShortUrlLengthHttps,\n\t}\n\tSetupRegularTweetSearchProcess(ts, tweetTime)\n\tlog.Println(\"Setup complete. Lets wait for the first trending project...\")\n\n\t// Waiting for tweets ...\n\tfor tweet := range ts.Channel {\n\t\t// Sometimes it happens that we won`t get a project.\n\t\t// In this situation we try to avoid empty tweets like ...\n\t\t//\t* https://twitter.com/TrendingGithub/status/628714326564696064\n\t\t//\t* https://twitter.com/TrendingGithub/status/628530032361795584\n\t\t//\t* https://twitter.com/TrendingGithub/status/628348405790711808\n\t\t// we will return here\n\t\t// We do this check here and not in tweets.go, because otherwise\n\t\t// a new tweet won`t be scheduled\n\t\tif len(tweet.ProjectName) <= 0 {\n\t\t\tlog.Println(\"No project found. No tweet sent.\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// In debug mode the twitter variable is not available, so we won`t tweet the tweet.\n\t\t// We will just output them.\n\t\t// This is a good development feature ;)\n\t\tif twitter.API == nil {\n\t\t\tlog.Printf(\"Tweet: %s (length: %d)\", tweet.Tweet, len(tweet.Tweet))\n\n\t\t} else {\n\t\t\tpostedTweet, err := twitter.Tweet(tweet.Tweet)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Tweet publishing: ❌ (%s)\\n\", err)\n\t\t\t} else {\n\t\t\t\tlog.Printf(\"Tweet publishing: ✅ (https://twitter.com/TrendingGithub/status/%s)\\n\", postedTweet.IdStr)\n\t\t\t}\n\t\t}\n\t\tts.MarkTweetAsAlreadyTweeted(tweet.ProjectName)\n\t}\n}", "func Create(pools *pools.Pool, port int) *http.Server {\n\tlgger := logger.Init(\"Engine Leader Web\", true, false, ioutil.Discard)\n\tvar host string\n\tif os.Getenv(\"ENV\") == \"dev\" {\n\t\thost = \"localhost\"\n\t} else {\n\t\thost = \"\"\n\t}\n\treturn &http.Server{\n\t\tAddr: fmt.Sprintf(\"%s:%d\", host, port),\n\t\tHandler: NewRouter(pools, lgger),\n\t\tReadTimeout: time.Duration(10000) * time.Millisecond, // TODO remove hardcoding\n\t\tWriteTimeout: time.Duration(10000) * time.Millisecond, // TODO remove hardcoding\n\t}\n}", "func InitNewWebsite(c *ishell.Context) {\n\n\t// Demande pour le nom du site web\n\tc.Print(\"Domain name of the website (ex : wafg.ca) : \")\n\n\tc.Print(\"Enter your full name : \")\n\n\t// Demande qu'elle page de base je veux d'enable\n\n\tc.Print(\"Admin account create with success\")\n\n}", "func Create(engine *leader.Leader, port int) *http.Server {\n\tgame.InitGames()\n\tlgger := logger.Init(\"BattleSnake Web\", true, false, ioutil.Discard)\n\tvar host string\n\tif os.Getenv(\"ENV\") == \"dev\" {\n\t\thost = \"localhost\"\n\t} else {\n\t\thost = \"\"\n\t}\n\treturn &http.Server{\n\t\tAddr: fmt.Sprintf(\"%s:%d\", host, port),\n\t\tHandler: web.NewRouter(engine, lgger),\n\t\tReadTimeout: time.Duration(500) * time.Millisecond, // TODO remove hardcoding\n\t\tWriteTimeout: time.Duration(500) * time.Millisecond, // TODO remove hardcoding\n\t}\n}", "func main() {\n\tuseSlingGet()\n\tuseSlingPost()\n\tuseSlingGetWithHeader()\n\tuseSlingGetWithQuery()\n\n}", "func (web Web) Run() {\n\tlog.Println(\"Starting webserver\")\n\n\t//Serve static files\n\tfs := http.FileServer(http.Dir(\"static/voipathon\"))\n\thttp.Handle(\"/\", fs)\n\n\tfsTestClient := http.FileServer(http.Dir(\"static/testclient\"))\n\thttp.Handle(\"/testclient/\", http.StripPrefix(\"/testclient/\", fsTestClient))\n\n\thttp.HandleFunc(\"/ws\", web.registerClient)\n\tlog.Println(\"Waiting for connections\")\n\tlog.Fatal(http.ListenAndServe(\":4242\", nil))\n}", "func SearchWeb(ctx context.Context, prodName string, w http.ResponseWriter, r *http.Request) ([]string, error) {\n\tvar (\n\t\tebayInfo webUtil = &ebayUtil{\n\t\t\tName: \"Ebay\",\n\t\t\tNumPerPage: 50,\n\t\t\tOnHTML: \"div[class='s-item__wrapper clearfix']\",\n\t\t\tParallel: 13,\n\t\t\tUserAgent: \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.75 Safari/537.36\",\n\t\t}\n\t\twatsonInfo webUtil = &watsonsUtil{\n\t\t\tName: \"Watsons\",\n\t\t\tNumPerPage: 64,\n\t\t\tOnHTML: \"e2-product-list\",\n\t\t\tParallel: 3,\n\t\t\tUserAgent: \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.75 Safari/537.36\",\n\t\t}\n\n\t\twebsites = []webUtil{\n\t\t\tebayInfo,\n\t\t\twatsonInfo,\n\t\t}\n\n\t\tresultJSON []string\n\t\tmu sync.Mutex\n\t\tErr error\n\t\tch = make(chan error, 2)\n\t\tcounter = 0\n\t)\n\n\tfor _, website := range websites {\n\t\tgo func(web webUtil) {\n\t\t\tcrawlWebsite(ctx, ch, &mu, web, prodName, &resultJSON, w)\n\t\t}(website)\n\t}\n\n\tfor err := range ch {\n\t\tcounter++\n\t\tif err != nil {\n\t\t\tif Err == nil {\n\t\t\t\tErr = err\n\t\t\t} else {\n\t\t\t\tErr = fmt.Errorf(\"%v;%v\", Err, err)\n\t\t\t}\n\t\t}\n\t\tif counter == len(websites) {\n\t\t\tbreak\n\t\t}\n\t}\n\tfmt.Println(\"Done err waiting\")\n\treturn resultJSON, Err\n}", "func (s WashingtonPostScraper) CreateNewWashingtonPostScraper() *WashingtonPostScraper {\n\tc := colly.NewCollector()\n\t// c := colly.NewCollector(colly.Debugger(&debug.LogDebugger{}))\n\tc.UserAgent = s.UserAgent()\n\tc.IgnoreRobotsTxt = false\n\n\t// Adding this wait so AJAX can load, might need to look at https://github.com/chromedp/chromedp in the future\n\tc.Limit(&colly.LimitRule{\n\t\tDelay: 5 * time.Second,\n\t})\n\n\tscraper := WashingtonPostScraper{\n\t\tcollector: c,\n\t}\n\treturn &scraper\n}", "func cloneSite(args []string) {\n\turl := args[0]\n\n\tif Serve == true {\n\t\t// grab the url from the\n\t\tif !parser.ValidateURL(url) && !parser.ValidateDomain(url) {\n\t\t\tfmt.Println(\"goclone <url>\")\n\t\t} else if parser.ValidateDomain(url) {\n\t\t\t// use the domain as the project name\n\t\t\tname := url\n\t\t\t// CreateProject\n\t\t\tprojectPath := file.CreateProject(name)\n\t\t\t// create the url\n\t\t\tvalidURL := parser.CreateURL(name)\n\t\t\t// Crawler\n\t\t\tcrawler.Crawl(validURL, projectPath)\n\t\t\t// Restructure html\n\t\t\thtml.LinkRestructure(projectPath)\n\t\t\terr := exec.Command(\"open\", \"http://localhost:5000\").Start()\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t\tserver.Serve(projectPath)\n\n\t\t} else if parser.ValidateURL(url) {\n\t\t\t// get the hostname\n\t\t\tname := parser.GetDomain(url)\n\t\t\t// create project\n\t\t\tprojectPath := file.CreateProject(name)\n\t\t\t// Crawler\n\t\t\tcrawler.Crawl(url, projectPath)\n\t\t\t// Restructure html\n\t\t\thtml.LinkRestructure(projectPath)\n\t\t\terr := exec.Command(\"open\", \"http://localhost:5000\").Start()\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t\tserver.Serve(projectPath)\n\t\t} else {\n\t\t\tfmt.Print(url)\n\t\t}\n\t} else {\n\t\t// grab the url from the\n\t\tif !parser.ValidateURL(url) && !parser.ValidateDomain(url) {\n\t\t\tfmt.Println(\"goclone <url>\")\n\t\t} else if parser.ValidateDomain(url) {\n\t\t\t// use the domain as the project name\n\t\t\tname := url\n\t\t\t// CreateProject\n\t\t\tprojectPath := file.CreateProject(name)\n\t\t\t// create the url\n\t\t\tvalidURL := parser.CreateURL(name)\n\t\t\t// Crawler\n\t\t\tcrawler.Crawl(validURL, projectPath)\n\t\t\t// Restructure html\n\t\t\thtml.LinkRestructure(projectPath)\n\t\t\tif Open {\n\t\t\t\t// automatically open project\n\t\t\t\terr := exec.Command(\"open\", projectPath+\"/index.html\").Start()\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t}\n\n\t\t} else if parser.ValidateURL(url) {\n\t\t\t// get the hostname\n\t\t\tname := parser.GetDomain(url)\n\t\t\t// create project\n\t\t\tprojectPath := file.CreateProject(name)\n\t\t\t// Crawler\n\t\t\tcrawler.Crawl(url, projectPath)\n\t\t\t// Restructure html\n\t\t\thtml.LinkRestructure(projectPath)\n\t\t\tif Open {\n\t\t\t\terr := exec.Command(\"open\", projectPath+\"/index.html\").Start()\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tfmt.Print(url)\n\t\t}\n\t}\n}", "func StartTweeting(twitter *Twitter, storageBackend storage.Pool) {\n\n\t// Setup tweet scheduling\n\tts := &TweetSearch{\n\t\tChannel: make(chan *Tweet),\n\t\tTrending: NewTrendingClient(),\n\t\tStorage: storageBackend,\n\t\tURLLength: twitter.Configuration.ShortUrlLengthHttps,\n\t}\n\tSetupRegularTweetSearchProcess(ts)\n\n\t// Waiting for tweets ...\n\tfor tweet := range ts.Channel {\n\t\t// Sometimes it happens that we won`t get a project.\n\t\t// In this situation we try to avoid empty tweets like ...\n\t\t//\t* https://twitter.com/TrendingGithub/status/628714326564696064\n\t\t//\t* https://twitter.com/TrendingGithub/status/628530032361795584\n\t\t//\t* https://twitter.com/TrendingGithub/status/628348405790711808\n\t\t// we will return here\n\t\t// We do this check here and not in tweets.go, because otherwise\n\t\t// a new tweet won`t be scheduled\n\t\tif len(tweet.ProjectName) <= 0 {\n\t\t\tlog.Print(\"No project found. No tweet sent.\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// In debug mode the twitter variable is not available, so we won`t tweet the tweet.\n\t\t// We will just output them.\n\t\t// This is a good development feature ;)\n\t\tif twitter.API == nil {\n\t\t\tlog.Printf(\"Tweet: %s (length: %d)\", tweet.Tweet, len(tweet.Tweet))\n\n\t\t} else {\n\t\t\tpostedTweet, err := twitter.Tweet(tweet.Tweet)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(err)\n\t\t\t} else {\n\t\t\t\tlog.Printf(\"Tweet %s posted\", postedTweet.IdStr)\n\t\t\t}\n\t\t}\n\t\tts.MarkTweetAsAlreadyTweeted(tweet.ProjectName)\n\t}\n}", "func Run() {\n\tadminHome := &admin.HomeController{Render: amber.New()}\n\tmemberHome := &member.HomeController{Render: amber.New()}\n\tmemberRSVP := &member.RSVPController{Render: amber.New()}\n\n\tr := httprouter.New()\n\tr.ServeFiles(\"/public/*filepath\", http.Dir(\"public/\"))\n\n\tr.GET(\"/admin\", adminHome.Action(adminHome.Index))\n\tr.GET(\"/member\", memberHome.Action(memberHome.Index))\n\tr.POST(\"/member\", memberHome.Action(memberHome.Create))\n\n\tr.GET(\"/rsvp\", memberRSVP.Action(memberRSVP.Index))\n\tr.GET(\"/rsvp/do\", memberRSVP.Action(memberRSVP.RSVP))\n\tr.POST(\"/rsvp/complete\", memberRSVP.Action(memberRSVP.Complete))\n\tr.POST(\"/rsvp\", memberRSVP.Action(memberRSVP.Login))\n\n\tn := negroni.New()\n\tn.Use(negroni.HandlerFunc(loggerMiddleware))\n\tn.Use(negroni.NewRecovery())\n\tn.UseHandler(r)\n\n\tPORT := os.Getenv(\"PORT\")\n\tif PORT == \"\" {\n\t\tPORT = \":3000\"\n\t}\n\tlog.Infoln(\"Starting server on\", PORT)\n\terr := http.ListenAndServe(PORT, n)\n\tif err != nil {\n\t\tlog.Fatalln(\"Error!\", err)\n\t}\n}", "func Home(w http.ResponseWriter, r *http.Request) { // Grab the span from the gorilla context. We do this so that we can grab\n\t// the span.Trace ID and link directly to the trace on the web-page itself!\n\tspan := context.Get(r, CtxSpanID).(appdash.SpanID)\n\n\t// We're going to make some API requests, so we create a HTTP client using\n\t// a appdash/httptrace transport here. The transport will inform Appdash of\n\t// the HTTP events occuring.\n\thttpClient := &http.Client{\n\t\tTransport: &httptrace.Transport{\n\t\t\tRecorder: appdash.NewRecorder(span, collector),\n\t\t\tSetName: true,\n\t\t},\n\t}\n\n\t// Make three API requests using our HTTP client.\n\tfor i := 0; i < 3; i++ {\n\t\tresp, err := httpClient.Get(\"/endpoint\")\n\t\tif err != nil {\n\t\t\tlog.Println(\"/endpoint:\", err)\n\t\t\tcontinue\n\t\t}\n\t\tresp.Body.Close()\n\t}\n\n\t// Render the page.\n\tfmt.Fprintf(w, `<!DOCTYPE html>\n\t\t\t\t\t\t\t\t\t\t<html>\n\t\t\t\t\t\t\t\t\t\t<head>\n\n\t\t\t\t\t\t\t\t\t\t <!-- Basic Page Needs\n\t\t\t\t\t\t\t\t\t\t –––––––––––––––––––––––––––––––––––––––––––––––––– -->\n\t\t\t\t\t\t\t\t\t\t <meta charset=\"utf-8\">\n\t\t\t\t\t\t\t\t\t\t <title>Test load</title>\n\t\t\t\t\t\t\t\t\t\t <meta name=\"description\" content=\"\">\n\t\t\t\t\t\t\t\t\t\t <meta name=\"author\" content=\"\">\n\n\t\t\t\t\t\t\t\t\t\t <!-- Mobile Specific Metas\n\t\t\t\t\t\t\t\t\t\t –––––––––––––––––––––––––––––––––––––––––––––––––– -->\n\t\t\t\t\t\t\t\t\t\t<meta name=\"viewport\" content=\"width=device-width, initial-scale=1\">\n\t\t\t\t\t\t\t\t\t\t<link href=\"https://bedbathandbeyond.qa.nrdecor.com/js/lib/jqwidgets/styles/jqx.base.css\" rel=\"stylesheet\" type=\"text/css\" id=\"jqw-styles\">\n\t\t\t\t\t\t\t\t\t\t<link href=\"https://bedbathandbeyond.qa.nrdecor.com/js/lib/jqwidgets/styles/jqx.office.css\" rel=\"stylesheet\" type=\"text/css\" id=\"jqw-styles\">\n\t\t\t\t\t\t\t\t\t\t<link href=\"https://bedbathandbeyond.qa.nrdecor.com/js/lib/justified-gallery/styles/justifiedGallery.css\" type=\"text/css\" rel=\"stylesheet\" >\n\n\n\t\t\t\t\t\t\t\t\t\t<link rel=\"stylesheet\" type=\"text/css\" href=\"https://bedbathandbeyond.qa.nrdecor.com/skin/frontend/enterprise/default/css/styles.css\" media=\"all\" />\n\t\t\t\t\t\t\t\t\t\t<link rel=\"stylesheet\" type=\"text/css\" href=\"https://bedbathandbeyond.qa.nrdecor.com/skin/frontend/enterprise/default/css/widgets.css\" media=\"all\" />\n\t\t\t\t\t\t\t\t\t\t<link rel=\"stylesheet\" type=\"text/css\" href=\"https://bedbathandbeyond.qa.nrdecor.com/skin/frontend/bbb2/common/css/common.css\" media=\"all\" />\n\t\t\t\t\t\t\t\t\t\t<link rel=\"stylesheet\" type=\"text/css\" href=\"https://bedbathandbeyond.qa.nrdecor.com/skin/frontend/bbb2/common/css/jquery-ui-1.9.2.custom.min.css\" media=\"all\" />\n\t\t\t\t\t\t\t\t\t\t<link rel=\"stylesheet\" type=\"text/css\" href=\"https://bedbathandbeyond.qa.nrdecor.com/skin/frontend/bbb2/web/css/web.css\" media=\"all\" />\n\t\t\t\t\t\t\t\t\t\t<link rel=\"stylesheet\" type=\"text/css\" href=\"https://bedbathandbeyond.qa.nrdecor.com/skin/frontend/bbb2/web/css/fonts.css\" media=\"all\" />\n\t\t\t\t\t\t\t\t\t\t<link rel=\"stylesheet\" type=\"text/css\" href=\"https://bedbathandbeyond.qa.nrdecor.com/skin/frontend/enterprise/default/css/print.css\" media=\"print\" />\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/prototype/prototype.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/lib/ccard.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/prototype/validation.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/scriptaculous/builder.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/scriptaculous/effects.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/scriptaculous/dragdrop.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/scriptaculous/controls.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/scriptaculous/slider.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/varien/js.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/varien/form.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/varien/menu.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\" src=\"https://bedbathandbeyond.qa.nrdecor.com/js/mage/translate.js\"></script>\n\t\t\t\t\t\t\t\t\t\t<script src=\"https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0-alpha1/jquery.min.js\"></script>\n\n\t\t\t\t\t\t\t\t\t\t <!-- Scripts\n\t\t\t\t\t\t\t\t\t\t –––––––––––––––––––––––––––––––––––––––––––––––––– -->\n\n\t\t\t\t\t\t\t\t\t\t<script type=\"text/javascript\">\n\t\t\t\t\t\t\t\t\t\t $(document).ready(function () {\n\t\t\t\t\t\t\t\t\t\t console.log(window.performance)//.getEntries())\n\t\t\t\t\t\t\t\t\t\t var arr = window.performance.getEntriesByType(\"resource\")\n\t\t\t\t\t\t\t\t\t\t jsonObj = [];\n\t\t\t\t\t\t\t\t\t\t console.log(jsonObj);\n\t\t\t\t\t\t\t\t\t\t $.each( arr, function( i, val ) {\n\t\t\t\t\t\t\t\t\t\t var name = val.name;\n\t\t\t\t\t\t\t\t\t\t var entryType = val.entryType;\n\t\t\t\t\t\t\t\t\t\t var startTime = val.fetchStart;\n\t\t\t\t\t\t\t\t\t\t var endTime = val.duration;\n\t\t\t\t\t\t\t\t\t\t var initiatorType = val.initiatorType;\n\n\t\t\t\t\t\t\t\t\t\t item = {}\n\t\t\t\t\t\t\t\t\t\t item [\"name\"] = name;\n\t\t\t\t\t\t\t\t\t\t item [\"entryType\"] = entryType;\n\t\t\t\t\t\t\t\t\t\t item [\"startTime\"] = startTime;\n\t\t\t\t\t\t\t\t\t\t item [\"endTime\"] = endTime;\n\t\t\t\t\t\t\t\t\t\t item [\"initiatorType\"] = initiatorType;\n\n\t\t\t\t\t\t\t\t\t\t jsonObj.push(item);\n\t\t\t\t\t\t\t\t\t\t });\n\t\t\t\t\t\t\t\t\t\t jsonString = JSON.stringify(jsonObj);\n\t\t\t\t\t\t\t\t\t\t console.log(jsonString);\n\t\t\t\t\t\t\t\t\t\t $.ajax({\n\t\t\t\t\t\t\t\t\t\t type: \"POST\",\n\t\t\t\t\t\t\t\t\t\t url: \"http://192.168.70.1:8699/endpoint\",\n\t\t\t\t\t\t\t\t\t\t data: jsonString\n\t\t\t\t\t\t\t\t\t\t // success: success,\n\t\t\t\t\t\t\t\t\t\t // dataType: dataType\n\t\t\t\t\t\t\t\t\t\t });\n\t\t\t\t\t\t\t\t\t\t });\n\t\t\t\t\t\t\t\t\t\t </script>\n\t\t\t\t\t\t\t\t\t\t</head>\n\t\t\t\t\t\t\t\t\t\t<body>\n\t\t\t\t\t\t\t\t\t\t<!-- Primary Page Layout\n\t\t\t\t\t\t\t\t\t\t –––––––––––––––––––––––––––––––––––––––––––––––––– -->\n\t\t\t\t\t\t\t\t\t\t <img id=\"image0\" src=\"http://flex.madebymufffin.com/images/inacup_donut.jpg\" alt=\"Smiley face\" height=\"42\" width=\"42\">\n\t\t\t\t\t\t\t\t\t\t<!-- End Document\n\t\t\t\t\t\t\t\t\t\t –––––––––––––––––––––––––––––––––––––––––––––––––– -->\n\t\t\t\t\t\t\t\t\t\t</body>\n\t\t\t\t\t\t\t\t\t\t</html>\n\t\t\t\t\t\t\t\t\t`)\n\tfmt.Fprintf(w, `<p><a href=\"http://localhost:8700/traces\" target=\"_\">View all traces</a></p>`)\n}", "func (w *web) Run() {\n\tw.ready()\n\n\taddr := fmt.Sprintf(\"%s:%s\", w.Host, w.Port)\n\tLogger.Info(\"Starting web server at: %s\", addr)\n\tw.instance.Run(standard.New(addr))\n}", "func main() {\n\tdb, err := sql.Open(\"sqlite3\", \"connect-sessions.db?cache=share&mode=twc&_foreign_keys=1&_journal_mode=wal\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tconfig := goconnect.NewDefaultConfig(goconnect.ClientConfig{\n\t\tHost: goconnect.StagingHost,\n\t\tClientID: \"telenordigital-connectexample-web\",\n\t\tPassword: \"\",\n\t\tLoginCompleteRedirectURI: \"/main.html\",\n\t\tLogoutCompleteRedirectURI: \"/\",\n\t})\n\t// Note: Ignoring errors since the store might already exist\n\tgoconnect.SQLStorePrepare(db)\n\n\tdbstore, err := goconnect.NewSQLStorage(db)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tconnect := goconnect.NewConnectIDWithStorage(config, dbstore)\n\n\t// The /css and /images endpoints\n\thttp.Handle(\"/css/\", http.StripPrefix(\"/css/\", http.FileServer(http.Dir(\"html/css\"))))\n\thttp.Handle(\"/images/\", http.StripPrefix(\"/images/\", http.FileServer(http.Dir(\"html/images\"))))\n\n\t// This the default start page\n\thttp.HandleFunc(\"/\", startPageHandler)\n\n\t// Main page - requires authentication\n\thttp.HandleFunc(\"/main.html\", connect.NewAuthHandlerFunc(mainPageHandler))\n\n\t// A protected resource - requires authentication\n\thttp.Handle(\"/extra/\", connect.NewAuthHandler(\n\t\thttp.StripPrefix(\"/extra/\", http.FileServer(http.Dir(\"html/extra\")))))\n\n\t// API endpoint - requires authentication\n\thttp.HandleFunc(\"/api/oneliner\", connect.NewAuthHandlerFunc(OneLinerHandlerFunc))\n\n\t// This isn't required since the /auth/info endpoint provides the same information but\n\t// this is just to show an example of how session information can be used in resources.\n\thttp.HandleFunc(\"/api/userinfo\", connect.NewAuthHandlerFunc(UserInfoHandlerFunc))\n\n\t// The ConnectID endpoints is added to /auth/. The actual endpoint depends on your\n\t// client configuration in ConnectID; ie what redirects are required.\n\thttp.Handle(\"/connect/\", connect.Handler())\n\n\t// Show the logged in user's properties.\n\thttp.HandleFunc(\"/connect/profile\", connect.SessionProfile)\n\n\tfmt.Println(\"Serving data on http://localhost:8080/\")\n\thttp.ListenAndServe(\":8080\", nil)\n}", "func main() {\n\t/*\n\t\tLoads the env variables\n\t*/\n\tgotenv.Load()\n\n\t/*\n\t\tGet the port no from .env file.\n\t\tConvert string to int\n\t\tIn case some error comes then process is stopped\n\t*/\n\tport := os.Getenv(\"WEBSITE_PORT\")\n\tdbUser := os.Getenv(\"USER_DB_USER\")\n\tdbPass := os.Getenv(\"USER_DB_PASS\")\n\tdbName := os.Getenv(\"USER_DB_NAME\")\n\tdbURL := os.Getenv(\"USER_DB_URL\")\n\n\t/*\n\t\tSetting Gin parameter and folders\n\t*/\n\tgin.SetMode(gin.ReleaseMode)\n\tgin.DefaultWriter = ioutil.Discard\n\n\tr = gin.Default()\n\tr.LoadHTMLGlob(\"templetes/html/*\")\n\tr.Static(\"/css\", \"templetes/css\")\n\tr.Static(\"/js\", \"templetes/js\")\n\tr.Static(\"/img\", \"templetes/img\")\n\tr.Static(\"/fonts\", \"templetes/fonts\")\n\n\t/*\n\t\tCalling function which will be connecting to Database\n\t\tIn case of error, we are stopping the execution.\n\n\t*/\n\terr := initDBConnection(dbUser, dbPass, dbURL, dbName)\n\tif err != nil {\n\t\tfmt.Printf(\"Error: %s\\n\", err.Error())\n\t\treturn\n\t}\n\tfmt.Println(\"DB connected\")\n\n\t/*\n\t\tcalling function to setup the routes\n\t*/\n\tgo initializeRoutes(port)\n\n\t/*\n\t\tStarting the server in the specified port\n\t*/\n\tfmt.Println(\"Web Portal is running on \" + port)\n\tr.Run(port)\n\tfmt.Println(\"-------Started server-------------\")\n}", "func start(w http.ResponseWriter, r *http.Request) {\n c := appengine.NewContext(r)\n u := user.Current(c)\n playerId := u.String()\n\n // Get or create the Game.\n game, err := getGame(c, \"nertz\")\n if err != nil {\n http.Error(w, err.Error(), 500)\n return\n }\n\n // Create a new Client, getting the channel token.\n token, err := game.makePlayer(c, playerId)\n if err != nil {\n http.Error(w, err.Error(), 500)\n return\n }\n\n url, err := user.LogoutURL(c, r.URL.String())\n if err != nil {\n http.Error(w, err.Error(), http.StatusInternalServerError)\n return\n }\n\n // Render the HTML template\n data := tmplData{ url, game.Id, token, }\n err = tmpl.Execute(w, data)\n if err != nil {\n http.Error(w, err.Error(), 500)\n return\n }\n}", "func StartWeb(hcfg Cfg, appcfg []Cfg) {\n\thostCfg = hcfg\n\tvar port = hcfg.Port\n\tc := cors.New(cors.Options{\n\t\tAllowedOrigins: []string{\"*\"},\n\t})\n\n\tlog.Println(\"Port .. \" + port)\n\trouter := http.NewServeMux()\n\trouter.Handle(\"/\"+hcfg.Project+\"/file/\", GetResource(hcfg))\n\tfor c := range appcfg {\n\t\trouter.Handle(\"/\"+appcfg[c].Project+\"/\", AppIndex(appcfg[c]))\n\t\trouter.Handle(\"/\"+appcfg[c].Project+\"/file/\", GetResource(appcfg[c]))\n\t\trouter.Handle(\"/\"+appcfg[c].Project+\"/dload\", Dload(appcfg[c]))\n\t}\n\trouter.Handle(\"/\", Index())\n\trouter.Handle(\"/config\", getConfig())\n\trouter.Handle(\"/validate\", Validate())\n\trouter.Handle(\"/transform\", Transform())\n\trouter.Handle(\"/verify\", DocVerify())\n\trouter.Handle(\"/rebuild\", Rebuild())\n\trouter.Handle(\"/rebuildall\", RebuildAll())\n\tflag.StringVar(&listenAddr, \"listen-addr\", port, \"server listen address\")\n\tflag.Parse()\n\tlogger := log.New(os.Stdout, \"http: \", log.LstdFlags)\n\tlogger.Println(\"Starting HTTP Server. .. \")\n\tnextRequestID := func() string {\n\t\treturn fmt.Sprintf(\"%d\", time.Now().UnixNano())\n\t}\n\tserver := &http.Server{\n\t\tAddr: listenAddr,\n\t\tHandler: tracing(nextRequestID)(logging(logger)(c.Handler(router))),\n\t\tErrorLog: logger,\n\t\tReadTimeout: 5 * time.Second,\n\t\tWriteTimeout: 10 * time.Second,\n\t\tIdleTimeout: 15 * time.Second,\n\t}\n\tdone := make(chan bool)\n\tquit := make(chan os.Signal, 1)\n\tsignal.Notify(quit, os.Interrupt)\n\tgo func() {\n\t\t<-quit\n\t\tlogger.Println(\"Server is shutting down...\")\n\t\tatomic.StoreInt32(&healthy, 0)\n\n\t\tctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)\n\t\tdefer cancel()\n\n\t\tserver.SetKeepAlivesEnabled(false)\n\t\tif err := server.Shutdown(ctx); err != nil {\n\t\t\tlogger.Fatalf(\"Could not gracefully shutdown the server: %v\\n\", err)\n\t\t}\n\t\tclose(done)\n\t}()\n\tlogger.Println(\"Server is ready to handle requests at\", listenAddr)\n\tatomic.StoreInt32(&healthy, 1)\n\tif err := server.ListenAndServe(); err != nil && err != http.ErrServerClosed {\n\t\tlogger.Fatalf(\"Could not listen on %s: %v\\n\", listenAddr, err)\n\t}\n\t<-done\n\tlogger.Println(\"Server stopped\")\n}", "func main() {\n\n\teventBus := utils.NewEventBus()\n\teventSource := models.NewEventSource(eventBus)\n\teventSource.Start()\n\tdefer eventSource.Stop()\n\n\tpiSource := models.NewProcessInfoSource(eventBus)\n\tpiSource.AddProcess(models.ProcessInfo{ID: \"P1\", Name: \"P1\", CPU: 10})\n\tpiSource.AddProcess(models.ProcessInfo{ID: \"P2\", Name: \"P2\", CPU: 10})\n\tpiSource.AddProcess(models.ProcessInfo{ID: \"P3\", Name: \"P3\", CPU: 30})\n\tpiSource.Start()\n\tdefer piSource.Stop()\n\n\ttaskRepo := models.NewTaskInfoRepo()\n\n\tmiSource := models.NewMachineInfoSource(eventBus)\n\tmiSource.InitWithFakeData()\n\tmiSource.Start()\n\tdefer miSource.Stop()\n\n\ttcx := ws.TestClientX{}\n\ttcx.Open(\"ws://localhost:8000/echo\")\n\tdefer tcx.Close()\n\n\tapp.Route(\"/\", &ui.UpdateButton{})\n\tapp.Route(\"/clusters\", views.NewClustersView(eventBus))\n\tapp.Route(\"/machines\", views.NewMachinesView(eventBus))\n\tapp.Route(\"/events\", views.NewEventsView(eventBus))\n\tapp.Route(\"/charts\", views.NewProcessesView(eventBus))\n\tapp.RouteWithRegexp(\"^/task.*\", ui.NewTaskDetail(taskRepo))\n\n\tapp.Run() // Launches the PWA.\n\n\tfmt.Println(\"DONE\")\n}", "func Run(\n\tendpoint string,\n\tisValidCrawlLink IsValidCrawlLinkFunction,\n\tconnectToDB ConnectToDBFunction,\n\taddEdgesIfDoNotExist AddEdgeFunction,\n\tgetNewNode GetNewNodeFunction,\n\tfilterPage FilterPageFunction,\n) {\n\t// first connect to db\n\tif err := connectToDB(); err != nil {\n\t\tlogFatal(\"Could not connect do db: %v\", err)\n\t}\n\t// get starting link if there isn't one already\n\tif endpoint == \"\" {\n\t\tlogMsg(\"Finding new node..\")\n\t\te, err := getNewNode()\n\t\tif err != nil {\n\t\t\tlogFatal(\"Could not find new starting node: %v\", err)\n\t\t} else {\n\t\t\tendpoint = e\n\t\t}\n\t\tlogMsg(\"New node found: %s\", e)\n\t}\n\t// parse out env\n\tmaxNodes, _ := strconv.Atoi(os.Getenv(\"MAX_APPROX_NODES\"))\n\tparallelism, _ := strconv.Atoi(os.Getenv(\"PARALLELISM\"))\n\tmsDelay, _ := strconv.Atoi(os.Getenv(\"MS_DELAY\"))\n\tCrawl(\n\t\tendpoint,\n\t\tint32(maxNodes),\n\t\tparallelism,\n\t\tmsDelay,\n\t\tisValidCrawlLink,\n\t\taddEdgesIfDoNotExist,\n\t\tfilterPage,\n\t)\n}", "func (m *Motifini) Run() error {\n\tvar err error\n\tm.Info.Println(\"Connecting to SecuritySpy:\", m.Conf.SecuritySpy.URL)\n\tif m.SSpy, err = securityspy.GetServer(m.Conf.SecuritySpy); err != nil {\n\t\treturn err\n\t}\n\tm.ProcessEventStream()\n\tdefer m.SSpy.Events.Stop(true)\n\n\tm.Info.Println(\"Opening Subscriber Database:\", m.Conf.Global.StateFile)\n\tif m.Subs, err = subscribe.GetDB(m.Conf.Global.StateFile); err != nil {\n\t\treturn errors.Wrap(err, \"sub state\")\n\t}\n\n\tm.Msgs = &messenger.Messenger{\n\t\tSSpy: m.SSpy,\n\t\tSubs: m.Subs,\n\t\tConf: m.Conf.Imessage,\n\t\tTempDir: m.Conf.Global.TempDir,\n\t\tInfo: log.New(os.Stdout, \"[MSGS] \", m.Info.Flags()),\n\t\tDebug: m.Debug,\n\t\tError: m.Error,\n\t}\n\tif err := messenger.New(m.Msgs); err != nil {\n\t\treturn err\n\t}\n\n\tm.HTTP = &webserver.Config{\n\t\tSSpy: m.SSpy,\n\t\tSubs: m.Subs,\n\t\tMsgs: m.Msgs,\n\t\tInfo: log.New(os.Stdout, \"[HTTP] \", m.Info.Flags()),\n\t\tDebug: m.Debug,\n\t\tError: m.Error,\n\t\tTempDir: m.Conf.Global.TempDir,\n\t\tAllowedTo: m.Conf.Global.AllowedTo,\n\t\tPort: m.Conf.Global.Port,\n\t}\n\tgo m.waitForSignal()\n\treturn webserver.Start(m.HTTP)\n}", "func main() {\n\tstories, err := reader.ReadJsonStory(\"./static/story/default.json\")\n\tif err != nil {\n\t\tlog.Panicln(err)\n\t}\n\n\tweb.Start(stories)\n}", "func RunWebCrawler() {\n\tCrawl(\"https://golang.org/\", 4, fetcher)\n}", "func (h *Handler) Web(web *web.Web) {\n\tweb.CreateRoutes(h.Router)\n}", "func (a *App) OpenWeb() {\n\ta.load()\n\tif !a.TodoList.IsSynced {\n\t\tfmt.Println(\"This list isn't synced! Use 'ultralist sync' to synchronize this list with ultralist.io.\")\n\t\treturn\n\t}\n\n\tfmt.Println(\"Opening this list on your browser...\")\n\topen.Start(\"https://app.ultralist.io/todolist/\" + a.EventLogger.CurrentSyncedList.UUID)\n}", "func (sb *SweetieBot) Connect() int {\n\tif sb.Debug { // The server does not necessarily tie a standard input to the program\n\t\tgo func() {\n\t\t\tvar input string\n\t\t\tfmt.Scanln(&input)\n\t\t\tatomic.StoreUint32(&sb.quit, QuitNow)\n\t\t}()\n\t}\n\n\tgo sb.deferProcessing()\n\tgo sb.idleCheckLoop()\n\tgo sb.deadlockDetector()\n\tgo sb.memberIngestionLoop()\n\tgo sb.ServeWeb()\n\tgo sb.buildMarkov()\n\n\terr := sb.DG.Open()\n\tif err == nil {\n\t\tfmt.Println(\"Connection established\")\n\t\tfor atomic.LoadUint32(&sb.quit) == QuitNone {\n\t\t\ttime.Sleep(800 * time.Millisecond)\n\t\t}\n\t\tbegin := time.Now().UTC().Unix()\n\t\tfor cur := time.Now().UTC().Unix(); (cur-begin) < MaxUpdateGrace && atomic.LoadUint32(&sb.quit) == QuitRaid; cur = time.Now().UTC().Unix() {\n\t\t\tquit := true\n\t\t\tfor _, g := range sb.Guilds {\n\t\t\t\tif cur-g.LastRaid < UpdateGrace {\n\t\t\t\t\tquit = false\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif quit {\n\t\t\t\tatomic.StoreUint32(&sb.quit, QuitNow)\n\t\t\t} else {\n\t\t\t\ttime.Sleep(2400 * time.Millisecond)\n\t\t\t}\n\t\t}\n\t} else {\n\t\tfmt.Println(\"Error opening websocket connection: \", err.Error())\n\t}\n\n\t/*if q, err := sb.DB.db.Query(\"SELECT DISTINCT Guild FROM members\"); err == nil {\n\t\tf, _ := os.Create(\"out.csv\")\n\t\tdefer f.Close()\n\t\tfor q.Next() {\n\t\t\tvar p uint64\n\t\t\tif err := q.Scan(&p); err == nil {\n\t\t\t\tif _, ok := sb.Guilds[NewDiscordGuild(p)]; !ok {\n\t\t\t\t\tfmt.Fprintf(f, \"%v,\", p)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}*/\n\n\tfmt.Println(\"Sweetiebot quitting\")\n\tsb.DG.Close()\n\tsb.DB.Close()\n\tsb.GuildsLock.Lock() // Prevents a race condition from sending a value to a closed channel\n\tclose(sb.memberChan)\n\tsb.GuildsLock.Unlock()\n\treturn BotVersion.Integer()\n}", "func Create(cfg map[string]interface{}) {\r\n\tgo func(c map[string]interface{}) {\r\n\t\tserver.CreateSingle(\"ws\", c, serveWs)\r\n\t}(cfg)\r\n\tgo func(c map[string]interface{}) {\r\n\t\tserver.CreateSingle(\"wss\", c, serveWs)\r\n\t}(cfg)\r\n}", "func (w *Web) Init(l *log.Logger) {\n\tif w.Port == \"\" {\n\t\tif !w.Secure {\n\t\t\tw.Port = \"80\"\n\t\t} else {\n\t\t\tw.Port = \"443\"\n\t\t}\n\t}\n\n\tw.sites = make(map[string]*Site)\n\tw.Logger = l\n\tw.L = l.TMsg\n\tw.E = l.TErr\n\n\tw.IdleTimeout = time.Second * 30\n\tw.ReadTimeout = time.Second * 10\n\tw.WriteTimeout = time.Second * 10\n\n\tif w.Secure {\n\t\tw.TLSConfig = &tls.Config{\n\t\t\tMinVersion: tls.VersionTLS12,\n\t\t\tNextProtos: []string{\"http/1.1\"},\n\t\t\tCurvePreferences: []tls.CurveID{tls.CurveP521, tls.CurveP384, tls.CurveP256},\n\t\t\tPreferServerCipherSuites: true,\n\t\t\tCipherSuites: []uint16{\n\t\t\t\ttls.TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,\n\t\t\t\ttls.TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,\n\t\t\t\ttls.TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305,\n\t\t\t\ttls.TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,\n\t\t\t\ttls.TLS_AES_128_GCM_SHA256,\n\t\t\t\ttls.TLS_AES_256_GCM_SHA384,\n\t\t\t\ttls.TLS_CHACHA20_POLY1305_SHA256,\n\t\t\t},\n\t\t}\n\t}\n\thttp.HandleFunc(\"/\", w.defaultHandler)\n}", "func Run() {\n\tbrowser.OpenURL(getURL())\n}", "func RunWeb() {\n\twebserver.New(Config).Run()\n}", "func main() {\n\t// Spin off the hub\n\thub := newHub()\n\tgo hub.run()\n\n\thttp.Handle(\"/frontend/dist/\", http.StripPrefix(\"/frontend/dist/\", http.FileServer(http.Dir(\"./frontend/dist/\"))))\n\thttp.Handle(\"/assets/\", http.StripPrefix(\"/assets/\", http.FileServer(http.Dir(\"./assets/\"))))\n\t// Serve index.html specifically\n\thttp.HandleFunc(\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\thttp.ServeFile(w, r, \"index.html\")\n\t})\n\thttp.HandleFunc(\"/api/socket\", func(w http.ResponseWriter, r *http.Request) {\n\t\tserveWs(hub, w, r)\n\t})\n\n\tport := \":4567\"\n\tlog.Println(\"Server listening at localhost\" + port)\n\thttp.ListenAndServe(port, nil)\n\n}", "func NewWeb(music *Music) *Web {\n\tw := &Web{\n\t\tmusic: music,\n\t\ttmpl: template.Must(template.New(\"tmpl\").Parse(webTemplates)),\n\t}\n\treturn w\n}", "func IndexWeb(c *gin.Context) {\n\tc.HTML(http.StatusOK, \"index.html\", gin.H{\n\t\t\"title\": \"Gin Demo Starter\",\n\t})\n}", "func (b *FromWebsocket) Run() {\n\tvar ws *websocket.Conn\n\tvar url string\n\tto, _ := time.ParseDuration(\"10s\")\n\tvar handshakeDialer = &websocket.Dialer{\n\t\tSubprotocols: []string{\"p1\", \"p2\"},\n\t\tHandshakeTimeout: to,\n\t}\n\tlistenWS := make(blocks.MsgChan)\n\twsHeader := http.Header{\"Origin\": {\"http://localhost/\"}}\n\n\ttoOut := make(blocks.MsgChan)\n\ttoError := make(chan error)\n\n\tfor {\n\t\tselect {\n\n\t\tcase msg := <-toOut:\n\t\t\tb.out <- msg\n\n\t\tcase ruleI := <-b.inrule:\n\t\t\tvar err error\n\t\t\t// set a parameter of the block\n\t\t\turl, err = util.ParseString(ruleI, \"url\")\n\t\t\tif err != nil {\n\t\t\t\tb.Error(err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif ws != nil {\n\t\t\t\tws.Close()\n\t\t\t}\n\n\t\t\tws, _, err = handshakeDialer.Dial(url, wsHeader)\n\t\t\tif err != nil {\n\t\t\t\tb.Error(\"could not connect to url\")\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tws.SetReadDeadline(time.Time{})\n\t\t\th := recvHandler{toOut, toError}\n\t\t\tgo h.recv(ws, listenWS)\n\n\t\tcase err := <-toError:\n\t\t\tb.Error(err)\n\n\t\tcase <-b.quit:\n\t\t\t// quit the block\n\t\t\treturn\n\t\tcase o := <-b.queryrule:\n\t\t\to <- map[string]interface{}{\n\t\t\t\t\"url\": url,\n\t\t\t}\n\t\tcase in := <-listenWS:\n\t\t\tb.out <- in\n\t\t}\n\t}\n}", "func New(db *app.DB) (http.Handler, *sandstorm.SessionBus) {\n\tr := mux.NewRouter()\n\n\tapiRouter := r.PathPrefix(\"/api\").Subrouter()\n\ts := api.NewStream(db)\n\tapiRouter.HandleFunc(\"/stream\", s.GetStream).Methods(\"GET\")\n\tb := api.NewBookmark(db)\n\tapiRouter.HandleFunc(\"/bookmark\", b.AddBookmark).Methods(\"POST\")\n\tapiRouter.HandleFunc(\"/bookmark/{id}\", b.RemoveBookmark).Methods(\"DELETE\")\n\td := api.NewDebug(db)\n\tapiRouter.HandleFunc(\"/debug\", d.GetDebug).Methods(\"GET\")\n\tme := api.NewMe(db)\n\tapiRouter.HandleFunc(\"/profile\", me.GetProfile).Methods(\"GET\")\n\tapiRouter.HandleFunc(\"/profile\", me.PutProfile).Methods(\"PUT\")\n\tself := api.NewSelf(db)\n\tapiRouter.HandleFunc(\"/self\", self.GetSelf).Methods(\"GET\")\n\tapiRouter.HandleFunc(\"/self\", self.PutSelf).Methods(\"PUT\")\n\n\tviewsRouter := r.PathPrefix(\"/views\").Subrouter()\n\tviewsRouter.HandleFunc(\"/title\", TitleHandler).Methods(\"GET\")\n\n\tsyncRouter := r.PathPrefix(\"/sync\").Subrouter()\n\tp := sync.NewPubsResource(db)\n\tsyncRouter.HandleFunc(\"/pubs\", p.GetPubs).Methods(\"GET\")\n\th := sync.NewHeadsResource(db)\n\tsyncRouter.HandleFunc(\"/heads\", h.GetHeads).Methods(\"GET\")\n\tf := sync.NewFeedResource(db)\n\tsyncRouter.HandleFunc(\"/feed/{id}\", f.GetFeed).Methods(\"GET\")\n\n\ta := sync.NewAnnounceResource(db)\n\t// Shouldn't really be a get, but is due to limitations of sandstorm\n\tsyncRouter.HandleFunc(\"/announce\", a.GetAnnouncement).Methods(\"GET\")\n\n\tr.Handle(\"/bundle.js\", http.FileServer(http.Dir(\"server/data/static/build\")))\n\tr.HandleFunc(\"/{path:.*}\", IndexHandler).Methods(\"GET\")\n\n\tgz := gziphandler.GzipHandler(r)\n\n\tif os.Getenv(\"SANDSTORM\") == \"1\" {\n\t\t// the sandstorm handler intercepts the sandstorm session ID and passes it to the Getter\n\t\t// So background requests are made with the sessionID that \"last touched\" the app\n\t\tss, bus := sandstorm.NewHandler(gz)\n\t\treturn ss, bus\n\t}\n\treturn gz, nil\n}", "func Run() {\n\tpool = setupDB()\n\n\t// Create plans\n\tplans = make(map[string]interface{})\n\tplans[\"small\"] = \"Small DB Instance - 2vCPU, 15.25 GiB RAM - $245/mo\"\n\n\tm := martini.Classic()\n\tm.Use(render.Renderer())\n\n\tm.Post(\"/v1/neptune/instance\", binding.Json(provisionspec{}), claimInstance)\n\tm.Delete(\"/v1/neptune/instance/:name\", deleteInstance)\n\tm.Get(\"/v1/neptune/url/:name\", getInstance)\n\tm.Get(\"/v1/neptune/plans\", func(r render.Render) { r.JSON(200, plans) })\n\tm.Post(\"/v1/neptune/tag\", binding.Json(tagspec{}), tagInstance)\n\n\tm.Run()\n}", "func spawnredditor(sr subreddit) {\t\n\t// Log start\n\tlog.Printf(\"Started listening to subreddit %s\",sr.Name)\n\n\t// Unwrap JSON\n\ttype structure struct {\n\t Data struct {\n\t \tChildren []struct {\n\t \t\tData struct {\n\t \t\t\tId \t\t\tstring\n\t \t\t\tSelf \t\tbool\t\t`json:\"is_self\"`\t\n\t \t\t\tUrl \t\tstring\n\t \t\t\tTs\t\t\tfloat64\t\t`json:\"created_utc\"`\n\t \t\t\tAuthor \t\tstring\n\t \t\t\tTitle \t\tstring\n\t \t\t\tSelftext\tstring\n\t \t\t\tPermalink\tstring\n\t \t\t}\n\t \t}\n\t }\n\t}\t\n\n\t// Define URI\n\turi := redditbaseurl + \"/r/\" + sr.Name + redditchannel\n\n\t// Loop endlessly\n\tfor {\t\n\t\t// Fetch JSON\n\t\traw, _ := fetchurl(uri)\t\t \t\t\n\n\t\t// Instantiate new structure\n\t\tvar s structure\n\n\t\t// Decode JSON\n\t\tjson.NewDecoder(raw.Body).Decode(&s)\t\t\t\t\t\n\n\t\t// Iterate through items\n\t\tfor _, child := range s.Data.Children {\n\t\t\t// See if it's newer than the last we knew\n\t\t\tif child.Data.Ts <= sr.Latest { continue }\n\n\t\t\t// Build proper item, map fields\n\t\t\titem := Item{\n\t\t\t\tid:\t\t\tchild.Data.Id,\n\t\t\t\tkind:\t\t\"reddit\",\t\t\n\t\t\t\tauthor:\t\tchild.Data.Author,\n\t\t\t\tpublished: \ttime.Unix(int64(child.Data.Ts),0),\n\t\t\t\ttitle:\t\tchild.Data.Title,\t\t\t\t\t\n\t\t\t\tbody: \t\tchild.Data.Selftext,\n\t\t\t}\n\n\t\t\t// If it's a posting instead of a link, start working on creating an Item\t\t\t\t\n\t\t\tif child.Data.Self {\n\t\t\t\t// Set Post URL\n\t\t\t\titem.url = child.Data.Url\t\t\t\n\n\t\t\t\t// Build post\n\t\t\t\tpost := Post{\n\t\t\t\t\titem: item,\n\t\t\t\t}\t\t\n\n\t\t\t\t// Send it off for processing\n\t\t\t\tgo AddPost(post)\t\t\t\t\n\n\t\t\t} else {\n\t\t\t\t// Set mention post url\n\t\t\t\titem.url = redditbaseurl + child.Data.Permalink\n\n\t\t\t\t// Build proper Mention\n\t\t\t\tmention := Mention{\n\t\t\t\t\titem: item,\n\t\t\t\t\ttarget: child.Data.Url,\t\t\t\n\t\t\t\t}\t\t\n\n\t\t\t\t// Send it off for processing\n\t\t\t\tgo AddMention(mention)\t\t\t\t\n\t\t\t}\n\t\t} \n\n\t\t// Set the timestamp to the newest we processed\n\t\tsr.Latest = s.Data.Children[0].Data.Ts\n\n\t\t// Wait until next update\n\t\ttime.Sleep(time.Minute)\n\t}\t\n}", "func (web *WebInterface) Run() error {\n\ttmplSearch, err := template.ParseFiles(path.Join(\"templates\", \"search.html\"))\n\tif err != nil {\n\t\treturn err\n\t}\n\ttmplResults, err := template.ParseFiles(path.Join(\"templates\", \"results.html\"))\n\tif err != nil {\n\t\treturn err\n\t}\n\thttp.HandleFunc(\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\ttmplSearch.Execute(w, nil)\n\t})\n\thttp.HandleFunc(\"/search\", func(w http.ResponseWriter, r *http.Request) {\n\t\tuserInput := r.FormValue(\"userInput\")\n\t\tlog.Debug().Str(\"text\", userInput).Msg(\"new search request\")\n\t\tif err != nil {\n\t\t\tlog.Err(err).Msg(\"error while reading index\")\n\t\t\thttp.Error(w, \"reading index error\", http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tsearchResults, err := web.searchFunc(userInput)\n\t\tif err != nil {\n\t\t\tlog.Err(err).Msg(\"error while searching\")\n\t\t\thttp.Error(w, \"searching error\", http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\ttmplResults.Execute(w, searchResults)\n\t})\n\treturn web.srv.ListenAndServe()\n}", "func Run(params *ContextParams) {\n\tr := createRouter(params)\n\n\tendless.DefaultHammerTime = 10 * time.Second\n\tendless.DefaultReadTimeOut = 295 * time.Second\n\tif err := endless.ListenAndServe(\":8080\", r); err != nil {\n\t\tlog.Infof(\"Server stopped: %s\", err)\n\t}\n}", "func main() {\n\tlog.SetFlags(log.LstdFlags | log.Lmicroseconds)\n\tlog.Printf(\"My peers are %v\", os.Getenv(\"PEERS\"))\n\tlog.Printf(\"traffic is %v\", os.Getenv(\"TRAFFIC\"))\n\tpeers := []*node.Peer{}\n\tfor _, s := range strings.Split(os.Getenv(\"PEERS\"), \" \") {\n\t\tp := &node.Peer{\n\t\t\tHost: fmt.Sprintf(\"node-%s\", s),\n\t\t\tPort: s}\n\t\tpeers = append(peers, p)\n\t}\n\n\n\tvar traffic = false\n\tif os.Getenv(\"TRAFFIC\") == \"1\" {\n\t\ttraffic = true\n\t}\n\n\tclientNode = client.NewClient(fmt.Sprintf(\"node-%s\", os.Getenv(\"PORT\")), os.Getenv(\"PORT\"), peers, uiChannel, nodeChannel, traffic)\n\n\terr := clientNode.SetupRPC()\n\tif err != nil {\n\t\tlog.Fatal(\"RPC setup error:\", err)\n\t}\n\terr = clientNode.Peer()\n\tif err != nil {\n\t\tlog.Fatal(\"Peering error:\", err)\n\t}\n\n\tfs := http.FileServer(http.Dir(\"../public\"))\n\thttp.Handle(\"/\", fs)\n\n\thttp.HandleFunc(\"/ws\", handleConnections)\n\thttp.HandleFunc(\"/disconnect\", handleDisconnect)\n\thttp.HandleFunc(\"/connect\", handleConnect)\n\thttp.HandleFunc(\"/getID\", handleGetID)\n\tgo handleMessages()\n\n\tgo func() {\n\t\terr := http.ListenAndServe(HttpPort, nil)\n\t\tif err != nil {\n\t\t\tlog.Fatal(\"ListenAndServe: \", err)\n\t\t}\n\t}()\n\n\tif traffic == true{\n\t\tclientNode.Start()\n\t}\n\n\tfor {\n\t\ttime.Sleep(time.Hour)\n\t}\n}", "func (b *Bot) startNewServers(newConfig *config.Config) {\n\tfor _, net := range newConfig.Networks() {\n\t\tif serverConf := b.conf.Network(net); nil == serverConf {\n\t\t\tserver, err := b.createServer(net, newConfig)\n\t\t\tif err != nil {\n\t\t\t\tb.botEnd <- fmt.Errorf(errFmtNewServer, net, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tb.servers[net] = server\n\n\t\t\tgo b.startServer(server, true, true)\n\t\t}\n\t}\n}", "func main() {\n\n\t//change to true when in production\n\tappCnf.InProduction = false\n\n\tsession = scs.New()\n\tsession.Lifetime = 24 * time.Hour\n\tsession.Cookie.Persist = true\n\tsession.Cookie.SameSite = http.SameSiteLaxMode\n\tsession.Cookie.Secure = appCnf.InProduction\n\n\tappCnf.Session = session\n\n\ttmplCache, err := render.CreateTemplateCache()\n\tif err != nil {\n\t\tfmt.Printf(\"Error crating template configuration, error %s \\n\", err)\n\t\t//fmt.Println(fmt.Sprintf(\"Error crating template configuration, error %s \\n\", err))\n\t}\n\tappCnf.TemplateCache = tmplCache\n\tappCnf.UseCache = false\n\n\trepo := handlers.NewRepo(&appCnf)\n\thandlers.NewHandlers(repo)\n\n\trender.NewTemplates(&appCnf)\n\n\tsrv := &http.Server{\n\t\tAddr: portNum,\n\t\tHandler: Routes(&appCnf),\n\t}\n\tfmt.Printf(\"Starting app on port %s for your pleasure \\n\", portNum)\n\t//fmt.Println(fmt.Sprintf(\"Starting app on port %s for your pleasure \\n\", portNum))\n\thttp.Handle(\"/static/\", http.StripPrefix(\"/static/images\", http.FileServer(http.Dir(\"static/images\"))))\n\terr = srv.ListenAndServe()\n\tlog.Fatal(err)\n\n}", "func mainHandler(w http.ResponseWriter, r *http.Request) {\n\n table := InitDB()\n if (table == nil) {\n fmt.Println(\"ERROR: SQL is NIL\")\n return\n }\n\n requests := libEvents(table, \"Guildhall Public\", \"request\")\n hosts := libEvents(table, \"Guildhall Public\", \"host\")\n\n strContents := readHtml(\"../index.html\", requests, hosts)\n fmt.Println(strContents)\n fmt.Fprint(w, strContents)\n}", "func (t *targetrunner) Run() error {\n\tconfig := cmn.GCO.Get()\n\tif err := t.si.Validate(); err != nil {\n\t\tcmn.ExitLogf(\"%v\", err)\n\t}\n\tt.httprunner.init(getstorstatsrunner(), config)\n\tt.owner.bmd = newBMDOwnerTgt()\n\n\tt.registerStats()\n\tt.httprunner.keepalive = gettargetkeepalive()\n\n\t// register object type and workfile type\n\tif err := fs.CSM.RegisterContentType(fs.ObjectType, &fs.ObjectContentResolver{}); err != nil {\n\t\tcmn.ExitLogf(\"%v\", err)\n\t}\n\tif err := fs.CSM.RegisterContentType(fs.WorkfileType, &fs.WorkfileContentResolver{}); err != nil {\n\t\tcmn.ExitLogf(\"%v\", err)\n\t}\n\n\tdryRunInit()\n\tt.gfn.local.tag, t.gfn.global.tag = \"local GFN\", \"global GFN\"\n\n\t// init meta-owners and load local instances\n\tt.owner.bmd.init() // BMD\n\tsmap, loaded := newSmap(), false // Smap\n\tif err := t.owner.smap.load(smap, config); err == nil {\n\t\tif errSmap := t.checkPresenceNetChange(smap); errSmap != nil {\n\t\t\tglog.Errorf(\"%s - proceeding anyway...\", errSmap)\n\t\t} else {\n\t\t\tloaded = true\n\t\t}\n\t} else if !os.IsNotExist(err) {\n\t\tglog.Errorf(\"%s: failed to load Smap (corruption?), err: %v\", t.si, err)\n\t}\n\t// insert self and always proceed starting up\n\tsmap.Tmap[t.si.ID()] = t.si\n\n\tcluster.InitTarget()\n\t//\n\t// join cluster\n\t//\n\tt.owner.smap.put(smap)\n\n\tif err := t.withRetry(t.joinCluster, \"join\", true /* backoff */); err != nil {\n\t\tif loaded {\n\t\t\tvar (\n\t\t\t\tsmapMaxVer int64\n\t\t\t\tprimaryURL string\n\t\t\t)\n\t\t\tif smapMaxVer, primaryURL = t.bcastHealth(smap); smapMaxVer > smap.version() {\n\t\t\t\tglog.Infof(\"%s: local copy of %s is older than v%d - retrying via %s\",\n\t\t\t\t\tt.si, smap, smapMaxVer, primaryURL)\n\t\t\t\terr = t.withRetry(t.joinCluster, \"join\", true, primaryURL)\n\t\t\t}\n\t\t}\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"%s failed to join cluster, err: %v\", t.si, err)\n\t\t\tglog.Errorf(\"%s is terminating\", t.si)\n\t\t\treturn err\n\t\t}\n\t}\n\n\tt.markNodeStarted()\n\n\tgo func() {\n\t\tt.pollClusterStarted(config.Timeout.CplaneOperation)\n\t\tt.markClusterStarted()\n\t}()\n\n\tt.detectMpathChanges()\n\n\t// init cloud\n\tt.cloud.init(t, config)\n\n\tt.authn = &authManager{\n\t\ttokens: make(map[string]*authRec),\n\t\trevokedTokens: make(map[string]bool),\n\t\tversion: 1,\n\t}\n\n\t// transactions\n\tt.transactions.init(t)\n\n\t//\n\t// REST API: register storage target's handler(s) and start listening\n\t//\n\ttransport.SetMux(cmn.NetworkPublic, t.publicServer.mux)\n\tif config.Net.UseIntraControl {\n\t\ttransport.SetMux(cmn.NetworkIntraControl, t.intraControlServer.mux)\n\t}\n\tif config.Net.UseIntraData {\n\t\ttransport.SetMux(cmn.NetworkIntraData, t.intraDataServer.mux)\n\t}\n\tt.initRecvHandlers()\n\n\tt.rebManager = reb.NewManager(t, config, getstorstatsrunner())\n\tec.Init(t, xaction.Registry)\n\n\taborted, _ := reb.IsRebalancing(cmn.ActResilver)\n\tif aborted {\n\t\tgo func() {\n\t\t\tglog.Infoln(\"resuming resilver...\")\n\t\t\tt.rebManager.RunResilver(\"\", false /*skipGlobMisplaced*/)\n\t\t}()\n\t}\n\n\tdsort.RegisterNode(t.owner.smap, t.owner.bmd, t.si, t.gmm, t, t.statsT)\n\tif err := t.httprunner.run(); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (agent *Agent) Run(domain string, filterRegex *regexp.Regexp, includeSubdomains bool, keys *session.Keys) (URLs chan scraping.URL) {\n\tURLs = make(chan scraping.URL)\n\n\tgo func() {\n\t\tdefer close(URLs)\n\n\t\tses, err := session.New(domain, filterRegex, includeSubdomains, 10, keys)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\twg := &sync.WaitGroup{}\n\n\t\t// Run each source in parallel on the target domain\n\t\tfor name, source := range agent.sources {\n\t\t\twg.Add(1)\n\n\t\t\tgo func(name string, source scraping.Source) {\n\t\t\t\tfor res := range source.Run(domain, ses, includeSubdomains) {\n\t\t\t\t\tURLs <- res\n\t\t\t\t}\n\n\t\t\t\twg.Done()\n\t\t\t}(name, source)\n\t\t}\n\n\t\twg.Wait()\n\n\t}()\n\n\treturn\n}", "func main() {\n\t// Create a HTTP client with Timeout\n\tclient := &http.Client {\n\t\tTimeout: 30 * time.Second,\n\t}\n\n\t// Define a Request in order to change the Header but\n\t// not send it yet.\n\trequest, err := http.NewRequest(\"GET\", \"http://webscraper.io/test-sites/e-commerce/allinone\", nil)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\t// Now that the Request is defined, override the Header!\n\trequest.Header.Set(\"User-Agent\", \"MyScraperBot v1.0 https://www.github.com/gitCMDR/\")\n\n\tfmt.Println(request)\n\t// Now let's make a Request\n\tresponse, err := client.Do(request)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer response.Body.Close()\n\n\n\t// Print the response to StdOut\n\t_, err = io.Copy(os.Stdout, response.Body)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n}", "func RunNew(args []string) {\n\n\t// Remove fragmenta backup from args list\n\targs = args[2:]\n\n\t// We expect two args left:\n\tif len(args) < 2 {\n\t\tlog.Printf(\"Both a project path and a project type or URL are required to create a new site\\n\")\n\t\treturn\n\t}\n\n\trepo := args[0]\n\tprojectPath, err := filepath.Abs(args[1])\n\tif err != nil {\n\t\tlog.Printf(\"Error expanding file path\\n\")\n\t\treturn\n\t}\n\n\tif !strings.HasPrefix(projectPath, filepath.Join(os.Getenv(\"GOPATH\"), \"src\")) {\n\t\tlog.Printf(\"WARNING: You should create your project in $GOPATH/src\\n\")\n\t}\n\n\tif fileExists(projectPath) {\n\t\tlog.Printf(\"A folder already exists at path %s\\n\", projectPath)\n\t\treturn\n\t}\n\n\tswitch repo {\n\tcase \"app\":\n\t\trepo = \"github.com/fragmenta/fragmenta-app\"\n\tcase \"cms\":\n\t\trepo = \"github.com/fragmenta/fragmenta-cms\"\n\t\t// TODO: Blog example does not exist yet\n\t\t//\tcase \"blog\":\n\t\t//\t\trepo = \"github.com/fragmenta/fragmenta-blog\"\n\tdefault:\n\t\t// TODO clean repo if it contains https or .git...\n\t}\n\n\t// Log fetching our files\n\tlog.Printf(\"Fetching from url: %s\\n\", repo)\n\n\t// Go get the project url, to make sure it is up to date, should use -u\n\t_, err = runCommand(\"go\", \"get\", repo)\n\tif err != nil {\n\t\tlog.Printf(\"Error calling go get %s\", err)\n\t\treturn\n\t}\n\n\t// Copy the pristine new site over\n\tgoProjectPath := filepath.Join(os.Getenv(\"GOPATH\"), \"src\", repo)\n\terr = copyNewSite(goProjectPath, projectPath)\n\tif err != nil {\n\t\tlog.Printf(\"Error copying project %s\", err)\n\t\treturn\n\t}\n\n\t// Generate config files\n\terr = generateConfig(projectPath)\n\tif err != nil {\n\t\tlog.Printf(\"Error generating config %s\", err)\n\t\treturn\n\t}\n\n\t// Generate a migration AND run it\n\terr = generateCreateSQL(projectPath)\n\tif err != nil {\n\t\tlog.Printf(\"Error generating migrations %s\", err)\n\t\treturn\n\t}\n\n\t// Output instructions to let them change setup first if they wish\n\tshowNewSiteHelp(projectPath)\n\n}", "func (wsv *web) run() {\n\tdefer wsv.doCloseDone.Done()\n\tdefer wsv.isRun.Store(false)\n\tdefer func() {\n\t\tif wsv.conf.Socket == \"\" {\n\t\t\treturn\n\t\t}\n\t\tif wsv.conf.Mode == \"unix\" || wsv.conf.Mode == \"unixpacket\" {\n\t\t\t_ = os.Remove(wsv.conf.Socket)\n\t\t}\n\t}()\n\n\t// Configure net/http web server\n\twsv.server = wsv.loadConfiguration()\n\tif wsv.err != nil {\n\t\treturn\n\t}\n\n\t// Configure keep alives of web server\n\tif wsv.conf.KeepAliveDisable {\n\t\twsv.server.SetKeepAlivesEnabled(false)\n\t}\n\t// Begin serve\n\twsv.err = wsv.server.Serve(wsv.listener)\n}", "func Run(apps map[string]interface{}, port string) {\n\trouter := mux.NewRouter()\n\tinitialize(apps, router)\n\n\tn := negroni.Classic()\n\tn.UseHandler(router)\n\tn.Run(\":\" + port)\n}", "func pingEngCore(bow *browser.Browser) {\n\n\tjar, err := cookiejar.New(nil)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tjar.SetCookies(bow.Url(), bow.SiteCookies())\n\n\n\t//Test that we can ping site without redirects\n\tclient := &http.Client{\n\t\tJar: jar,\n\t}\n\n\t//Url that we should use to test pingback \n\tpingUrl := bow.Url().String()\n\tif err != nil {\n\t\tpanic(err)\t\n\t}\n\t\n\t//timeoutStart := time.Now()\n\tvar totalPing int64 = 0\n\tvar numPing int64 = 0\n\t//var avgPing int64 = 0\n\n\t//Loop while our session hasn't expired\n\tfor err == nil {\n\t\t//lag timer start\n\t\tstart := time.Now()\n\n\t\tres, err := client.Get(pingUrl)\n\t\tif err != nil {\n\t\t\tmsg := \"Error:\" + err.Error()\n\t\t\t//c <- msg\n\t\t\tlog.Println(msg)\n\t\t\tfmt.Println(msg)\n\t\t\tbreak\t\n\t\t} else {\n\t\t\tlag := time.Since(start)\n\t\t\t//var msg string\n\t\t\tcurrPing := PingRecord{\n\t\t\t\tUrl: pingUrl,\n\t\t\t}\t\t\n\t\t\t\n\t\t\tnumPing++\n\t\t\ttotalPing = totalPing + int64(lag) \n\n\t\t\t//\trunning slow\n\t\t\tif lag > time.Duration(300)*time.Second {\n\t\t\t\t//msg = pingUrl + \" lag: \" + lag.String()\n\t\t\t\t//currPing.lagTime = lag.String()\t\n\t\t\t}\t\n\n\t\t\t//msg = pingUrl + \", lag: \" + lag.String()\n\t\t\t//c <- msg\n\t\n\t\t\tcurrPing.LagTime = lag.Nanoseconds()\n\t\t\tc <- &currPing\t\n\t\n\t\t\tres.Body.Close()\n\t\t}\n\t}\n\t\n\t//timeoutPeriod := time.Since(timeoutStart)\n\t//avgPing = totalPing/numPing\n\t\n\t//This is a bit hacky\n\t//msg := \"Timeout Period was:\" + timeoutPeriod.String() + \" average ping lag: \" + strconv.FormatInt(avgPing, 10) \n\t//c <- msg\n\tclose(c) \n}", "func StartBridgeNetwork(reader io.Reader, writer io.Writer) {\n\thttpAddr := \":7777\"\n\thttpsAddr := \":7778\"\n\n\tproxy := goproxy.NewProxyHTTPServer()\n\tproxy.Verbose = true\n\tproxy.BridgeID = \"TestBridge\"\n\tproxy.BridgeLinkNames = \"create_test-wait_test\"\n\tproxy.DataChannelReader = reader\n\tproxy.DataChannelWriter = writer\n\n\tif proxy.Verbose {\n\t\tlog.Printf(\"Server starting up! - configured to listen on http interface %s and https interface %s\", httpAddr, httpsAddr)\n\t}\n\n\tproxy.NonProxyHandler = http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tif req.Host == \"\" {\n\t\t\tfmt.Fprintln(w, \"Cannot handle requests without Host header, e.g., HTTP 1.0\")\n\t\t\treturn\n\t\t}\n\t\treq.URL.Scheme = \"http\"\n\t\treq.URL.Host = req.Host\n\t\tproxy.ServeHTTP(w, req)\n\t})\n\n\tproxy.OnRequest(goproxy.ReqHostMatches(regexp.MustCompile(\"^.*$\"))).HandleConnect(goproxy.AlwaysMitm)\n\n\tproxy.OnRequest(goproxy.ReqHostMatches(regexp.MustCompile(\"^.*:80$\"))).HijackConnect(func(req *http.Request, client net.Conn, ctx *goproxy.ProxyCtx) {\n\t\tdefer func() {\n\t\t\tif e := recover(); e != nil {\n\t\t\t\tctx.Logf(\"error connecting to remote: %v\", e)\n\t\t\t\tclient.Write([]byte(\"HTTP/1.1 500 Cannot reach destination\\r\\n\\r\\n\"))\n\t\t\t}\n\t\t\tclient.Close()\n\t\t}()\n\t\tclientBuf := bufio.NewReadWriter(bufio.NewReader(client), bufio.NewWriter(client))\n\t\tremote, err := connectDial(proxy, \"tcp\", req.URL.Host)\n\t\torPanic(err)\n\t\tremoteBuf := bufio.NewReadWriter(bufio.NewReader(remote), bufio.NewWriter(remote))\n\t\tfor {\n\t\t\treq, err := http.ReadRequest(clientBuf.Reader)\n\t\t\torPanic(err)\n\t\t\torPanic(req.Write(remoteBuf))\n\t\t\torPanic(remoteBuf.Flush())\n\t\t\tresp, err := http.ReadResponse(remoteBuf.Reader, req)\n\t\t\torPanic(err)\n\t\t\torPanic(resp.Write(clientBuf.Writer))\n\t\t\torPanic(clientBuf.Flush())\n\t\t}\n\t})\n\n\tvar httpServer *http.Server\n\tgo func() {\n\t\thttpServer = &http.Server{\n\t\t\tAddr: httpAddr,\n\t\t\tHandler: proxy,\n\t\t}\n\t\terr := httpServer.ListenAndServe()\n\t\tif err == nil {\n\t\t\tlog.Fatalf(\"Error ListenAndServe %v\", err)\n\t\t}\n\t}()\n\t// listen to the TLS ClientHello but make it a CONNECT request instead\n\tln, err := net.Listen(\"tcp\", httpsAddr)\n\tif err != nil {\n\t\tlog.Fatalf(\"Error listening for https connections - %v\", err)\n\t}\n\tfor {\n\t\tc, err := ln.Accept()\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Error accepting new connection - %v\", err)\n\t\t\tcontinue\n\t\t}\n\t\tgo func(c net.Conn) {\n\t\t\ttlsConn, err := vhost.TLS(c)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Error accepting new connection - %v\", err)\n\t\t\t}\n\t\t\tif tlsConn.Host() == \"\" {\n\t\t\t\tlog.Printf(\"Cannot support non-SNI enabled clients\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tconnectReq := &http.Request{\n\t\t\t\tMethod: \"CONNECT\",\n\t\t\t\tURL: &url.URL{\n\t\t\t\t\tOpaque: tlsConn.Host(),\n\t\t\t\t\tHost: net.JoinHostPort(tlsConn.Host(), \"443\"),\n\t\t\t\t},\n\t\t\t\tHost: tlsConn.Host(),\n\t\t\t\tHeader: make(http.Header),\n\t\t\t\tRemoteAddr: c.RemoteAddr().String(),\n\t\t\t}\n\t\t\tresp := dumbResponseWriter{tlsConn}\n\t\t\tproxy.ServeHTTP(resp, connectReq)\n\t\t}(c)\n\t}\n}", "func WebStartUp() {\n\thttp.HandleFunc(\"/\", servePage)\n\thttp.ListenAndServe(\":8080\", nil)\n}", "func (d *Bootstrap) Startup(net libp2p.Network) {\n\tgo func() {\n\t\ttime.Sleep(2 * time.Second)\n\t\tfor _, addr := range d.Addrs {\n\t\t\ts, err := net.NewSession(addr)\n\t\t\tif err != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ts.Send(Ping{IsServer: s.GetSelfAddr().IsServer()})\n\t\t}\n\t}()\n}", "func web(srvChan chan string, channel, nick, hostname string, args []string) {\n\tmessage := \"NOTICE \" + channel + \" :https://anex.us/\"\n\tlog.Println(message)\n\tsrvChan <- message\n}", "func (s *WebServer) Run(addr string) error {\n\tinitHandlers(s)\n\texpvar.Publish(\"Goroutines\", expvar.Func(func() interface{} {\n\t\treturn runtime.NumGoroutine()\n\t}))\n\n\thttp.Handle(\"/prom\", s.hub.Metrics.getHandler())\n\n\tsock, err := net.Listen(\"tcp\", addr)\n\tif err != nil {\n\t\treturn err\n\t}\n\tgo func() {\n\t\tfmt.Println(\"HTTP now available at\", addr)\n\t\tlog.Fatal(http.Serve(sock, nil))\n\t}()\n\treturn nil\n}", "func (a *WsApp) Run(addr string) {\n\n\tif addr == \"\" {\n\t\taddr = \"8082\"\n\t}\n\n\thttp.HandleFunc(\"/steam\", a.addrsSteam)\n\t// http.HandleFunc(\"/steam_blocks\", a.blocksSteam)\n}", "func (client Client) Web(parameters Parameters) (WebResultContainer, error) {\n\tvar result webResultWrapper\n\n\tif err := client.search(parameters.GetURI(SearchTypeWeb), &result); err != nil {\n\t\treturn WebResultContainer{}, err\n\t}\n\n\treturn result.Data, nil\n}", "func doJoin(w http.ResponseWriter, r *http.Request) {\n\ts, err := newServerCGI(w, r)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\thost, path, port := s.extractHost(\"join\")\n\thost = s.remoteIP(host)\n\tif host == \"\" {\n\t\treturn\n\t}\n\tn, err := node.MakeNode(host, path, port)\n\tif err != nil || !n.IsAllowed() {\n\t\treturn\n\t}\n\tif _, err := n.Ping(); err != nil {\n\t\treturn\n\t}\n\tsuggest := s.NodeManager.ReplaceNodeInList(n)\n\tif suggest == nil {\n\t\tfmt.Fprintln(s.wr, \"WELCOME\")\n\t\treturn\n\t}\n\tfmt.Fprintf(s.wr, \"WELCOME\\n%s\\n\", suggest.Nodestr)\n}", "func main() {\n\tprocessor := CrawlerProcessor{}\n\tc := semweb.NewCrawler(&processor)\n\tc.Run(\"https://twitter.com/mickael\")\n\tfmt.Println(\"Profiles:\")\n\tfor _, url := range processor.profileUrls {\n\t\tfmt.Println(url)\n\t}\n}", "func Run(configPath string) {\n\tif configPath == \"\" {\n\t\tconfigPath = \"data/config.yml\"\n\t}\n\tsetConfiguration(configPath)\n\tconf := config.GetConfig()\n\tweb := router.Setup()\n\tlog.Println(\"text similarity running on port \" + conf.Server.Port)\n\t_ = web.Run(\":\" + conf.Server.Port)\n}", "func main() {\n\n\t// get any cmd line flags\n\tconfigFile := flag.String(\"config\", \"config.json\", \"path to the config file\")\n\tAppInstance := flag.String(\"instance\", \"\", \"instance name\")\n\tAppCluster := flag.String(\"cluster\", \"\", \"Name for the cluster\")\n\tAWSRegion := flag.String(\"awsregion\", \"\", \"What AWS region we should look for resources in\")\n\tAWSProfile := flag.String(\"awsprofile\", \"\", \"What AWS security profile from the user/.aws/credentials file to use\")\n\tDBPath := flag.String(\"dbpath\", \"\", \"Path to the db\")\n\tflag.Parse()\n\n\t// read our config\n\tourConfig = webber.LoadConfig(*configFile)\n\n\t//set any cmd line overrides\n\tif len(*AWSRegion) > 0 {\n\t\tourConfig.AWSRegion = *AWSRegion\n\t}\n\tif len(*AWSProfile) > 0 {\n\t\tourConfig.AWSProfile = *AWSProfile\n\t}\n\tif len(*DBPath) > 0 {\n\t\tourConfig.DBPath = *DBPath\n\t}\n\n\t\n\t////////////////////////////\n\t// set up our logger\n\t// fill out the AppInfo struct so the logger knows who it is writting logs for:\n\tapp := logger.AppInfo{Name:ourConfig.AppName, Version:ourConfig.AppVersion, Instance:*AppInstance,Cluster:*AppCluster}\n\tfhLogger := logger.NewFirehoseLogger(app, ourConfig.AWSRegion, ourConfig.AWSProfile, ourConfig.LoggerFirehoseDeliveryStream)\n\tlogger.StdLogger = fhLogger\n\tlogger.StdLogger.StdOutOn(true)\n\tlogger.StdLogger.LOG(logger.INFO, \"\", \"SimonLB is starting up\", nil)\n\n\t// connect to our db\n\tdbSession, dbErr := mgo.Dial(ourConfig.DBPath)\n\tif ( dbErr != nil ) {\n\t\t// can't connect to our db\n\t\tlogger.StdLogger.LOG(logger.CRITICAL, \"\", fmt.Sprintf(\"Can't connect to db: %s\", dbErr), nil)\n\t\tos.Exit(1)\n\t}\n\tcDb = wtmcache.NewDb(dbSession, \"simonlb\")\n\twebber.CreateSessionDbCollection(cDb, ourConfig.SessionCollName)\n\twtmcache.CreateAutoIncDbCollection(cDb)\n\twtmcache.EnsureAutoIncrement(\"simonusers\")\n\twtmcache.EnsureAutoIncrement(\"tourneys\")\n\twtmcache.EnsureAutoIncrement(\"teams\")\n\tCreateUserDbCollection(cDb)\n\tCreateTeamDbCollection(cDb)\n\tCreateTourneyDbCollection(cDb)\n\n\thttpClient = webber.NewHttpClient(nil);\n\n\t// create an App Server\n\tas := webber.NewAppServer(ourConfig)\n\n\t//////////////////////////////////\n\t// create a couple of handlers\n\n\t// create our auth handler and assign it to <apibase>/auth\n\tauths := NewAuthServer(ourConfig.ApiBase + \"/auth\")\n\tas.RegisterHandler(auths)\n\n\t// create our teams handler and assign it to <apibase>/teams\n\tteams := NewTeamServer(ourConfig.ApiBase + \"/teams\")\n\tas.RegisterHandler(teams)\n\n\t// create an AWS session\n\tvar awsErr error\n\tawsSession, awsErr = session.NewSession(&aws.Config{Region:aws.String(ourConfig.AWSRegion)})\n\tif (awsErr != nil ) {\n\t\tlogger.StdLogger.LOG(logger.ERROR, \"\", fmt.Sprintf(\"Unable to create AWS session: %s\", awsErr.Error()), nil)\n\t}\n\n\t// now start the server\n\thttp.HandleFunc(\"/\", as.Handler)\n\thttp.ListenAndServe(ourConfig.Port, nil)\n\t\n}", "func newSession(snowflakes SnowflakeCollector) (net.PacketConn, *smux.Session, error) {\n\tclientID := turbotunnel.NewClientID()\n\n\t// We build a persistent KCP session on a sequence of ephemeral WebRTC\n\t// connections. This dialContext tells RedialPacketConn how to get a new\n\t// WebRTC connection when the previous one dies. Inside each WebRTC\n\t// connection, we use EncapsulationPacketConn to encode packets into a\n\t// stream.\n\tdialContext := func(ctx context.Context) (net.PacketConn, error) {\n\t\tlog.Printf(\"redialing on same connection\")\n\t\t// Obtain an available WebRTC remote. May block.\n\t\tconn := snowflakes.Pop()\n\t\tif conn == nil {\n\t\t\treturn nil, errors.New(\"handler: Received invalid Snowflake\")\n\t\t}\n\t\tlog.Println(\"---- Handler: snowflake assigned ----\")\n\t\t// Send the magic Turbo Tunnel token.\n\t\t_, err := conn.Write(turbotunnel.Token[:])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\t// Send ClientID prefix.\n\t\t_, err = conn.Write(clientID[:])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn NewEncapsulationPacketConn(dummyAddr{}, dummyAddr{}, conn), nil\n\t}\n\tpconn := turbotunnel.NewRedialPacketConn(dummyAddr{}, dummyAddr{}, dialContext)\n\n\t// conn is built on the underlying RedialPacketConn—when one WebRTC\n\t// connection dies, another one will be found to take its place. The\n\t// sequence of packets across multiple WebRTC connections drives the KCP\n\t// engine.\n\tconn, err := kcp.NewConn2(dummyAddr{}, nil, 0, 0, pconn)\n\tif err != nil {\n\t\tpconn.Close()\n\t\treturn nil, nil, err\n\t}\n\t// Permit coalescing the payloads of consecutive sends.\n\tconn.SetStreamMode(true)\n\t// Set the maximum send and receive window sizes to a high number\n\t// Removes KCP bottlenecks: https://gitlab.torproject.org/tpo/anti-censorship/pluggable-transports/snowflake/-/issues/40026\n\tconn.SetWindowSize(65535, 65535)\n\t// Disable the dynamic congestion window (limit only by the\n\t// maximum of local and remote static windows).\n\tconn.SetNoDelay(\n\t\t0, // default nodelay\n\t\t0, // default interval\n\t\t0, // default resend\n\t\t1, // nc=1 => congestion window off\n\t)\n\t// On the KCP connection we overlay an smux session and stream.\n\tsmuxConfig := smux.DefaultConfig()\n\tsmuxConfig.Version = 2\n\tsmuxConfig.KeepAliveTimeout = 10 * time.Minute\n\tsess, err := smux.Client(conn, smuxConfig)\n\tif err != nil {\n\t\tconn.Close()\n\t\tpconn.Close()\n\t\treturn nil, nil, err\n\t}\n\n\treturn pconn, sess, err\n}", "func main() {\n\tstart := time.Now()\n\tlog.Println(\"cnweb.main Iniitalizing cnweb\")\n\tctx := context.Background()\n\tvar err error\n\tb, err = initApp(ctx)\n\tif err != nil {\n\t\tlog.Printf(\"main() error for initApp, will retry on subsequent HTTP requests: %v\", err)\n\t}\n\n\turlPrefix := b.webConfig.GetVar(\"URLPrefix\")\n\tlog.Printf(\"main: urlPrefix: %s\", urlPrefix)\n\tif urlPrefix != \"/\" {\n\t\thttp.HandleFunc(\"/\", displayHome)\n\t}\n\thttp.HandleFunc(\"/#\", findHandler)\n\thttp.HandleFunc(\"/find/\", findHandler)\n\thttp.HandleFunc(\"/findadvanced/\", findFullText)\n\thttp.HandleFunc(\"/findsubstring\", findSubstring)\n\thttp.HandleFunc(\"/findtm\", translationMemory)\n\thttp.HandleFunc(\"/healthcheck\", healthcheck)\n\thttp.HandleFunc(\"/loggedin/changepassword\", changePasswordFormHandler)\n\thttp.HandleFunc(\"/library\", library)\n\thttp.HandleFunc(\"/loggedin/login\", loginHandler)\n\thttp.HandleFunc(\"/loggedin/login_form\", loginFormHandler)\n\thttp.HandleFunc(\"/loggedin/logout_form\", logoutForm)\n\thttp.HandleFunc(\"/loggedin/logout\", logoutHandler)\n\thttp.HandleFunc(\"/loggedin/session\", sessionHandler)\n\thttp.HandleFunc(\"/loggedin/portal\", portalHandler)\n\thttp.HandleFunc(\"/loggedin/portal_library/\", portalLibraryHandler)\n\thttp.HandleFunc(\"/loggedin/request_reset\", requestResetHandler)\n\thttp.HandleFunc(\"/loggedin/request_reset_form\", requestResetFormHandler)\n\thttp.HandleFunc(\"/loggedin/reset_password\", resetPasswordFormHandler)\n\thttp.HandleFunc(\"/loggedin/reset_password_submit\", resetPasswordHandler)\n\thttp.HandleFunc(\"/loggedin/submitcpwd\", changePasswordHandler)\n\tif b != nil {\n\t\tinitTranslationClients(b)\n\t} else {\n\t\tlog.Println(\"cnweb.man b == nil\")\n\t}\n\thttp.HandleFunc(\"/translateprocess\", processTranslation)\n\thttp.HandleFunc(\"/translate\", translationHome)\n\thttp.HandleFunc(\"/words/\", wordDetail)\n\n\t// If serving static HTML content\n\tstaticBucket := b.webConfig.GetVar(\"StaticBucket\")\n\tif len(staticBucket) > 0 {\n\t\tlog.Println(\"main: initializing GCS static content handler\")\n\t\tgcsClient, err := storage.NewClient(ctx)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"main error getting GCS client %v\", err)\n\t\t} else {\n\t\t\tsh := httphandling.NewGcsHandler(gcsClient, staticBucket, b.sessionEnforcer)\n\t\t\turlPrefix := b.webConfig.GetVar(\"URLPrefix\")\n\t\t\tlog.Printf(\"main: urlPrefix: %s\", urlPrefix)\n\t\t\tif len(urlPrefix) > 0 {\n\t\t\t\thttp.Handle(urlPrefix, sh)\n\t\t\t}\n\t\t\thttp.Handle(\"/web/\", http.StripPrefix(\"/web/\", sh))\n\t\t}\n\t} else {\n\t\tlog.Println(\"main: initializing local file static content handler\")\n\t\tsh := httphandling.NewStaticHandler(b.sessionEnforcer)\n\t\thttp.Handle(\"/web/\", http.StripPrefix(\"/web/\", sh))\n\t}\n\n\tportStr := \":\" + strconv.Itoa(config.GetPort())\n\tstartupTime := time.Since(start)\n\tlog.Printf(\"cnweb.main Started in %d millis, http server running at http://localhost%s\", startupTime.Milliseconds(), portStr)\n\terr = http.ListenAndServe(portStr, nil)\n\tif err != nil {\n\t\tlog.Printf(\"main() error for starting server: %v\", err)\n\t\tos.Exit(1)\n\t}\n}", "func main() {\n\tpulumi.Run(func(ctx *pulumi.Context) error {\n\t\tconfig := config.New(ctx, \"\")\n\n\t\tcount := config.GetInt(\"count\")\n\t\tif count == 0 {\n\t\t\tcount = 2\n\t\t}\n\t\tusername := config.Require(\"username\")\n\t\tpassword := config.Require(\"password\")\n\n\t\t// All resources will share a resource group.\n\t\trg, err := core.NewResourceGroup(ctx, \"server-rg\", nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Create a network and subnet for all VMs.\n\t\tnetwork, err := network.NewVirtualNetwork(ctx, \"server-network\", &network.VirtualNetworkArgs{\n\t\t\tResourceGroupName: rg.Name,\n\t\t\tAddressSpaces: pulumi.StringArray{pulumi.String(\"10.0.0.0/16\")},\n\t\t\tSubnets: network.VirtualNetworkSubnetArray{\n\t\t\t\tnetwork.VirtualNetworkSubnetArgs{\n\t\t\t\t\tName: pulumi.String(\"default\"),\n\t\t\t\t\tAddressPrefix: pulumi.String(\"10.0.1.0/24\"),\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\n\t\tsubnetID := network.Subnets.Index(pulumi.Int(0)).Id().ApplyT(func(val *string) (string, error) {\n\t\t\tif val == nil {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\treturn *val, nil\n\t\t}).(pulumi.StringOutput)\n\n\t\t// Now, allocate a few websever VMs -- by default, just 2, but this is configurable.\n\t\tvar ipAddresses pulumi.StringArray\n\t\tfor i := 0; i < count; i++ {\n\t\t\tserver, err := NewWebserver(ctx, fmt.Sprintf(\"ws-%v\", i), &WebserverArgs{\n\t\t\t\tUsername: pulumi.String(username),\n\t\t\t\tPassword: pulumi.String(password),\n\t\t\t\tBootScript: pulumi.String(fmt.Sprintf(`#!/bin/bash\necho \"Hello, from Server %v!\" > index.html\nnohup python -m SimpleHTTPServer 80 &`, i)),\n\t\t\t\tResourceGroupName: rg.Name,\n\t\t\t\tSubnetID: subnetID,\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tipAddresses = append(ipAddresses, server.GetIPAddress(ctx))\n\t\t}\n\n\t\tctx.Export(\"ipAddresses\", ipAddresses)\n\t\treturn nil\n\t})\n}", "func New(strs ...interface{}) (*Browser, error) {\n\tvar err error\n\tb := &Browser{}\n\tfor _, str := range strs {\n\t\tif errr := rpc.Register(str); errr != nil {\n\t\t\treturn nil, errr\n\t\t}\n\t}\n\tport := js.Global.Get(\"window\").Get(\"location\").Get(\"port\").String()\n\tb.s, err = websocket.Dial(\"ws://localhost:\" + port + \"/ws-client\") // Blocks until connection is established\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Println(\"connected to ws-client\")\n\tgo jsonrpc.ServeConn(b.s)\n\n\tb.c, err = websocket.Dial(\"ws://localhost:\" + port + \"/ws-server\") // Blocks until connection is established\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Println(\"connected to ws-server\")\n\tb.Client = jsonrpc.NewClient(b.c)\n\treturn b, nil\n}", "func serveWs(hub *Hub, w http.ResponseWriter, r *http.Request) {\n\tconn, err := upgrader.Upgrade(w, r, nil)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\tclient := &Client{hub: hub, conn: conn, send: make(chan []byte, 256)}\n\tclient.hub.register <- client\n\n\t// Allow collection of memory referenced by the caller by doing all work in\n\t// new goroutines.\n\tgo client.writePump()\n\tgo client.readPump()\n}", "func index(w http.ResponseWriter, r *http.Request) {\n\n\tsession := sessions.Start(w, r)\n\tif len(session.GetString(\"username\")) == 0 {\n\n\t\thttp.Redirect(w, r, \"/login\", 301)\n\n\t}\n\n\tcurrentUser := QueryUser(session.GetString(\"username\"))\n\n\tfeatured_rows, _ := db.Query(\"SELECT id, title, icon, banner FROM communities WHERE is_featured = 1 LIMIT 4\")\n\tvar featured []community\n\n\tfor featured_rows.Next() {\n\n\t\tvar row = community{}\n\n\t\terr = featured_rows.Scan(&row.ID, &row.Title, &row.Icon, &row.Banner)\n\t\tif err != nil {\n\n\t\t\tfmt.Println(err)\n\n\t\t}\n\n\t\tfeatured = append(featured, row)\n\n\t}\n\tfeatured_rows.Close()\n\n\tcommunity_rows, _ := db.Query(\"SELECT id, title, icon, banner FROM communities ORDER BY id DESC LIMIT 6\")\n\tvar communities []community\n\n\tfor community_rows.Next() {\n\n\t\tvar row = community{}\n\n\t\terr = community_rows.Scan(&row.ID, &row.Title, &row.Icon, &row.Banner)\n\t\tif err != nil {\n\n\t\t\tfmt.Println(err)\n\n\t\t}\n\t\tcommunities = append(communities, row)\n\n\t}\n\tcommunity_rows.Close()\n\n\tpjax := r.Header.Get(\"X-PJAX\") == \"\"\n\n\tvar data = map[string]interface{}{\n\t\t\"Title\": \"Communities\",\n\t\t\"Pjax\": pjax,\n\t\t\"CurrentUser\": currentUser,\n\t\t\"Featured\": featured,\n\t\t\"Communities\": communities,\n\t}\n\n\terr := templates.ExecuteTemplate(w, \"index.html\", data)\n\tif err != nil {\n\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\n\t}\n\treturn\n\n}", "func main() {\n\tapp := cli.NewApp()\n\tapp.Name = \"pepper gui\"\n\tapp.Usage = \"User interface for pepper\"\n\tapp.Flags = []cli.Flag{\n\t\tcli.StringFlag{\n\t\t\tName: \"ip\",\n\t\t\tValue: \"127.0.0.1\",\n\t\t\tUsage: \"IP to bind\",\n\t\t},\n\t\tcli.IntFlag{\n\t\t\tName: \"port\",\n\t\t\tValue: 6480,\n\t\t\tUsage: \"port to bind\",\n\t\t},\n\t}\n\tapp.Action = func(c *cli.Context) {\n\t\trouter := httprouter.New()\n\t\t// prod (tag dev is not present)\n\t\tif assets == nil {\n\t\t\t// rices boxes\n\t\t\thttpbox := rice.MustFindBox(\"../assets/http\")\n\t\t\tassets = httpbox.HTTPBox()\n\t\t}\n\n\t\t// assets\n\t\trouter.Handler(\"GET\", \"/assets/*res\", http.StripPrefix(\"/assets/\", http.FileServer(assets)))\n\n\t\t// index\n\t\trouter.HandlerFunc(\"GET\", \"/\", handlerIndex)\n\n\t\t// http server\n\t\thttpAddr := c.String(\"ip\") + \":\" + strconv.FormatInt(int64(c.Int(\"port\")), 10)\n\t\tbrowserAddr := httpAddr\n\t\tif c.String(\"ip\") == \"0.0.0.0\" {\n\t\t\tbrowserAddr = \"127.0.0.1:\" + strconv.FormatInt(int64(c.Int(\"port\")), 10)\n\t\t}\n\n\t\tn := negroni.New(negroni.NewRecovery())\n\t\tn.UseHandler(router)\n\t\tselect {\n\t\tcase <-time.After(1 * time.Second):\n\t\t\tbrowser.OpenURL(browserAddr)\n\t\t}\n\t\tlog.Println(\"GUI HTTP server listening on \" + httpAddr)\n\t\tlog.Fatalln(http.ListenAndServe(httpAddr, n))\n\t}\n\n\tapp.Run(os.Args)\n\n}", "func NewTraceWeb(ts *TraceServer, tt *TestTracker) *TraceWeb {\n\treturn &TraceWeb{\n\t\tts: ts,\n\t\ttt: tt,\n\t}\n}", "func Start() {\n\tIPs := DNS.GetAddr(Config.BootstrapDomainName)\n\tnodeContext = NC.NewNodeContext()\n\tnodeContext.SetLocalName(nameService.GetLocalName())\n\tnodeContext.LocalIp, _ = DNS.ExternalIP()\n\tmp = MP.NewMessagePasser(nodeContext.LocalName)\n\tstreamer = Streamer.NewStreamer(mp, nodeContext)\n\n\n\t// We use for loop to connect with all supernode one-by-one,\n\t// if a connection to one supernode fails, an error message\n\t// will be sent by messagePasser, and this message is further\n\t// processed in error handler.\n\t// init_fail: used in hello phase\n\t// exit: used when all supernode cannot be connected.\n\tmp.AddMappings([]string{\"exit\", \"init_fail\", \"super_fail\", \"ack\", \"loadtrack_result\"})\n\n\t// Initialize all the package structs\n\n\t// Define all the channel names and the binded functions\n\t// TODO: Register your channel name and binded eventhandlers here\n\t// The map goes as map[channelName][eventHandler]\n\t// All the messages with type channelName will be put in this channel by messagePasser\n\t// Then the binded handler of this channel will be called with the argument (*Message)\n\n\tchannelNames := map[string]func(*MP.Message){\n\t\t\"election_assign\": joinAssign,\n\t\t\"error\" : errorHandler,\n\n\t\t// The streaming related handlers goes here\n\t\t\"streaming_election\": streamer.HandleElection,\n\t\t\"streaming_join\": streamer.HandleJoin,\n\t\t\"streaming_data\": streamer.HandleStreamerData,\n\t\t\"streaming_stop\": streamer.HandleStop,\n\t\t\"streaming_assign\": streamer.HandleAssign,\n\t\t\"streaming_new_program\": streamer.HandleNewProgram,\n\t\t\"streaming_stop_program\": streamer.HandleStopProgram,\n\t\t\"streaming_quit\": streamer.HandleChildQuit,\n\t}\n\n\t// Init and listen\n\tfor channelName, handler := range channelNames {\n\t\t// Init all the channels listening on\n\t\tmp.Messages[channelName] = make(chan *MP.Message)\n\t\t// Bind all the functions listening on the channel\n\t\tgo listenOnChannel(channelName, handler)\n\t}\n\tgo nodeJoin(IPs)\n\tgo NodeCLIInterface(streamer)\n\twebInterface(streamer, nodeContext)\n\texitMsg := <- mp.Messages[\"exit\"]\n\tvar exitData string\n\tMP.DecodeData(&exitData, exitMsg.Data)\n\tfmt.Printf(\"Node: receiving force exit message [%s], node exit\\n\", exitData);\n}", "func New(token string, loader func(*GuildInfo) []Module) *SweetieBot {\n\tpath, _ := GetCurrentDir()\n\tselfhoster := &Selfhost{SelfhostBase{BotVersion.Integer()}, AtomicBool{0}, sync.Map{}}\n\trand.Seed(time.Now().UTC().Unix())\n\n\thostfile, gerr := ioutil.ReadFile(\"selfhost.json\")\n\tif gerr != nil {\n\n\t\tdefer func() {\n\t\t\tif r := recover(); r != nil {\n\t\t\t\tfmt.Println(\"Fatal error, press enter to exit: \", r)\n\t\t\t\tinput := \"\"\n\t\t\t\tfmt.Scanln(&input)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t}()\n\t\tInstall(path, selfhoster)\n\t}\n\n\tsb := &SweetieBot{\n\t\tToken: token,\n\t\tSelfName: \"Sweetie Bot\",\n\t\tAppName: \"Sweetie Bot\",\n\t\tDebugChannels: make(map[DiscordGuild]DiscordChannel),\n\t\tGuilds: make(map[DiscordGuild]*GuildInfo),\n\t\tLastMessages: make(map[DiscordChannel]int64),\n\t\tMaxConfigSize: 1000000,\n\t\tMaxUniqueItems: 25000,\n\t\tStartTime: time.Now().UTC().Unix(),\n\t\theartbeat: 4294967290,\n\t\tloader: loader,\n\t\tmemberChan: make(chan *GuildInfo, 2500),\n\t\tdeferChan: make(chan deferPair, 2000),\n\t\tSelfhoster: selfhoster,\n\t\tWebSecure: false,\n\t\tWebDomain: \"localhost\",\n\t\tWebPort: \":80\",\n\t\tchangelog: map[int]string{\n\t\t\tAssembleVersion(1, 0, 1, 9): \"- Introduced the !decrement command. This will decrement a counter by 1.\",\n\t\t\tAssembleVersion(1, 0, 1, 8): \"- Actually fix new user detection by requesting the necessary privileged intent\\n- Fixed quote mention problems caused by discordgo deleting the entire member list on reconnecting.\",\n\t\t\tAssembleVersion(1, 0, 1, 7): \"- Attempt to fix new user detection problems\",\n\t\t\tAssembleVersion(1, 0, 1, 6): \"- Remove useless silver checks\",\n\t\t\tAssembleVersion(1, 0, 1, 5): \"- Update discordgo\",\n\t\t\tAssembleVersion(1, 0, 1, 4): \"- Update discordgo\",\n\t\t\tAssembleVersion(1, 0, 1, 3): \"- Fix discordgo\",\n\t\t\tAssembleVersion(1, 0, 1, 2): \"- Update API\",\n\t\t\tAssembleVersion(1, 0, 1, 1): \"- Fix dumb crash\",\n\t\t\tAssembleVersion(1, 0, 1, 0): \"- Fix certain messages to use the correct prefix\\n- Clean removed members from the schedule\\n- Added keep highest n dice to !showroll\\n- Updated the transcripts one last time\\n- !fight and !ship no longer require a database\\n- Added basic mobile support to the website\\n- Support channel now locked to patreon supporters\\n- SWEETIE BOT IS NOW ARCHIVED. No new features will be added.\",\n\t\t\tAssembleVersion(1, 0, 0, 3): \"- Sanitize invalid command notification because nobody fucking reads the documentation.\\n- Fixed embed field titles.\\n- Made sweetiebot more persistent when added to a server.\\n- WARNING: Large servers may soon be restricted to Patreon members due to immense load put on the bot.\",\n\t\t\tAssembleVersion(1, 0, 0, 2): \"- Removed selfhosting support until further notice\\n- Sweetiebot now builds static version of the website.\",\n\t\t\tAssembleVersion(1, 0, 0, 1): \"- You can no longer !ban or !silence mods or admins.\\n- !import now accepts server IDs instead of just names.\\n- Using Member Role silencing is now optional when setting up a new server (but still highly recommended).\",\n\t\t\tAssembleVersion(1, 0, 0, 0): \"- Fixed hidden newuserrole dependency.\\n- Introduced Member role silencing, which solves rate limiting problems during raids. To enable this, use !help SetMemberRole for more information.\",\n\t\t\tAssembleVersion(0, 9, 9, 34): \"- Fixed resilencer race condition\\n- Fixed alias crash bug\\n- Improved documentation.\",\n\t\t\tAssembleVersion(0, 9, 9, 33): \"- Changed how language override works\",\n\t\t\tAssembleVersion(0, 9, 9, 32): \"- Security hotfix\",\n\t\t\tAssembleVersion(0, 9, 9, 31): \"- Fixed crash bug\\n- Added language override file\\n- No longer allow leaving and rejoining a server to clear silence status\",\n\t\t\tAssembleVersion(0, 9, 9, 30): \"- Fixed permissions error on import\",\n\t\t\tAssembleVersion(0, 9, 9, 29): \"- Moved markov chain to in-memory representation.\\n- Cleaned up database\\n- Add exponential backoff option to bored module.\\n- Detects if it can't send an embed to a channel and sends a warning instead.\\n- All edits show up in message log search\\n- Changed how !wipe parses it's arguments. Please check !help wipe\\n- Added season 8 transcripts\\n- Any administrator can now grant server silver.\",\n\t\t\tAssembleVersion(0, 9, 9, 28): \"- Increase max rules\\n- add mismatched parentheses check.\",\n\t\t\tAssembleVersion(0, 9, 9, 27): \"- Update documentation\\n- Updated the server and dependencies\\n- Optimized tag queries with lots of OR statements\\n- Added 'new member' role to the users module, a role that is added to all new members for a limited time after joining.\",\n\t\t\tAssembleVersion(0, 9, 9, 26): \"- Fixed crash in !search.\\n- Fixed installer SQL script.\\n- !setconfig timezonelocation now verifies the value and warns you if it is invalid.\\n- Omitting the channel in !wipe now simply defaults to the current channel.\",\n\t\t\tAssembleVersion(0, 9, 9, 25): \"- Changed !autosilence command to !raidsilence and migrated any existing aliases.\\n- The bot now tells the user if a PM failed to be sent.\\n- The bot now yells at you if you haven't set it up on the server yet.\\n- Added a silence timeout even though this is a bad idea becuase you all wanted it so damn bad.\\n- Added a counter module for all your counting needs.\\n- Setting a config string value to \\\"\\\" will now actually delete the string value.\",\n\t\t\tAssembleVersion(0, 9, 9, 24): \"- Fix updater issue on linux\\n- provide zip files instead of raw files for downloads\\n- Fix timezones on windows without go installations\\n- more idiotproofing\",\n\t\t\tAssembleVersion(0, 9, 9, 23): \"- Fixed crash in RolesModule\",\n\t\t\tAssembleVersion(0, 9, 9, 22): \"- Fixed crash in RolesModule and FilterModule\",\n\t\t\tAssembleVersion(0, 9, 9, 21): \"- Put lastmessages back on a lock for better performance\\n- Fix docker-specific bugs, add docker self-hosting image, because docker is cool now.\",\n\t\t\tAssembleVersion(0, 9, 9, 20): \"- Improve locking situation, begin concentrated effort to find and eliminate deadlocks via stacktraces\\n- The bot now yells at you if you try to set your timezone to Etc/GMT±00\",\n\t\t\tAssembleVersion(0, 9, 9, 19): \"- Fix installer and silver permissions handling\\n- Removed polls module, replaced with !poll command in the Misc module that analyzes emoji reaction polls instead.\",\n\t\t\tAssembleVersion(0, 9, 9, 18): \"- Fix database cleanup to preserve banned user comments.\",\n\t\t\tAssembleVersion(0, 9, 9, 17): \"- Fix potential detection failure in deadlock detector.\",\n\t\t\tAssembleVersion(0, 9, 9, 16): \"- Fix installer on linux\\n- Upgrade transcripts in database\\n- Better 64-bit support\",\n\t\t\tAssembleVersion(0, 9, 9, 15): \"- No longer attempts to track embed message updates\\n- Ignores new member join messages and other special messages\\n- Re-added echoembed command\\n- Autosilencing now include a reason for the silence\\n- Filters can now add pressure when triggered, and can be configured to not remove the message at all. Check the documentation for details\\n- Filters are no longer applied to bots/mods/admins.\\n- Ownership changes are properly tracked\\n- RemoveEvent now works on repeating events\\n- Sweetiebot now accepts escaped user pings and role mentions in the form <\\\\@12345> or <\\\\@&12345>. This won't ping the role/user, but still allows you to specify an exact ID.\\n- Now has a 200ms delay before deleting messages to prevent ghost messages.\\n- Ensure any user who sends a message will always have their username as an alias, even if it was missed before.\",\n\t\t\tAssembleVersion(0, 9, 9, 14): \"- Fuck Daylight Savings\\n- Also, fuck timezones\\n- Prevent silenced members from using emoji reactions.\\n- Removed main instance status loop (still available on selfhost instances)\\n- Can no longer search for a user that is not in your server. If you need to search for a banned user, ping them using the ID or specify username#1234. This makes searches much faster.\",\n\t\t\tAssembleVersion(0, 9, 9, 13): \"- Made some error messages more clear\\n- Fixed database cleanup functions\\n- Sweetiebot now deletes all information about guilds she hasn't been on for 3 days.\",\n\t\t\tAssembleVersion(0, 9, 9, 12): \"- Fix crash on !setfilter\",\n\t\t\tAssembleVersion(0, 9, 9, 11): \"- Merged !showroll command\\n- Prevented setting your default server to one that isn't set up.\",\n\t\t\tAssembleVersion(0, 9, 9, 10): \"- Sweetiebot no longer inserts !quote and !drop into the bored commands after restarting, unless the bored commands are empty; if you need to disable bored, disable the module instead.\\n- Exorcised demons from three servers with corrupted channel information.\\n- Filters now applied to invalid commands.\",\n\t\t\tAssembleVersion(0, 9, 9, 9): \"- Fix lastseen values\\n- Fix missing access error message when sweetie doesn't have read message history permissions.\",\n\t\t\tAssembleVersion(0, 9, 9, 8): \"- Restore old functionality of !echo\\n- say whether a user was autosilenced upon joining.\",\n\t\t\tAssembleVersion(0, 9, 9, 7): \"- Added !createroll\\n- !setconfig now accepts arbitrary strings, without quotes, in basic and [map] settings. Quotes are still required for [list] and [maplist] settings. Deletion NO LONGER USES \\\"\\\" in [map] settings. Simply pass nothing to delete a key.\\n- Fixed display problem in !getconfig, which now displays lists in alphabetical order.\",\n\t\t\tAssembleVersion(0, 9, 9, 6): \"- Update to go v1.9.3\\n- Improve database error handling.\\n- Fix chatlog race condition.\",\n\t\t\tAssembleVersion(0, 9, 9, 5): \"- Message logging is now deferred to a single thread to prevent database deadlocking.\\n- Username lookup now does fuzzy lookups on all aliases\\n- Only retains 10 most used aliases.\",\n\t\t\tAssembleVersion(0, 9, 9, 4): \"- Fix crash\",\n\t\t\tAssembleVersion(0, 9, 9, 3): \"- Sweetie Bot no longer tracks presence updates, because they were the cause of the database slowdowns. This means !lastseen will only operate on last message sent.\\n- Fixed !search.\\n- Added !assignrole\",\n\t\t\tAssembleVersion(0, 9, 9, 2): \"- Attempt #2 at fixing the database :U\",\n\t\t\tAssembleVersion(0, 9, 9, 1): \"- Database restructuring and optimizations\",\n\t\t\tAssembleVersion(0, 9, 9, 0): \"- Sweetie Bot now supports selfhosting and gives all patreon supporters access to paid features (chat logs and higher database limits). To get the new features, make sure you've linked your Patreon and Discord accounts. Check the GitHub readme for more instructions.\\n- Help now hides disabled or restricted commands from users.\\n- The bot name is no longer hardcoded: it will use whatever nickname it has on the server or the bot name given to the selfhost instance.\\n- Removed echoembed command\\n- Made listguilds and dumptables restricted commands\\n- renamed AlertRole to ModRole and Search.MaxResults to Miscellaneous.MaxSearchResults, moved TrackUserLeft to Users and added NotifyChannel to track users instead of using !autosilence.\\n- Spoiler module and Emote module have been replaced by a Filter module. If you were using these modules, your existing configuration was migrated to the new module. Anti-Spam was also renamed to Spam and Help/About renamed to Information.\\n- Removed !addset/!removeset/!searchset. Use !addstatus/!removestatus/etc. or !addfilter/!removefilter/etc. instead.\\n- Scheduler module no longer hides episode names outside of spoiler channels.\\n- Server owners and admins can now use any command in any channel.\\n- getconfig/setconfig now accept/display role names and channel names in addition to pings, and enforce valid types on all inputs. Using !getconfig on a category now displays all options in that category. All commands now accept simply writing out the role name, channel name, or user name.\\n- !addbirthday was changed to be easier to use, and now adds the birthday using the user's timezone.\\n- This is a total rewrite of sweetiebot, so if you find any bugs, or you think you should have paid features but you don't, please visit sweetiebot's support channel: https://discord.gg/t2gVQvN\",\n\t\t\tAssembleVersion(0, 9, 8, 23): \"- Remove !getpressure restriction.\\n- Limit results of !searchtags\",\n\t\t\tAssembleVersion(0, 9, 8, 22): \"- Prevent race-condition crashing set management.\\n- Force boolean configuration values to take only true or false.\",\n\t\t\tAssembleVersion(0, 9, 8, 21): \"- Made !userinfo more persistent at trying to find a match.\",\n\t\t\tAssembleVersion(0, 9, 8, 20): \"- Changed !searchtag to !searchtags because it's more consistent. Feel free to alias it back.\",\n\t\t\tAssembleVersion(0, 9, 8, 19): \"- Change how !remove works, use !remove * <item> to remove something with spaces from all tags.\\n- !pick now requires tags to be one argument, but supports * to pick from all tags.\\n-!searchtag can now take * in the tag argument to search all tags\",\n\t\t\tAssembleVersion(0, 9, 8, 18): \"- Fix specific tag search allowing tags from other servers to leak\",\n\t\t\tAssembleVersion(0, 9, 8, 17): \"- !tags now truncates output to 50 items unless user is a moderator\\n- More information added to !add and !tags\\n- Fixed bug with !remove\\n- Allow more lines to be returned before switching to private message\",\n\t\t\tAssembleVersion(0, 9, 8, 16): \"- All servers now have audit logs.\\n- Collections are now tags in the database, supporting complex tag searching. Use !tags and !searchtags to explore tags. Built-in collections are now managed through !addset, !removeset, and !searchset.\\n- Ignore LockWaitTimeout errors\",\n\t\t\tAssembleVersion(0, 9, 8, 15): \"- Return all possible !wipe errors\",\n\t\t\tAssembleVersion(0, 9, 8, 14): \"- Reduce database pressure on startup\",\n\t\t\tAssembleVersion(0, 9, 8, 13): \"- Fix crash on startup.\\n- Did more code refactoring, fixed several spelling errors.\",\n\t\t\tAssembleVersion(0, 9, 8, 12): \"- Do bulk member insertions in single batch to reduce database pressure.\\n- Removed bestpony command\\n- Did large internal code refactor\",\n\t\t\tAssembleVersion(0, 9, 8, 11): \"- User left now lists username+discriminator instead of pinging them to avoid @invalid-user problems.\\n- Add ToS to !about\\n- Bot now detects when it's about to be rate limited and combines short messages into a single large message. Helps keep bot responsive during huge raids.\\n- Fixed race condition in spam module.\",\n\t\t\tAssembleVersion(0, 9, 8, 10): \"- !setup can now be run by any user with the administrator role.\\n- Sweetie splits up embed messages if they have more than 25 fields.\\n- Added !getraid and !banraid commands\\n- Replaced !wipewelcome with generic !wipe command\\n- Added LinePressure, which adds pressure for each newline in a message\\n- Added TrackUserLeft, which will send a message when a user leaves in addition to when they join.\",\n\t\t\tAssembleVersion(0, 9, 8, 9): \"- Moved several options to outside files to make self-hosting simpler to set up\",\n\t\t\tAssembleVersion(0, 9, 8, 8): \"- !roll returns errors now.\\n- You can now change the command prefix to a different ascii character - no, you can't set it to an emoji. Don't try.\",\n\t\t\tAssembleVersion(0, 9, 8, 7): \"- Account creation time included on join message.\\n- Specifying the config category is now optional. For example, !setconfig rules 3 \\\"blah\\\" works.\",\n\t\t\tAssembleVersion(0, 9, 8, 6): \"- Support a lot more time formats and make time format more obvious.\",\n\t\t\tAssembleVersion(0, 9, 8, 5): \"- Augment discordgo with maps instead of slices, and switch to using standard discordgo functions.\",\n\t\t\tAssembleVersion(0, 9, 8, 4): \"- Update discordgo.\",\n\t\t\tAssembleVersion(0, 9, 8, 3): \"- Allow deadlock detector to respond to deadlocks in the underlying discordgo library.\\n- Fixed guild user count.\",\n\t\t\tAssembleVersion(0, 9, 8, 2): \"- Simplify sweetiebot setup\\n- Setting autosilence now resets the lockdown timer\\n- Sweetiebot won't restore the verification level if it was manually changed by an administrator.\",\n\t\t\tAssembleVersion(0, 9, 8, 1): \"- Switch to fork of discordgo to fix serious connection error handling issues.\",\n\t\t\tAssembleVersion(0, 9, 8, 0): \"- Attempts to register if she is removed from a server.\\n- Silencing has been redone to minimize rate-limiting problems.\\n- Sweetie now tracks the first time someone posts a message, used in the \\\"bannewcomers\\\" command, which bans everyone who sent their first message in the past two minutes (configurable).\\n- Sweetie now attempts to engage a lockdown when a raid is detected by temporarily increasing the server verification level. YOU MUST GIVE THE BOT \\\"MANAGE SERVER\\\" PERMISSIONS FOR THIS TO WORK! This can be disabled by setting Spam.LockdownDuration to 0.\",\n\t\t\tAssembleVersion(0, 9, 7, 9): \"- Discard Group DM errors from legacy conversations.\",\n\t\t\tAssembleVersion(0, 9, 7, 8): \"- Correctly deal with rare edge-case on !userinfo queries.\",\n\t\t\tAssembleVersion(0, 9, 7, 7): \"- Sweetiebot sends an autosilence change message before she starts silencing raiders, to ensure admins get immediate feedback even if discord is being slow.\",\n\t\t\tAssembleVersion(0, 9, 7, 6): \"- Sweetiebot now ignores other bots by default. To revert this, run '!setconfig basic.listentobots true' and she will listen to them again, but will never attempt to silence them.\\n- Removed legacy timezones\\n- Spam messages are limited to 300 characters in the log.\",\n\t\t\tAssembleVersion(0, 9, 7, 5): \"- Compensate for discordgo being braindead and forgetting JoinedAt dates.\",\n\t\t\tAssembleVersion(0, 9, 7, 4): \"- Update discordgo API.\",\n\t\t\tAssembleVersion(0, 9, 7, 3): \"- Fix permissions issue.\",\n\t\t\tAssembleVersion(0, 9, 7, 2): \"- Fix ignoring admins in anti-spam.\",\n\t\t\tAssembleVersion(0, 9, 7, 1): \"- Fixed an issue with out-of-date guild objects not including all server members.\",\n\t\t\tAssembleVersion(0, 9, 7, 0): \"- Groups have been removed and replaced with user-assignable roles. All your groups have automatically been migrated to roles. If there was a name-collision with an existing role, your group name will be prefixed with 'sb-', which you can then resolve yourself. Use '!help roles' to get usage information about the new commands.\",\n\t\t\tAssembleVersion(0, 9, 6, 9): \"- Sweetiebot no longer logs her own actions in the audit log\",\n\t\t\tAssembleVersion(0, 9, 6, 8): \"- Sweetiebot now has a deadlock detector and will auto-restart if she detects that she is not responding to !about\\n- Appending @ to the end of a name or server is no longer necessary. If sweetie finds an exact match to your query, she will always use that.\",\n\t\t\tAssembleVersion(0, 9, 6, 7): \"- Sweetiebot no longer attempts to track edited messages for spam detection. This also fixes a timestamp bug with pinned messages.\",\n\t\t\tAssembleVersion(0, 9, 6, 6): \"- Sweetiebot now automatically sets Silence permissions on newly created channels. If you have a channel that silenced members should be allowed to speak in, make sure you've set it as the welcome channel via !setconfig users.welcomechannel #yourchannel\",\n\t\t\tAssembleVersion(0, 9, 6, 5): \"- Fix spam detection error for edited messages.\",\n\t\t\tAssembleVersion(0, 9, 6, 4): \"- Enforce max DB connections to try to mitigate connection problems\",\n\t\t\tAssembleVersion(0, 9, 6, 3): \"- Extreme spam could flood SB with user updates, crashing the database. She now throttles user updates to help prevent this.\\n- Anti-spam now uses discord's message timestamp, which should prevent false positives from network problems\\n- Sweetie will no longer silence mods for spamming under any circumstance.\",\n\t\t\tAssembleVersion(0, 9, 6, 2): \"- Renamed !quickconfig to !setup, added a friendly PM to new servers to make initial setup easier.\",\n\t\t\tAssembleVersion(0, 9, 6, 1): \"- Fix !bestpony crash\",\n\t\t\tAssembleVersion(0, 9, 6, 0): \"- Sweetiebot is now self-repairing and can function without a database, although her functionality is EXTREMELY limited in this state.\",\n\t\t\tAssembleVersion(0, 9, 5, 9): \"- MaxRemoveLookback no longer relies on the database and can now be used in any server. However, it only deletes messages from the channel that was spammed in.\",\n\t\t\tAssembleVersion(0, 9, 5, 8): \"- You can now specify per-channel pressure overrides via '!setconfig spam.maxchannelpressure <channel> <pressure>'.\",\n\t\t\tAssembleVersion(0, 9, 5, 7): \"- You can now do '!pick collection1+collection2' to pick a random item from multiple collections.\\n- !fight <monster> is now sanitized.\\n- !silence now tells you when someone already silenced will be unsilenced, if ever.\",\n\t\t\tAssembleVersion(0, 9, 5, 6): \"- Prevent idiots from setting status.cooldown to 0 and breaking everything.\",\n\t\t\tAssembleVersion(0, 9, 5, 5): \"- Fix crash on invalid command limits.\",\n\t\t\tAssembleVersion(0, 9, 5, 4): \"- Added ignorerole for excluding certain users from spam detection.\\n- Adjusted unsilence to force bot to assume user is unsilenced so it can be used to fix race conditions.\",\n\t\t\tAssembleVersion(0, 9, 5, 3): \"- Prevent users from aliasing existing commands.\",\n\t\t\tAssembleVersion(0, 9, 5, 2): \"- Show user account creation date in userinfo\\n- Added !SnowflakeTime command\",\n\t\t\tAssembleVersion(0, 9, 5, 1): \"- Allow !setconfig to edit float values\",\n\t\t\tAssembleVersion(0, 9, 5, 0): \"- Completely overhauled Anti-Spam module. Sweetie now analyzes message content and tracks text pressure users exert on the chat. See !help anti-spam for details, or !getconfig spam for descriptions of the new configuration options. Your old MaxImages and MaxPings settings were migrated over to ImagePressure and PingPressure, respectively.\",\n\t\t\tAssembleVersion(0, 9, 4, 5): \"- Escape nicknames correctly\\n- Sweetiebot no longer tracks per-server nickname changes, only username changes.\\n- You can now use the format username#1234 in user arguments.\",\n\t\t\tAssembleVersion(0, 9, 4, 4): \"- Fix locks, update endpoint calls, improve antispam response.\",\n\t\t\tAssembleVersion(0, 9, 4, 3): \"- Emergency revert of last changes\",\n\t\t\tAssembleVersion(0, 9, 4, 2): \"- Spammer killing is now asynchronous and should have fewer duplicate alerts.\",\n\t\t\tAssembleVersion(0, 9, 4, 1): \"- Attempt to make sweetiebot more threadsafe.\",\n\t\t\tAssembleVersion(0, 9, 4, 0): \"- Reduced number of goroutines, made updating faster.\",\n\t\t\tAssembleVersion(0, 9, 3, 9): \"- Added !getaudit command for server admins.\\n- Updated documentation for consistency.\",\n\t\t\tAssembleVersion(0, 9, 3, 8): \"- Removed arbitrary limit on spam message detection, replaced with sanity limit of 600.\\n- Sweetiebot now automatically detects invalid spam.maxmessage settings and removes them instead of breaking your server.\\n- Replaced a GuildMember call with an initial state check to eliminate lag and some race conditions.\",\n\t\t\tAssembleVersion(0, 9, 3, 7): \"- If a collection only has one item, just display the item.\\n- If you put \\\"!\\\" into CommandRoles[<command>], it will now allow any role EXCEPT the roles specified to use <command>. This behaves the same as the channel blacklist function.\",\n\t\t\tAssembleVersion(0, 9, 3, 6): \"- Add log option to autosilence.\\n- Ensure you actually belong to the server you set as your default.\",\n\t\t\tAssembleVersion(0, 9, 3, 5): \"- Improve help messages.\",\n\t\t\tAssembleVersion(0, 9, 3, 4): \"- Prevent cross-server message sending exploit, without destroying all private messages this time.\",\n\t\t\tAssembleVersion(0, 9, 3, 3): \"- Emergency revert change.\",\n\t\t\tAssembleVersion(0, 9, 3, 2): \"- Prevent cross-server message sending exploit.\",\n\t\t\tAssembleVersion(0, 9, 3, 1): \"- Allow sweetiebot to be executed as a user bot.\",\n\t\t\tAssembleVersion(0, 9, 3, 0): \"- Make argument parsing more consistent\\n- All commands that accepted a trailing argument without quotes no longer strip quotes out. The quotes will now be included in the query, so don't put them in if you don't want them!\\n- You can now escape '\\\"' inside an argument via '\\\\\\\"', which will work even if discord does not show the \\\\ character.\",\n\t\t\tAssembleVersion(0, 9, 2, 3): \"- Fix echoembed crash when putting in invalid parameters.\",\n\t\t\tAssembleVersion(0, 9, 2, 2): \"- Update help text.\",\n\t\t\tAssembleVersion(0, 9, 2, 1): \"- Add !joingroup warning to deal with breathtaking stupidity of zootopia users.\",\n\t\t\tAssembleVersion(0, 9, 2, 0): \"- Remove !lastping\\n- Help now lists modules with no commands\",\n\t\t\tAssembleVersion(0, 9, 1, 1): \"- Fix crash in !getconfig\",\n\t\t\tAssembleVersion(0, 9, 1, 0): \"- Renamed config options\\n- Made things more clear for new users\\n- Fixed legacy importable problem\\n- Fixed command saturation\\n- Added botchannel notification\\n- Changed getconfig behavior for maps\",\n\t\t\tAssembleVersion(0, 9, 0, 4): \"- To protect privacy, !listguilds no longer lists servers that do not have Basic.Importable set to true.\\n- Remove some more unnecessary sanitization\",\n\t\t\tAssembleVersion(0, 9, 0, 3): \"- Don't sanitize links already in code blocks\",\n\t\t\tAssembleVersion(0, 9, 0, 2): \"- Alphabetize collections because Tawmy is OCD\",\n\t\t\tAssembleVersion(0, 9, 0, 1): \"- Update documentation\\n- Simplify !collections output\",\n\t\t\tAssembleVersion(0, 9, 0, 0): \"- Completely restructured Sweetie Bot into a module-based architecture\\n- Disabling/Enabling a module now disables/enables all its commands\\n- Help now includes information about modules\\n- Collections command is now pretty\",\n\t\t\tAssembleVersion(0, 8, 17, 2): \"- Added ability to hide negative rules because Tawmy is weird\",\n\t\t\tAssembleVersion(0, 8, 17, 1): \"- Added echoembed command\",\n\t\t\tAssembleVersion(0, 8, 17, 0): \"- Sweetiebot can now send embeds\\n- Made about message pretty\",\n\t\t\tAssembleVersion(0, 8, 16, 3): \"- Update discordgo structs to account for breaking API change.\",\n\t\t\tAssembleVersion(0, 8, 16, 2): \"- Enable sweetiebot to tell dumbasses that they are dumbasses.\",\n\t\t\tAssembleVersion(0, 8, 16, 1): \"- !add can now add to multiple collections at the same time.\",\n\t\t\tAssembleVersion(0, 8, 16, 0): \"- Alphabetized the command list\",\n\t\t\tAssembleVersion(0, 8, 15, 4): \"- ReplaceMentions now breaks role pings (but does not resolve them)\",\n\t\t\tAssembleVersion(0, 8, 15, 3): \"- Use database to resolve users to improve responsiveness\",\n\t\t\tAssembleVersion(0, 8, 15, 2): \"- Improved !vote error messages\",\n\t\t\tAssembleVersion(0, 8, 15, 1): \"- Quickconfig actually sets silentrole now\",\n\t\t\tAssembleVersion(0, 8, 15, 0): \"- Use 64-bit integer conversion\",\n\t\t\tAssembleVersion(0, 8, 14, 6): \"- Allow adding birthdays on current day\\n-Update avatar change function\",\n\t\t\tAssembleVersion(0, 8, 14, 5): \"- Allow exact string matching on !import\",\n\t\t\tAssembleVersion(0, 8, 14, 4): \"- Added !import\\n- Added Importable option\\n- Make !collections more useful\",\n\t\t\tAssembleVersion(0, 8, 14, 3): \"- Allow pinging multiple groups via group1+group2\",\n\t\t\tAssembleVersion(0, 8, 14, 2): \"- Fix !createpoll unique option key\\n- Add !addoption\",\n\t\t\tAssembleVersion(0, 8, 14, 1): \"- Clean up !poll\",\n\t\t\tAssembleVersion(0, 8, 14, 0): \"- Added !poll, !vote, !createpoll, !deletepoll and !results commands\",\n\t\t\tAssembleVersion(0, 8, 13, 1): \"- Fixed !setconfig rules\",\n\t\t\tAssembleVersion(0, 8, 13, 0): \"- Added changelog\\n- Added !rules command\",\n\t\t\tAssembleVersion(0, 8, 12, 0): \"- Added temporary silences\",\n\t\t\tAssembleVersion(0, 8, 11, 5): \"- Added \\\"dumbass\\\" to Sweetie Bot's vocabulary\",\n\t\t\tAssembleVersion(0, 8, 11, 4): \"- Display channels in help for commands\",\n\t\t\tAssembleVersion(0, 8, 11, 3): \"- Make defaultserver an independent command\",\n\t\t\tAssembleVersion(0, 8, 11, 2): \"- Add !defaultserver command\",\n\t\t\tAssembleVersion(0, 8, 11, 1): \"- Fix !autosilence behavior\",\n\t\t\tAssembleVersion(0, 8, 11, 0): \"- Replace mentions in !search\\n- Add temporary ban to !ban command\",\n\t\t\tAssembleVersion(0, 8, 10, 0): \"- !ping now accepts newlines\\n- Added build version to make moonwolf happy\",\n\t\t\tAssembleVersion(0, 8, 9, 0): \"- Add silence message for Tawmy\\n- Make silence message ping user\\n- Fix #27 (Sweetie Bot explodes if you search nothing)\\n- Make !lastseen more reliable\",\n\t\t\tAssembleVersion(0, 8, 8, 0): \"- Log all commands sent to SB in DB-enabled servers\",\n\t\t\tAssembleVersion(0, 8, 7, 0): \"- Default to main server for PMs if it exists\\n- Restrict PM commands to the server you belong in (fix #26)\\n- Make spam deletion lookback configurable\\n- Make !quickconfig complain if permissions are wrong\\n- Add giant warning label for Tawmy\\n- Prevent parse time crash\\n- Make readme more clear on how things work\\n- Sort !listguild by user count\\n- Fallback to search all users if SB can't find one in the current server\",\n\t\t\tAssembleVersion(0, 8, 6, 0): \"- Add full timezone support\\n- Deal with discord's broken permissions\\n- Improve timezone help messages\",\n\t\t\tAssembleVersion(0, 8, 5, 0): \"- Add !userinfo\\n- Fix #15 (Lock down !removeevent)\\n- Fix guildmember query\\n- Use nicknames in more places\",\n\t\t\tAssembleVersion(0, 8, 4, 0): \"- Update readme, remove disablebored\\n- Add delete command\",\n\t\t\tAssembleVersion(0, 8, 3, 0): \"- Actually seed random number generator because Cloud is a FUCKING IDIOT\\n- Allow newlines in commands\\n- Bored module is now fully programmable\\n- Display user ID in !aka\\n- Hopefully stop sweetie from being an emo teenager\\n- Add additional stupid proofing\\n- Have bored commands override all restrictions\",\n\t\t\tAssembleVersion(0, 8, 2, 0): \"- Enable multi-server message logging\\n- Extend !searchquote\\n- Attach !lastping to current server\\n- Actually make aliases work with commands\",\n\t\t\tAssembleVersion(0, 8, 1, 0): \"- Add dynamic collections\\n- Add quotes\\n- Prevent !aka command from spawning evil twins\\n- Add !removealias\\n- Use nicknames where possible\\n- Fix off by one error\\n- Sanitize !search output \",\n\t\t\tAssembleVersion(0, 8, 0, 0): \"- Appease the dark gods of discord's API\\n- Allow sweetiebot to track nicknames\\n- update help\\n- Include nickname in searches\",\n\t\t},\n\t}\n\n\tjson.Unmarshal(hostfile, sb)\n\tsb.Token = strings.TrimSpace(sb.Token)\n\tsb.EmptyGuild = NewGuildInfo(sb, &discordgo.Guild{})\n\n\tsb.EmptyGuild.Config.FillConfig()\n\tsb.EmptyGuild.Config.SetupDone = true\n\tsb.EmptyGuild.Modules = sb.loader(sb.EmptyGuild)\n\tsort.Sort(moduleArray(sb.EmptyGuild.Modules))\n\n\tfor _, v := range sb.EmptyGuild.Modules {\n\t\tsb.EmptyGuild.RegisterModule(v)\n\t\tfor _, command := range v.Commands() {\n\t\t\tif command.Info().ServerIndependent {\n\t\t\t\tsb.EmptyGuild.AddCommand(command, v)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Load language override\n\tif configHelpFile, err := ioutil.ReadFile(\"confighelp.json\"); err == nil {\n\t\tif err = json.Unmarshal(configHelpFile, &ConfigHelp); err != nil {\n\t\t\tfmt.Println(\"Error loading config help replacement file: \", err)\n\t\t}\n\t}\n\n\tif stringsFile, err := ioutil.ReadFile(\"strings.json\"); err == nil {\n\t\tif err = json.Unmarshal(stringsFile, &StringMap); err != nil {\n\t\t\tfmt.Println(\"Error loading strings replacement file: \", err)\n\t\t}\n\t}\n\n\tdb, err := dbLoad(&emptyLog{}, \"mysql\", strings.TrimSpace(sb.DBAuth))\n\tsb.DB = db\n\tif !db.Status.Get() {\n\t\tfmt.Println(\"Database connection failure - running in No Database mode: \", err.Error())\n\t} else {\n\t\terr = sb.DB.LoadStatements()\n\t\tif err == nil {\n\t\t\tfmt.Println(\"Finished loading database statements\")\n\t\t} else {\n\t\t\tfmt.Println(\"Loading database statements failed: \", err)\n\t\t\tfmt.Println(\"DATABASE IS BADLY FORMATTED OR CORRUPT - TERMINATING SWEETIE BOT!\")\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tvar dg *discordgo.Session\n\tif sb.IsUserMode {\n\t\tdg, err = discordgo.New(sb.Token)\n\t\tfmt.Println(\"Started SweetieBot on a user account.\")\n\t} else {\n\t\tdg, err = discordgo.New(\"Bot \" + sb.Token)\n\t\tdg.Identify.Intents = discordgo.MakeIntent(discordgo.IntentsAllWithoutPrivileged | discordgo.IntentsGuildMembers)\n\t}\n\tsb.DG = &DiscordGoSession{*dg}\n\n\tif err != nil {\n\t\tfmt.Println(\"Error creating discord session\", err.Error())\n\t\treturn nil\n\t}\n\tsb.DG.LogLevel = discordgo.LogWarning\n\n\tsb.DG.AddHandler(sb.OnReady)\n\tsb.DG.AddHandler(sb.MessageCreate)\n\tsb.DG.AddHandler(sb.MessageUpdate)\n\tsb.DG.AddHandler(sb.MessageDelete)\n\tsb.DG.AddHandler(sb.UserUpdate)\n\tsb.DG.AddHandler(sb.GuildUpdate)\n\tsb.DG.AddHandler(sb.GuildMemberAdd)\n\tsb.DG.AddHandler(sb.GuildMemberRemove)\n\tsb.DG.AddHandler(sb.GuildMemberUpdate)\n\tsb.DG.AddHandler(sb.GuildMembersChunk)\n\tsb.DG.AddHandler(sb.GuildBanAdd)\n\tsb.DG.AddHandler(sb.GuildBanRemove)\n\tsb.DG.AddHandler(sb.GuildRoleDelete)\n\tsb.DG.AddHandler(sb.GuildCreate)\n\tsb.DG.AddHandler(sb.ChannelCreate)\n\treturn sb\n}", "func Home(w http.ResponseWriter, r *http.Request) {\n\t// Start a new root Span and therefore a new trace.\n\tspan := opentracing.StartSpan(r.URL.Path)\n\tdefer span.Finish()\n\n\t// OpenTracing allows for arbritary tags to be added to a Span.\n\tspan.SetTag(\"Request.Host\", r.Host)\n\tspan.SetTag(\"Request.Address\", r.RemoteAddr)\n\taddHeaderTags(span, r.Header)\n\n\t// Baggage Items are similar to tags, however they are propagated to all\n\t// children spans, so this will show up in the API calls.\n\tspan.SetBaggageItem(\"User\", os.Getenv(\"USER\"))\n\n\t// We're going to make some API request, so we use the default HTTP client\n\t// to send HTTP requests with trace information placed inside the headers.\n\thttpClient := http.DefaultClient\n\n\t// Make three API requests using our HTTP client.\n\tfor i := 0; i < 3; i++ {\n\t\treq, err := http.NewRequest(\"GET\", \"http://localhost:8699/endpoint\", nil)\n\t\tif err != nil {\n\t\t\tlog.Println(\"/endpoint:\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// We inject the span into the request headers before making the request.\n\t\tcarrier := opentracing.HTTPHeadersCarrier(req.Header)\n\t\tspan.Tracer().Inject(span.Context(), opentracing.HTTPHeaders, carrier)\n\t\tresp, err := httpClient.Do(req)\n\t\tif err != nil {\n\t\t\tlog.Println(\"/endpoint:\", err)\n\n\t\t\t// Log the error to the span.\n\t\t\tspan.LogEvent(err.Error())\n\t\t\tcontinue\n\t\t}\n\n\t\tspan.SetTag(\"Response.Status\", resp.Status)\n\t\tresp.Body.Close()\n\t}\n\n\t// Render the page.\n\tfmt.Fprintf(w, `<p>Three API requests have been made!</p>`)\n\tfmt.Fprintf(w, `<p><a href=\"http://localhost:8700/traces\" target=\"_\">View the trace</a></p>`)\n}", "func registerWebsite(w http.ResponseWriter, r *http.Request) {\r\n\tdb, err = gorm.Open(\"sqlite3\", \"website.db\")\r\n\tif err != nil {\r\n\t\tpanic(\"Could not connect to database\")\r\n\t}\r\n\tdefer db.Close()\r\n\r\n\tdecoder := json.NewDecoder(r.Body)\r\n\tvar reqBody regWebReqBody\r\n\terr := decoder.Decode(&reqBody)\r\n\tif err != nil {\r\n\t\tpanic(err)\r\n\t}\r\n\r\n\tfor _, website := range reqBody.Websites {\r\n\r\n\t\tvar web Website\r\n\t\tif db.Where(\"URL = ?\", website.URL).First(&web).RecordNotFound() {\r\n\t\t\tdb.Create(&website)\r\n\t\t\t_ = setCron(website)\r\n\t\t}\r\n\t}\r\n\r\n\tfmt.Fprintf(w, \"Website(s) Successfully Created.\")\r\n}", "func Start(logDir string) error {\n\t// InitalBot then return 我嘗試在切開一點點看看\n\tslackAPI, lineAPI, telegramAPI := InitialBots()\n\n\t// Run echo web server\n\te := echo.New()\n\te.Debug = true\n\n\t// Middleware\n\te.Use(middleware.Logger())\n\te.Use(middleware.Recover())\n\n\t// setup access logger\n\tlogPath := filepath.Join(logDir, \"httpd.log\")\n\tif _, err := os.Stat(logPath); os.IsNotExist(err) {\n\t\tif _, err := os.Create(logPath); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tf, err := os.OpenFile(logPath, os.O_APPEND|os.O_WRONLY, 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\te.Use(middleware.LoggerWithConfig(middleware.LoggerConfig{\n\t\tFormat: \"[${time_rfc3339}] ${status} ${method} ${path} (${remote_ip}) ${latency_human}\\n\",\n\t\tOutput: f,\n\t}))\n\n\t// Routes\n\te.GET(\"/health\", health())\n\te.POST(\"/api/slack\", slackHandler(slackAPI, lineAPI, telegramAPI))\n\te.POST(\"/api/line\", lineHandler(slackAPI, lineAPI, telegramAPI))\n\te.POST(\"/api/telegram\", telegramHandler(slackAPI, lineAPI, telegramAPI))\n\n\tbindURL := fmt.Sprintf(\"%s:%s\", viper.GetString(\"bind\"), viper.GetString(\"port\"))\n\tlog.Infof(\"Listening on %s\", bindURL)\n\n\treturn e.Start(bindURL)\n}", "func SetUpService(webServiceEndpoint string,healthCheckEndpoint string,subscriptionServiceUrl string,googleSubscriptionsUrl string,clientId string, clientSecret string, callbackUrl string, issuer string, sessionKey string, cloudCommerceProcurementUrl string, partnerId string, finishUrl string, finishUrlTitle string, testMode string) error {\n\thandler := GetSubscriptionFrontendHandler(subscriptionServiceUrl,googleSubscriptionsUrl,clientId, clientSecret, callbackUrl, issuer, sessionKey, cloudCommerceProcurementUrl, partnerId, finishUrl, finishUrlTitle)\n\n\thealthCheck := mux.NewRouter()\n\thealthCheck.Methods(http.MethodGet).Path(\"/healthz\").HandlerFunc(handler.Healthz)\n\tgo http.ListenAndServe(\":\"+healthCheckEndpoint, healthCheck)\n\n\twebService := mux.NewRouter()\n\tif testModeBool,err := strconv.ParseBool(testMode); err==nil && testModeBool {\n\t\twebService.Methods(http.MethodGet).Path(\"/resetsaas\").HandlerFunc(handler.ResetSaas)\n\t\twebService.Methods(http.MethodGet).Path(\"/signupsaastest\").HandlerFunc(handler.SignupSaasTest)\n\t}\n\twebService.Methods(http.MethodGet).Path(\"/signupprod/{accountId}\").HandlerFunc(handler.SignupProd)\n\twebService.Methods(http.MethodPost).Path(\"/signupsaas\").HandlerFunc(handler.SignupSaas)\n\twebService.Methods(http.MethodGet).Path(\"/login\").HandlerFunc(handler.Auth0Login)\n\twebService.Methods(http.MethodGet).Path(\"/callback\").HandlerFunc(handler.Auth0Callback)\n\twebService.Methods(http.MethodPost).Path(\"/finishSaas\").HandlerFunc(handler.FinishSaas)\n\twebService.Methods(http.MethodPost).Path(\"/finishProd\").HandlerFunc(handler.FinishProd)\n\n\twebService.Methods(http.MethodGet).Path(\"/healthz\").HandlerFunc(handler.Healthz)\n\n\twebService.HandleFunc(\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\thttp.Redirect(w, r, \"https://www.cloudbees.com\", http.StatusFound)\n\t})\n\n\treturn http.ListenAndServe(\":\"+webServiceEndpoint, webService)\n}", "func Run() {\n\tgetRoutes()\n\trouter.Run(\":5000\")\n}", "func (server Server) Create_Open_Html() {\n\tlayernames := []string{}\n\tfor _,i := range server.Mbtiles {\n\t\tlayernames = append(layernames,Get_Vector_Layers(i)...)\n\t}\n\tfor _,i := range server.Geobufs {\n\t\tlayernames = append(layernames,i.Config_Dynamic.LayerName)\n\t}\n\n\tlayer_parts := []string{}\n\tfor _,layername := range layernames {\n\t\tlayer_parts = append(layer_parts,Get_Part_Layer(layername))\n\t}\n\tmiddle := strings.Join(layer_parts,\"\\n\")\n\tstart,end := Start_End()\n\n\ttotal := start + \"\\n\" + middle + \"\\n\" + end\n\n\tioutil.WriteFile(\"index.html\",[]byte(total),0677)\n\n\texec.Command(\"open\",\"index.html\").Run()\n\n}", "func SetUpDebate(server *gin.Engine, driver neo4j.Driver) {\n\t// -------soso: this is for profile page\n\tserver.GET(\"/allDebateAnswer/:email\", CheckAuthToken(func(c *gin.Context, _ string) {\n\t\temail := c.Param(\"email\")\n\t\tresult, err := queries.LoadDebateAnswers(driver, email)\n\t\tif err != nil {\n\t\t\tc.String(500, \"Internal server error\")\n\t\t\treturn\n\t\t} else if result == nil {\n\t\t\tc.String(404, \"Not found\")\n\t\t\treturn\n\t\t}\n\t\tc.JSON(200, result)\n\t}))\n\t// -------sosoend\n\n\tneo4jDriver2 = driver\n\n\t//fmt.Println(\"here1\")\n\n\tsetQuestionDBConstraint()\n\tsetUpQuestions()\n\n\t// endpoints\n\tserver.POST(\"/question\", getQuestion)\n\tserver.POST(\"/answer\", postAnswer)\n\tserver.POST(\"/answerExists\", checkAnswer)\n\tserver.POST(\"/getAnswer\", getAnswer)\n\tserver.POST(\"/getAnswers\", getAnswers)\n\tserver.POST(\"/getRating\", getRating)\n\tserver.POST(\"/addRating\", addRating)\n\tserver.POST(\"/getUsersRating\", getUsersRating)\n\n\tc := cron.New()\n\t// Job runs every day at 12am\n\t//c.AddFunc(\"19 17 * * *\", func() {\n\tc.AddFunc(\"0 0 * * *\", func() {\n\t\tcurrentTime := time.Now()\n\t\tday := currentTime.Day() - 1\n\t\t//day := currentTime.Day()\n\t\tdayString := strconv.Itoa(day)\n\t\tendDigitString := dayString[len(dayString)-1:]\n\t\tupdateACS(\"fanalyst\" + endDigitString)\n\t\tupdateACS(\"analyst\" + endDigitString)\n\t\tupdateACS(\"proanalyst\" + endDigitString)\n\t\tupdateACS(\"expertanalyst\" + endDigitString)\n\t})\n\tc.Start()\n}", "func trainSet(poolName string) {\n\t// MAKE SURE EMPTY POOL EXISTS\n\n\tvar urls []string\n\tsuffix := \"?api_key=hRnTyyGav36S6bxcJzhY37kQktPrHs33mmTUXjti&include_participants=1&include_matches=1\"\n\n\t// Non-prefixed ones\n\tfor i := 3; i < 10; i++ {\n\t\tfor j := 0; j < 10; j++ {\n\n\t\t\t// Special cases\n\t\t\tif i == 3 && j == 0 {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif i == 3 && j == 7 {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\turl := fmt.Sprintf(\"https://api.challonge.com/v1/tournaments/4qm%d%d%s\", i, j, suffix)\n\t\t\turls = append(urls, url)\n\t\t}\n\t}\n\n\t//100 - 106\n\tfor i := 0; i < 5; i++ {\n\t\turl := fmt.Sprintf(\"https://api.challonge.com/v1/tournaments/4qm10%d%s\", i, suffix)\n\t\turls = append(urls, url)\n\t}\n\n\t// Prefixed ones\n\tfor i := 0; i < 3; i++ {\n\t\tfor j := 0; j < 10; j++ {\n\n\t\t\tif i == 0 && j < 5 {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif i == 2 && j == 2 {\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\turl := fmt.Sprintf(\"https://api.challonge.com/v1/tournaments/4qs-4qm1%d%d%s\", i, j, suffix)\n\t\t\turls = append(urls, url)\n\t\t}\n\t}\n\n\tfor _, url := range urls {\n\t\tprintln(url)\n\t}\n\n\t// Calculate glicko for each tournament URL\n\tfor _, url := range urls {\n\t\tpool, err := readPool(poolName)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tprintln(\"CALCULATING \", url)\n\n\t\tglicko := calculateGlicko(getTournament(url), pool)\n\n\t\t// Merge new ranked players with ones that didn't participate\n\t\taddPlayer := true\n\t\tfor _, p := range pool.Players {\n\t\t\tfor j, gp := range glicko {\n\t\t\t\tif p.Name == gp.Name {\n\t\t\t\t\taddPlayer = false\n\t\t\t\t\tglicko[j].ResetAbsences() // Player attended\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif addPlayer {\n\t\t\t\tp.IncreaseAbsence()\n\t\t\t\tglicko = append(glicko, p)\n\t\t\t}\n\t\t\taddPlayer = true\n\t\t}\n\n\t\tpool.Players = glicko\n\t\tsort.Slice(pool.Players, func(i, j int) bool {\n\t\t\treturn pool.Players[i].Glicko.R > pool.Players[j].Glicko.R\n\t\t})\n\n\t\tj, err := json.Marshal(pool)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\n\t\t_, err = saveJson(j, poolName)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\n\t\tprintln(\"SUCCESS FOR \", url)\n\t}\n\n}", "func SetupWebserver() *echo.Echo {\n\te := echo.New()\n\te.Use(middleware.CORSWithConfig(middleware.CORSConfig{\n\t\tAllowOrigins: []string{\"*\"},\n\t\tAllowMethods: []string{echo.GET, echo.HEAD, echo.PUT, echo.PATCH, echo.POST, echo.DELETE},\n\t}), middleware.GzipWithConfig(middleware.GzipConfig{\n\t\tLevel: 1,\n\t}))\n\te.GET(\"/\", func(c echo.Context) error {\n\t\treturn c.HTML(200, string(\"Joseph Stock API Server\"))\n\t})\n\te.GET(\"/hsicomponents\", func(c echo.Context) error {\n\t\tstockSymbols, err := aastocks.GetHSIConstituentsCodes()\n\t\tif err == nil {\n\t\t\tresult := []string{}\n\t\t\tfor _, obj := range stockSymbols {\n\t\t\t\tresult = append(result, obj.Symbol)\n\t\t\t}\n\t\t\treturn c.JSON(200, result)\n\t\t}\n\t\treturn c.JSON(500, ErrorWrapper{err.Error()})\n\t})\n\te.GET(\"/hsicomponents/quote\", func(c echo.Context) error {\n\t\tq, err := CLIUtils.GetQuotesOfHSIComponents()\n\t\tif err == nil {\n\t\t\treturn c.JSON(200, q)\n\t\t}\n\t\treturn c.JSON(500, ErrorWrapper{err.Error()})\n\t})\n\te.GET(\"/quote/:symbol\", handlers.QuoteSymbolHandler)\n\te.GET(\"/history/:symbol\", handlers.SymbolHistoryHandler)\n\n\tsetupTradingviewAPI(e)\n\tsetupTradingviewUI(e)\n\n\treturn e\n}", "func TakeOver() {\n\tlis, err := net.Listen(\"tcp\", *address)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tlog.Println(\"listening on\", lis.Addr().String())\n\n\tif err := http.Serve(lis, nil); err != nil {\n\t\tpanic(err)\n\t}\n}", "func main() {\n\tflag.Parse()\n\n\tdebug := *flagDebug\n\n\tif debug {\n\t\tlog.Println(\"Starting in debug mode\")\n\t\tfile.Resolve(func() (string, error) {\n\t\t\treturn os.Getenv(\"TOBA_PATH\"), nil\n\t\t})\n\t}\n\n\tc, err := config.Load()\n\tweb.ExitIfError(err)\n\n\tc.HTTP.SyncFileAccess = debug\n\n\tif debug {\n\t\tc.HTTP.FromFolder = \"static\"\n\t}\n\n\tlicense, err := license.Load()\n\tweb.ExitIfError(err)\n\n\tif flagFiles != nil {\n\t\tfolder := *flagFiles\n\t\tif folder != \"\" {\n\t\t\tc.HTTP.FromFolder = folder\n\t\t}\n\t}\n\n\tweb.ExitIfError(err)\n\tweb.ExitIfError(db.Initialize(c.Database))\n\n\tif license.Valid {\n\t\tserveLicensed(c)\n\t} else {\n\t\tserveSetup(c)\n\t}\n\t// TODO: add timeout properties to server\n}", "func main() {\n\n\tgo memcache.MemCache()\n\tgo db.DBConnections()\n\n\te := echo.New()\n\n\te.Use(middleware.CORS())\n\te.Use(middleware.Secure())\n\te.Use(middleware.LoggerWithConfig(middleware.LoggerConfig{\n\t\tFormat: \"method=${method}, uri=${uri}, status=${status}\\n\",\n\t}))\n\n\te.GET(\"/\", func(c echo.Context) error {\n\t\treturn c.Redirect(http.StatusTemporaryRedirect, \"/a/index\")\n\t})\n\te.GET(\"/login\", func(c echo.Context) error {\n\t\treturn c.File(\"src/tabula-rasa/web/public/login.html\")\n\t})\n\te.POST(\"/signin\", resource.AccountResource{}.Login)\n\te.POST(\"signup\", resource.AccountResource{}.Register)\n\n\ta := e.Group(\"/a\")\n\ta.GET(\"/signout\", resource.AccountResource{}.Logout)\n\ta.Use(authMiddleware)\n\ta.Static(\"/static\", \"src/tabula-rasa/web\")\n\ta.GET(\"/index\", func(c echo.Context) error {\n\t\treturn c.File(\"src/tabula-rasa/web/public/index.html\")\n\t})\n\t// TODO add auth\n\tapi := e.Group(\"/rest\")\n\tapi.GET(\"/:name/:typ\", resource.TableResource{}.Get)\n\tapi.POST(\"/:name/:typ\", resource.TableResource{}.Save)\n\tapi.DELETE(\"/:name/:typ/:id\", resource.TableResource{}.Delete)\n\n\tif *sslcert != \"\" && *sslkey != \"\" {\n\t\te.Logger.Fatal(e.StartTLS(fmt.Sprintf(\":%v\", *port),\n\t\t\t*sslcert, *sslkey))\n\t} else {\n\t\te.Logger.Fatal(e.Start(fmt.Sprintf(\":%v\", *port)))\n\t}\n}", "func (rest *RestServer) Run() {\n\tgin.SetMode(gin.ReleaseMode)\n\trouter := gin.Default()\n\n\tbasicAuth := router.Group(\"/\", gin.BasicAuth(rest.User))\n\tbasicAuth.GET(\"/clients\", func(c *gin.Context) {\n\t\tpager := NewHttpPager(c.Request)\n\t\tprovider := new(SliceDataProvider)\n\t\tlist := rest.Srv.Monitor.Clients()\n\t\tprovider.Pager = pager\n\t\tprovider.SetModels(list)\n\t\tb := make(gmqtt.ClientList, 0)\n\t\tprovider.Models(&b)\n\t\tobj := PageAbleObj{\n\t\t\tModels: b,\n\t\t\tPage: pager.Page + 1,\n\t\t\tPageSize: pager.PageSize,\n\t\t\tCurrentCount: provider.Count(),\n\t\t\tTotalCount: pager.TotalCount,\n\t\t\tTotalPage: pager.PageCount(),\n\t\t}\n\t\tc.JSON(http.StatusOK, obj)\n\n\t})\n\tbasicAuth.GET(\"/client/:id\", func(c *gin.Context) {\n\n\t\tmodel, exist := rest.Srv.Monitor.GetClient(c.Param(\"id\"))\n\t\tif !exist {\n\t\t\tc.String(http.StatusNotFound, \"%s\", \"Client Not Found\")\n\t\t\treturn\n\t\t}\n\t\tc.JSON(http.StatusOK, model)\n\t})\n\tbasicAuth.GET(\"/sessions\", func(c *gin.Context) {\n\t\tpager := NewHttpPager(c.Request)\n\t\tprovider := new(SliceDataProvider)\n\t\tlist := rest.Srv.Monitor.Sessions()\n\t\tprovider.Pager = pager\n\t\tprovider.SetModels(list)\n\t\tb := make(gmqtt.SessionList, 0)\n\t\tprovider.Models(&b)\n\t\tobj := PageAbleObj{\n\t\t\tModels: b,\n\t\t\tPage: pager.Page + 1,\n\t\t\tPageSize: pager.PageSize,\n\t\t\tCurrentCount: provider.Count(),\n\t\t\tTotalCount: pager.TotalCount,\n\t\t\tTotalPage: pager.PageCount(),\n\t\t}\n\t\tc.JSON(http.StatusOK, obj)\n\t})\n\tbasicAuth.GET(\"/session/:id\", func(c *gin.Context) {\n\t\tmodel, ok := rest.Srv.Monitor.GetSession(c.Param(\"id\"))\n\t\tif !ok {\n\t\t\tc.String(http.StatusNotFound, \"%s\", \"Session Not Found\")\n\t\t\treturn\n\t\t}\n\t\tc.JSON(http.StatusOK, model)\n\t})\n\tbasicAuth.GET(\"/subscriptions/:id\", func(c *gin.Context) {\n\t\tpager := NewHttpPager(c.Request)\n\t\tprovider := new(SliceDataProvider)\n\t\tlist := rest.Srv.Monitor.ClientSubscriptions(c.Param(\"id\"))\n\t\tprovider.Pager = pager\n\t\tprovider.SetModels(list)\n\t\tb := make(gmqtt.SubscriptionList, 0)\n\t\tprovider.Models(&b)\n\t\tobj := PageAbleObj{\n\t\t\tModels: b,\n\t\t\tPage: pager.Page + 1,\n\t\t\tPageSize: pager.PageSize,\n\t\t\tCurrentCount: provider.Count(),\n\t\t\tTotalCount: pager.TotalCount,\n\t\t\tTotalPage: pager.PageCount(),\n\t\t}\n\t\tc.JSON(http.StatusOK, obj)\n\t})\n\tbasicAuth.GET(\"/subscriptions\", func(c *gin.Context) {\n\t\tpager := NewHttpPager(c.Request)\n\t\tprovider := new(SliceDataProvider)\n\t\tlist := rest.Srv.Monitor.Subscriptions()\n\t\tprovider.Pager = pager\n\t\tprovider.SetModels(list)\n\t\tb := make(gmqtt.SubscriptionList, 0)\n\t\tprovider.Models(&b)\n\t\tobj := PageAbleObj{\n\t\t\tModels: b,\n\t\t\tPage: pager.Page + 1,\n\t\t\tPageSize: pager.PageSize,\n\t\t\tCurrentCount: provider.Count(),\n\t\t\tTotalCount: pager.TotalCount,\n\t\t\tTotalPage: pager.PageCount(),\n\t\t}\n\t\tc.JSON(http.StatusOK, obj)\n\t})\n\tbasicAuth.POST(\"/subscribe\", func(c *gin.Context) {\n\t\tqosParam := c.PostForm(\"qos\")\n\t\ttopic := c.PostForm(\"topic\")\n\t\tcid := c.PostForm(\"clientID\")\n\t\tqos, err := strconv.Atoi(qosParam)\n\t\tif err != nil {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", packets.ErrInvalQos)\n\t\t\treturn\n\t\t}\n\t\tif qos < 0 || qos > 2 {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", packets.ErrInvalQos)\n\t\t\treturn\n\t\t}\n\n\t\tif !packets.ValidTopicFilter([]byte(topic)) {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", packets.ErrInvalTopicFilter)\n\t\t\treturn\n\t\t}\n\n\t\tif cid == \"\" {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", \"invalid clientID\")\n\t\t\treturn\n\t\t}\n\t\trest.Srv.Subscribe(cid, []packets.Topic{\n\t\t\t{Qos: uint8(qos), Name: topic},\n\t\t})\n\t\tc.JSON(http.StatusOK, OkResponse{\n\t\t\tCode: 0,\n\t\t\tResult: make([]interface{}, 0),\n\t\t})\n\t})\n\tbasicAuth.POST(\"/unsubscribe\", func(c *gin.Context) {\n\t\ttopic := c.PostForm(\"topic\")\n\t\tcid := c.PostForm(\"clientID\")\n\n\t\tif !packets.ValidTopicFilter([]byte(topic)) {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", packets.ErrInvalTopicFilter)\n\t\t\treturn\n\t\t}\n\n\t\tif cid == \"\" {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", \"invalid clientID\")\n\t\t\treturn\n\t\t}\n\t\trest.Srv.UnSubscribe(cid, []string{topic})\n\t\tc.JSON(http.StatusOK, OkResponse{\n\t\t\tCode: 0,\n\t\t\tResult: make([]interface{}, 0),\n\t\t})\n\t})\n\tbasicAuth.POST(\"/publish\", func(c *gin.Context) {\n\t\tqosParam := c.PostForm(\"qos\")\n\t\ttopic := c.PostForm(\"topic\")\n\t\tpayload := c.PostForm(\"payload\")\n\t\tqos, err := strconv.Atoi(qosParam)\n\t\tif err != nil {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", packets.ErrInvalQos)\n\t\t\treturn\n\t\t}\n\t\tif qos < 0 || qos > 2 {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", packets.ErrInvalQos)\n\t\t\treturn\n\t\t}\n\t\tif !packets.ValidTopicName([]byte(topic)) {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", packets.ErrInvalTopicName)\n\t\t\treturn\n\t\t}\n\t\tif !packets.ValidUTF8([]byte(payload)) {\n\t\t\tc.String(http.StatusBadRequest, \"%s\", packets.ErrInvalUtf8)\n\t\t\treturn\n\t\t}\n\t\tpub := &packets.Publish{\n\t\t\tQos: uint8(qos),\n\t\t\tTopicName: []byte(topic),\n\t\t\tPayload: []byte(payload),\n\t\t}\n\t\trest.Srv.Publish(pub)\n\t\tc.JSON(http.StatusOK, OkResponse{\n\t\t\tCode: 0,\n\t\t\tResult: make([]interface{}, 0),\n\t\t})\n\n\t})\n\n\trouter.Run(rest.Addr)\n\n}", "func (tc *textileClient) start(ctx context.Context, cfg config.Config) error {\n\ttc.cfg = cfg\n\tauth := common.Credentials{}\n\tvar opts []grpc.DialOption\n\n\topts = append(opts, grpc.WithInsecure())\n\topts = append(opts, grpc.WithPerRPCCredentials(auth))\n\n\tvar threads *threadsClient.Client\n\tvar buckets *bucketsClient.Client\n\tvar netc *nc.Client\n\n\t// by default it goes to local threads now\n\thost := \"127.0.0.1:3006\"\n\n\tlog.Debug(\"Creating buckets client in \" + host)\n\tif b, err := bucketsClient.NewClient(host, opts...); err != nil {\n\t\tcmd.Fatal(err)\n\t} else {\n\t\tbuckets = b\n\t}\n\n\tlog.Debug(\"Creating threads client in \" + host)\n\tif t, err := threadsClient.NewClient(host, opts...); err != nil {\n\t\tcmd.Fatal(err)\n\t} else {\n\t\tthreads = t\n\t}\n\n\tif n, err := nc.NewClient(host, opts...); err != nil {\n\t\tcmd.Fatal(err)\n\t} else {\n\t\tnetc = n\n\t}\n\n\ttc.bucketsClient = buckets\n\ttc.threads = threads\n\ttc.netc = netc\n\n\ttc.isRunning = true\n\n\t// Attempt to connect to the Hub\n\t_, err := tc.getHubCtx(ctx)\n\tif err != nil {\n\t\tlog.Error(\"Could not connect to Textile Hub. Starting in offline mode.\", err)\n\t} else {\n\t\ttc.isConnectedToHub = true\n\t}\n\n\ttc.Ready <- true\n\treturn nil\n}", "func Setup() *celerity.Server {\n\tdb, _ := gorm.Open(\"sqlite3\", \"./data.sqlite3\")\n\tdb.LogMode(true)\n\n\tdb.AutoMigrate(\n\t\tmodels.User{},\n\t\tmodels.Room{},\n\t\tmodels.Message{},\n\t\tmodels.Session{},\n\t)\n\n\tvar lobbyCount int\n\n\tdb.Model(&models.Room{}).Where(\"slug = ?\", \"lobby\").Count(&lobbyCount)\n\tif lobbyCount == 0 {\n\t\tdb.Create(&models.Room{\n\t\t\tSlug: \"lobby\",\n\t\t\tName: \"The Lobby\",\n\t\t\tTopic: \"Welcome to the lobby\",\n\t\t})\n\t}\n\n\th := Handler{db}\n\n\tserver := celerity.New()\n\n\tserver.Use(middleware.RequestLogger())\n\tserver.Pre(middleware.CORS())\n\n\tserver.Channel(\"chat\", \"/ws/:token\", h.SocketHandler)\n\n\tsecured := server.Scope(\"/\")\n\tsecured.Use(middleware.Aegis(middleware.AegisConfig{\n\t\tAdapter: AegisAdapter{db},\n\t}))\n\tsecured.POST(\"/rooms\", h.CreateRoom)\n\tsecured.GET(\"/rooms/:slug\", h.GetRoom)\n\tsecured.GET(\"/validate\", h.ValidateSession)\n\n\tserver.POST(\"/signup\", h.CreateUser)\n\tserver.POST(\"/login\", h.Login)\n\n\tserver.ServePath(\"/\", \"./frontend/build\")\n\tserver.ServeFile(\"*\", \"./frontend/build/index.html\")\n\n\treturn server\n}", "func main() {\n\n\tlog.SetOutput(os.Stdout)\n\n\ttoCrawl, _ := url.Parse(\"http://www.monzo.com\")\n\tvar filter crawler.Restriction = func(url *url.URL) bool {\n\t\treturn url.Host == toCrawl.Host\n\t}\n\tvar op1 crawler.Operation = func(in *url.URL) *url.URL {\n\t\tif in != nil {\n\t\t\thashIndex := strings.Index(in.String(), \"#\")\n\t\t\tif hashIndex > 0 {\n\t\t\t\tout, err := url.Parse(in.String()[:hashIndex])\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn in\n\t\t\t\t}\n\t\t\t\treturn out\n\t\t\t}\n\t\t}\n\t\treturn in\n\t}\n\n\twg := sync.WaitGroup{}\n\twg.Add(1)\n\tout := make(chan model.CrawlerOutput, 100)\n\tgo func() {\n\t\tdefer wg.Done()\n\t\tfor each := range out {\n\t\t\tconsumer.CreateSiteMap(path, each.URL, each.PageLinks, each.ResponseBody)\n\t\t}\n\t}()\n\tdone := make(chan struct{})\n\n\tc := crawler.NewCrawler(nil, crawler.Setting{\n\t\tRestrictions: []crawler.Restriction{filter},\n\t\tOperation: op1,\n\t\tWaitTimes: 100 * time.Millisecond,\n\t\tWorkers: 10,\n\t\tGetResponseBody: true,\n\t})\n\tgo c.Crawl(toCrawl, out, done)\n\n\tselect {\n\tcase <-time.After(10 * time.Second):\n\t\tdone <- struct{}{}\n\t}\n\twg.Wait()\n}", "func serveWs(hub *Hub, w http.ResponseWriter, r *http.Request) {\n\tconn, err := upgrader.Upgrade(w, r, nil)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn\n\t}\n\tclient := &Client{hub: hub, conn: conn, send: make(chan []byte, 256)}\n\tclient.hub.register <- client\n\tgo client.writePump()\n\tclient.readPump()\n}", "func (uplink *SenseUplink) Connect(ctx context.Context, url string, headers http.Header, cookieJar http.CookieJar, allowUntrusted bool, timeout time.Duration, reconnect bool) error {\n\tif uplink.Global != nil {\n\t\tuplink.Global.DisconnectFromServer()\n\t\tuplink.Global = nil\n\t}\n\n\tdialer := enigma.Dialer{\n\t\tMockMode: uplink.MockMode,\n\t\tInterceptors: []enigma.Interceptor{\n\t\t\t(&enigmainterceptors.MetricsHandler{\n\t\t\t\tLog: uplink.LogMetric,\n\t\t\t}).MetricsInterceptor,\n\t\t\tuplink.retryInterceptor,\n\t\t},\n\t\tTLSClientConfig: &tls.Config{\n\t\t\tInsecureSkipVerify: allowUntrusted,\n\t\t},\n\t}\n\tif cookieJar != nil {\n\t\tdialer.Jar = cookieJar\n\t}\n\tdialer.TrafficLogger = uplink.Traffic\n\n\tonUnexpectedDisconnect := func() {\n\t\tif helpers.IsContextTriggered(uplink.ctx) {\n\t\t\treturn\n\t\t}\n\t\tuplink.executeFailedConnectFuncs()\n\t}\n\n\tsetupDialer(&dialer, timeout, uplink.logEntry, onUnexpectedDisconnect, uplink.MaxFrameSize)\n\n\t// TODO somehow get better values for connect time\n\tstartTimestamp := time.Now()\n\tglobal, err := dialer.Dial(ctx, url, headers)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Error connecting to Sense\")\n\t}\n\tpostDialTimestamp := time.Now()\n\tuplink.Global = global\n\n\t// Setup\n\tconnectMsgChan := global.SessionMessageChannel(globals.EventTopics...)\n\tdefer global.CloseSessionMessageChannel(connectMsgChan)\n\n\tif err := uplink.trafficMetrics.Update(startTimestamp, postDialTimestamp, 0, 0); err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\n\tif uplink.MockMode {\n\t\treturn nil\n\t}\n\n\ttopicshandler := NewTopicsHandler(connectMsgChan)\n\ttopicshandler.Start(uplink.logEntry)\n\n\t// setup logging of traffic metrics for pushed events\n\tgo func() {\n\t\tsessionChan := global.SessionMessageChannel()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event, ok := <-sessionChan:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\t// Metrics not triggered for pushed, update metrics here\n\t\t\t\tif err := uplink.trafficMetrics.UpdateReceived(time.Now(), int64(len(event.Content))); err != nil {\n\t\t\t\t\tuplink.logEntry.LogError(err)\n\t\t\t\t}\n\t\t\tcase <-uplink.ctx.Done():\n\t\t\t\treturn\n\t\t\tcase <-global.Closed():\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\tselect {\n\tcase <-topicshandler.OnConnectedReceived:\n\tcase <-time.After(timeout):\n\t\treturn errors.Errorf(\"websocket connected, but no state created or attach (timeout)\")\n\tcase <-uplink.ctx.Done():\n\t\treturn errors.Errorf(\"websocket connected, but no state created or attach (context cancelled)\")\n\tcase <-global.Closed():\n\t\t// TODO currently there's no way to get close state, but subsequent EngineVersion will get the error message\n\t\t// once a method to get close state (requires changes to enigma-go) this should be checked and returned here instead\n\t\t// return errors.Errorf(\"websocket unexpectedly closed during connection attempt\")\n\t}\n\n\t// send a quick request, after this OnConnected and EventTopicOnAuthenticationInformation has been done and websocket possibly force closed\n\tversion, connectErr := global.EngineVersion(uplink.ctx)\n\n\t// By now topics should be received, first check topic errors, before errors on version message\n\tif err := topicshandler.IsErrorState(reconnect, uplink.logEntry); err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\n\tif connectErr != nil {\n\t\treturn errors.Wrap(connectErr, \"websocket connected, but got error on requesting version\")\n\t}\n\tuplink.logEntry.LogInfo(\"EngineVersion\", version.ComponentVersion)\n\n\treturn nil\n}", "func Run(env *Env) error {\n\n\te := echo.New()\n\te.SetDebug(true)\n\te.Use(mw.Logger())\n\te.Use(mw.Recover())\n\n\t// Render HTML\n\n\ttemplates, err := template.ParseGlob(filepath.Join(env.TemplateDir, \"*.tmpl\"))\n\tif err != nil {\n\t\treturn err\n\t}\n\te.SetRenderer(&renderer{templates})\n\n\t// add all the application globals we'll need\n\n\tauth := &defaultAuthenticator{}\n\tsession := newSession(env.Config)\n\n\te.Use(func(h echo.HandlerFunc) echo.HandlerFunc {\n\t\treturn func(c *echo.Context) error {\n\t\t\tc.Set(storeContextKey, env.Store)\n\t\t\tc.Set(mailerContextKey, env.Mailer)\n\t\t\tc.Set(feedparserContextKey, env.Feedparser)\n\t\t\tc.Set(cacheContextKey, env.Cache)\n\t\t\tc.Set(configContextKey, env.Config)\n\t\t\tc.Set(sessionContextKey, session)\n\t\t\tc.Set(authenticatorContextKey, auth)\n\t\t\treturn h(c)\n\t\t}\n\t})\n\n\t// catch sql no rows errors and return as a 404\n\te.Use(func(h echo.HandlerFunc) echo.HandlerFunc {\n\t\treturn func(c *echo.Context) error {\n\t\t\terr := h(c)\n\t\t\tif err == store.ErrNoRows {\n\t\t\t\treturn echo.NewHTTPError(http.StatusNotFound)\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t})\n\n\t// static configuration\n\te.Static(env.StaticURL, env.StaticDir)\n\n\tconfigureRoutes(e)\n\n\t// add CSRF protection\n\ts := e.Server(fmt.Sprintf(\":%v\", env.Port))\n\ts.Handler = nosurf.NewPure(e)\n\thttp2.ConfigureServer(s, nil)\n\treturn s.ListenAndServe()\n\n}", "func Run(){\n\tRestApiImplPtr.RegisterApi()\n\tRestServerImplPtr.StartRestServer();\n}", "func RunDev(plugin plugin.YomoObjectPlugin, endpoint string) {\n\n\tgo func() {\n\t\tlog.SetPrefix(fmt.Sprintf(\"[%s:%v]\", plugin.Name(), os.Getpid()))\n\t\tlog.Printf(\"plugin service start... [%s]\", endpoint)\n\n\t\t// binding plugin\n\t\tpluginStream := framework.NewObjectPlugin(plugin)\n\n\t\t// decoding\n\t\tdeStream1 := txtkv.NewObjectDecoder(plugin.Observed())\n\n\t\t//过滤\n\t\tdeStream2 := txtkv.NewFilterDecoder(plugin.Observed())\n\n\t\t// encoding\n\t\tenStream := txtkv.NewObjectEncoder(plugin.Observed())\n\n\t\tdeStream := io.MultiWriter(deStream1.Writer, deStream2.Writer)\n\n\t\tgo func() { io.CopyN(pluginStream.Writer, deStream1.Reader, 1024) }() // nolint\n\t\tgo func() { io.CopyN(enStream.Writer, pluginStream.Reader, 1024) }() // nolint\n\t\tgo func() { io.CopyN(enStream.Writer, deStream2.Reader, 1024) }() // nolint\n\n\t\t// activation service\n\t\tframework.NewServer(endpoint, deStream, enStream.Reader)\n\t}()\n\n\tyomoEchoClient, err := util.QuicClient(\"echo.cella.fun:11521\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tyomoPluginClient, err := util.QuicClient(endpoint)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tgo io.Copy(yomoPluginClient, yomoEchoClient)\n\tgo io.Copy(os.Stdout, yomoPluginClient)\n\n\tfor {\n\t\ttime.Sleep(time.Second)\n\t\tyomoEchoClient.Write([]byte(\"ping\"))\n\t}\n\n}", "func New() (*SingleCrawler, error) {\n\n defer glog.Flush()\n\n var crawler SingleCrawler\n startURL := *UrlPtr\n maxp := *MaxpPtr\n maxc := *MaxcPtr\n maxt := *MaxtPtr\n Filename := *OutfilePtr\n NumWorkers := *NumwPtr\n\n // validate the user input URL and decide if it's okay to use\n if govalidator.IsURL(startURL) == false {\n glog.Error(\"The starting URL is invalid. Please enter a valid URL.\")\n return nil, errors.New(\"Bad starting URL.\")\n }\n if maxp < 0 || maxc < 0 || maxt < 0 {\n glog.Error(\"Please pass in values > = 0 for max constraints (max print, max pages, max time). Please pass > 0 for the number of workers.\")\n return nil, errors.New(\"Bad values for maxprint, maxpages, maxtime, or NumWorkers\")\n }\n if NumWorkers <= 0 || NumWorkers > MAX_WORKERS {\n glog.Error(\"Number of workes is invalid. Must be > 0, and less that MAX_WORKERS.\")\n return nil, errors.New(\"Bad value for NumWorkers\")\n }\n if len(Filename) >= 255 {\n glog.Error(\"Filename can't be larger than 255 characters. Trimming Filename.\")\n Filename = Filename[0:100]\n }\n\n\n crawler.MAX_PAGES = maxc\n crawler.PRINT_LIMIT = maxp\n crawler.NumPages = 0\n crawler.NumWorkers = NumWorkers\n crawler.MAX_TIME = time.Duration(maxt) * time.Second\n crawler.Sitemap = make( [] Page, crawler.MAX_PAGES)\n \n\n // Parse the URL - make sure it's ok to use\n domain, err := url.Parse(startURL)\n if err != nil {\n glog.Error(\"Error parsing domain of starting URL\")\n return nil, errors.New(\"Unable to parse domain of start URL.\")\n }\n err = DomainCheck( domain )\n if err != nil {\n glog.Error(\"Error parsing domain of starting URL\")\n return nil, err\n }\n crawler.Site = domain\n \n if Filename != \"\" {\n crawler.Filename = Filename\n } else {\n crawler.Filename = crawler.Site.Host + \".txt\"\n if len( crawler.Filename ) >= 255 {\n crawler.Filename = crawler.Filename[0:100]\n }\n }\n\n if err = IsOk( &crawler ); err!=nil{\n return nil, err\n }\n\n return &crawler, nil\n\n}", "func homePage(w http.ResponseWriter, r *http.Request) {\n\tnCalls += 1\n\tfmt.Fprintf(w, \"Welcome to the tut.rest REST website!\")\n\tfmt.Printf(\"Endpoint(%3v): REST website\\n\", nCalls)\n}", "func (s *Source) Run(ctx context.Context, domain string, session *subscraping.Session) <-chan subscraping.Result {\n\tresults := make(chan subscraping.Result)\n\n\tgo func() {\n\t\tdefer close(results)\n\n\t\tif session.Keys.Spyse == \"\" {\n\t\t\treturn\n\t\t}\n\n\t\tclient, err := spyse.NewClient(session.Keys.Spyse, nil)\n\t\tif err != nil {\n\t\t\tresults <- subscraping.Result{Source: s.Name(), Type: subscraping.Error, Error: err}\n\t\t\treturn\n\t\t}\n\n\t\tdomainSvc := spyse.NewDomainService(client)\n\n\t\tvar searchDomain = \".\" + domain\n\t\tvar subdomainsSearchParams spyse.QueryBuilder\n\n\t\tsubdomainsSearchParams.AppendParam(spyse.QueryParam{\n\t\t\tName: domainSvc.Params().Name.Name,\n\t\t\tOperator: domainSvc.Params().Name.Operator.EndsWith,\n\t\t\tValue: searchDomain,\n\t\t})\n\n\t\ttotalResults, err := domainSvc.SearchCount(ctx, subdomainsSearchParams.Query)\n\t\tif err != nil {\n\t\t\tresults <- subscraping.Result{Source: s.Name(), Type: subscraping.Error, Error: err}\n\t\t\treturn\n\t\t}\n\n\t\tif totalResults == 0 {\n\t\t\treturn\n\t\t}\n\n\t\t// The default \"Search\" method returns only first 10 000 subdomains\n\t\t// To obtain more than 10 000 subdomains the \"Scroll\" method should be using\n\t\t// Note: The \"Scroll\" method is only available for \"PRO\" customers, so we need to check\n\t\t// quota.IsScrollSearchEnabled param\n\t\tif totalResults > searchMethodResultsLimit && client.Account().IsScrollSearchEnabled {\n\t\t\tvar scrollID string\n\t\t\tvar scrollResults *spyse.DomainScrollResponse\n\n\t\t\tfor {\n\t\t\t\tscrollResults, err = domainSvc.ScrollSearch(ctx, subdomainsSearchParams.Query, scrollID)\n\t\t\t\tif err != nil {\n\t\t\t\t\tresults <- subscraping.Result{Source: s.Name(), Type: subscraping.Error, Error: err}\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif len(scrollResults.Items) > 0 {\n\t\t\t\t\tscrollID = scrollResults.SearchID\n\n\t\t\t\t\tfor i := range scrollResults.Items {\n\t\t\t\t\t\tresults <- subscraping.Result{\n\t\t\t\t\t\t\tSource: s.Name(),\n\t\t\t\t\t\t\tType: subscraping.Subdomain,\n\t\t\t\t\t\t\tValue: scrollResults.Items[i].Name,\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tvar limit = 100\n\t\t\tvar searchResults []spyse.Domain\n\n\t\t\tfor offset := 0; int64(offset) < totalResults && int64(offset) < searchMethodResultsLimit; offset += limit {\n\t\t\t\tsearchResults, err = domainSvc.Search(ctx, subdomainsSearchParams.Query, limit, offset)\n\t\t\t\tif err != nil {\n\t\t\t\t\tresults <- subscraping.Result{Source: s.Name(), Type: subscraping.Error, Error: err}\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tfor i := range searchResults {\n\t\t\t\t\tresults <- subscraping.Result{\n\t\t\t\t\t\tSource: s.Name(),\n\t\t\t\t\t\tType: subscraping.Subdomain,\n\t\t\t\t\t\tValue: searchResults[i].Name,\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn results\n}", "func New(db database.Database, st storage.Storage, s *discordgo.Session,\n\tcmd shireikan.Handler, lct *lctimer.LifeCycleTimer, config *config.Config,\n\tpmw *middleware.PermissionsMiddleware, ota *onetimeauth.OneTimeAuth) (ws *WebServer, err error) {\n\n\tws = new(WebServer)\n\n\tif !strings.HasPrefix(config.WebServer.PublicAddr, \"http\") {\n\t\tprotocol := \"http\"\n\t\tif config.WebServer.TLS != nil && config.WebServer.TLS.Enabled {\n\t\t\tprotocol += \"s\"\n\t\t}\n\t\tconfig.WebServer.PublicAddr = fmt.Sprintf(\"%s://%s\", protocol, config.WebServer.PublicAddr)\n\t}\n\n\tif config.WebServer.APITokenKey == \"\" {\n\t\tconfig.WebServer.APITokenKey, err = random.GetRandBase64Str(32)\n\t} else if len(config.WebServer.APITokenKey) < 32 {\n\t\terr = errors.New(\"APITokenKey must have at leats a length of 32 characters\")\n\t}\n\tif err != nil {\n\t\treturn\n\t}\n\n\tws.config = config\n\tws.db = db\n\tws.st = st\n\tws.session = s\n\tws.cmdhandler = cmd\n\tws.pmw = pmw\n\tws.ota = ota\n\tws.rlm = NewRateLimitManager()\n\tws.af = NewAntiForgery()\n\tws.router = routing.New()\n\tws.server = &fasthttp.Server{\n\t\tHandler: ws.router.HandleRequest,\n\t}\n\n\tws.auth, err = NewAuth(db, s, lct, []byte(config.WebServer.APITokenKey))\n\tif err != nil {\n\t\treturn\n\t}\n\n\tws.dcoauth = discordoauth.NewDiscordOAuth(\n\t\tconfig.Discord.ClientID,\n\t\tconfig.Discord.ClientSecret,\n\t\tconfig.WebServer.PublicAddr+endpointAuthCB,\n\t\tws.auth.LoginFailedHandler,\n\t\tws.auth.LoginSuccessHandler,\n\t)\n\n\tws.registerHandlers()\n\n\treturn\n}", "func (p *FactoidPlugin) registerWeb() {\n\tr := chi.NewRouter()\n\tr.HandleFunc(\"/api\", p.serveAPI)\n\tr.HandleFunc(\"/req\", p.serveQuery)\n\tr.HandleFunc(\"/\", p.serveQuery)\n\tp.b.RegisterWebName(r, \"/factoid\", \"Factoid\")\n}", "func New(s *service.Service) http.Handler {\n h := &handler{s}\n api := way.NewRouter()\n api.HandleFunc(\"POST\", \"/login\", h.login)\n api.HandleFunc(\"POST\", \"/send_magic_link\", h.sendMagicLink)\n api.HandleFunc(\"GET\", \"/auth_redirect\", h.authRedirect)\n api.HandleFunc(\"GET\", \"/user\", h.authUser)\n api.HandleFunc(\"POST\", \"/users/:username/toggle_follow\", h.toggleFollow)\n api.HandleFunc(\"PUT\", \"/user/avatar\", h.updateAvatar)\n api.HandleFunc(\"POST\", \"/users\", h.createUser)\n api.HandleFunc(\"GET\", \"/users\", h.users)\n api.HandleFunc(\"GET\", \"/users/:username\", h.user)\n api.HandleFunc(\"GET\", \"/users/:username/followers\", h.followers)\n api.HandleFunc(\"GET\", \"/users/:username/posts\", h.posts)\n api.HandleFunc(\"GET\", \"/users/:username/followees\", h.followees)\n\n api.HandleFunc(\"POST\", \"/posts\", h.createPost)\n api.HandleFunc(\"GET\", \"/posts/:post_id\", h.post)\n api.HandleFunc(\"POST\", \"/posts/:post_id/toggle_like\", h.togglePostLike)\n api.HandleFunc(\"POST\", \"/posts/:post_id/comments\", h.createComment)\n api.HandleFunc(\"GET\", \"/posts/:post_id/comments\", h.comments)\n\n api.HandleFunc(\"POST\", \"/comments/:comment_id/toggle_like\", h.toggleCommentLike)\n api.HandleFunc(\"GET\", \"/timeline\", h.timeline)\n api.HandleFunc(\"POST\", \"/posts/:post_id/toggle_subscription\", h.togglePostSubscription)\n\n api.HandleFunc(\"GET\", \"/notifications\", h.notifications)\n api.HandleFunc(\"POST\", \"/notifications/:notification_id/mark_as_read\", h.markNotificationAsRead)\n api.HandleFunc(\"POST\", \"/mark_notifications_as_read\", h.markAllNotificationsAsRead)\n\n fs := http.FileServer(&spaFileSystem{http.Dir(\"public\")})\n r := way.NewRouter()\n r.Handle(\"*\", \"/api...\", http.StripPrefix(\"/api\", h.withAuth(api)))\n r.Handle(\"GET\", \"/...\", fs)\n return r\n}" ]
[ "0.61486995", "0.58042717", "0.55214345", "0.5442607", "0.5282036", "0.52660775", "0.52590233", "0.52099", "0.51979893", "0.5196904", "0.518868", "0.5142204", "0.5105513", "0.5049837", "0.50302064", "0.50073564", "0.4975622", "0.49714735", "0.49605906", "0.49463308", "0.49382287", "0.49377906", "0.49364895", "0.493464", "0.4926555", "0.49252194", "0.4907649", "0.4901659", "0.489885", "0.48968387", "0.48735845", "0.48690662", "0.48668924", "0.48632184", "0.4849236", "0.48486802", "0.48233953", "0.48233625", "0.48097938", "0.4809483", "0.47931594", "0.47919855", "0.47896102", "0.47596127", "0.47594744", "0.47529066", "0.4730679", "0.47296733", "0.47174272", "0.4716456", "0.47134173", "0.4711399", "0.46984679", "0.4698113", "0.46945608", "0.46942487", "0.46911088", "0.46883604", "0.46808922", "0.4679529", "0.46754795", "0.46747065", "0.4674124", "0.4672648", "0.46707976", "0.4669044", "0.46589127", "0.46565682", "0.46464983", "0.46441472", "0.46416906", "0.46383816", "0.46316317", "0.46307275", "0.46231422", "0.4622489", "0.46204865", "0.46135974", "0.46116894", "0.46098536", "0.4608271", "0.46074677", "0.4606928", "0.46057045", "0.46038932", "0.46021137", "0.45926812", "0.45909476", "0.4588517", "0.45842263", "0.45833203", "0.4583073", "0.45773098", "0.45767677", "0.45753926", "0.4572902", "0.4572425", "0.45712", "0.45708993", "0.45673114", "0.45669323" ]
0.0
-1
InitImplDefaults initializes safe defaults for all fields in Spitter.
func (spitterImpl *SpitterImpl) InitImplDefaults() { spitterImpl.SpiderlingImpl.InitImplDefaults() spitterImpl.spittingWebToNestImpl = nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (jobImpl *JobImpl) InitImplDefaults() {\n\tjobImpl.GameObjectImpl.InitImplDefaults()\n\n\tjobImpl.carryLimitImpl = 0\n\tjobImpl.damageImpl = 0\n\tjobImpl.energyImpl = 0\n\tjobImpl.movesImpl = 0\n\tjobImpl.rangeImpl = 0\n\tjobImpl.shieldImpl = 0\n\tjobImpl.titleImpl = \"\"\n\tjobImpl.unitCostImpl = 0\n}", "func (unitImpl *UnitImpl) InitImplDefaults() {\n\tunitImpl.GameObjectImpl.InitImplDefaults()\n\n\tunitImpl.actedImpl = true\n\tunitImpl.dashXImpl = 0\n\tunitImpl.dashYImpl = 0\n\tunitImpl.energyImpl = 0\n\tunitImpl.genariumImpl = 0\n\tunitImpl.isBusyImpl = true\n\tunitImpl.jobImpl = nil\n\tunitImpl.legendariumImpl = 0\n\tunitImpl.movesImpl = 0\n\tunitImpl.mythiciteImpl = 0\n\tunitImpl.ownerImpl = nil\n\tunitImpl.protectorImpl = nil\n\tunitImpl.rariumImpl = 0\n\tunitImpl.shieldImpl = 0\n\tunitImpl.xImpl = 0\n\tunitImpl.yImpl = 0\n}", "func (d *Default) Init(params map[interface{}]interface{}) error {\n\treturn nil\n}", "func init() {\n\tResetDefaults()\n}", "func (scl *SimpleConfigurationLayer) setDefaults() {\n\tscl.Changelog = ent.NewChangelogConfiguration()\n\tscl.CommitMessageConventions = ent.NewCommitMessageConventions()\n\tscl.Git = ent.NewGitConfiguration()\n\tsvra := make(map[string]*ent.Attachment)\n\tscl.ReleaseAssets = &svra\n\tscl.ReleaseTypes = ent.NewReleaseTypes()\n\tsvsc := make(map[string]*ent.ServiceConfiguration)\n\tscl.Services = &svsc\n\tscl.Substitutions = ent.NewSubstitutions()\n}", "func init() {\n\tviper.SetDefault(\"attestation-type\", constants.DefaultAttestationType)\n\tviper.SetDefault(\"poll-interval-minutes\", constants.PollingIntervalMinutes)\n\n\t//Set default values for TLS\n\tviper.SetDefault(\"tls-cert-file\", constants.DefaultTLSCertFile)\n\tviper.SetDefault(\"tls-key-file\", constants.DefaultTLSKeyFile)\n\tviper.SetDefault(\"tls-common-name\", constants.DefaultIHUBTlsCn)\n\tviper.SetDefault(\"tls-san-list\", constants.DefaultTLSSan)\n\n\t//Set default values for log\n\tviper.SetDefault(\"log-max-length\", constants.DefaultLogEntryMaxlength)\n\tviper.SetDefault(\"log-enable-stdout\", true)\n\tviper.SetDefault(\"log-level\", \"info\")\n}", "func (cfg *SupervisorConfig) InitDefaults() {\n\tif cfg.WatchTick == 0 {\n\t\tcfg.WatchTick = time.Second\n\t}\n}", "func (d *Default) Init() error {\n\treturn nil\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescUsername is the schema descriptor for username field.\n\tuserDescUsername := userFields[0].Descriptor()\n\t// user.DefaultUsername holds the default value on creation for the username field.\n\tuser.DefaultUsername = userDescUsername.Default.(string)\n\t// userDescName is the schema descriptor for name field.\n\tuserDescName := userFields[1].Descriptor()\n\t// user.DefaultName holds the default value on creation for the name field.\n\tuser.DefaultName = userDescName.Default.(string)\n\t// userDescSurname is the schema descriptor for surname field.\n\tuserDescSurname := userFields[2].Descriptor()\n\t// user.DefaultSurname holds the default value on creation for the surname field.\n\tuser.DefaultSurname = userDescSurname.Default.(string)\n}", "func (o *GetComponentByIDParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (c *AuthConfig) init() {\n\tif c.Provisioners == nil {\n\t\tc.Provisioners = provisioner.List{}\n\t}\n\tif c.Template == nil {\n\t\tc.Template = &ASN1DN{}\n\t}\n\tif c.Backdate == nil {\n\t\tc.Backdate = &provisioner.Duration{\n\t\t\tDuration: DefaultBackdate,\n\t\t}\n\t}\n}", "func InitDefault(o interface{}) {\n\tt := reflect.TypeOf(o).Elem()\n\tv := reflect.ValueOf(o).Elem()\n\n\tfieldCount := t.NumField()\n\n\tfor i := 0; i < fieldCount; i++ {\n\t\tfield := t.Field(i)\n\n\t\tif v.Field(i).Kind() == reflect.Struct {\n\t\t\tInitDefault(v.Field(i).Addr().Interface())\n\t\t\tcontinue\n\t\t}\n\n\t\tif defaultValue, ok := field.Tag.Lookup(\"default\"); ok {\n\n\t\t\tswitch defaultValue {\n\t\t\tcase \"UNSETFLOAT\":\n\t\t\t\tv.Field(i).SetFloat(UNSETFLOAT)\n\t\t\tcase \"UNSETINT\":\n\t\t\t\tv.Field(i).SetInt(UNSETINT)\n\t\t\tcase \"-1\":\n\t\t\t\tv.Field(i).SetInt(-1)\n\t\t\tcase \"true\":\n\t\t\t\tv.Field(i).SetBool(true)\n\t\t\tdefault:\n\t\t\t\tlog.Panic(\"Unknown defaultValue\", zap.Reflect(\"default\", v))\n\t\t\t}\n\t\t}\n\n\t}\n}", "func (cfg *Config) InitDefaults() {\n\tif cfg.NumWorkers == 0 {\n\t\tcfg.NumWorkers = uint64(runtime.NumCPU())\n\t}\n\n\tif cfg.AllocateTimeout == 0 {\n\t\tcfg.AllocateTimeout = time.Minute\n\t}\n\n\tif cfg.DestroyTimeout == 0 {\n\t\tcfg.DestroyTimeout = time.Minute\n\t}\n\tif cfg.Supervisor == nil {\n\t\treturn\n\t}\n\tcfg.Supervisor.InitDefaults()\n}", "func SetDefaults(cfg Provider) {\n\tcfg.SetTypeByDefaultValue(true)\n\n\tcfg.SetDefault(\"hmac.secret\", \"\")\n\n\t// Bangs\n\tcfg.SetDefault(\"bangs.path\", \"../bangs\")\n\n\t// Brand\n\tcfg.SetDefault(\"brand.name\", \"Jive Search\")\n\tcfg.SetDefault(\"brand.tagline\", \"A search engine that doesn't track you.\")\n\tcfg.SetDefault(\"brand.logo\", \"\")\n\tcfg.SetDefault(\"brand.small_logo\",\n\t\t`<svg xmlns=\"http://www.w3.org/2000/svg\" width=\"115px\" height=\"48px\">\n\t\t\t<defs>\n\t\t\t\t<style>\n\t\t\t\t\t#logo{\n\t\t\t\t\t\tfont-size:20px;\n\t\t\t\t\t} \n\t\t\t\t</style>\n\t\t\t</defs>\n\t\t\t<g>\n\t\t\t\t<text id=\"logo\" x=\"0\" y=\"37\" fill=\"#000\">Jive Search</text>\n\t\t\t</g>\n\t\t</svg>`)\n\n\t// Server\n\tport := 8000\n\tcfg.SetDefault(\"server.host\", fmt.Sprintf(\"http://127.0.0.1:%d\", port))\n\n\t// Frontend Cache\n\tcfg.SetDefault(\"cache.instant\", 1*time.Second)\n\tcfg.SetDefault(\"cache.search\", 1*time.Second)\n\n\t// languages are in the order of preference\n\t// empty slice = all languages\n\t// Note: the crawler and frontend packages (for now) don't support language config yet.\n\t// See note in search/document/document.go\n\tcfg.SetDefault(\"languages\", []string{}) // e.g. JIVESEARCH_LANGUAGES=\"en fr de\"\n\n\t// Elasticsearch\n\tcfg.SetDefault(\"elasticsearch.url\", \"http://127.0.0.1:9200\")\n\tcfg.SetDefault(\"elasticsearch.search.index\", \"test-search\")\n\tcfg.SetDefault(\"elasticsearch.search.type\", \"document\")\n\n\tcfg.SetDefault(\"elasticsearch.bangs.index\", \"test-bangs\")\n\tcfg.SetDefault(\"elasticsearch.bangs.type\", \"bang\")\n\n\tcfg.SetDefault(\"elasticsearch.image.index\", \"test-images\")\n\tcfg.SetDefault(\"elasticsearch.image.type\", \"image\")\n\n\tcfg.SetDefault(\"elasticsearch.query.index\", \"test-queries\")\n\tcfg.SetDefault(\"elasticsearch.query.type\", \"query\")\n\n\tcfg.SetDefault(\"elasticsearch.robots.index\", \"test-robots\")\n\tcfg.SetDefault(\"elasticsearch.robots.type\", \"robots\")\n\n\t// PostgreSQL\n\t// Note: there is a security concern if postgres password is stored in env variable\n\t// but setting it as an env var w/in systemd nullifies this.\n\tcfg.SetDefault(\"postgresql.host\", \"localhost\")\n\tcfg.SetDefault(\"postgresql.user\", \"jivesearch\")\n\tcfg.SetDefault(\"postgresql.password\", \"mypassword\")\n\tcfg.SetDefault(\"postgresql.database\", \"jivesearch\")\n\n\t// Redis\n\tcfg.SetDefault(\"redis.host\", \"\")\n\tcfg.SetDefault(\"redis.port\", 6379)\n\n\t// crawler defaults\n\ttme := 5 * time.Minute\n\tcfg.SetDefault(\"crawler.useragent.full\", \"https://github.com/jivesearch/jivesearch\")\n\tcfg.SetDefault(\"crawler.useragent.short\", \"jivesearchbot\")\n\tcfg.SetDefault(\"crawler.time\", tme.String())\n\tcfg.SetDefault(\"crawler.since\", 30*24*time.Hour)\n\tcfg.SetDefault(\"crawler.seeds\", []string{\n\t\t\"https://moz.com/top500/domains\",\n\t\t\"https://domainpunch.com/tlds/topm.php\",\n\t\t\"https://www.wikipedia.org/\",\n\t},\n\t)\n\n\tworkers := 100\n\tcfg.SetDefault(\"crawler.workers\", workers)\n\tcfg.SetDefault(\"crawler.max.bytes\", 1024000) // 1MB...too little? too much??? Rememer <script> tags can take up a lot of bytes.\n\tcfg.SetDefault(\"crawler.timeout\", 25*time.Second)\n\tcfg.SetDefault(\"crawler.max.queue.links\", 100000)\n\tcfg.SetDefault(\"crawler.max.links\", 100)\n\tcfg.SetDefault(\"crawler.max.domain.links\", 10000)\n\tcfg.SetDefault(\"crawler.truncate.title\", 100)\n\tcfg.SetDefault(\"crawler.truncate.keywords\", 25)\n\tcfg.SetDefault(\"crawler.truncate.description\", 250)\n\n\t// image nsfw scoring and metadata\n\tcfg.SetDefault(\"nsfw.host\", \"http://127.0.0.1:8080\")\n\tcfg.SetDefault(\"nsfw.workers\", 10)\n\tcfg.SetDefault(\"nsfw.since\", now().AddDate(0, -1, 0))\n\n\t// ProPublica API\n\tcfg.SetDefault(\"propublica.key\", \"my_key\")\n\n\t// useragent for fetching api's, images, etc.\n\tcfg.SetDefault(\"useragent\", \"https://github.com/jivesearch/jivesearch\")\n\n\t// stackoverflow API settings\n\tcfg.SetDefault(\"stackoverflow.key\", \"app key\") // https://stackapps.com/apps/oauth/\n\n\t// FedEx package tracking API settings\n\tcfg.SetDefault(\"fedex.account\", \"account\")\n\tcfg.SetDefault(\"fedex.password\", \"password\")\n\tcfg.SetDefault(\"fedex.key\", \"key\")\n\tcfg.SetDefault(\"fedex.meter\", \"meter\")\n\n\t// Maps\n\tcfg.SetDefault(\"mapbox.key\", \"key\")\n\n\t// MaxMind geolocation DB\n\tcfg.SetDefault(\"maxmind.database\", \"/usr/share/GeoIP/GeoLite2-City.mmdb\")\n\n\t// Search Providers\n\tcfg.SetDefault(\"yandex.key\", \"key\")\n\tcfg.SetDefault(\"yandex.user\", \"user\")\n\n\t// UPS package tracking API settings\n\tcfg.SetDefault(\"ups.user\", \"user\")\n\tcfg.SetDefault(\"ups.password\", \"password\")\n\tcfg.SetDefault(\"ups.key\", \"key\")\n\n\t// USPS package tracking API settings\n\tcfg.SetDefault(\"usps.user\", \"user\")\n\tcfg.SetDefault(\"usps.password\", \"password\")\n\n\t// OpenWeatherMap API settings\n\tcfg.SetDefault(\"openweathermap.key\", \"key\")\n\n\t// wikipedia settings\n\ttruncate := 250\n\tcfg.SetDefault(\"wikipedia.truncate\", truncate) // chars\n\n\t// command flags\n\tcmd := cobra.Command{}\n\tcmd.Flags().Int(\"workers\", workers, \"number of workers\")\n\tcfg.BindPFlag(\"crawler.workers\", cmd.Flags().Lookup(\"workers\"))\n\tcmd.Flags().Duration(\"time\", tme, \"duration the crawler should run\")\n\tcfg.BindPFlag(\"crawler.time\", cmd.Flags().Lookup(\"time\"))\n\n\tcmd.Flags().Int(\"port\", port, \"server port\")\n\tcfg.BindPFlag(\"frontend.port\", cmd.Flags().Lookup(\"port\"))\n\n\t// control debug output\n\tcmd.Flags().Bool(\"debug\", false, \"turn on debug output\")\n\tcfg.BindPFlag(\"debug\", cmd.Flags().Lookup(\"debug\"))\n\n\t// change search provider\n\tcmd.Flags().String(\"provider\", \"\", \"choose search provider\")\n\tcfg.BindPFlag(\"search.provider\", cmd.Flags().Lookup(\"provider\"))\n\n\t// wikipedia dump file settings\n\tcmd.Flags().String(\"dir\", \"\", \"path to save wiki dump files\")\n\tcfg.BindPFlag(\"wikipedia.dir\", cmd.Flags().Lookup(\"dir\"))\n\n\tcmd.Flags().Bool(\"wikidata\", true, \"include wikidata\")\n\tcfg.BindPFlag(\"wikipedia.wikidata\", cmd.Flags().Lookup(\"wikidata\"))\n\n\tcmd.Flags().Bool(\"wikipedia\", true, \"include wikipedia\")\n\tcfg.BindPFlag(\"wikipedia.wikipedia\", cmd.Flags().Lookup(\"wikipedia\"))\n\n\tcmd.Flags().Bool(\"wikiquote\", true, \"include wikiquote\")\n\tcfg.BindPFlag(\"wikipedia.wikiquote\", cmd.Flags().Lookup(\"wikiquote\"))\n\n\tcmd.Flags().Bool(\"wiktionary\", true, \"include wiktionary\")\n\tcfg.BindPFlag(\"wikipedia.wiktionary\", cmd.Flags().Lookup(\"wiktionary\"))\n\n\tcmd.Flags().Int(\"truncate\", truncate, \"number of characters to extract from text\")\n\tcfg.BindPFlag(\"wikipedia.truncate\", cmd.Flags().Lookup(\"truncate\"))\n\n\tcmd.Flags().Bool(\"delete\", true, \"delete file after parsed\")\n\tcfg.BindPFlag(\"wikipedia.delete\", cmd.Flags().Lookup(\"delete\"))\n\n\tcfg.BindPFlag(\"wikipedia.workers\", cmd.Flags().Lookup(\"workers\"))\n\n\tif err := cmd.Execute(); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (c *InterfaceConfig) FillDefaults() {\n\tc.HTTPAddr = \":8080\"\n\tc.RPCAddr = \":7080\"\n}", "func setDefaultValues(cfg *config.ConfigToml) {\n\n\t//Algorithm\n\tif cfg.Algorithm.Name == \"\" {\n\t\tcfg.Algorithm.Name = defaultAlgorithm\n\t}\n\n\t//API\n\tif cfg.API.Port == 0 {\n\t\tcfg.API.Port = defaultAPIPort\n\t}\n\tif cfg.API.User == \"\" {\n\t\tcfg.API.User = defaultAPIUser\n\t}\n\n\t//DATABASE\n\tif cfg.Database.Database == \"\" {\n\t\tcfg.Database.Database = defaultDBname\n\t}\n\tif cfg.Database.DatabaseType == \"\" {\n\t\tcfg.Database.DatabaseType = defaultDBType\n\t}\n\tif cfg.Database.IP == \"\" {\n\t\tcfg.Database.IP = defaultDBIP\n\t}\n\tif cfg.Database.User == \"\" {\n\t\tcfg.Database.User = defaultDBSetupUser\n\t}\n\tif cfg.Database.Password == \"\" {\n\t\tcfg.Database.Password = defaultDBSetupPassword\n\t}\n\tif cfg.Database.Port == 0 {\n\t\tcfg.Database.Port = defaultDBPort\n\t}\n\n\t//Project\n\tif cfg.Project.Name == \"\" {\n\t\tcfg.Project.Name = defaultProjectName\n\t}\n\tif cfg.Project.Description == \"\" {\n\t\tcfg.Project.Description = defaultProjectDescription\n\t}\n\n\t//Modules\n\tif cfg.Modules == nil {\n\t\tcfg.Modules = make(map[string]config.Module, 0)\n\t}\n}", "func Defaults(obj interface{}) {\n\tif reflect.TypeOf(obj).Kind() != reflect.Ptr {\n\t\tlog.Panic(\"panic at [yaronsumel/filler] : obj kind passed to Defaults should be Ptr\")\n\t}\n\tv := reflect.TypeOf(obj).Elem()\n\ts := reflect.ValueOf(obj).Elem()\n\tfor i := 0; i < v.NumField(); i++ {\n\t\tcurrentField := v.Field(i)\n\t\ttag := currentField.Tag.Get(defaultsTagName)\n\t\tif tag == emptyTag || tag == ignoreTag {\n\t\t\tcontinue\n\t\t}\n\t\tif s.FieldByName(currentField.Name).CanSet() {\n\t\t\tswitch s.FieldByName(currentField.Name).Kind() {\n\t\t\tcase reflect.Int:\n\t\t\t\ti, _ := strconv.Atoi(tag)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(i))\n\t\t\tcase reflect.Int8:\n\t\t\t\ti, _ := strconv.ParseInt(tag, 10, 8)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(i))\n\t\t\tcase reflect.Int16:\n\t\t\t\ti, _ := strconv.ParseInt(tag, 10, 16)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(i))\n\t\t\tcase reflect.Int32:\n\t\t\t\ti, _ := strconv.ParseInt(tag, 10, 32)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(i))\n\t\t\tcase reflect.Int64:\n\t\t\t\ti, _ := strconv.ParseInt(tag, 10, 64)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(i))\n\t\t\tcase reflect.Uint:\n\t\t\t\tu, _ := strconv.ParseUint(tag, 10, 32)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(u))\n\t\t\tcase reflect.Uint8:\n\t\t\t\tu, _ := strconv.ParseUint(tag, 10, 8)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(u))\n\t\t\tcase reflect.Uint16:\n\t\t\t\tu, _ := strconv.ParseUint(tag, 10, 16)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(u))\n\t\t\tcase reflect.Uint32:\n\t\t\t\tu, _ := strconv.ParseUint(tag, 10, 32)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(u))\n\t\t\tcase reflect.Uint64:\n\t\t\t\tu, _ := strconv.ParseUint(tag, 10, 64)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(u))\n\t\t\tcase reflect.String:\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(tag))\n\t\t\tcase reflect.Bool:\n\t\t\t\tb, _ := strconv.ParseBool(tag)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(b))\n\t\t\tcase reflect.Float32:\n\t\t\t\tf, _ := strconv.ParseFloat(tag, 32)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(f))\n\t\t\tcase reflect.Float64:\n\t\t\t\tf, _ := strconv.ParseFloat(tag, 64)\n\t\t\t\ts.FieldByName(currentField.Name).Set(reflect.ValueOf(f))\n\t\t\t}\n\t\t}\n\t}\n}", "func (o *GetCurrentGenerationParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func SetDefaults() {\n\tviper.SetDefault(Debug, false)\n\tviper.SetDefault(APIHost, \"0.0.0.0\")\n\tviper.SetDefault(APIPort, 8080)\n\tviper.SetDefault(SSLEnabled, false)\n\tviper.SetDefault(SSLGenerateLetsencrypt, false)\n\tviper.SetDefault(UploadsMaximumSize, 100000)\n\tviper.SetDefault(LogPath, \"./logs\")\n\tviper.SetDefault(LogLevel, \"info\")\n\tviper.SetDefault(LogDeleteAfterDays, 30)\n\tviper.SetDefault(DatabaseType, \"mysql\")\n\tviper.SetDefault(DatabaseHost, \"localhost\")\n\tviper.SetDefault(DatabasePort, 3306)\n\tviper.SetDefault(DatabaseName, \"forum\")\n\tviper.SetDefault(DatabaseUsername, \"root\")\n\tviper.SetDefault(DatabasePassword, \"\")\n}", "func (o *FetchIntegrationFormParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *ConfigGetParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (s *NotifTeams) SetDefaults() {\n\t// noop\n}", "func setDefaultValues() {\n\tviper.SetDefault(constant.RESOURCE_FILES_MANDATORY, util.ResourceFiles_Mandatory)\n\tviper.SetDefault(constant.RESOURCE_FILES_OPTIONAL, util.ResourceFiles_Optional)\n\tviper.SetDefault(constant.RESOURCE_FILES_SKIP, util.ResourceFiles_Skip)\n\tviper.SetDefault(constant.PLATFORM_VERSIONS, util.PlatformVersions)\n}", "func (p *Params) applySensibleDefaults() {\n\tif p.Name == \"\" {\n\t\tp.Name = DefaultAtlasName\n\t}\n\tif p.Width == 0 {\n\t\tp.Width = DefaultAtlasWidth\n\t}\n\tif p.Height == 0 {\n\t\tp.Height = DefaultAtlasHeight\n\t}\n}", "func setupDefaults() {\n\tclient.DefaultClient = grpcCli.NewClient()\n\tserver.DefaultServer = grpcSvr.NewServer()\n\tnetwork.DefaultNetwork = mucpNet.NewNetwork()\n\tmetrics.DefaultMetricsReporter = noopMet.New()\n\n\t// setup rpc implementations after the client is configured\n\tauth.DefaultAuth = authSrv.NewAuth()\n\tbroker.DefaultBroker = brokerSrv.NewBroker()\n\tevents.DefaultStream = eventsSrv.NewStream()\n\tevents.DefaultStore = eventsSrv.NewStore()\n\tregistry.DefaultRegistry = registrySrv.NewRegistry()\n\trouter.DefaultRouter = routerSrv.NewRouter()\n\tstore.DefaultStore = storeSrv.NewStore()\n\tstore.DefaultBlobStore = storeSrv.NewBlobStore()\n\truntime.DefaultRuntime = runtimeSrv.NewRuntime()\n}", "func (o *GetBundleByKeyParams) SetDefaults() {\n\tvar (\n\t\tauditDefault = string(\"NONE\")\n\n\t\tincludedDeletedDefault = bool(false)\n\t)\n\n\tval := GetBundleByKeyParams{\n\t\tAudit: &auditDefault,\n\t\tIncludedDeleted: &includedDeletedDefault,\n\t}\n\n\tval.timeout = o.timeout\n\tval.Context = o.Context\n\tval.HTTPClient = o.HTTPClient\n\t*o = val\n}", "func (o *QueryFirewallFieldsParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (self *MapTypeDef) Init() *MapTypeDef {\n\tif self.Keys == \"\" {\n\t\tself.Keys = \"String\"\n\t}\n\tif self.Items == \"\" {\n\t\tself.Items = \"Any\"\n\t}\n\treturn self\n}", "func initDefaultValue() {\n\tif configPath == \"\" {\n\t\tconfigPath = constant.DefaultConfigPath\n\t}\n\n\tif apiConfigPath == \"\" {\n\t\tapiConfigPath = constant.DefaultApiConfigPath\n\t}\n\n\tif logConfigPath == \"\" {\n\t\tlogConfigPath = constant.DefaultLogConfigPath\n\t}\n\n\tif logLevel == \"\" {\n\t\tlogLevel = constant.DefaultLogLevel\n\t}\n\n\tif limitCpus == \"\" {\n\t\tlimitCpus = constant.DefaultLimitCpus\n\t}\n\n\tif logFormat == \"\" {\n\t\tlogFormat = constant.DefaultLogFormat\n\t}\n}", "func (o *UpdateFlowParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescApikey is the schema descriptor for apikey field.\n\tuserDescApikey := userFields[5].Descriptor()\n\t// user.DefaultApikey holds the default value on creation for the apikey field.\n\tuser.DefaultApikey = userDescApikey.Default.(func() uuid.UUID)\n\t// userDescActivated is the schema descriptor for activated field.\n\tuserDescActivated := userFields[8].Descriptor()\n\t// user.DefaultActivated holds the default value on creation for the activated field.\n\tuser.DefaultActivated = userDescActivated.Default.(bool)\n\t// userDescLocked is the schema descriptor for locked field.\n\tuserDescLocked := userFields[9].Descriptor()\n\t// user.DefaultLocked holds the default value on creation for the locked field.\n\tuser.DefaultLocked = userDescLocked.Default.(bool)\n\t// userDescUpdatedAt is the schema descriptor for updated_at field.\n\tuserDescUpdatedAt := userFields[10].Descriptor()\n\t// user.DefaultUpdatedAt holds the default value on creation for the updated_at field.\n\tuser.DefaultUpdatedAt = userDescUpdatedAt.Default.(func() time.Time)\n\t// userDescCreatedAt is the schema descriptor for created_at field.\n\tuserDescCreatedAt := userFields[11].Descriptor()\n\t// user.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuser.DefaultCreatedAt = userDescCreatedAt.Default.(func() time.Time)\n\t// userDescID is the schema descriptor for id field.\n\tuserDescID := userFields[0].Descriptor()\n\t// user.DefaultID holds the default value on creation for the id field.\n\tuser.DefaultID = userDescID.Default.(func() uuid.UUID)\n}", "func (o *ListSSHKeysParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (c *FileConfig) InitDefaults() {\n\tif c.Level == \"\" {\n\t\tc.Level = DefaultFileLevel\n\t}\n\tif c.Size == 0 {\n\t\tc.Size = DefaultFileSizeMiB\n\t}\n\tif c.MaxAge == 0 {\n\t\tc.MaxAge = DefaultFileMaxAgeDays\n\t}\n\tif c.MaxBackups == 0 {\n\t\tc.MaxBackups = DefaultFileMaxBackups\n\t}\n\tif c.FlushInterval == nil {\n\t\ts := DefaultFileFlushSeconds\n\t\tc.FlushInterval = &s\n\t}\n}", "func (c *Configuration) setDefaults() {\n\tfor i := range c.TiCommunityBlunderbuss {\n\t\tc.TiCommunityBlunderbuss[i].setDefaults()\n\t}\n\n\tfor i := range c.TiCommunityCherrypicker {\n\t\tc.TiCommunityCherrypicker[i].setDefaults()\n\t}\n\n\tfor i := range c.TiCommunityTars {\n\t\tc.TiCommunityTars[i].setDefaults()\n\t}\n\n\tif len(c.LogLevel) == 0 {\n\t\tc.LogLevel = defaultLogLevel.String()\n\t}\n}", "func (o *ServiceInstanceGetParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *FreezeParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (c *Config) FillDefaults(data any) {\n\telement := reflect.ValueOf(data).Elem()\n\tfor i := 0; i < element.NumField(); i++ {\n\t\tfield := element.Type().Field(i)\n\t\tfieldName := field.Name\n\t\tfieldDefault := field.Tag.Get(\"default\")\n\t\tfieldValue := element.FieldByName(fieldName).Interface()\n\t\t// log.Printf(\"\\n field = %+v, fieldName = %+v, default = %+v, value = %+v\", field, fieldName, fieldDefault, fieldValue)\n\t\tif fieldValue == \"\" && fieldDefault != \"\" {\n\t\t\telement.FieldByName(fieldName).SetString(fieldDefault)\n\t\t}\n\t}\n}", "func (o *GetKeyBlockByHeightParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *GetPropertyDescriptorParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *EditParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func init() {\n\tpetFields := schema.Pet{}.Fields()\n\t_ = petFields\n\t// petDescHeight is the schema descriptor for height field.\n\tpetDescHeight := petFields[0].Descriptor()\n\t// pet.HeightValidator is a validator for the \"height\" field. It is called by the builders before save.\n\tpet.HeightValidator = petDescHeight.Validators[0].(func(int) error)\n\t// petDescWeight is the schema descriptor for weight field.\n\tpetDescWeight := petFields[1].Descriptor()\n\t// pet.WeightValidator is a validator for the \"weight\" field. It is called by the builders before save.\n\tpet.WeightValidator = petDescWeight.Validators[0].(func(float64) error)\n\t// petDescChip is the schema descriptor for chip field.\n\tpetDescChip := petFields[7].Descriptor()\n\t// pet.DefaultChip holds the default value on creation for the chip field.\n\tpet.DefaultChip = petDescChip.Default.(func() uuid.UUID)\n}", "func (c *Config) InitDefaults() {\n\tc.File.InitDefaults()\n\tc.Console.InitDefaults()\n}", "func (o *ValidateParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (c *Config) FillDefaults() {\n\tc.Cookies.FillDefaults()\n\tc.Interfaces.FillDefaults()\n}", "func (o *TestEndpointParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *GetClockParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (p *Provider) SetDefaults() {\n\tp.Provider.SetDefaults()\n\tp.Endpoints = []string{\"127.0.0.1:6379\"}\n}", "func (o *GetTasksGetPhpParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (s *Service) Init() {\n\ts.Comments = make([]string, 0)\n\ts.Attributes = make([]*Attribute, 0)\n\ts.Methods = make([]*Method, 0)\n}", "func (o *Entry) Defaults() {\n if o.Type == \"\" {\n o.Type = \"ipv4\"\n }\n\n if o.ToInterface == \"\" {\n o.ToInterface = \"any\"\n }\n\n if o.Service == \"\" {\n o.Service = \"any\"\n }\n\n if len(o.SourceAddresses) == 0 {\n o.SourceAddresses = []string{\"any\"}\n }\n\n if len(o.DestinationAddresses) == 0 {\n o.DestinationAddresses = []string{\"any\"}\n }\n\n if o.SatType == \"\" {\n o.SatType = None\n }\n\n if o.DatType == \"\" {\n o.DatType = DatTypeStatic\n }\n}", "func (o *PcloudSapGetParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (ic *InstanceCreate) defaults() {\n\tif _, ok := ic.mutation.CreatedAt(); !ok {\n\t\tv := instance.DefaultCreatedAt()\n\t\tic.mutation.SetCreatedAt(v)\n\t}\n\tif _, ok := ic.mutation.UpdatedAt(); !ok {\n\t\tv := instance.DefaultUpdatedAt()\n\t\tic.mutation.SetUpdatedAt(v)\n\t}\n\tif _, ok := ic.mutation.ID(); !ok {\n\t\tv := instance.DefaultID()\n\t\tic.mutation.SetID(v)\n\t}\n}", "func (o *GetDistrictForSchoolParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *GetRepository15Params) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *CreateMigrationInvoiceParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (d *Default) Init(c *viper.Viper) error {\n\tlog.Println(\"Init called\")\n\treturn nil\n}", "func init() {\n\tfor group, values := range defaultConfigs {\n\t\tcore.RegisterConfig(group, values)\n\t}\n\tcore.RegisterService(\"indicator\", indicator.Configs, &indicator.IndicatorServiceFactory{})\n\tcore.RegisterService(\"executor\", executor.Configs, &executor.ExecutorServiceFactory{})\n}", "func init() {\n\t// common\n\tviper.SetDefault(\"log.level\", \"info\")\n\n\t// ethereum\n\tviper.SetDefault(\"ethereum.addr\", \"https://cloudflare-eth.com\")\n\tviper.SetDefault(\"ethereum.wss\", false)\n\n\t// grpc admin\n\tviper.SetDefault(\"grpc.host\", \"0.0.0.0\")\n\tviper.SetDefault(\"grpc.port\", 9090)\n\tviper.SetDefault(\"grpc.timeout\", \"120s\")\n\n\t// cache\n\tviper.SetDefault(\"cachesize\", 100)\n}", "func (c *Config) SetDefaults() {\n\tc.ClientID = \"owl-shop\"\n\tc.TopicReplicationFactor = 3\n\n\tc.SASL.SetDefaults()\n}", "func (d *DefaultDriver) Init(sched, nodeDriver, token, storageProvisioner, csiGenericDriverConfigMap string) error {\n\tStorageProvisioner = DefaultStorageProvisioner\n\treturn nil\n}", "func (o *PetListParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *RegisterApplicationParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (pd *pymtData) Init(p payment.PymtUpsert) {\n\tpd.Type = p.Type\n\tpd.Attrs = p.Attributes\n}", "func (c *Config) ApplyDefaults(globals integrations.Globals) error {\n\tc.Common.ApplyDefaults(globals.SubsystemOpts.Metrics.Autoscrape)\n\tif id, err := c.Identifier(globals); err == nil {\n\t\tc.Common.InstanceKey = &id\n\t}\n\treturn nil\n}", "func (fpsc *FloorPlanScaleCreate) defaults() {\n\tif _, ok := fpsc.mutation.CreateTime(); !ok {\n\t\tv := floorplanscale.DefaultCreateTime()\n\t\tfpsc.mutation.SetCreateTime(v)\n\t}\n\tif _, ok := fpsc.mutation.UpdateTime(); !ok {\n\t\tv := floorplanscale.DefaultUpdateTime()\n\t\tfpsc.mutation.SetUpdateTime(v)\n\t}\n}", "func (o *GetFqdnCacheParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func init() {\n\t// set reasonable defaults\n\tsetDefaults()\n\n\t// override defaults with configuration read from configuration file\n\tviper.AddConfigPath(\"$GOPATH/src/github.com/xlab-si/emmy/config\")\n\terr := loadConfig(\"defaults\", \"yml\")\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n}", "func (o *GetGCParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (t *TopicConfig) SetDefaults() {\n\tif t.Spec.MigrationConfig == nil {\n\t\tt.Spec.MigrationConfig = &TopicMigrationConfig{}\n\t}\n\n\tif t.Spec.MigrationConfig.PartitionBatchSize == 0 {\n\t\t// Migration partitions one at a time\n\t\tt.Spec.MigrationConfig.PartitionBatchSize = 1\n\t}\n\n\tif t.Spec.PlacementConfig.Picker == \"\" {\n\t\tt.Spec.PlacementConfig.Picker = PickerMethodRandomized\n\t}\n}", "func init() {\n\tinitCfgDir()\n\tinitCreds()\n}", "func (c *ConsoleConfig) InitDefaults() {\n\tif c.Level == \"\" {\n\t\tc.Level = DefaultConsoleLevel\n\t}\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescCreatedAt is the schema descriptor for created_at field.\n\tuserDescCreatedAt := userFields[2].Descriptor()\n\t// user.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuser.DefaultCreatedAt = userDescCreatedAt.Default.(func() time.Time)\n\t// userDescUpdatedAt is the schema descriptor for updated_at field.\n\tuserDescUpdatedAt := userFields[3].Descriptor()\n\t// user.DefaultUpdatedAt holds the default value on creation for the updated_at field.\n\tuser.DefaultUpdatedAt = userDescUpdatedAt.Default.(func() time.Time)\n\t// user.UpdateDefaultUpdatedAt holds the default value on update for the updated_at field.\n\tuser.UpdateDefaultUpdatedAt = userDescUpdatedAt.UpdateDefault.(func() time.Time)\n\t// userDescID is the schema descriptor for id field.\n\tuserDescID := userFields[0].Descriptor()\n\t// user.IDValidator is a validator for the \"id\" field. It is called by the builders before save.\n\tuser.IDValidator = userDescID.Validators[0].(func(int) error)\n}", "func (o *RegenerateDeployKeyParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *ServiceBindingBindingParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *GetCatalogXMLParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *GetRouteByIDParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (it *Interceptor) SetDefaults(ctx context.Context) {\n\tif !contexts.IsUpgradeViaDefaulting(ctx) {\n\t\treturn\n\t}\n\tif _, ok := it.GetLabels()[\"server/type\"]; !ok {\n\t\t// if server type is not set its assumed that running server is http\n\t\tit.Labels = map[string]string{\n\t\t\t\"server/type\": \"http\",\n\t\t}\n\t}\n}", "func (element *ElasticsearchTopologyElement) fillDefaults() {\n\tif element.ZoneCount == 0 {\n\t\telement.ZoneCount = DefaultDataZoneCount\n\t}\n}", "func (tb *TriggerBinding) SetDefaults(ctx context.Context) {}", "func Init(c Config) error {\n\terr := c.validate()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tapiKEY = c.APIKey\n\tlog = c.Logger\n\tlistMapper = c.ListMapper\n\n\treturn nil\n}", "func (s *TLSSpec) SetDefaults(defaultCASecretName string) {\n\tif s.GetCASecretName() == \"\" {\n\t\t// Note that we don't check for nil here, since even a specified, but empty\n\t\t// string should result in the default value.\n\t\ts.CASecretName = util.NewType[string](defaultCASecretName)\n\t}\n\tif s.GetTTL() == \"\" {\n\t\t// Note that we don't check for nil here, since even a specified, but zero\n\t\t// should result in the default value.\n\t\ts.TTL = NewDuration(defaultTLSTTL)\n\t}\n}", "func (o *CredentialProviderAPI) Init() {\n\tif len(o.handlers) == 0 {\n\t\to.initHandlerCache()\n\t}\n}", "func (o *VectorThumbnailParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func init() {\n\tprizeFields := schema.Prize{}.Fields()\n\t_ = prizeFields\n\t// prizeDescSocialUserID is the schema descriptor for social_user_id field.\n\tprizeDescSocialUserID := prizeFields[0].Descriptor()\n\t// prize.DefaultSocialUserID holds the default value on creation for the social_user_id field.\n\tprize.DefaultSocialUserID = prizeDescSocialUserID.Default.(string)\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescLatestHeroScore is the schema descriptor for latest_hero_score field.\n\tuserDescLatestHeroScore := userFields[1].Descriptor()\n\t// user.DefaultLatestHeroScore holds the default value on creation for the latest_hero_score field.\n\tuser.DefaultLatestHeroScore = userDescLatestHeroScore.Default.(int)\n\t// userDescBetterHeroScore is the schema descriptor for better_hero_score field.\n\tuserDescBetterHeroScore := userFields[2].Descriptor()\n\t// user.DefaultBetterHeroScore holds the default value on creation for the better_hero_score field.\n\tuser.DefaultBetterHeroScore = userDescBetterHeroScore.Default.(int)\n\t// userDescSocialUserID is the schema descriptor for social_user_id field.\n\tuserDescSocialUserID := userFields[4].Descriptor()\n\t// user.DefaultSocialUserID holds the default value on creation for the social_user_id field.\n\tuser.DefaultSocialUserID = userDescSocialUserID.Default.(string)\n\t// userDescSocialAvatarURL is the schema descriptor for social_avatar_url field.\n\tuserDescSocialAvatarURL := userFields[5].Descriptor()\n\t// user.DefaultSocialAvatarURL holds the default value on creation for the social_avatar_url field.\n\tuser.DefaultSocialAvatarURL = userDescSocialAvatarURL.Default.(string)\n\t// userDescSocialEmail is the schema descriptor for social_email field.\n\tuserDescSocialEmail := userFields[6].Descriptor()\n\t// user.DefaultSocialEmail holds the default value on creation for the social_email field.\n\tuser.DefaultSocialEmail = userDescSocialEmail.Default.(string)\n\t// userDescSocialName is the schema descriptor for social_name field.\n\tuserDescSocialName := userFields[7].Descriptor()\n\t// user.DefaultSocialName holds the default value on creation for the social_name field.\n\tuser.DefaultSocialName = userDescSocialName.Default.(string)\n\t// userDescSocialType is the schema descriptor for social_type field.\n\tuserDescSocialType := userFields[8].Descriptor()\n\t// user.DefaultSocialType holds the default value on creation for the social_type field.\n\tuser.DefaultSocialType = userDescSocialType.Default.(string)\n\t// userDescSocialPayload is the schema descriptor for social_payload field.\n\tuserDescSocialPayload := userFields[9].Descriptor()\n\t// user.DefaultSocialPayload holds the default value on creation for the social_payload field.\n\tuser.DefaultSocialPayload = userDescSocialPayload.Default.(string)\n\t// userDescHeroPlayed is the schema descriptor for hero_played field.\n\tuserDescHeroPlayed := userFields[10].Descriptor()\n\t// user.DefaultHeroPlayed holds the default value on creation for the hero_played field.\n\tuser.DefaultHeroPlayed = userDescHeroPlayed.Default.(uint)\n\t// userDescHeroRepeat is the schema descriptor for hero_repeat field.\n\tuserDescHeroRepeat := userFields[11].Descriptor()\n\t// user.DefaultHeroRepeat holds the default value on creation for the hero_repeat field.\n\tuser.DefaultHeroRepeat = userDescHeroRepeat.Default.(uint)\n\tuseractiverecordFields := schema.UserActiveRecord{}.Fields()\n\t_ = useractiverecordFields\n\t// useractiverecordDescScore is the schema descriptor for score field.\n\tuseractiverecordDescScore := useractiverecordFields[3].Descriptor()\n\t// useractiverecord.DefaultScore holds the default value on creation for the score field.\n\tuseractiverecord.DefaultScore = useractiverecordDescScore.Default.(int)\n\t// useractiverecordDescIsFinished is the schema descriptor for is_finished field.\n\tuseractiverecordDescIsFinished := useractiverecordFields[4].Descriptor()\n\t// useractiverecord.DefaultIsFinished holds the default value on creation for the is_finished field.\n\tuseractiverecord.DefaultIsFinished = useractiverecordDescIsFinished.Default.(uint)\n}", "func (o *GetIconParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *ListHetznerSizesParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func init() {\n\t// set default values for tls\n\tviper.SetDefault(\"tls-cert-file\", constants.TLSCertPath)\n\tviper.SetDefault(\"tls-key-file\", constants.TLSKeyPath)\n\tviper.SetDefault(\"san-list\", constants.DefaultTlsSan)\n\n\t// set default values for log\n\tviper.SetDefault(\"log-max-length\", constants.DefaultLogEntryMaxlength)\n\tviper.SetDefault(\"log-enable-stdout\", true)\n\tviper.SetDefault(\"log-level\", \"info\")\n\n\t// set default values for server\n\tviper.SetDefault(\"server-port\", constants.DefaultPort)\n\tviper.SetDefault(\"server-read-timeout\", constants.DefaultReadTimeout)\n\tviper.SetDefault(\"server-read-header-timeout\", constants.DefaultReadHeaderTimeout)\n\tviper.SetDefault(\"server-write-timeout\", constants.DefaultWriteTimeout)\n\tviper.SetDefault(\"server-idle-timeout\", constants.DefaultIdleTimeout)\n\tviper.SetDefault(\"server-max-header-bytes\", constants.DefaultMaxHeaderBytes)\n\n\tviper.SetDefault(\"cms-ca-cert-validity\", constants.DefaultCACertValidity)\n\tviper.SetDefault(\"cms-ca-organization\", constants.DefaultOrganization)\n\tviper.SetDefault(\"cms-ca-locality\", constants.DefaultLocality)\n\tviper.SetDefault(\"cms-ca-province\", constants.DefaultProvince)\n\tviper.SetDefault(\"cms-ca-country\", constants.DefaultCountry)\n\n\tviper.SetDefault(\"aas-tls-cn\", constants.DefaultAasTlsCn)\n\tviper.SetDefault(\"aas-jwt-cn\", constants.DefaultAasJwtCn)\n\tviper.SetDefault(\"aas-tls-san\", constants.DefaultTlsSan)\n\n\tviper.SetDefault(\"token-duration-mins\", constants.DefaultTokenDurationMins)\n}", "func (pc *PetCreate) defaults() {\n\tif _, ok := pc.mutation.Age(); !ok {\n\t\tv := pet.DefaultAge\n\t\tpc.mutation.SetAge(v)\n\t}\n\tif _, ok := pc.mutation.Trained(); !ok {\n\t\tv := pet.DefaultTrained\n\t\tpc.mutation.SetTrained(v)\n\t}\n}", "func (o *BulkStartCPUGeneratorsParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func initializeDefaults() error {\n\tif err := AddDefaultProfiles(); err != nil {\n\t\tlogger.Get().Error(\"Default Storage profiles create failed: %v\", err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *GetIngredientVersionRevisionParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescUserName is the schema descriptor for user_name field.\n\tuserDescUserName := userFields[0].Descriptor()\n\t// user.DefaultUserName holds the default value on creation for the user_name field.\n\tuser.DefaultUserName = userDescUserName.Default.(string)\n\t// userDescEmail is the schema descriptor for email field.\n\tuserDescEmail := userFields[1].Descriptor()\n\t// user.DefaultEmail holds the default value on creation for the email field.\n\tuser.DefaultEmail = userDescEmail.Default.(string)\n\t// userDescTelNum is the schema descriptor for tel_num field.\n\tuserDescTelNum := userFields[2].Descriptor()\n\t// user.DefaultTelNum holds the default value on creation for the tel_num field.\n\tuser.DefaultTelNum = userDescTelNum.Default.(string)\n\t// userDescPassword is the schema descriptor for password field.\n\tuserDescPassword := userFields[3].Descriptor()\n\t// user.DefaultPassword holds the default value on creation for the password field.\n\tuser.DefaultPassword = userDescPassword.Default.(string)\n\t// userDescPasswordStr is the schema descriptor for password_str field.\n\tuserDescPasswordStr := userFields[4].Descriptor()\n\t// user.DefaultPasswordStr holds the default value on creation for the password_str field.\n\tuser.DefaultPasswordStr = userDescPasswordStr.Default.(string)\n\t// userDescRegType is the schema descriptor for reg_type field.\n\tuserDescRegType := userFields[5].Descriptor()\n\t// user.DefaultRegType holds the default value on creation for the reg_type field.\n\tuser.DefaultRegType = userDescRegType.Default.(int8)\n\t// userDescRegisterIP is the schema descriptor for register_ip field.\n\tuserDescRegisterIP := userFields[7].Descriptor()\n\t// user.DefaultRegisterIP holds the default value on creation for the register_ip field.\n\tuser.DefaultRegisterIP = userDescRegisterIP.Default.(int)\n\t// userDescTelStatus is the schema descriptor for tel_status field.\n\tuserDescTelStatus := userFields[8].Descriptor()\n\t// user.DefaultTelStatus holds the default value on creation for the tel_status field.\n\tuser.DefaultTelStatus = userDescTelStatus.Default.(int8)\n\t// userDescStatus is the schema descriptor for status field.\n\tuserDescStatus := userFields[9].Descriptor()\n\t// user.DefaultStatus holds the default value on creation for the status field.\n\tuser.DefaultStatus = userDescStatus.Default.(int8)\n\t// userDescCreatedAt is the schema descriptor for created_at field.\n\tuserDescCreatedAt := userFields[10].Descriptor()\n\t// user.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuser.DefaultCreatedAt = userDescCreatedAt.Default.(func() time.Time)\n\tuserextendFields := schema.UserExtend{}.Fields()\n\t_ = userextendFields\n\t// userextendDescRealName is the schema descriptor for real_name field.\n\tuserextendDescRealName := userextendFields[1].Descriptor()\n\t// userextend.DefaultRealName holds the default value on creation for the real_name field.\n\tuserextend.DefaultRealName = userextendDescRealName.Default.(string)\n\t// userextendDescIDNumber is the schema descriptor for id_number field.\n\tuserextendDescIDNumber := userextendFields[2].Descriptor()\n\t// userextend.DefaultIDNumber holds the default value on creation for the id_number field.\n\tuserextend.DefaultIDNumber = userextendDescIDNumber.Default.(string)\n\t// userextendDescAge is the schema descriptor for age field.\n\tuserextendDescAge := userextendFields[3].Descriptor()\n\t// userextend.DefaultAge holds the default value on creation for the age field.\n\tuserextend.DefaultAge = userextendDescAge.Default.(int8)\n\t// userextendDescSex is the schema descriptor for sex field.\n\tuserextendDescSex := userextendFields[4].Descriptor()\n\t// userextend.DefaultSex holds the default value on creation for the sex field.\n\tuserextend.DefaultSex = userextendDescSex.Default.(int8)\n\t// userextendDescBirth is the schema descriptor for birth field.\n\tuserextendDescBirth := userextendFields[5].Descriptor()\n\t// userextend.DefaultBirth holds the default value on creation for the birth field.\n\tuserextend.DefaultBirth = userextendDescBirth.Default.(int)\n\t// userextendDescIcon is the schema descriptor for icon field.\n\tuserextendDescIcon := userextendFields[6].Descriptor()\n\t// userextend.DefaultIcon holds the default value on creation for the icon field.\n\tuserextend.DefaultIcon = userextendDescIcon.Default.(string)\n\t// userextendDescNickName is the schema descriptor for nick_name field.\n\tuserextendDescNickName := userextendFields[7].Descriptor()\n\t// userextend.DefaultNickName holds the default value on creation for the nick_name field.\n\tuserextend.DefaultNickName = userextendDescNickName.Default.(string)\n\t// userextendDescImei is the schema descriptor for imei field.\n\tuserextendDescImei := userextendFields[8].Descriptor()\n\t// userextend.DefaultImei holds the default value on creation for the imei field.\n\tuserextend.DefaultImei = userextendDescImei.Default.(string)\n\t// userextendDescOaid is the schema descriptor for oaid field.\n\tuserextendDescOaid := userextendFields[9].Descriptor()\n\t// userextend.DefaultOaid holds the default value on creation for the oaid field.\n\tuserextend.DefaultOaid = userextendDescOaid.Default.(string)\n\t// userextendDescDeviceID is the schema descriptor for device_id field.\n\tuserextendDescDeviceID := userextendFields[10].Descriptor()\n\t// userextend.DefaultDeviceID holds the default value on creation for the device_id field.\n\tuserextend.DefaultDeviceID = userextendDescDeviceID.Default.(string)\n\t// userextendDescSystemName is the schema descriptor for system_name field.\n\tuserextendDescSystemName := userextendFields[11].Descriptor()\n\t// userextend.DefaultSystemName holds the default value on creation for the system_name field.\n\tuserextend.DefaultSystemName = userextendDescSystemName.Default.(string)\n\t// userextendDescSystemVersion is the schema descriptor for system_version field.\n\tuserextendDescSystemVersion := userextendFields[12].Descriptor()\n\t// userextend.DefaultSystemVersion holds the default value on creation for the system_version field.\n\tuserextend.DefaultSystemVersion = userextendDescSystemVersion.Default.(string)\n\t// userextendDescAdid is the schema descriptor for adid field.\n\tuserextendDescAdid := userextendFields[13].Descriptor()\n\t// userextend.DefaultAdid holds the default value on creation for the adid field.\n\tuserextend.DefaultAdid = userextendDescAdid.Default.(string)\n\t// userextendDescGameID is the schema descriptor for game_id field.\n\tuserextendDescGameID := userextendFields[14].Descriptor()\n\t// userextend.DefaultGameID holds the default value on creation for the game_id field.\n\tuserextend.DefaultGameID = userextendDescGameID.Default.(string)\n\t// userextendDescThirdPlatformID is the schema descriptor for third_platform_id field.\n\tuserextendDescThirdPlatformID := userextendFields[15].Descriptor()\n\t// userextend.DefaultThirdPlatformID holds the default value on creation for the third_platform_id field.\n\tuserextend.DefaultThirdPlatformID = userextendDescThirdPlatformID.Default.(int8)\n\t// userextendDescCreatedAt is the schema descriptor for created_at field.\n\tuserextendDescCreatedAt := userextendFields[16].Descriptor()\n\t// userextend.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuserextend.DefaultCreatedAt = userextendDescCreatedAt.Default.(func() time.Time)\n}", "func (o *AddOrUpdateNodePoolConfigItemParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (p *Provider) SetDefaults() {\n\tp.RootKey = \"traefik\"\n}", "func (s *NotifTeams) SetDefaults() {\n\ts.RenderFacts = utl.NewTrue()\n\ts.TemplateBody = NotifTeamsDefaultTemplateBody\n}", "func (o *PostDeviceUpsertParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (o *PetCreateParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (g *GenOpts) EnsureDefaults() error {\n\tif g.defaultsEnsured {\n\t\treturn nil\n\t}\n\n\tg.templates = templates.ShallowClone()\n\n\tg.templates.LoadDefaults()\n\n\tif g.LanguageOpts == nil {\n\t\tg.LanguageOpts = DefaultLanguageFunc()\n\t}\n\n\tDefaultSectionOpts(g)\n\n\t// set defaults for flattening options\n\tif g.FlattenOpts == nil {\n\t\tg.FlattenOpts = &analysis.FlattenOpts{\n\t\t\tMinimal: true,\n\t\t\tVerbose: true,\n\t\t\tRemoveUnused: false,\n\t\t\tExpand: false,\n\t\t}\n\t}\n\n\tif g.DefaultScheme == \"\" {\n\t\tg.DefaultScheme = defaultScheme\n\t}\n\n\tif g.DefaultConsumes == \"\" {\n\t\tg.DefaultConsumes = runtime.JSONMime\n\t}\n\n\tif g.DefaultProduces == \"\" {\n\t\tg.DefaultProduces = runtime.JSONMime\n\t}\n\n\t// always include validator with models\n\tg.IncludeValidator = true\n\n\tif g.Principal == \"\" {\n\t\tg.Principal = iface\n\t\tg.PrincipalCustomIface = false\n\t}\n\n\tg.defaultsEnsured = true\n\treturn nil\n}", "func (o *MetroclusterInterconnectGetParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func Init(as service.AuthSvc, ss service.SerialSvc, sl, el log.MLogger) {\n\tauthSvc, serialSvc, systemLog, errorLog = as, ss, sl, el\n}" ]
[ "0.6839208", "0.6786656", "0.6145108", "0.60509324", "0.6035193", "0.59128183", "0.58774954", "0.57771736", "0.57705206", "0.573478", "0.567347", "0.56701875", "0.5638894", "0.5609368", "0.5605868", "0.5598546", "0.5596984", "0.55728424", "0.5521497", "0.5506418", "0.55063957", "0.5501024", "0.5498065", "0.5490916", "0.5468591", "0.54598385", "0.54483", "0.54446024", "0.5437369", "0.5433655", "0.5429043", "0.5426972", "0.54151744", "0.5395931", "0.5393133", "0.5386719", "0.53751564", "0.53727424", "0.53647774", "0.53630996", "0.5361135", "0.5360727", "0.53549325", "0.53503174", "0.5350214", "0.5349245", "0.5334951", "0.5332632", "0.5331161", "0.5318599", "0.5310227", "0.5308889", "0.5308204", "0.53056705", "0.52986217", "0.52787006", "0.5271157", "0.52641046", "0.5256994", "0.5256422", "0.5256077", "0.5255452", "0.5245789", "0.52435005", "0.52351594", "0.5235063", "0.5232961", "0.5223492", "0.5218721", "0.52175945", "0.52157444", "0.5206276", "0.520528", "0.5203527", "0.5194667", "0.5192108", "0.5186853", "0.51843876", "0.518389", "0.51832086", "0.5173169", "0.5169973", "0.51661783", "0.51651984", "0.51596856", "0.5159217", "0.5154546", "0.51515275", "0.51497483", "0.514752", "0.51448387", "0.5144129", "0.5143008", "0.51373816", "0.5130945", "0.512001", "0.5112713", "0.5108555", "0.5107382", "0.510737" ]
0.7516784
0
DeltaMerge merges the delta for a given attribute in Spitter.
func (spitterImpl *SpitterImpl) DeltaMerge( deltaMerge base.DeltaMerge, attribute string, delta interface{}, ) (bool, error) { merged, err := spitterImpl.SpiderlingImpl.DeltaMerge( deltaMerge, attribute, delta, ) if merged || err != nil { return merged, err } spidersDeltaMerge, ok := deltaMerge.(DeltaMerge) if !ok { return false, errors.New( "deltaMerge is not the expected type of: " + "'spiders.impl.DeltaMerge'", ) } switch attribute { case "spittingWebToNest": spitterImpl.spittingWebToNestImpl = spidersDeltaMerge.Nest(delta) return true, nil } return false, nil // no errors in delta merging }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (unitImpl *UnitImpl) DeltaMerge(\n\tdeltaMerge base.DeltaMerge,\n\tattribute string,\n\tdelta interface{},\n) (bool, error) {\n\tmerged, err := unitImpl.GameObjectImpl.DeltaMerge(\n\t\tdeltaMerge,\n\t\tattribute,\n\t\tdelta,\n\t)\n\tif merged || err != nil {\n\t\treturn merged, err\n\t}\n\n\tstardashDeltaMerge, ok := deltaMerge.(DeltaMerge)\n\tif !ok {\n\t\treturn false, errors.New(\n\t\t\t\"deltaMerge is not the expected type of: \" +\n\t\t\t\t\"'stardash.impl.DeltaMerge'\",\n\t\t)\n\t}\n\n\tswitch attribute {\n\tcase \"acted\":\n\t\tunitImpl.actedImpl = stardashDeltaMerge.Boolean(delta)\n\t\treturn true, nil\n\tcase \"dashX\":\n\t\tunitImpl.dashXImpl = stardashDeltaMerge.Float(delta)\n\t\treturn true, nil\n\tcase \"dashY\":\n\t\tunitImpl.dashYImpl = stardashDeltaMerge.Float(delta)\n\t\treturn true, nil\n\tcase \"energy\":\n\t\tunitImpl.energyImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"genarium\":\n\t\tunitImpl.genariumImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"isBusy\":\n\t\tunitImpl.isBusyImpl = stardashDeltaMerge.Boolean(delta)\n\t\treturn true, nil\n\tcase \"job\":\n\t\tunitImpl.jobImpl = stardashDeltaMerge.Job(delta)\n\t\treturn true, nil\n\tcase \"legendarium\":\n\t\tunitImpl.legendariumImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"moves\":\n\t\tunitImpl.movesImpl = stardashDeltaMerge.Float(delta)\n\t\treturn true, nil\n\tcase \"mythicite\":\n\t\tunitImpl.mythiciteImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"owner\":\n\t\tunitImpl.ownerImpl = stardashDeltaMerge.Player(delta)\n\t\treturn true, nil\n\tcase \"protector\":\n\t\tunitImpl.protectorImpl = stardashDeltaMerge.Unit(delta)\n\t\treturn true, nil\n\tcase \"rarium\":\n\t\tunitImpl.rariumImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"shield\":\n\t\tunitImpl.shieldImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"x\":\n\t\tunitImpl.xImpl = stardashDeltaMerge.Float(delta)\n\t\treturn true, nil\n\tcase \"y\":\n\t\tunitImpl.yImpl = stardashDeltaMerge.Float(delta)\n\t\treturn true, nil\n\t}\n\n\treturn false, nil // no errors in delta merging\n}", "func (jobImpl *JobImpl) DeltaMerge(\n\tdeltaMerge base.DeltaMerge,\n\tattribute string,\n\tdelta interface{},\n) (bool, error) {\n\tmerged, err := jobImpl.GameObjectImpl.DeltaMerge(\n\t\tdeltaMerge,\n\t\tattribute,\n\t\tdelta,\n\t)\n\tif merged || err != nil {\n\t\treturn merged, err\n\t}\n\n\tstardashDeltaMerge, ok := deltaMerge.(DeltaMerge)\n\tif !ok {\n\t\treturn false, errors.New(\n\t\t\t\"deltaMerge is not the expected type of: \" +\n\t\t\t\t\"'stardash.impl.DeltaMerge'\",\n\t\t)\n\t}\n\n\tswitch attribute {\n\tcase \"carryLimit\":\n\t\tjobImpl.carryLimitImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"damage\":\n\t\tjobImpl.damageImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"energy\":\n\t\tjobImpl.energyImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"moves\":\n\t\tjobImpl.movesImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"range\":\n\t\tjobImpl.rangeImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"shield\":\n\t\tjobImpl.shieldImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\tcase \"title\":\n\t\tjobImpl.titleImpl = stardashDeltaMerge.String(delta)\n\t\treturn true, nil\n\tcase \"unitCost\":\n\t\tjobImpl.unitCostImpl = stardashDeltaMerge.Int(delta)\n\t\treturn true, nil\n\t}\n\n\treturn false, nil // no errors in delta merging\n}", "func (a *AttributeDefinition) Merge(other *AttributeDefinition) *AttributeDefinition {\n\tif other == nil {\n\t\treturn a\n\t}\n\tif a == nil {\n\t\treturn other\n\t}\n\tleft := a.Type.(Object)\n\tright := other.Type.(Object)\n\tif left == nil || right == nil {\n\t\tpanic(\"cannot merge non object attributes\") // bug\n\t}\n\tfor n, v := range right {\n\t\tleft[n] = v\n\t}\n\treturn a\n}", "func (inj *Injector) Merge(i *Injector, override bool) error {\n\tfor name, v := range i.data {\n\t\tif _, ok := inj.data[name]; ok && !override {\n\t\t\treturn ErrInjectorSetTwicePointer(v)\n\t\t}\n\t\tinj.data[name] = v\n\t}\n\treturn nil\n}", "func (mb *MutableBag) Merge(bags ...*MutableBag) error {\n\t// first step is to make sure there are no redundant definitions of the same attribute\n\tkeys := make(map[string]bool)\n\tfor _, bag := range bags {\n\t\tif bag == nil {\n\t\t\tcontinue\n\t\t}\n\t\tfor k := range bag.values {\n\t\t\tif keys[k] {\n\t\t\t\treturn fmt.Errorf(\"conflicting value for attribute %s\", k)\n\t\t\t}\n\t\t\tkeys[k] = true\n\t\t}\n\t}\n\n\t// now that we know there are no conflicting definitions, do the actual merging...\n\tfor _, bag := range bags {\n\t\tif bag == nil {\n\t\t\tcontinue\n\t\t}\n\t\tfor k, v := range bag.values {\n\t\t\tmb.values[k] = copyValue(v)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (i *instance) merge(seq int, dep map[paxi.ID]int) {\n\tif seq > i.seq {\n\t\ti.seq = seq\n\t\ti.changed = true\n\t}\n\tfor id, d := range dep {\n\t\tif d > i.dep[id] {\n\t\t\ti.dep[id] = d\n\t\t\ti.changed = true\n\t\t}\n\t}\n}", "func (a Annotation) Merge(other schema.Annotation) schema.Annotation {\n\tvar ant Annotation\n\tswitch other := other.(type) {\n\tcase Annotation:\n\t\tant = other\n\tcase *Annotation:\n\t\tif other != nil {\n\t\t\tant = *other\n\t\t}\n\tdefault:\n\t\treturn a\n\t}\n\tif tag := ant.StructTag; tag != \"\" {\n\t\ta.StructTag = tag\n\t}\n\treturn a\n}", "func(t *TargImp) Merge(other Target) (Target, os.Error) {\n\tx := other.(*TargImp)\n\tif x.Name() != t.Name() {\n\t\treturn nil, os.NewError(\"cannot merge targets with different names\")\n\t}\n\t\n\tfor y := 0; y < x.dependlen; y++ {\n\t\tif !t.isDependent(x.dependencies[y]) {\n\t\t\tt.dependencies[t.dependlen] = x.dependencies[y]\n\t\t\tt.dependlen++\n\t\t}\n\t}\n\t\n\treturn t, nil\n}", "func (d *Release) merge() map[string]interface{} {\n\tif len(d.dep) > 0 {\n\t\treturn d.dep\n\t}\n\tif d.src = d.ref.ToDeploy(d.env1, d.env2); len(d.src) == 0 {\n\t\t// No variable in this project for these environments\n\t\treturn nil\n\t}\n\td.dst = d.fetch()\n\td.dep = make(map[string]interface{})\n\tfor k, sv := range d.src {\n\t\tdv, ok := d.dst[k]\n\t\tswitch {\n\t\tcase !ok:\n\t\t\td.dep[k] = sv\n\t\t\td.task.Add++\n\t\tcase sv == nil:\n\t\t\td.dep[k] = sv\n\t\t\td.task.Del++\n\t\tcase sv != dv:\n\t\t\td.dep[k] = sv\n\t\t\td.task.Upd++\n\t\tdefault:\n\t\t\td.task.NoOp++\n\t\t}\n\t}\n\treturn d.dep\n}", "func (entry *Entry) Merge(merge *Entry) {\n\tfor name, value := range merge.fields {\n\t\tentry.SetField(name, value)\n\t}\n}", "func (o *Object) Merge(other *Object) *Object {\n\tres := o\n\tfor _, nat := range *other {\n\t\tres.Set(nat.Name, DupAtt(nat.Attribute))\n\t}\n\treturn res\n}", "func DeltaMerge(tx *sql.Tx, targetTable, tempTable, conditional string) error {\n\tif tx == nil || targetTable == \"\" || tempTable == \"\" || conditional == \"\" {\n\t\treturn nil\n\t}\n\tdeleteQuery := fmt.Sprintf(`\n\t\t\tDELETE FROM %v\n\t\t\tUSING %v\n\t\t\tWHERE %v\n\t`, targetTable, tempTable, conditional)\n\n\tif _, err := tx.Exec(deleteQuery); err != nil {\n\t\treturn err\n\t}\n\n\tinsertQuery := fmt.Sprintf(\"INSERT INTO %v SELECT DISTINCT * FROM %v\", targetTable, tempTable)\n\tif _, err := tx.Exec(insertQuery); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (h HyperLogLog) Merge(ctx context.Context, keys ...string) error {\n\treq := newRequestSize(1+len(keys), \"\\r\\n$7\\r\\nPFMERGE\\r\\n$\")\n\treq.addStringAndStrings(h.name, keys)\n\treturn h.c.cmdSimple(ctx, req)\n}", "func (c *Aggregator) Merge(oa export.Aggregator, desc *metric.Descriptor) error {\n\to, _ := oa.(*Aggregator)\n\tif o == nil {\n\t\treturn aggregator.NewInconsistentAggregatorError(c, oa)\n\t}\n\tc.value.AddNumber(desc.NumberKind(), o.value)\n\treturn nil\n}", "func (m *Merge) merge(travVals *traversedValues) error {\n\tswitch travVals.data.Kind() {\n\tcase reflect.Struct:\n\t\tfor i := 0; i < travVals.data.NumField(); i++ {\n\t\t\tnewTravVals := newTraversedValues()\n\t\t\tnewTravVals.data = travVals.data.Field(i)\n\t\t\tfor _, s := range travVals.sources {\n\t\t\t\tnewTravVals.sources = append(newTravVals.sources, &traversedSource{\n\t\t\t\t\tname: s.name,\n\t\t\t\t\tdata: s.data.Field(i),\n\t\t\t\t})\n\t\t\t}\n\t\t\tnewTravVals.joinFieldPath(travVals.fieldPath, travVals.data.Type().Field(i).Name)\n\n\t\t\terr := m.merge(newTravVals)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\tcase reflect.Ptr:\n\t\tswitch travVals.data.Type().Elem().Kind() {\n\t\tcase reflect.Struct, reflect.Map, reflect.Slice, reflect.Array, reflect.Interface, reflect.Ptr:\n\t\t\treturn errors.Wrapf(errUnsupportedType, \"merge: `%s`\", travVals.data.Type().Elem().Kind())\n\t\tcase reflect.Invalid:\n\t\t\treturn errReallyWrong\n\t\tdefault:\n\t\t\tselectedValue, selectedSource, err := m.selectValue(travVals)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\ttravVals.data.Set(selectedValue)\n\n\t\t\tm.mergedMeta = append(m.mergedMeta, MergedMeta{\n\t\t\t\tFieldPath: travVals.fieldPath,\n\t\t\t\tSource: selectedSource,\n\t\t\t\tValue: fmt.Sprintf(\"%+v\", selectedValue.Elem()),\n\t\t\t})\n\t\t}\n\tcase reflect.Invalid:\n\t\treturn errReallyWrong\n\tdefault:\n\t\treturn errors.Wrapf(errUnsupportedType, \"merge: `%s`\", travVals.data.Kind())\n\t}\n\n\treturn nil\n}", "func (m groupSalaryStatsMap) merge(key string, src *groupSalaryStats) {\n\tif s, ok := m[key]; ok {\n\t\ts.salaries.merge(&src.salaries)\n\t\treturn\n\t}\n\tm[key] = src\n}", "func (r *AggregateMetadata) Merge(other AggregateMetadata) {\n\tfor k, v := range other {\n\t\t(*r)[k] += v\n\t}\n}", "func NewMerger(kvFlusher kv.Flusher) (kv.Merger, error) {\n\tmetaFlusher, err := NewFlusher(kvFlusher)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &merger{\n\t\tmetaFlusher: metaFlusher,\n\t\tkvFlusher: kvFlusher,\n\t}, nil\n}", "func (me *Sac) Merge(i interface{}) *Sac {\n\n\tswitch reflect.TypeOf(i).Kind() {\n\tcase reflect.Struct:\n\t\tif s, ok := i.(Sac); ok {\n\t\t\tme.Merge(s.Data)\n\t\t} else {\n\t\t\tme.Merge(structs.Map(i))\n\t\t}\n\tcase reflect.Map:\n\t\tm := i.(map[string]interface{})\n\t\tfor key, val := range m {\n\t\t\tif _, exists := me.Data[key]; exists {\n\t\t\t\tpanic(\"Merge field already exists:\" + key)\n\t\t\t} else {\n\t\t\t\tme.Data[key] = val\n\t\t\t}\n\t\t}\n\tdefault:\n\t\tpanic(\"Can't merge something that is not struct or map\")\n\t}\n\n\treturn me\n}", "func (attr SGRAttr) Merge(other SGRAttr) SGRAttr {\n\tif other&SGRAttrClear != 0 {\n\t\tattr = SGRClear\n\t}\n\tattr |= other & SGRAttrMask\n\tif c, set := other.FG(); set {\n\t\tattr = attr.SansFG() | c.FG()\n\t}\n\tif c, set := other.BG(); set {\n\t\tattr = attr.SansBG() | c.BG()\n\t}\n\treturn attr\n}", "func (c *Aggregator) Merge(oa aggregator.Aggregator, desc *sdkapi.Descriptor) error {\n\to, _ := oa.(*Aggregator)\n\tif o == nil {\n\t\treturn aggregator.NewInconsistentAggregatorError(c, oa)\n\t}\n\n\tc.samples = combine(c.samples, o.samples)\n\treturn nil\n}", "func (sp *Properties) Merge(merge api.Properties) {\n\tfor _, id := range merge.Order() {\n\t\tprop, _ := merge.Get(id)\n\t\tsp.Add(prop)\n\t}\n}", "func (v *VersionVector) Merge(other *VersionVector) {\n\tother.REach(func(actor string, t LamportTime) {\n\t\tv.Witness(actor, t)\n\t})\n}", "func (t *Telemetry) Merge(b *Telemetry) *Telemetry {\n\tconfig := *t\n\n\tif b.StatsdAddress != \"\" {\n\t\tconfig.StatsdAddress = b.StatsdAddress\n\t}\n\n\treturn &config\n}", "func (m Meta) Merge(other Change) (otherx, cx Change) {\n\tif m.Change != nil {\n\t\tother, m.Change = m.Change.Merge(other)\n\t}\n\treturn other, m\n}", "func (m Metadata) Merge(other Metadata) Metadata {\n\tresult := m.Copy()\n\tfor k, v := range other {\n\t\tresult[k] = v // other takes precedence\n\t}\n\treturn result\n}", "func (l *AttributeList) Merge(other AttributeList) AttributeList {\n\tif l.Len() == 0 {\n\t\treturn other\n\t}\n\tif other.Len() == 0 {\n\t\treturn *l\n\t}\n\tl.last.Next = other.first\n\tl.last = other.last\n\tl.length += other.Len()\n\treturn *l\n}", "func Merge(base, overrides interface{}) error {\n\tbaseBytes, err := json.Marshal(base)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert current object to byte sequence\")\n\t}\n\n\toverrideBytes, err := json.Marshal(overrides)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert current object to byte sequence\")\n\t}\n\n\tpatchMeta, err := strategicpatch.NewPatchMetaFromStruct(base)\n\tif err != nil {\n\t\treturn errors.WrapIf(err, \"failed to produce patch meta from struct\")\n\t}\n\tpatch, err := strategicpatch.CreateThreeWayMergePatch(overrideBytes, overrideBytes, baseBytes, patchMeta, true)\n\tif err != nil {\n\t\treturn errors.WrapIf(err, \"failed to create three way merge patch\")\n\t}\n\n\tmerged, err := strategicpatch.StrategicMergePatchUsingLookupPatchMeta(baseBytes, patch, patchMeta)\n\tif err != nil {\n\t\treturn errors.WrapIf(err, \"failed to apply patch\")\n\t}\n\n\tvalueOfBase := reflect.Indirect(reflect.ValueOf(base))\n\tinto := reflect.New(valueOfBase.Type())\n\tif err := json.Unmarshal(merged, into.Interface()); err != nil {\n\t\treturn err\n\t}\n\tif !valueOfBase.CanSet() {\n\t\treturn errors.New(\"unable to set unmarshalled value into base object\")\n\t}\n\tvalueOfBase.Set(reflect.Indirect(into))\n\treturn nil\n}", "func (p *Merger) Merge() error {\n\tsourceYml := map[interface{}]interface{}{}\n\tfor _, file := range p.files {\n\t\ttemp, err := p.load(file)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := p.appendUpdate(sourceYml, temp); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tout, err := yaml.Marshal(&sourceYml)\n\tif err != nil {\n\t\treturn err\n\t} else {\n\t\tioutil.WriteFile(p.target, out, os.ModePerm)\n\t}\n\treturn nil\n}", "func (def *Definition) Merge(w window.Span, v interface{}, prev window.State) error {\n\tf := merger.Cons()\n\tif def.g.merger != nil {\n\t\tf = merger.Fold(def.g.merger)\n\t}\n\treturn def.g.window.Merge(w, v, prev, f)\n}", "func (mdl *Model) Merge(model std.Model) error {\n\treturn nil\n}", "func (wb *WriteBatch) Merge(key, value []byte) {\n\tcKey := bytesToChar(key)\n\tcValue := bytesToChar(value)\n\tC.rocksdb_writebatch_merge(wb.c, cKey, C.size_t(len(key)), cValue, C.size_t(len(value)))\n}", "func (tm *merger) Merge(tagKeyID uint32, dataBlocks [][]byte) error {\n\tmaxSequenceID := uint32(0) // target sequence of tag value id\n\t// 1. prepare tagKeyMetas\n\tvar tagKeyMetas []TagKeyMeta\n\tfor _, dataBlock := range dataBlocks {\n\t\ttagKeyMeta, err := newTagKeyMeta(dataBlock)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif maxSequenceID < tagKeyMeta.TagValueIDSeq() {\n\t\t\tmaxSequenceID = tagKeyMeta.TagValueIDSeq()\n\t\t}\n\t\ttagKeyMetas = append(tagKeyMetas, tagKeyMeta)\n\t}\n\t// 2. iterator trie data, then merge the tag values\n\tfor _, tagKeyMeta := range tagKeyMetas {\n\t\titr, err := tagKeyMeta.PrefixIterator(nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor itr.Valid() {\n\t\t\ttm.metaFlusher.FlushTagValue(cloneSlice(itr.Key()), encoding.ByteSlice2Uint32(itr.Value()))\n\t\t\titr.Next()\n\t\t}\n\t}\n\tif err := tm.metaFlusher.FlushTagKeyID(tagKeyID, maxSequenceID); err != nil {\n\t\treturn err\n\t}\n\treturn tm.metaFlusher.commitTagKeyID()\n}", "func (m *Metadata) Merge(src proto.Message) {\n\tmetadata, ok := src.(*Metadata)\n\tif !ok {\n\t\treturn\n\t}\n\t*m = *metadata\n\t// Manually clone expiry timestamp as proto.Clone\n\t// cannot cope with values that contain unexported\n\t// attributes (as time.Time does)\n\tif metadata.Expires != nil {\n\t\texpires := *metadata.Expires\n\t\tm.Expires = &expires\n\t}\n\tif len(metadata.Labels) != 0 {\n\t\tm.Labels = make(map[string]string)\n\t\tfor k, v := range metadata.Labels {\n\t\t\tm.Labels[k] = v\n\t\t}\n\t}\n}", "func (s *EnvironmentSource) Merge(ko *koanf.Koanf, _ *schema.StructValidator) (err error) {\n\treturn ko.Merge(s.koanf)\n}", "func (a *API) Merge(other API) {\n\tif a.Short == \"\" {\n\t\ta.Short = other.Short\n\t}\n\n\tif a.Long == \"\" {\n\t\ta.Long = other.Long\n\t}\n\n\ta.Operations = append(a.Operations, other.Operations...)\n}", "func (s *FKCascadeRuntimeStats) Merge(other execdetails.RuntimeStats) {\n\ttmp, ok := other.(*FKCascadeRuntimeStats)\n\tif !ok {\n\t\treturn\n\t}\n\ts.Total += tmp.Total\n\ts.Keys += tmp.Keys\n}", "func (b Binding) Merge(b2 Binding) Binding {\n\n\tresult := Binding{}\n\n\tfor k, v := range b {\n\t\tresult[k] = v\n\t}\n\n\tfor k, v := range b2 {\n\t\tresult[k] = v\n\t}\n\n\treturn result\n}", "func Merge(original interface{}, update interface{}) (merged interface{}, err error) {\n\n\tif original == nil {\n\t\treturn update, nil\n\t}\n\n\tif update == nil {\n\t\treturn original, nil\n\t}\n\n\t// call the recursive merge\n\treturn merge(original, update)\n}", "func (so *Operations) Merge(merge api.Operations) {\n\tfor _, id := range merge.Order() {\n\t\top, _ := merge.Get(id)\n\t\tso.Add(op)\n\t}\n}", "func (s *CommandLineSource) Merge(ko *koanf.Koanf, val *schema.StructValidator) (err error) {\n\treturn ko.Merge(s.koanf)\n}", "func (c *Aggregator) Merge(oa export.Aggregator, desc *metric.Descriptor) error {\n\to, _ := oa.(*Aggregator)\n\tif o == nil {\n\t\treturn aggregator.NewInconsistentAggregatorError(c, oa)\n\t}\n\n\tc.state.sum.AddNumber(desc.NumberKind(), o.state.sum)\n\tc.state.count += o.state.count\n\n\tfor i := 0; i < len(c.state.bucketCounts); i++ {\n\t\tc.state.bucketCounts[i] += o.state.bucketCounts[i]\n\t}\n\treturn nil\n}", "func (mb *MutableBag) Merge(bag *MutableBag) {\n\tfor k, v := range bag.values {\n\t\t// the input bags cannot override values already in the destination bag\n\t\tif !mb.Contains(k) {\n\t\t\tmb.values[k] = copyValue(v)\n\t\t}\n\t}\n}", "func Merge(dits []Iterator, cb func([]*Meta) error) error {\n\treturn mergeByKey(dits, metaInputID, cb)\n}", "func (c *DogStatsDConfig) Merge(o *DogStatsDConfig) *DogStatsDConfig {\n\tif c == nil {\n\t\tif o == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn o.Copy()\n\t}\n\n\tif o == nil {\n\t\treturn c.Copy()\n\t}\n\n\tr := c.Copy()\n\n\tif o.Address != nil {\n\t\tr.Address = StringCopy(o.Address)\n\t}\n\n\tif o.Period != nil {\n\t\tr.Period = TimeDurationCopy(o.Period)\n\t}\n\n\treturn r\n}", "func (e EdgeMetadatas) Merge(other EdgeMetadatas) EdgeMetadatas {\n\tcp := e.Copy()\n\tfor k, v := range other {\n\t\tcp[k] = cp[k].Merge(v)\n\t}\n\treturn cp\n}", "func (el *Elements) SumMerge(left Elements, mult float64) {\n\tfor _, v := range left {\n\t\tif ndx, exists := (*el).Index(v.Name); exists {\n\t\t\t(*el)[ndx].Value += v.Value * mult\n\t\t} else {\n\t\t\t(*el).Add(v.Name, v.Value*mult)\n\t\t}\n\t}\n}", "func (e *dataUsageEntry) merge(other dataUsageEntry) {\n\te.Objects += other.Objects\n\te.Versions += other.Versions\n\te.Size += other.Size\n\tors := other.ReplicationStats\n\tempty := replicationStats{}\n\tif ors != nil && *ors != empty {\n\t\tif e.ReplicationStats == nil {\n\t\t\te.ReplicationStats = &replicationStats{}\n\t\t}\n\t\te.ReplicationStats.PendingSize += other.ReplicationStats.PendingSize\n\t\te.ReplicationStats.FailedSize += other.ReplicationStats.FailedSize\n\t\te.ReplicationStats.ReplicatedSize += other.ReplicationStats.ReplicatedSize\n\t\te.ReplicationStats.ReplicaSize += other.ReplicationStats.ReplicaSize\n\t\te.ReplicationStats.PendingCount += other.ReplicationStats.PendingCount\n\t\te.ReplicationStats.FailedCount += other.ReplicationStats.FailedCount\n\n\t}\n\n\tfor i, v := range other.ObjSizes[:] {\n\t\te.ObjSizes[i] += v\n\t}\n}", "func mergeIfMergable(obj reflect.Value, src reflect.Value) (reflect.Value, bool) {\n\tvar out reflect.Value\n\n\t// Look for the .WithDefaults method.\n\tmeth, ok := obj.Type().MethodByName(\"Merge\")\n\tif !ok {\n\t\treturn out, false\n\t}\n\n\t// Verify the signature matches our Mergable psuedointerface:\n\t// - two inputs (the receiver), and one output\n\t// - input types match output type exactly (disallow the usual pointer receiver semantics)\n\tif meth.Type.NumIn() != 2 || meth.Type.NumOut() != 1 {\n\t\treturn out, false\n\t}\n\tif meth.Type.In(0) != meth.Type.In(1) || meth.Type.In(0) != meth.Type.Out(0) {\n\t\treturn out, false\n\t}\n\n\t// Psuedointerface matches, call the .Merge method.\n\tout = meth.Func.Call([]reflect.Value{obj, src})[0]\n\n\treturn out, true\n}", "func Merge(src interface{}, dst interface{}, ignore ...string) interface{} {\n\tsrcStruct := ST.New(src)\n\tdstStruct := ST.New(dst)\nmain:\n\tfor _, field := range srcStruct.Fields() {\n\t\t_, ok := dstStruct.FieldOk(field.Name())\n\t\tif !ok {\n\n\t\t\tcontinue\n\t\t}\n\t\tfor _, ign := range ignore {\n\t\t\t//skip field if it's on the ignore list\n\t\t\tif ign == field.Name() {\n\t\t\t\tcontinue main\n\t\t\t}\n\t\t}\n\t\terr := dstStruct.Field(field.Name()).Set(field.Value())\n\t\tif nil != err {\n\t\t\tlog.Error(\"Failed to assign value from to field \", field.Name(), \" got error \", err)\n\t\t}\n\t}\n\treturn dst\n}", "func merge(original interface{}, update interface{}) (merged interface{}, err error) {\n\n\tlogger.Info(\"Merging\", \"original\", original, \"update\", update)\n\n\tswitch O := original.(type) {\n\n\tcase map[string]interface{}:\n\t\tU, ok := update.(map[string]interface{})\n\t\tif !ok {\n\t\t\treturn nil, errors.New(\"update is not mS like original\")\n\t\t}\n\t\tlogger.Info(\"mS entering\")\n\t\tfor key, val := range U {\n\t\t\tlogger.Debug(\"in merge mS-U\", \"key\", key, \"val\", val, \"curr\", O[key])\n\t\t\tif curr, exists := O[key]; exists {\n\t\t\t\ttmp, err := merge(curr, val)\n\t\t\t\tlogger.Debug(\"after merge mS\", \"tmp\", tmp, \"err\", err)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, errors.Wrap(err, \"in merge mS\")\n\t\t\t\t}\n\t\t\t\tO[key] = tmp\n\t\t\t} else {\n\t\t\t\tO[key] = val\n\t\t\t}\n\t\t}\n\t\tlogger.Info(\"mS returning\", \"O\", O)\n\t\treturn O, nil\n\n\tcase []interface{}:\n\t\tU, ok := update.([]interface{})\n\t\tif !ok {\n\t\t\treturn nil, errors.New(\"update is not aI like original\")\n\t\t}\n\t\t// logger.Warn(\"O\", \"data\", O)\n\t\t// logger.Warn(\"U\", \"data\", U)\n\n\t\tlogger.Info(\"aI entering\")\n\t\t// turn update into map\n\t\tUM := map[string]interface{}{}\n\t\tfor i, elem := range U {\n\t\t\tswitch E := elem.(type) {\n\n\t\t\tcase map[string]interface{}:\n\t\t\t\tname, ok := E[\"name\"]\n\t\t\t\tif !ok {\n\t\t\t\t\treturn nil, errors.New(\"original array objects must have names to be merged\")\n\t\t\t\t}\n\t\t\t\tUM[name.(string)] = E\n\n\t\t\tcase string:\n\t\t\t\tUM[E] = E\n\n\t\t\tdefault:\n\t\t\t\tlogger.Error(\"original unknown elem type in aI\", \"i\", i, \"elem\", elem)\n\t\t\t\treturn nil, errors.New(\"original unknown elem type in aI\")\n\t\t\t}\n\t\t}\n\n\t\tfor i, elem := range O {\n\t\t\t// logger.Crit(\"O-loop\", \"i\", i, \"elem\", elem)\n\t\t\tswitch E := elem.(type) {\n\n\t\t\tcase map[string]interface{}:\n\t\t\t\tiname, ok := E[\"name\"]\n\t\t\t\tif !ok {\n\t\t\t\t\treturn nil, errors.New(\"original array objects must have names to be merged\")\n\t\t\t\t}\n\n\t\t\t\tname := iname.(string)\n\t\t\t\t// logger.Error(\"Name\", \"name\", name)\n\n\t\t\t\tcurr, exists := UM[name]\n\t\t\t\tif exists {\n\t\t\t\t\ttmp, err := merge(elem, curr)\n\t\t\t\t\t// this is correct, the var names curr and elem are backwards...\n\t\t\t\t\t// busy fixing a bug\n\t\t\t\t\t// logger.Crit(\"merging with existing element\", \"key\", name, \"val\", curr, \"curr\", elem)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn nil, errors.Wrap(err, \"in merge MS\")\n\t\t\t\t\t}\n\t\t\t\t\tO[i] = tmp\n\t\t\t\t\tdelete(UM, name)\n\t\t\t\t}\n\t\t\tcase string:\n\t\t\t\t_, exists := UM[E]\n\t\t\t\tif exists {\n\t\t\t\t\tdelete(UM, E)\n\t\t\t\t}\n\n\t\t\tdefault:\n\t\t\t\tlogger.Error(\"original unknown elem type in aI\", \"i\", i, \"elem\", elem)\n\t\t\t\treturn nil, errors.New(\"original unknown elem type in aI\")\n\t\t\t}\n\t\t}\n\t\t// merge\n\t\tlogger.Info(\"aI\")\n\n\t\t// turn back into array\n\t\tOA := []interface{}{}\n\t\tfor _, val := range O {\n\t\t\tOA = append(OA, val)\n\t\t}\n\t\tfor _, elem := range U {\n\t\t\tswitch E := elem.(type) {\n\n\t\t\tcase map[string]interface{}:\n\t\t\t\tname, ok := E[\"name\"]\n\t\t\t\tif !ok {\n\t\t\t\t\treturn nil, errors.New(\"original array objects must have names to be merged\")\n\t\t\t\t}\n\t\t\t\t_, exists := UM[name.(string)]\n\t\t\t\tif exists {\n\t\t\t\t\tOA = append(OA, elem)\n\t\t\t\t}\n\n\t\t\tcase string:\n\t\t\t\t_, exists := UM[E]\n\t\t\t\tif exists {\n\t\t\t\t\tOA = append(OA, elem)\n\t\t\t\t}\n\n\t\t\t}\n\t\t}\n\n\t\t// logger.Error(\"OA\", \"data\", OA)\n\n\t\tlogger.Info(\"aI returning\", \"OA\", OA)\n\t\treturn OA, nil\n\n\tcase string:\n\t\treturn update, nil\n\n\tdefault:\n\t\treturn nil, errors.New(\"unmergable original\" + fmt.Sprintf(\"%t, %+v\", original, original))\n\n\t}\n\n\tlogger.Crit(\"Shouldn't get here (end of merge function)\")\n\treturn nil, errors.New(\"PANIC, should not get here\")\n}", "func Merge(dataA *interface{}, dataB *interface{}, arrayAppend bool) {\n\tmergeRecursive(dataA, dataB, arrayAppend, nil, \"\")\n}", "func (s *DealService) Merge(ctx context.Context, id int, opt *DealsMergeOptions) (*Response, error) {\n\turi := fmt.Sprintf(\"/deals/%v/merge\", id)\n\treq, err := s.client.NewRequest(http.MethodPut, uri, nil, opt)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.client.Do(ctx, req, nil)\n}", "func (j *JobScaling) Merge(b *JobScaling) *JobScaling {\n\tconfig := *j\n\n\tif b.Enabled {\n\t\tconfig.Enabled = b.Enabled\n\t}\n\n\tif b.ConsulToken != \"\" {\n\t\tconfig.ConsulToken = b.ConsulToken\n\t}\n\n\tif b.ConsulKeyLocation != \"\" {\n\t\tconfig.ConsulKeyLocation = b.ConsulKeyLocation\n\t}\n\n\treturn &config\n}", "func merge(dest *config, src config) {\n\tmergo.Merge(dest, src, mergo.WithOverride)\n}", "func MergeLogging(v bool) Option {\n\treturn func(o *options) {\n\t\to.mergeLogging = v\n\t}\n}", "func (res *ruleMatchResults) merge(other ruleMatchResults) *ruleMatchResults {\n\tif res.cutoverNanos < other.cutoverNanos {\n\t\tres.cutoverNanos = other.cutoverNanos\n\t}\n\tres.pipelines = append(res.pipelines, other.pipelines...)\n\treturn res\n}", "func Merge(spell *Spell, line string) string {\n\t//\"allowedClasses\":{\"myst\":true,\"tech\":true,\"wysh\":true}\n\tlog.Infof(\"%#v\", spell)\n\tclassStanza := regexp.MustCompile(`\"allowedClasses\":.*?}`).FindString(line)\n\tnewClassStanza := fmt.Sprintf(`\"allowedClasses\":{\"myst\":%t,\"tech\":%t,\"wysh\":%t}`,spell.Mystic,spell.Technomancer,spell.Witchwarper)\n\tnewLine := strings.Replace(line, classStanza,newClassStanza, 1)\n\tif spell.Level != 0{\n\t\tlevelStanza := regexp.MustCompile(`\"level\":[0-9]`).FindString(line)\n\t\tnewLevelStanza := fmt.Sprintf(`\"level\":%d`,spell.Level)\n\t\tnewLine = strings.Replace(newLine, levelStanza,newLevelStanza,1)\n\t}\n\tschoolStanza := regexp.MustCompile(`\"school\":.*?,`).FindString(line)\n\tnewSchoolStanza := fmt.Sprintf(`\"school\":\"%s\",`, spell.School)\n\tnewLine = strings.Replace(newLine, schoolStanza,newSchoolStanza,1)\n\n\treturn newLine\n}", "func (c *client) Merge(org, repo string, pr int, details MergeDetails) error {\n\tdurationLogger := c.log(\"Merge\", org, repo, pr, details)\n\tdefer durationLogger()\n\n\tge := githubError{}\n\tec, err := c.request(&request{\n\t\tmethod: http.MethodPut,\n\t\tpath: fmt.Sprintf(\"/repos/%s/%s/pulls/%d/merge\", org, repo, pr),\n\t\torg: org,\n\t\trequestBody: &details,\n\t\texitCodes: []int{200, 405, 409},\n\t}, &ge)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif ec == 405 {\n\t\tif strings.Contains(ge.Message, \"Base branch was modified\") {\n\t\t\treturn UnmergablePRBaseChangedError(ge.Message)\n\t\t}\n\t\tif strings.Contains(ge.Message, \"You're not authorized to push to this branch\") {\n\t\t\treturn UnauthorizedToPushError(ge.Message)\n\t\t}\n\t\tif strings.Contains(ge.Message, \"Merge commits are not allowed on this repository\") {\n\t\t\treturn MergeCommitsForbiddenError(ge.Message)\n\t\t}\n\t\treturn UnmergablePRError(ge.Message)\n\t} else if ec == 409 {\n\t\treturn ModifiedHeadError(ge.Message)\n\t}\n\n\treturn nil\n}", "func Merge(transformers ...Transformer) Transformer {\n\treturn merged(transformers)\n}", "func (e FlatMap) Merge(env ...FlatMap) (res FlatMap) {\n\tres = FlatMap{}\n\tfor _, e1 := range append([]FlatMap{e}, env...) {\n\t\tfor k, v := range e1 {\n\t\t\tres[k] = v\n\t\t}\n\t}\n\treturn\n}", "func (p params) merge(source params) params {\n\tfor newKey, newValues := range source {\n\t\tif _, ok := p[newKey]; ok {\n\t\t\tp[newKey] = append(p[newKey], newValues...)\n\t\t} else {\n\t\t\tp[newKey] = newValues\n\t\t}\n\t}\n\treturn p\n}", "func (s *service) Merge(ID string, p *Payment) (*Payment, error) {\n\to, err := s.Get(ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tp.ID = ID\n\tif err := s.merger.Merge(o, p); err != nil {\n\t\treturn nil, log.Errors(ErrMergeFailed, err)\n\t}\n\n\tif err := s.update(o); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn o, nil\n}", "func (m Metaitems) Merge(items Metaitems, strategy MetadataStrategy) Metaitems {\nOUTER:\n\tfor _, newItem := range items {\n\t\tfor _, oldItem := range m {\n\t\t\tif oldItem.Source == newItem.Source {\n\t\t\t\tif oldItem.Metadata != nil {\n\t\t\t\t\toldItem.Metadata = oldItem.Metadata.Merge(newItem.Metadata, strategy)\n\t\t\t\t}\n\t\t\t\tcontinue OUTER\n\t\t\t}\n\t\t}\n\n\t\tm = append(m, newItem)\n\t}\n\n\treturn m\n}", "func (m *MockPullRequestClient) Merge(org, repo string, pr int, details github.MergeDetails) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Merge\", org, repo, pr, details)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (gh *GithubRequireAheadPrecondition) Merge(deployment common.Deployment, ctx *GithubRequireAheadContext) (string, error) {\n\tbase := deployment.Ref()\n\thead := *ctx.RepoDetails.DefaultBranch\n\tcommitMessage := fmt.Sprintf(\"Merge '%v' into '%v'\", head, base)\n\tcommit, err := ctx.RepoClient.Merge(base, head, commitMessage)\n\tif err != nil {\n\t\treturn \"\", err\n\t} else {\n\t\treturn *commit.SHA, nil\n\t}\n}", "func (ops *SimpleOperations) Merge(merge Operations) {\n\tops.safe()\n\n\tfor _, operation := range merge.Order() {\n\t\tmergeOperation, _ := merge.Get(operation)\n\t\tops.Add(mergeOperation)\n\t}\n}", "func mergeProllyTableData(ctx *sql.Context, tm *TableMerger, finalSch schema.Schema, mergeTbl *doltdb.Table, valueMerger *valueMerger) (*doltdb.Table, *MergeStats, error) {\n\titer, err := threeWayDiffer(ctx, tm, valueMerger)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tlr, err := tm.leftTbl.GetRowData(ctx)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tleftEditor := durable.ProllyMapFromIndex(lr).Rewriter(finalSch.GetKeyDescriptor(), finalSch.GetValueDescriptor())\n\n\tai, err := mergeTbl.GetArtifacts(ctx)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tartEditor := durable.ProllyMapFromArtifactIndex(ai).Editor()\n\n\tkeyless := schema.IsKeyless(tm.leftSch)\n\n\tpri, err := newPrimaryMerger(leftEditor, tm, valueMerger, finalSch)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tsec, err := newSecondaryMerger(ctx, tm, valueMerger, finalSch)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tconflicts, err := newConflictMerger(ctx, tm, artEditor)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tcheckValidator, err := newCheckValidator(ctx, tm, valueMerger, finalSch, artEditor)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\t// validator shares an artifact editor with conflict merge\n\tuniq, err := newUniqValidator(ctx, finalSch, tm, valueMerger, artEditor)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tnullChk, err := newNullValidator(ctx, finalSch, tm, valueMerger, artEditor, leftEditor, sec.leftMut)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\ts := &MergeStats{\n\t\tOperation: TableModified,\n\t}\n\tfor {\n\t\tdiff, err := iter.Next(ctx)\n\t\tif errors.Is(err, io.EOF) {\n\t\t\tbreak\n\t\t} else if err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\tcnt, err := uniq.validateDiff(ctx, diff)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\ts.ConstraintViolations += cnt\n\n\t\tcnt, err = nullChk.validateDiff(ctx, diff)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\ts.ConstraintViolations += cnt\n\t\tif cnt > 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\tcnt, err = checkValidator.validateDiff(ctx, diff)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\ts.ConstraintViolations += cnt\n\n\t\tswitch diff.Op {\n\t\tcase tree.DiffOpDivergentModifyConflict, tree.DiffOpDivergentDeleteConflict:\n\t\t\t// In this case, a modification or delete was made to one side, and a conflicting delete or modification\n\t\t\t// was made to the other side, so these cannot be automatically resolved.\n\t\t\ts.DataConflicts++\n\t\t\terr = conflicts.merge(ctx, diff, nil)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\tcase tree.DiffOpRightAdd:\n\t\t\ts.Adds++\n\t\t\terr = pri.merge(ctx, diff, tm.rightSch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\t\terr = sec.merge(ctx, diff, tm.rightSch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\tcase tree.DiffOpRightModify:\n\t\t\ts.Modifications++\n\t\t\terr = pri.merge(ctx, diff, tm.rightSch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\t\terr = sec.merge(ctx, diff, tm.rightSch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\tcase tree.DiffOpRightDelete:\n\t\t\ts.Deletes++\n\t\t\terr = pri.merge(ctx, diff, tm.rightSch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\t\terr = sec.merge(ctx, diff, tm.rightSch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\tcase tree.DiffOpDivergentModifyResolved:\n\t\t\t// In this case, both sides of the merge have made different changes to a row, but we were able to\n\t\t\t// resolve them automatically.\n\t\t\ts.Modifications++\n\t\t\terr = pri.merge(ctx, diff, nil)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\t\terr = sec.merge(ctx, diff, nil)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\tcase tree.DiffOpConvergentAdd, tree.DiffOpConvergentModify, tree.DiffOpConvergentDelete:\n\t\t\t// In this case, both sides of the merge have made the same change, so no additional changes are needed.\n\t\t\tif keyless {\n\t\t\t\ts.DataConflicts++\n\t\t\t\terr = conflicts.merge(ctx, diff, nil)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, nil, err\n\t\t\t\t}\n\t\t\t}\n\t\tdefault:\n\t\t\t// Currently, all changes are applied to the left-side of the merge, so for any left-side diff ops,\n\t\t\t// we can simply ignore them since that data is already in the destination (the left-side).\n\t\t}\n\t}\n\n\tfinalRows, err := pri.finalize(ctx)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tleftIdxs, rightIdxs, err := sec.finalize(ctx)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tfinalIdxs, err := mergeProllySecondaryIndexes(ctx, tm, leftIdxs, rightIdxs, finalSch, finalRows, conflicts.ae)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tfinalArtifacts, err := conflicts.finalize(ctx)\n\n\t// collect merged data in |finalTbl|\n\tfinalTbl, err := mergeTbl.UpdateRows(ctx, finalRows)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tfinalTbl, err = finalTbl.SetIndexSet(ctx, finalIdxs)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tfinalTbl, err = finalTbl.SetArtifacts(ctx, finalArtifacts)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn finalTbl, s, nil\n}", "func (p *StringBuilder) Merge(builder *StringBuilder) {\n\tif builder == nil {\n\t\treturn\n\t}\n\tp.buffer = append(p.buffer, builder.buffer...)\n}", "func (s *YAMLFileSource) Merge(ko *koanf.Koanf, _ *schema.StructValidator) (err error) {\n\treturn ko.Merge(s.koanf)\n}", "func (c *ConfigureOpener) Merge(other ConfigureOpener) {\n\tif c.ErrorThresholdPercentage == 0 {\n\t\tc.ErrorThresholdPercentage = other.ErrorThresholdPercentage\n\t}\n\tif c.RequestVolumeThreshold == 0 {\n\t\tc.RequestVolumeThreshold = other.RequestVolumeThreshold\n\t}\n\tif c.Now == nil {\n\t\tc.Now = other.Now\n\t}\n\tif c.RollingDuration == 0 {\n\t\tc.RollingDuration = other.RollingDuration\n\t}\n\tif c.NumBuckets == 0 {\n\t\tc.NumBuckets = other.NumBuckets\n\t}\n}", "func (d UserData) MergeWith(other m.UserData) {\n\td.ModelData.MergeWith(other.Underlying())\n}", "func merge(rw http.ResponseWriter, req *http.Request) {\r\n\t// Decode the POST body\r\n\tdecoder := json.NewDecoder(req.Body)\r\n\tvar model distributed.MergeRequestModel\r\n\terr := decoder.Decode(&model)\r\n\tif err != nil {\r\n\t\tpanic(err)\r\n\t}\r\n\r\n\tleft := model.Left\r\n\tright := model.Right\r\n\tsize, i, j := len(left)+len(right), 0, 0\r\n\tslice := make([]int, size, size)\r\n\r\n\tfor k := 0; k < size; k++ {\r\n\t\tif i > len(left)-1 && j <= len(right)-1 {\r\n\t\t\tslice[k] = right[j]\r\n\t\t\tj++\r\n\t\t} else if j > len(right)-1 && i <= len(left)-1 {\r\n\t\t\tslice[k] = left[i]\r\n\t\t\ti++\r\n\t\t} else if left[i] < right[j] {\r\n\t\t\tslice[k] = left[i]\r\n\t\t\ti++\r\n\t\t} else {\r\n\t\t\tslice[k] = right[j]\r\n\t\t\tj++\r\n\t\t}\r\n\t}\r\n\r\n\tresponseModel := distributed.MergeResponseModel{slice}\r\n\r\n\t// Parse the response model into a writable format\r\n\tresString, err := json.Marshal(responseModel)\r\n\r\n\trw.Write(resString)\r\n}", "func mergeSingle(m *meta.Meta, merges *mergeList) []meta.MergeUpdate {\n\tvar results []meta.MergeUpdate\n\tfor _, tn := range merges.tn {\n\t\tresult := m.Merge(tn.table, tn.nmerge)\n\t\tresults = append(results, result)\n\t}\n\treturn results\n}", "func (pc ParseContext) Merge(o ParseContext) {\n\tfor k, vs := range o.Args {\n\t\tpc.Args[k] = append(pc.Args[k], vs...)\n\t}\n\n\tfor k, vs := range o.Opts {\n\t\tpc.Opts[k] = append(pc.Opts[k], vs...)\n\t}\n}", "func MetaMerge(a, b map[string]interface{}) map[string]interface{} {\n\tfor k, v := range b {\n\t\t// don't over-write\n\t\tif _, ok := a[k]; !ok {\n\t\t\ta[k] = v\n\t\t}\n\t}\n\treturn a\n}", "func Merge(parameters ...Parameters) (Parameters, error) {\n\ts := Parameters(map[string]interface{}{})\n\tfor _, parameter := range parameters {\n\t\tif err := mergo.Merge(&s, parameter, mergo.WithOverride, mergo.WithAppendSlice); err != nil {\n\t\t\treturn s, errors.Wrap(err, \"cannot merge parameters\")\n\t\t}\n\t}\n\treturn s, nil\n}", "func Merge(target, source interface{}, opt *Options) error {\n\tvT := reflect.ValueOf(target)\n\tvS := reflect.ValueOf(source)\n\n\tif target != nil && vT.Type() == valType {\n\t\tvT = vT.Interface().(reflect.Value)\n\t}\n\tif source != nil && vS.Type() == valType {\n\t\tvS = vS.Interface().(reflect.Value)\n\t}\n\n\tif vT.Kind() != reflect.Ptr {\n\t\treturn errors.New(\"target must be a pointer\")\n\t}\n\n\tif !reflect.Indirect(vT).IsValid() {\n\t\treturn errors.New(\"target can not be zero value\")\n\t}\n\n\t// use defaults if none are provided\n\tif opt == nil {\n\t\topt = NewOptions()\n\t}\n\n\tif opt.mergeFuncs == nil {\n\t\treturn errors.New(\"invalid options, use NewOptions() to generate and then modify as needed\")\n\t}\n\n\t//make a copy here so if there is an error mid way, the target stays in tact\n\tcp := vT.Elem()\n\n\tmerged, err := merge(cp, reflect.Indirect(vS), opt)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !isSettable(vT.Elem(), merged) {\n\t\treturn fmt.Errorf(\"Merge failed: expected merged result to be %v but got %v\",\n\t\t\tvT.Elem().Type(), merged.Type())\n\t}\n\n\tvT.Elem().Set(merged)\n\treturn nil\n}", "func (d *dataUsageCache) merge(other dataUsageCache) {\n\texistingRoot := d.root()\n\totherRoot := other.root()\n\tif existingRoot == nil && otherRoot == nil {\n\t\treturn\n\t}\n\tif otherRoot == nil {\n\t\treturn\n\t}\n\tif existingRoot == nil {\n\t\t*d = other.clone()\n\t\treturn\n\t}\n\tif other.Info.LastUpdate.After(d.Info.LastUpdate) {\n\t\td.Info.LastUpdate = other.Info.LastUpdate\n\t}\n\texistingRoot.merge(*otherRoot)\n\teHash := d.rootHash()\n\tfor key := range otherRoot.Children {\n\t\tentry := other.Cache[key]\n\t\tflat := other.flatten(entry)\n\t\texisting := d.Cache[key]\n\t\t// If not found, merging simply adds.\n\t\texisting.merge(flat)\n\t\td.replaceHashed(dataUsageHash(key), &eHash, existing)\n\t}\n}", "func (m *Merge) Merge(data interface{}, s *Sources) ([]MergedMeta, error) {\n\t// all types of data & sources must be the same\n\tif err := m.checkDataTypes(data, s); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := m.mergeSources(data, s.Sources()); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn m.mergedMeta, nil\n}", "func Merge(dest interface{}, source interface{}) error {\n\topts := make([]func(*mergo.Config), 0)\n\n\t// lists are always overridden - we don't append merged lists since it generally makes things more complicated\n\topts = append(opts, mergo.WithOverride)\n\n\terr := mergo.Merge(dest, source, opts...)\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\treturn nil\n}", "func (f *First) Merge(ctx *sql.Context, buffer, partial sql.Row) error {\n\treturn nil\n}", "func mergeRule(gen, old *bf.CallExpr) *bf.CallExpr {\n\tgenRule := bf.Rule{Call: gen}\n\toldRule := bf.Rule{Call: old}\n\tmerged := *old\n\tmerged.List = nil\n\tmergedRule := bf.Rule{Call: &merged}\n\n\t// Copy unnamed arguments from the old rule without merging. The only rule\n\t// generated with unnamed arguments is go_prefix, which we currently\n\t// leave in place.\n\t// TODO: maybe gazelle should allow the prefix to be changed.\n\tfor _, a := range old.List {\n\t\tif b, ok := a.(*bf.BinaryExpr); ok && b.Op == \"=\" {\n\t\t\tbreak\n\t\t}\n\t\tmerged.List = append(merged.List, a)\n\t}\n\n\t// Merge attributes from the old rule. Preserve comments on old attributes.\n\t// Assume generated attributes have no comments.\n\tfor _, k := range oldRule.AttrKeys() {\n\t\toldAttr := oldRule.AttrDefn(k)\n\t\tif !mergeableFields[k] || shouldKeep(oldAttr) {\n\t\t\tmerged.List = append(merged.List, oldAttr)\n\t\t\tcontinue\n\t\t}\n\n\t\toldExpr := oldAttr.Y\n\t\tgenExpr := genRule.Attr(k)\n\t\tmergedExpr, err := mergeExpr(genExpr, oldExpr)\n\t\tif err != nil {\n\t\t\t// TODO: add a verbose mode and log errors like this.\n\t\t\tmergedExpr = genExpr\n\t\t}\n\t\tif mergedExpr != nil {\n\t\t\tmergedAttr := *oldAttr\n\t\t\tmergedAttr.Y = mergedExpr\n\t\t\tmerged.List = append(merged.List, &mergedAttr)\n\t\t}\n\t}\n\n\t// Merge attributes from genRule that we haven't processed already.\n\tfor _, k := range genRule.AttrKeys() {\n\t\tif mergedRule.Attr(k) == nil {\n\t\t\tmergedRule.SetAttr(k, genRule.Attr(k))\n\t\t}\n\t}\n\n\treturn &merged\n}", "func (m MessageDescriptorMap) Merge(other MessageDescriptorMap) {\n\tfor id, other := range other {\n\t\tif m[id] == nil {\n\t\t\tm[id] = other\n\t\t} else {\n\t\t\tif other.Translations[defaultLanguage] != m[id].Translations[defaultLanguage] {\n\t\t\t\tm[id].updated = true\n\t\t\t}\n\t\t\tfor language, translation := range other.Translations {\n\t\t\t\tif language == defaultLanguage {\n\t\t\t\t\tcontinue // This one is set from the Define.\n\t\t\t\t}\n\t\t\t\tm[id].Translations[language] = translation\n\t\t\t}\n\t\t}\n\t}\n}", "func (n *Notification) Merge(b *Notification) *Notification {\n\tconfig := *n\n\n\tif b.ClusterIdentifier != \"\" {\n\t\tconfig.ClusterIdentifier = b.ClusterIdentifier\n\t}\n\n\tif b.ClusterScalingUID != \"\" {\n\t\tconfig.ClusterScalingUID = b.ClusterScalingUID\n\t}\n\n\tif b.PagerDutyServiceKey != \"\" {\n\t\tconfig.PagerDutyServiceKey = b.PagerDutyServiceKey\n\t}\n\n\treturn &config\n}", "func makeMerger(name string, extensions []string, converter jsonConverter) *Merger {\n\treturn &Merger{\n\t\tName: name,\n\t\tExtensions: extensions,\n\t\tMerge: makeToJSONMergeFunc(converter),\n\t}\n}", "func (s *FKCheckRuntimeStats) Merge(other execdetails.RuntimeStats) {\n\ttmp, ok := other.(*FKCheckRuntimeStats)\n\tif !ok {\n\t\treturn\n\t}\n\ts.Total += tmp.Total\n\ts.Check += tmp.Check\n\ts.Lock += tmp.Lock\n\ts.Keys += tmp.Keys\n}", "func (a *Addresses) Merge(b *Addresses) *Addresses {\n\tresult := *a\n\n\tif b.HTTP != \"\" {\n\t\tresult.HTTP = b.HTTP\n\t}\n\treturn &result\n}", "func Merge(mta *MTA, mtaExt *EXT) error {\n\terr := chain().\n\t\textendMap(&mta.Parameters, mta.ParametersMetaData, mtaExt.Parameters, mergeRootParametersErrorMsg, mtaExt.ID).\n\t\terr\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = mergeModules(*mta, mtaExt.Modules); err != nil {\n\t\treturn errors.Wrapf(err, mergeExtErrorMsg, mtaExt.ID)\n\t}\n\n\tif err = mergeResources(*mta, mtaExt.Resources); err != nil {\n\t\treturn errors.Wrapf(err, mergeExtErrorMsg, mtaExt.ID)\n\t}\n\n\treturn nil\n}", "func (g *GitLocal) Merge(dir string, commitish string) error {\n\treturn g.GitCLI.Merge(dir, commitish)\n}", "func NewMerger() merger {\n\treturn merger{}\n}", "func Merge(datas interfaces.Sortable) interfaces.Sortable {\n\tmergeSort(datas, 0, datas.Len()-1)\n\treturn datas\n}", "func (t Tags) Merge(key string, value ...string) {\n\tfor _, v := range value {\n\t\tcurrent := t.GetAll(key)\n\t\tfound := false\n\t\tfor _, cv := range current {\n\t\t\tif v == cv {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tt.Add(key, v)\n\t\t}\n\t}\n}", "func (s *segment) merge(oth *segment) {\n\ts.pkt.Data().Merge(oth.pkt.Data())\n\ts.dataMemSize = s.pkt.MemSize()\n\toth.dataMemSize = oth.pkt.MemSize()\n}", "func (t TagSet) Merge(more TagSet) TagSet {\n\tmerged := t[:]\n\treturn append(merged, more...)\n}", "func (m *MockClient) Merge(org, repo string, pr int, details github.MergeDetails) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Merge\", org, repo, pr, details)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (m *Message) UnmarshalMergeJSON(js []byte) error {\n\treturn m.UnmarshalMergeJSONPB(&jsonpb.Unmarshaler{}, js)\n}", "func merge(dst, src Yaml) {\n\tfor k, v := range src {\n\t\tdst[k] = v\n\t}\n}", "func (recipient *BulkRecipient) AddMergeData(key string, value string) *BulkRecipient {\n\tif recipient.MergeData == nil {\n\t\trecipient.MergeData = make(map[string]string)\n\t}\n\trecipient.MergeData[key] = value\n\n\treturn recipient\n}", "func (g *Gitlab) MergePullRequest(ctx context.Context, pullReq scm.PullRequest) error {\n\tpr := pullReq.(pullRequest)\n\n\tshouldRemoveSourceBranch := true\n\t_, _, err := g.glClient.MergeRequests.AcceptMergeRequest(pr.targetPID, pr.iid, &gitlab.AcceptMergeRequestOptions{\n\t\tShouldRemoveSourceBranch: &shouldRemoveSourceBranch,\n\t}, gitlab.WithContext(ctx))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}" ]
[ "0.6245515", "0.61435264", "0.56309336", "0.5469952", "0.54615533", "0.54158205", "0.53108287", "0.5222636", "0.51804227", "0.5180254", "0.5049148", "0.4957558", "0.4927703", "0.48822376", "0.4880821", "0.48743123", "0.48371434", "0.48234692", "0.48100564", "0.4801188", "0.47924456", "0.47749627", "0.47535196", "0.47370833", "0.4682917", "0.46700048", "0.46685594", "0.46644884", "0.46432963", "0.46428013", "0.46375602", "0.4636351", "0.4615403", "0.46093643", "0.45973706", "0.45933405", "0.458391", "0.4579835", "0.45738596", "0.4542364", "0.4533074", "0.45293835", "0.45282012", "0.4508866", "0.45010254", "0.44889444", "0.4472007", "0.44600338", "0.44543213", "0.4444601", "0.44443682", "0.44390327", "0.4431672", "0.44181934", "0.44031662", "0.44028577", "0.4401898", "0.4400917", "0.43980387", "0.4385888", "0.4382026", "0.43814546", "0.43783504", "0.43664712", "0.43656093", "0.43621516", "0.43542355", "0.43472403", "0.4346597", "0.4346381", "0.43449652", "0.43389362", "0.4336883", "0.43356445", "0.43338764", "0.4329719", "0.4315992", "0.431437", "0.4305865", "0.43041325", "0.42953748", "0.42918527", "0.4285976", "0.4279205", "0.42710206", "0.42640525", "0.42628077", "0.42603636", "0.4256772", "0.42479447", "0.42458904", "0.4237322", "0.42348137", "0.42347515", "0.42232728", "0.42181337", "0.42147794", "0.42141494", "0.4209779", "0.4198484" ]
0.6708245
0
PingHandler returns a Handler for Ping Request
func PingHandler(formatter *render.Render) http.HandlerFunc { return func(w http.ResponseWriter, req *http.Request) { formatter.JSON(w, http.StatusOK, struct{ Message string }{"Payments API is up and running !!"}) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *Conn) PingHandler() func(appData string) error {\n\treturn c.handlePing\n}", "func NewPingHandler(config *config.Config) *PingHandler {\n\treturn &PingHandler{cf: config}\n}", "func (app *Application) PingHandler() http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, request *http.Request) {\n\t\t// memStats := &runtime.MemStats{}\n\t\t// runtime.ReadMemStats(memStats)\n\n\t\tcontent, err := json.MarshalIndent(map[string]interface{}{\n\t\t\t\"response\": \"pong\",\n\t\t\t\"numGoroutine\": runtime.NumGoroutine(),\n\t\t\t// \"memStats\": memStats,\n\t\t\t\"goVersion\": runtime.Version(),\n\t\t\t\"version\": app.Config.Version,\n\t\t\t\"stats\": app.stats.Data(),\n\t\t}, \"\", \" \")\n\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(content)\n\t})\n}", "func PingHandler(w http.ResponseWriter, r *http.Request) {\n\tstart := time.Now()\n\tresponse := make(map[string]interface{})\n\tresponse[\"status\"] = \"ok\"\n\tresult := make(map[string]interface{})\n\tresult[\"result\"] = \"Pong\"\n\tresponse[\"data\"] = result\n\tstatus := SendJsonResponseFromInterface(w, r, response)\n\tLigneous.Info(fmt.Sprintf(\"%v %v %v [%v]\", r.RemoteAddr, r.URL.Path, time.Since(start), status))\n}", "func (h *Handler) servePing(w http.ResponseWriter, r *http.Request) {}", "func pingHandler(w http.ResponseWriter, r *http.Request) {\n\tlogrus.Debug(\"/_ping request received\")\n\tw.WriteHeader(http.StatusOK)\n\tw.Write([]byte{'O', 'K'})\n}", "func PingHandler(w http.ResponseWriter, r *http.Request) {\n\tjob := HttpRequest{w: w, r: r}\n\tvar data map[string]interface{}\n\tdata = make(map[string]interface{})\n\tdata[\"status\"] = \"success\"\n\tresult := make(map[string]interface{})\n\tresult[\"result\"] = \"pong\"\n\tresult[\"registered\"] = startTime.UTC()\n\tresult[\"uptime\"] = time.Since(startTime).Seconds()\n\tresult[\"num_cores\"] = runtime.NumCPU()\n\tdata[\"data\"] = result\n\tjs := job.MarshalJsonFromStruct(data)\n\tjob.SendJsonResponse(js)\n}", "func PingMessageHandler(px *Proxy) daemon.Handler {\n\treturn func(w daemon.ResponseWriteCloser, msg daemon.Messager) {\n\t\tif msg.Type() != daemon.PingMsgType {\n\t\t\tpx.Printf(\"Mux error, dispatch %s message to ping message handler\\n\", msg.Type())\n\t\t\treturn\n\t\t}\n\n\t\tpx.ResetPingTimer()\n\n\t\tw.Write(&daemon.PongMessage{Value: \"PONG\"})\n\t\tpx.Debugln(\"Send pong message\")\n\t}\n}", "func PingHandler(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tpong, err := client.Ping().Result()\n\tif err != nil {\n\t\thttp.Error(w, \"Service Unavailable\", http.StatusInternalServerError)\n\t}\n\tjson.NewEncoder(w).Encode(models.OutResponse{Message: pong, Status: http.StatusOK})\n}", "func (c *RawConnectionMock) PingHandler() func(appData string) error {\n\targs := c.Called()\n\treturn args.Get(0).(func(appData string) error)\n}", "func pingHandler(c *gin.Context) {\n\tvar data TargetPayload\n\tif err := c.Bind(&data); err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": err.Error()})\n\t\treturn\n\t}\n\n\tpinger, err := ping.NewPinger(data.Target)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": err.Error()})\n\t\treturn\n\t}\n\n\tpinger.SetPrivileged(true)\n\tpinger.Count = 3\n\tpinger.Run()\n\tstats := pinger.Statistics()\n\tc.JSON(http.StatusOK, gin.H{\"status\": stats})\n}", "func (srv *Server) Ping(w http.ResponseWriter, r *http.Request) {\n\tsrv.wrap(srv.handlePing)(w, r)\n}", "func (m *Monitor) PingHandler(w http.ResponseWriter, r *http.Request) {\n\tif !websocket.IsWebSocketUpgrade(r) {\n\t\tlog.Println(\"No upgrade requested\")\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(\"400 - Request websocket upgrade\"))\n\t\treturn\n\t}\n\tconn, upgradeErr := upgrader.Upgrade(w, r, nil)\n\tif upgradeErr != nil {\n\t\tlog.Print(\"Could not upgrade connection: \", upgradeErr)\n\t\treturn\n\t}\n\tdefer conn.Close()\n\t// Get pings\n\tsub := m.Broker.Subscribe()\n\tdefer m.Broker.Unsubscribe(sub)\n\tfor {\n\t\t// On receive, send on conn\n\t\tselect {\n\t\tcase update := <-sub:\n\t\t\t// Publish to websocket connection\n\t\t\tconn.WriteJSON(update)\n\t\tdefault:\n\t\t}\n\t}\n\tlog.Print(\"Exit\")\n}", "func pingHandler(w http.ResponseWriter, r *http.Request) {\n\tw.Write([]byte(\"pong\\n\"))\n}", "func pingHandler(formatter *render.Render) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, req *http.Request) {\n\t\tformatter.JSON(w, http.StatusOK, struct{ Ping string }{\"OK\"})\n\t}\n}", "func (p *Peer) pingHandler() {\n\tpingTicker := time.NewTicker(pingInterval)\n\tdefer pingTicker.Stop()\n\nout:\n\tfor {\n\t\tselect {\n\t\tcase <-pingTicker.C:\n\t\t\tnonce, err := wire.RandomUint64()\n\t\t\tif err != nil {\n\t\t\t\tlog.Errorf(\"Not sending ping to %s: %v\", p, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tp.QueueMessage(wire.NewMsgPing(nonce), nil)\n\n\t\tcase <-p.quit:\n\t\t\tbreak out\n\t\t}\n\t}\n}", "func (s *SimpleHandler) Ping() (*computing.StatusOfService, error) {\n\tfmt.Print(\"Pong!\\n\")\n\treturn &computing.StatusOfService{Version: \"0.0.1\", Network: \"9090\"}, nil\n}", "func (h *Handler) Ping() error {\n\tresult := h.ring.Ping()\n\tif result.Val() != \"PONG\" {\n\t\treturn ErrPing\n\t}\n\n\treturn nil\n}", "func (h *Handler) Ping(w http.ResponseWriter, r *http.Request) {\n\tresponse.Message(200, w, \"Pong\")\n}", "func (s *Server) handlePing(request []byte) {\n\tvar payload serverutil.MsgPing\n\tif err := getPayload(request, &payload); err != nil {\n\t\tlog.Panic(err)\n\t}\n\taddr := payload.AddrSender.String()\n\tp, _ := s.GetPeer(addr)\n\tp.IncreaseBytesReceived(uint64(len(request)))\n\ts.AddPeer(p)\n\ts.Log(true, \"Ping received from :\", addr)\n\ts.sendPong(payload.AddrSender)\n\n}", "func pingHandler(w http.ResponseWriter, _ *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"text/plain\")\n\tw.Write([]byte(\"pong\"))\n}", "func pingHandler(w http.ResponseWriter, r *http.Request) {\n\tlogrus.Info(\"Check Availability\")\n\n\th := `<!DOCTYPE html><html><head><title>Availability Check</title>\n<style type=\"text/css\">\ndiv {font-size: 16px; font-weight: bold;}\n</style></head><body>\n<div>PONG</div>\n</body></html>`\n\n\t// return http code 200\n\tw.WriteHeader(http.StatusOK)\n\t// output\n\tfmt.Fprint(w, gohtml.Format(h))\n}", "func (s *ProxyService) Ping(req *rpc.PingRequest, server rpc.Hello_PingServer) error {\n\treturn nil\n}", "func (c *Conn) SetPingHandler(h func(appData string) error) {\n\tif h == nil {\n\t\th = func(message string) error {\n\t\t\terr := c.WriteControl(PongMessage, []byte(message), time.Now().Add(writeWait))\n\t\t\tif err == ErrCloseSent {\n\t\t\t\treturn nil\n\t\t\t} else if e, ok := err.(net.Error); ok && e.Temporary() {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t}\n\tc.handlePing = h\n}", "func (s *status) HandlePing(c *gin.Context) {\n\tc.String(http.StatusOK, \"pong\")\n}", "func (c *RawConnectionMock) SetPingHandler(h func(appData string) error) {\n\tc.Called(h)\n}", "func RegisterPingEndpoint(mux mux) {\n\tmux.Handle(\"/ping\", adaptCheckToHandler(PingHealthz.Check))\n}", "func pingHandler(server *Server, client *Client, msg ircmsg.IrcMessage) bool {\n\tclient.Send(nil, server.name, \"PONG\", msg.Params...)\n\treturn false\n}", "func (con *GuestHandler) Ping(c echo.Context) (err error) {\n\t// Server is up and running, return OK!\n\treturn c.String(http.StatusOK, \"Pong\")\n}", "func (s *Server) handlePing(p Peer, ping *payload.Ping) error {\n\terr := p.HandlePing(ping)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = s.requestBlocksOrHeaders(p)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn p.EnqueueP2PMessage(NewMessage(CMDPong, payload.NewPing(s.chain.BlockHeight(), s.id)))\n}", "func (a *api) h_GET_ping(c *gin.Context) {\n\ta.logger.Debug(\"GET /ping\")\n\t//\terr := a.EmSender.Send(\"[email protected]\", \"Hello Dima\", \"How are U?/nDima.\")\n\t//\tif err != nil {\n\t//\t\ta.logger.Error(\"Could not send err=\", err)\n\t//\t}\n\tc.String(http.StatusOK, \"pong URL conversion is \"+composeURI(c.Request, \"\"))\n}", "func (s *Session) AddPingHandler(fn func(Ping, *Session)) {\n\ts.handlers.pingHandler = fn\n}", "func (s *Server) Ping(ctx context.Context, in *service.Empty) (*service.Empty, error) {\n\treturn &service.Empty{}, nil\n}", "func (c *PingMiddleware) ServeHTTP(w http.ResponseWriter, r *http.Request, next traffic.NextMiddlewareFunc) (http.ResponseWriter, *http.Request) {\n if r.URL.Path == \"/ping\" {\n fmt.Fprint(w, \"pong\\n\")\n\n return w, r\n }\n\n if nextMiddleware := next(); nextMiddleware != nil {\n arw := w.(*traffic.AppResponseWriter)\n arw.SetVar(\"ping\", \"pong\")\n w, r = nextMiddleware.ServeHTTP(w, r, next)\n }\n\n return w, r\n}", "func Handler() *healthHandler {\n\treturn handler\n}", "func (s *Server) Ping(w http.ResponseWriter, r *http.Request) {\n\tif _, err := w.Write([]byte(\"pong\")); err != nil {\n\t\ts.Log.WithField(\"err\", err).Error(\"failed to write http response\")\n\t}\n}", "func (p *KiteHTTPPinger) Ping() Status {\n\tres, err := p.Client.Get(p.Address)\n\tif err != nil {\n\t\treturn Failure\n\t}\n\tdefer res.Body.Close()\n\n\tresData, err := ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\treturn Failure\n\t}\n\n\tif string(resData) != kiteHTTPResponse {\n\t\treturn Failure\n\t}\n\n\treturn Success\n}", "func (h *Hub) Ping(ctx context.Context, _ *pb.PingRequest) (*pb.PingReply, error) {\n\tlog.G(h.ctx).Info(\"handling Ping request\")\n\treturn &pb.PingReply{}, nil\n}", "func (c *Conn) PongHandler() func(appData string) error {\n\treturn c.handlePong\n}", "func pingHandler(formatter *render.Render) http.HandlerFunc {\n return func(w http.ResponseWriter, req *http.Request) {\n formatter.JSON(w, http.StatusOK, struct{ Ping string }{\"Pong\"})\n }\n}", "func (s *Server) Ping(context.Context, *empty.Empty) (*empty.Empty, error) {\n\treturn &empty.Empty{}, nil\n}", "func (s *Server) Ping(ctx context.Context, _ *yages.Empty) (*yages.Content, error) {\n\treturn &yages.Content{Text: \"pong\"}, nil\n}", "func Ping(w http.ResponseWriter, _ *http.Request) {\n\tfmt.Fprint(w, \"OK\")\n}", "func NewPing(eventID *OutgoingEventID) *Ping {\n\treturn &Ping{\n\t\tOutgoingEvent: OutgoingEvent{\n\t\t\tID: eventID.Next(),\n\t\t\tTypedEvent: event.TypedEvent{Type: \"ping\"},\n\t\t},\n\t}\n}", "func (s *Server) Ping(ctx context.Context, ping *pingpong.PingRequest) (*pingpong.PongResponse, error) {\n\treturn &pingpong.PongResponse{\n\t\tOk: true,\n\t}, nil\n}", "func Ping(w http.ResponseWriter, r *http.Request) {\n\turl := r.FormValue(\"url\")\n\n\tfmt.Println(\"Ping\", url)\n\n\tresp, err := http.Head(url)\n\tif err != nil {\n\t\thttp.Error(w, \"Unable to connect\", 500)\n\t\treturn\n\t}\n\n\tw.WriteHeader(resp.StatusCode)\n}", "func NewHandler() Handler {\n\th := &basicHandler{\n\t\tlivenessChecks: make(map[string]Check),\n\t\treadinessChecks: make(map[string]Check),\n\t}\n\th.Handle(\"/live\", http.HandlerFunc(h.LiveEndpoint))\n\th.Handle(\"/ready\", http.HandlerFunc(h.ReadyEndpoint))\n\treturn h\n}", "func (p *Proxy) Handler() http.Handler {\n\tmux := goji.NewMux()\n\n\tmux.HandleFuncC(pat.Get(\"/healthcheck\"), func(c context.Context, w http.ResponseWriter, r *http.Request) {\n\t\tw.Write([]byte(\"ok\\n\"))\n\t})\n\n\tmux.Handle(pat.Post(\"/import\"), handleProxy(p))\n\n\tmux.Handle(pat.Get(\"/debug/pprof/cmdline\"), http.HandlerFunc(pprof.Cmdline))\n\tmux.Handle(pat.Get(\"/debug/pprof/profile\"), http.HandlerFunc(pprof.Profile))\n\tmux.Handle(pat.Get(\"/debug/pprof/symbol\"), http.HandlerFunc(pprof.Symbol))\n\tmux.Handle(pat.Get(\"/debug/pprof/trace\"), http.HandlerFunc(pprof.Trace))\n\t// TODO match without trailing slash as well\n\tmux.Handle(pat.Get(\"/debug/pprof/*\"), http.HandlerFunc(pprof.Index))\n\n\treturn mux\n}", "func handler() http.Handler {\n\th := http.NewServeMux()\n\th.HandleFunc(\"/healthcheck\", healthFunc)\n\treturn h\n}", "func NewPing(pingService postmand.PingService, logger *zap.Logger) *Ping {\n\treturn &Ping{\n\t\tpingService: pingService,\n\t\tlogger: logger,\n\t}\n}", "func (h *Handler) HealthHandler(w http.ResponseWriter, r *http.Request) {\n\tif err := h.DB.Ping(); err != nil {\n\t\tlog.Printf(\"health check failed: %s\\n\", err)\n\t\twriteHTTPResponse(w, http.StatusInternalServerError, map[string]string{\"message\": \"I'm unhealthy\"})\n\t} else {\n\t\twriteHTTPResponse(w, http.StatusOK, map[string]string{\"message\": \"I'm healthy\"})\n\t}\n}", "func init() {\n\trouter().GET(\"/ping\", pingHandler)\n}", "func PulseHandler(db *postgres.DB) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\terr := db.Ping()\n\t\tif err != nil {\n\t\t\thttp.Error(w, \"failed to connect to DB\", http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tfmt.Fprintf(w, \"ok\")\n\t})\n}", "func (s *server) Ping(ctx context.Context, in *tt.NodeAddress) (*tt.Response, error) {\n\tip := net.ParseIP(in.IP)\n\tif ip == nil {\n\t\treturn nil, errors.New(\"not a valid ip address to ping\")\n\t}\n\tn, err := smudge.CreateNodeByIP(net.ParseIP(in.IP), uint16(in.Port))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = smudge.PingNode(n)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &tt.Response{Code: tt.Response_OK}, nil\n}", "func Ping(addr string) (PingResponse, error) {\n\treturn PingTimeout(addr, DEFAULT_TIMEOUT)\n}", "func (h *PingHandler) Run(ctx scope.Context) error {\n\treturn nil\n}", "func NewPylonHealthHandler(p *Pylon) http.HandlerFunc {\r\n\treturn func(w http.ResponseWriter, r *http.Request) {\r\n\t\tt, err := template.New(\"PylonHealthTemplate\").Parse(pylonTemplate)\r\n\t\tif err != nil {\r\n\t\t\tlogError(err.Error())\r\n\t\t}\r\n\t\tif err := t.Execute(w, getRenders(p)); err != nil {\r\n\t\t\tlogError(\"Could not render the HTML template\")\r\n\t\t}\r\n\t\tlogDebug(\"Served heath page HTML\")\r\n\t}\r\n}", "func Handler(t time.Time) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\terr := db.DB().Ping()\n\t\tif err != nil {\n\t\t\terr = errors.Wrap(err, errDbPing.Error())\n\t\t\tlog.Printf(\"%+v\", err)\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t}\n\t\t// Construct response body.\n\t\tresp := struct {\n\t\t\tDbAvailable bool `json:\"db_available\"`\n\t\t\tStartupTime string `json:\"startup_time\"`\n\t\t\tStatus string `json:\"status\"`\n\t\t}{\n\t\t\tDbAvailable: true,\n\t\t\tStartupTime: t.Format(time.RFC3339),\n\t\t\tStatus: statusOkMsg,\n\t\t}\n\t\tb, err := json.Marshal(resp)\n\t\tif err != nil {\n\t\t\terr = errors.Wrap(err, errMarshalResp.Error())\n\t\t\tlog.Printf(\"%+v\", err)\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t} else {\n\t\t\t_, err = w.Write(b)\n\t\t\tif err != nil {\n\t\t\t\terr = errors.Wrap(err, errWriteResp.Error())\n\t\t\t\tlog.Printf(\"%+v\", err)\n\t\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\t}\n\t\t}\n\t})\n}", "func (s *Service) Ping(ctx context.Context, e *empty.Empty) (*empty.Empty, error) {\n\treturn &empty.Empty{}, nil\n}", "func (b *Backend) Ping(ctx context.Context, req *dashboard.Hello) (*dashboard.Pong, error) {\n\tif req.GetMessage() == \"Expect-Error\" {\n\t\treturn nil, newStatus(codes.Canceled, \"operation canceled because client sent Expect-Error\").err()\n\t}\n\treturn &dashboard.Pong{\n\t\tReply: req.GetMessage(),\n\t}, nil\n}", "func ping_handler(ip_string string) string{\n\n\t_, err := net.ResolveIPAddr(\"ip4:icmp\", ip_string)\n\tif err != nil{\n\t\tlog.Println(err)\n\t\treturn \"Fail\"\n\t}\n\n\tcommand_line := \"ping\"\n\tdata, err := exec.Command(command_line,ip_string).Output()\n\tif err != nil{\n\t\tlog.Println(err)\n\t\treturn \"Fail\"\n\t}\n\n\tdata_split := strings.Split(string(data),\"\\n\")\n\tre_fail := regexp.MustCompile(\"failed\")\n\tfor counter := range data_split{\n\t\tif len(re_fail.FindString(data_split[counter])) > 0{\n\t\t\treturn \"Fail\"\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn \"OK\"\n}", "func (r *BreakerRPC) Ping(c rcontext.Context, arg *BreakerReq, res *BreakerReply) (err error) {\n\tif rand.Int31n(100) < 40 {\n\t\treturn ecode.ServerErr\n\t}\n\tres.Success = true\n\treturn\n}", "func (c *RawConnectionMock) PongHandler() func(appData string) error {\n\targs := c.Called()\n\treturn args.Get(0).(func(appData string) error)\n}", "func pingResponseHandler(_ *PingMessageTask, pdu *libcoap.Pdu) {\n\tlog.WithField(\"Type\", pdu.Type).WithField(\"Code\", pdu.Code).Debug(\"Ping Ack\")\n}", "func (s *Status) Ping(args struct{}, reply *struct{}) error {\n\treturn nil\n}", "func (m *Miner) Ping(ctx context.Context, _ *pb.PingRequest) (*pb.PingReply, error) {\n\tlog.G(m.ctx).Info(\"got ping request from Hub\")\n\treturn &pb.PingReply{}, nil\n}", "func (c *Client) Ping() error {\n\treturn c.request(\"GET\", \"/ping\", nil)\n}", "func NewHandler() *Handler {\n\th := &Handler{\n\t\trouter: httprouter.New(),\n\t\tLogger: log15.New(),\n\t}\n\th.router.Handler(\"GET\", status.Path, status.Handler(h.healthStatus))\n\th.router.GET(\"/status\", h.handleGetStatus)\n\th.router.POST(\"/stop\", h.handlePostStop)\n\treturn h\n}", "func (handler *PreflightCheckHandler) Handler() util.Handler {\n\treturn handler.Handle\n}", "func newHealthServiceHandler(db *sqlx.DB) (http.Handler, *healthz.StatusChecker) {\n\tstatus := healthz.NewStatusChecker(healthz.Healthy)\n\thealthMux := healthz.NewHealthServiceHandler(healthz.NewCheckers(), healthz.NewCheckers(status, healthz.NewPingChecker(db)))\n\n\treturn healthMux, status\n}", "func (this *PreferenceController) Ping(writer http.ResponseWriter, request *http.Request) *result.WebResult {\n\n\treturn this.Success(core.VERSION)\n\n}", "func (this *PreferenceController) Ping(writer http.ResponseWriter, request *http.Request) *result.WebResult {\n\n\treturn this.Success(core.VERSION)\n\n}", "func (this *Protocol) HandlePing(args *PingArgs, reply *PingReply) error {\n\tif this.peerList != nil {\n\t\tthis.peerList.DiscoveredPeer(args.Me)\n\t}\n\treturn nil\n}", "func (MainController) Ping(c *gin.Context) {\n\tc.JSON(200, gin.H{\n\t\t\"message\": \"ashwat watches fifty shades of grey\",\n\t})\n}", "func Ping(w http.ResponseWriter, r *http.Request) {\n\t_, err := w.Write([]byte(\"pong\"))\n\tif err != nil {\n\t\tlog.Print(err.Error())\n\t}\n}", "func (r *Router) handler(h Handler) http.Handler {\n\treturn &handler{h}\n}", "func (r *Router) handler(h Handler) http.Handler {\n\treturn &handler{h}\n}", "func (a API) Ping(cmd *None) (e error) {\n\tRPCHandlers[\"ping\"].Call <-API{a.Ch, cmd, nil}\n\treturn\n}", "func GetHealthHandler(ghp probe.GetHealthParams) middleware.Responder {\n\treturn probe.NewGetHealthOK()\n}", "func (p *Ping) Ping(target p2pcrypto.PublicKey, msg string) (string, error) {\n\tvar response string\n\treqid := crypto.NewUUID()\n\tping := &pb.Ping{\n\t\tReqID: reqid[:],\n\t\tReq: true,\n\t\tMessage: msg,\n\t}\n\tpchan, err := p.sendRequest(target, reqid, ping)\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\ttimer := time.NewTimer(PingTimeout)\n\tselect {\n\tcase res := <-pchan:\n\t\tresponse = res.Message\n\t\tp.pendMuxtex.Lock()\n\t\tdelete(p.pending, reqid)\n\t\tp.pendMuxtex.Unlock()\n\tcase <-timer.C:\n\t\treturn response, errPingTimedOut\n\t}\n\n\treturn response, nil\n}", "func Ping(c *gin.Context) {\n\tc.JSON(200, gin.H{\n\t\t\"message\": \"pong\",\n\t})\n}", "func (h *Health) Handler() http.Handler {\n\treturn http.HandlerFunc(h.HandlerFunc)\n}", "func Ping(addr string) (PingResponse, error) {\n\treturn ping(addr, DefaultTimeout)\n}", "func newHealthHandler(s *server) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tif r.Method != \"GET\" {\n\t\t\tw.WriteHeader(http.StatusMethodNotAllowed)\n\t\t\treturn\n\t\t}\n\n\t\tctx := context.TODO()\n\t\tif r.URL.Path == \"/local\" || r.URL.Path == \"/local/\" {\n\t\t\thandleLocalStatus(ctx, s, w, r)\n\t\t\treturn\n\t\t}\n\n\t\tif r.URL.Path == \"/history\" || r.URL.Path == \"/history/\" {\n\t\t\thandleHistory(ctx, s, w, r)\n\t\t\treturn\n\t\t}\n\n\t\tstatus, err := s.Status(ctx, nil)\n\t\tif err != nil {\n\t\t\troundtrip.ReplyJSON(w, http.StatusServiceUnavailable, map[string]string{\"error\": err.Error()})\n\t\t\treturn\n\t\t}\n\n\t\thttpStatus := http.StatusOK\n\t\tif isDegraded(*status.GetStatus()) {\n\t\t\thttpStatus = http.StatusServiceUnavailable\n\t\t}\n\n\t\troundtrip.ReplyJSON(w, httpStatus, status.GetStatus())\n\t}\n}", "func (s *Server) Ping(ctx context.Context, in *PingMessage) (*PingMessage, error) {\n\tlog.Printf(\"Receive message %s\\n\", in.Ping)\n\n\treturn &PingMessage{Ping: \"pong\"}, nil\n}", "func TestPingRoute(t *testing.T) {\n\t// Build the routes.\n\trouter := setupRouter()\n\n\t// Exploit a standard library Writer that records what you write to it.\n\tw := httptest.NewRecorder()\n\t// Construct an http request that will exercise a handled route.\n\treq, _ := http.NewRequest(\"GET\", \"/ping\", nil)\n\t// Instruct the router to handle this request.\n\trouter.ServeHTTP(w, req)\n\n\t// Scrutinise what the writer has recorded.\n\tassert.Equal(t, 200, w.Code) // Response code OK\n\tassert.Equal(t, \"pong\", w.Body.String()) // Body content.\n}", "func NewHandler(service Service) Handler {\n\treturn Handler{service: service}\n}", "func NewHandler(bouncer security.BouncerService,\n\tstatus *portainer.Status,\n\tdemoService *demo.Service,\n\tdataStore dataservices.DataStore,\n\tupgradeService upgrade.Service) *Handler {\n\n\th := &Handler{\n\t\tRouter: mux.NewRouter(),\n\t\tdataStore: dataStore,\n\t\tdemoService: demoService,\n\t\tstatus: status,\n\t\tupgradeService: upgradeService,\n\t}\n\n\trouter := h.PathPrefix(\"/system\").Subrouter()\n\n\tadminRouter := router.PathPrefix(\"/\").Subrouter()\n\tadminRouter.Use(bouncer.AdminAccess)\n\n\tadminRouter.Handle(\"/upgrade\", httperror.LoggerHandler(h.systemUpgrade)).Methods(http.MethodPost)\n\n\tauthenticatedRouter := router.PathPrefix(\"/\").Subrouter()\n\tauthenticatedRouter.Use(bouncer.AuthenticatedAccess)\n\n\tauthenticatedRouter.Handle(\"/version\", http.HandlerFunc(h.version)).Methods(http.MethodGet)\n\tauthenticatedRouter.Handle(\"/nodes\", httperror.LoggerHandler(h.systemNodesCount)).Methods(http.MethodGet)\n\tauthenticatedRouter.Handle(\"/info\", httperror.LoggerHandler(h.systemInfo)).Methods(http.MethodGet)\n\n\tpublicRouter := router.PathPrefix(\"/\").Subrouter()\n\tpublicRouter.Use(bouncer.PublicAccess)\n\n\tpublicRouter.Handle(\"/status\", httperror.LoggerHandler(h.systemStatus)).Methods(http.MethodGet)\n\n\t// Deprecated /status endpoint, will be removed in the future.\n\th.Handle(\"/status\",\n\t\tbouncer.PublicAccess(httperror.LoggerHandler(h.statusInspectDeprecated))).Methods(http.MethodGet)\n\th.Handle(\"/status/version\",\n\t\tbouncer.AuthenticatedAccess(http.HandlerFunc(h.versionDeprecated))).Methods(http.MethodGet)\n\th.Handle(\"/status/nodes\",\n\t\tbouncer.AuthenticatedAccess(httperror.LoggerHandler(h.statusNodesCountDeprecated))).Methods(http.MethodGet)\n\n\treturn h\n}", "func makeHealthHandler() func(http.ResponseWriter, *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tswitch r.Method {\n\t\tcase http.MethodGet:\n\t\t\tif atomic.LoadInt32(&acceptingConnections) == 0 || lockFilePresent() == false {\n\t\t\t\tw.WriteHeader(http.StatusServiceUnavailable)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tw.WriteHeader(http.StatusOK)\n\t\t\tw.Write([]byte(\"OK\"))\n\n\t\t\tbreak\n\t\tdefault:\n\t\t\tw.WriteHeader(http.StatusMethodNotAllowed)\n\t\t}\n\t}\n}", "func NewHandler(entrypoint, network, address string) http.Handler {\n\tconnFactory := gofast.SimpleConnFactory(network, address)\n\tpool := gofast.NewClientPool(\n\t\tgofast.SimpleClientFactory(connFactory),\n\t\t10,\n\t\t60*time.Second,\n\t)\n\th := gofast.NewHandler(\n\t\tgofast.NewFileEndpoint(entrypoint)(gofast.BasicSession),\n\t\tpool.CreateClient,\n\t)\n\treturn h\n}", "func NewPingService(p Pinger) *PingService {\n\treturn &PingService{Pinger: p}\n}", "func BindPing(app *app.App) {\n\tapp.Bind(PingCommand, pingCmd)\n}", "func NewHandler(next http.Handler) Handler {\n\treturn Handler{\n\t\tNext: next,\n\t\tLogger: JSONLogger,\n\t\tSkip: SkipHealthEndpoint,\n\t}\n}", "func (client *activeClient) Ping(c *ishell.Context) error {\n\treturn client.RPC.Call(\"API.Ping\", void, &void)\n}", "func Handler() (Response, error) {\n\treturn NewResponse(\"Hello, World\"), nil\n}", "func (s *Service) Ping(c context.Context) (err error) {\n\treturn\n}", "func (r *Route) handler(h Handler) http.Handler {\n\treturn &handler{h}\n}", "func HealthHandler() http.Handler {\n\treturn &healthHandler{}\n}", "func ping(c *bm.Context) {\n\tc.JSON(nil, srv.Ping(c))\n}", "func NewPoolingHandler(validator auth.Validator, handler MessagesRetriever, logger log.FieldLogger) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tlogger.Info(\"received new pooling request\")\n\t\tw.Header().Set(\"Access-Control-Allow-Origin\", r.Header.Get(\"Origin\"))\n\t\tw.Header().Set(\"Access-Control-Allow-Credentials\", \"true\")\n\n\t\tcookie, err := r.Cookie(config.TokenName)\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\tlogger.Warn(\"request with no token\")\n\t\t\treturn\n\t\t}\n\n\t\tusername, err := validator.Validate(cookie.Value)\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\tlogger.Warn(\"validate error\\t\", err)\n\t\t\treturn\n\t\t}\n\t\tlogger.Info(\"validated user\\t\", username)\n\n\t\tmsgs := handler.GetUserMessages(username)\n\t\tif len(msgs) == 0 {\n\t\t\tw.WriteHeader(http.StatusNoContent)\n\t\t\tlogger.Warn(\"no messages found for user:\\t\", username)\n\t\t\treturn\n\t\t}\n\t\tlogger.Debug(fmt.Sprintf(\"found %v messages for user:\\t%s\", len(msgs), username))\n\n\t\tcontent, err := json.Marshal(msgs)\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\tlogger.Warn(\"err while marshaling the messages:\\t\", err)\n\t\t\treturn\n\t\t}\n\n\t\t_, err = w.Write(content)\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\tlogger.Warn(\"err while writing the messages to the respond:\\t\", err)\n\t\t\treturn\n\t\t}\n\n\t}\n}" ]
[ "0.7720891", "0.76226425", "0.74403113", "0.73901325", "0.7134259", "0.71059835", "0.70995325", "0.7083865", "0.7082013", "0.70737785", "0.701905", "0.70126235", "0.69297284", "0.68595296", "0.6839317", "0.6711285", "0.6630836", "0.6571368", "0.6465587", "0.6418863", "0.6384126", "0.6268736", "0.6244683", "0.6235231", "0.6212272", "0.6166612", "0.6115513", "0.6083123", "0.60428303", "0.6005201", "0.6000954", "0.5997073", "0.59965944", "0.59956974", "0.59063524", "0.58958197", "0.58751976", "0.58508533", "0.58250266", "0.58231014", "0.58151823", "0.57991064", "0.5786223", "0.5778561", "0.5774965", "0.57607037", "0.57482123", "0.5739548", "0.5737286", "0.5733693", "0.57334405", "0.57150954", "0.5710583", "0.57067436", "0.56867987", "0.5676566", "0.56505704", "0.5649459", "0.56405413", "0.5609511", "0.5605195", "0.56039876", "0.560252", "0.55985004", "0.5598339", "0.55941516", "0.557795", "0.5569838", "0.5565967", "0.555821", "0.5542941", "0.5542941", "0.55381703", "0.5535848", "0.55310273", "0.552447", "0.552447", "0.55159104", "0.5515801", "0.55122936", "0.55065936", "0.54931235", "0.54809874", "0.5480602", "0.54667187", "0.5460067", "0.5453946", "0.54530334", "0.54497373", "0.5448379", "0.5445333", "0.54354346", "0.5432278", "0.543187", "0.5427062", "0.54137963", "0.5401744", "0.5400784", "0.5399679", "0.5399383" ]
0.6801965
15
MakePaymentHandler returns a Handler for making a payment
func MakePaymentHandler(formatter *render.Render) http.HandlerFunc { return func(w http.ResponseWriter, req *http.Request) { log.Printf("Make payment handler function") fmt.Println("Make payment handler router function") var payment model.Payment _ = json.NewDecoder(req.Body).Decode(&payment) transactionID, _ := service.MakePayment(&payment) formatter.JSON(w, http.StatusOK, struct { Success bool Message string TransactionID int }{ true, "Fee Payment Transaction Recorded Successfully", transactionID, }) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func MakePaymentHandlers(r *mux.Router, n negroni.Negroni, service payment.UseCase) {\n\tpaymentsRouter := r.PathPrefix(\"/v1/payments\").Subrouter()\n\tpaymentsRouter.Handle(\"\", n.With(\n\t\tnegroni.Wrap(paymentFindAll(service)),\n\t)).Methods(\"GET\", \"OPTIONS\")\n\n\tpaymentsRouter.Handle(\"/{paymentID}\", n.With(\n\t\tnegroni.Wrap(paymentFind(service)),\n\t)).Methods(\"GET\", \"OPTIONS\")\n\n\tpaymentsRouter.Handle(\"\", n.With(\n\t\tnegroni.Wrap(paymentCreate(service)),\n\t)).Methods(\"POST\", \"OPTIONS\")\n\n\tpaymentsRouter.Handle(\"/{paymentID}\", n.With(\n\t\tnegroni.Wrap(paymentUpdate(service)),\n\t)).Methods(\"PUT\", \"OPTIONS\")\n\n\tpaymentsRouter.Handle(\"/{paymentID}\", n.With(\n\t\tnegroni.Wrap(paymentDelete(service)),\n\t)).Methods(\"DELETE\", \"OPTIONS\")\n}", "func (h CreatePaymentRequestHandler) Handle(params paymentrequestop.CreatePaymentRequestParams) middleware.Responder {\n\t// TODO: authorization to create payment request\n\n\treturn h.AuditableAppContextFromRequestWithErrors(params.HTTPRequest,\n\t\tfunc(appCtx appcontext.AppContext) (middleware.Responder, error) {\n\n\t\t\tpayload := params.Body\n\t\t\tif payload == nil {\n\t\t\t\terr := apperror.NewBadDataError(\"Invalid payment request: params Body is nil\")\n\t\t\t\terrPayload := payloads.ClientError(handlers.SQLErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\t\t\t\tappCtx.Logger().Error(err.Error(), zap.Any(\"payload\", errPayload))\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestBadRequest().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tappCtx.Logger().Info(\"primeapi.CreatePaymentRequestHandler info\", zap.String(\"pointOfContact\", params.Body.PointOfContact))\n\n\t\t\tmoveTaskOrderIDString := payload.MoveTaskOrderID.String()\n\t\t\tmtoID, err := uuid.FromString(moveTaskOrderIDString)\n\t\t\tif err != nil {\n\t\t\t\tappCtx.Logger().Error(\"Invalid payment request: params MoveTaskOrderID cannot be converted to a UUID\",\n\t\t\t\t\tzap.String(\"MoveTaskOrderID\", moveTaskOrderIDString), zap.Error(err))\n\t\t\t\t// create a custom verrs for returning a 422\n\t\t\t\tverrs :=\n\t\t\t\t\t&validate.Errors{Errors: map[string][]string{\n\t\t\t\t\t\t\"move_id\": {\"id cannot be converted to UUID\"},\n\t\t\t\t\t},\n\t\t\t\t\t}\n\t\t\t\terrPayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tisFinal := false\n\t\t\tif payload.IsFinal != nil {\n\t\t\t\tisFinal = *payload.IsFinal\n\t\t\t}\n\n\t\t\tpaymentRequest := models.PaymentRequest{\n\t\t\t\tIsFinal: isFinal,\n\t\t\t\tMoveTaskOrderID: mtoID,\n\t\t\t}\n\n\t\t\t// Build up the paymentRequest.PaymentServiceItems using the incoming payload to offload Swagger data coming\n\t\t\t// in from the API. These paymentRequest.PaymentServiceItems will be used as a temp holder to process the incoming API data\n\t\t\tvar verrs *validate.Errors\n\t\t\tpaymentRequest.PaymentServiceItems, verrs, err = h.buildPaymentServiceItems(appCtx, payload)\n\n\t\t\tif err != nil || verrs.HasAny() {\n\n\t\t\t\tappCtx.Logger().Error(\"could not build service items\", zap.Error(err))\n\t\t\t\t// TODO: do not bail out before creating the payment request, we need the failed record\n\t\t\t\t// we should create the failed record and store it as failed with a rejection\n\t\t\t\terrPayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(errPayload), err\n\t\t\t}\n\n\t\t\tcreatedPaymentRequest, err := h.PaymentRequestCreator.CreatePaymentRequestCheck(appCtx, &paymentRequest)\n\t\t\tif err != nil {\n\t\t\t\tappCtx.Logger().Error(\"Error creating payment request\", zap.Error(err))\n\t\t\t\tswitch e := err.(type) {\n\t\t\t\tcase apperror.InvalidCreateInputError:\n\t\t\t\t\tverrs := e.ValidationErrors\n\t\t\t\t\tdetail := err.Error()\n\t\t\t\t\tpayload := payloads.ValidationError(detail, h.GetTraceIDFromRequest(params.HTTPRequest), verrs)\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(payload), err\n\n\t\t\t\tcase apperror.NotFoundError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.NotFoundMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestNotFound().WithPayload(payload), err\n\t\t\t\tcase apperror.ConflictError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.ConflictErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestConflict().WithPayload(payload), err\n\t\t\t\tcase apperror.InvalidInputError:\n\t\t\t\t\tpayload := payloads.ValidationError(err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest), &validate.Errors{})\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestUnprocessableEntity().WithPayload(payload), err\n\t\t\t\tcase apperror.QueryError:\n\t\t\t\t\tif e.Unwrap() != nil {\n\t\t\t\t\t\t// If you can unwrap, log the internal error (usually a pq error) for better debugging\n\t\t\t\t\t\tappCtx.Logger().Error(\"primeapi.CreatePaymentRequestHandler query error\", zap.Error(e.Unwrap()))\n\t\t\t\t\t}\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestInternalServerError().WithPayload(\n\t\t\t\t\t\tpayloads.InternalServerError(nil, h.GetTraceIDFromRequest(params.HTTPRequest))), err\n\n\t\t\t\tcase *apperror.BadDataError:\n\t\t\t\t\tpayload := payloads.ClientError(handlers.BadRequestErrMessage, err.Error(), h.GetTraceIDFromRequest(params.HTTPRequest))\n\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestBadRequest().WithPayload(payload), err\n\t\t\t\tdefault:\n\t\t\t\t\tappCtx.Logger().Error(\"Payment Request\",\n\t\t\t\t\t\tzap.Any(\"payload\", payload))\n\t\t\t\t\treturn paymentrequestop.NewCreatePaymentRequestInternalServerError().WithPayload(\n\t\t\t\t\t\tpayloads.InternalServerError(nil, h.GetTraceIDFromRequest(params.HTTPRequest))), err\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturnPayload := payloads.PaymentRequest(createdPaymentRequest)\n\t\t\tappCtx.Logger().Info(\"Successful payment request creation for mto ID\", zap.String(\"moveID\", moveTaskOrderIDString))\n\t\t\treturn paymentrequestop.NewCreatePaymentRequestCreated().WithPayload(returnPayload), nil\n\t\t})\n}", "func (fn CreatePaymentRequestHandlerFunc) Handle(params CreatePaymentRequestParams) middleware.Responder {\n\treturn fn(params)\n}", "func paymentCreate(service payment.UseCase) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tdefer r.Body.Close()\n\t\tvar p *entity.Payment\n\t\terr := json.NewDecoder(r.Body).Decode(&p)\n\t\tif err != nil {\n\t\t\trespondWithError(w, http.StatusBadRequest, \"Invalid request payload\")\n\t\t\treturn\n\t\t}\n\t\tp.ID, err = service.Store(p)\n\t\tif err != nil {\n\t\t\trespondWithError(w, http.StatusInternalServerError, err.Error())\n\t\t\treturn\n\t\t}\n\t\trespondWithJSON(w, http.StatusCreated, p)\n\t})\n}", "func (s *Server) handleDashboardPayment() http.HandlerFunc {\n\tvar o sync.Once\n\tvar tpl *template.Template\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tctx, logger := GetLogger(s.getCtx(r))\n\t\to.Do(func() {\n\t\t\ttpl = s.loadWebTemplateDashboard(ctx, \"payment.html\")\n\t\t})\n\t\tctx, provider, data, errs, ok := s.createTemplateDataDashboard(w, r.WithContext(ctx), tpl, true)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tdata[TplParamActiveNav] = provider.GetURLBookings()\n\n\t\t//load the booking\n\t\tidStr := r.FormValue(URLParams.BookID)\n\t\tctx, book, ok := s.loadTemplateBook(w, r.WithContext(ctx), tpl, data, errs, idStr, true, false)\n\t\tif !ok {\n\t\t\ts.SetCookieErr(w, Err)\n\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\t\tdata[TplParamFormAction] = book.GetURLPayment()\n\n\t\t//check if a payment is supported, otherwise view the order\n\t\tif !book.SupportsPayment() {\n\t\t\thttp.Redirect(w, r.WithContext(ctx), book.GetURLView(), http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\n\t\t//check if already paid, in which case just view the payment\n\t\tif book.IsPaid() {\n\t\t\thttp.Redirect(w, r.WithContext(ctx), book.GetURLPaymentView(), http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\n\t\t//load the service\n\t\tnow := data[TplParamCurrentTime].(time.Time)\n\t\tctx, _, ok = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, book.Service.ID, now)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\n\t\t//check the method\n\t\tif r.Method == http.MethodGet {\n\t\t\tdata[TplParamDesc] = \"\"\n\t\t\tdata[TplParamEmail] = book.Client.Email\n\t\t\tdata[TplParamName] = book.Client.Name\n\t\t\tdata[TplParamPhone] = book.Client.Phone\n\t\t\tdata[TplParamPrice] = book.ComputeServicePrice()\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\n\t\t//read the form\n\t\tdesc := r.FormValue(URLParams.Desc)\n\t\temail := r.FormValue(URLParams.Email)\n\t\tname := r.FormValue(URLParams.Name)\n\t\tphone := r.FormValue(URLParams.Phone)\n\t\tpriceStr := r.FormValue(URLParams.Price)\n\n\t\t//prepare the data\n\t\tdata[TplParamDesc] = desc\n\t\tdata[TplParamEmail] = email\n\t\tdata[TplParamName] = name\n\t\tdata[TplParamPhone] = phone\n\t\tdata[TplParamPrice] = priceStr\n\n\t\t//validate the form\n\t\tform := &PaymentForm{\n\t\t\tEmailForm: EmailForm{\n\t\t\t\tEmail: strings.TrimSpace(email),\n\t\t\t},\n\t\t\tNameForm: NameForm{\n\t\t\t\tName: name,\n\t\t\t},\n\t\t\tPhone: FormatPhone(phone),\n\t\t\tPrice: priceStr,\n\t\t\tDescription: desc,\n\t\t\tClientInitiated: false,\n\t\t\tDirectCapture: false,\n\t\t}\n\t\tok = s.validateForm(w, r.WithContext(ctx), tpl, data, errs, form, true)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\n\t\t//save the payment\n\t\tctx, payment, err := s.savePaymentBooking(ctx, provider, book, form, now)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(\"save payment\", \"error\", err)\n\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\n\t\t//queue the email\n\t\tpaymentUI := s.createPaymentUI(payment)\n\t\tctx, err = s.queueEmailInvoice(ctx, provider.Name, paymentUI)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(\"queue email invoice\", \"error\", err)\n\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\n\t\t//success\n\t\ts.SetCookieMsg(w, MsgPaymentSuccess)\n\t\thttp.Redirect(w, r.WithContext(ctx), book.GetURLView(), http.StatusSeeOther)\n\t}\n}", "func (s *Server) handleDashboardPaymentView() http.HandlerFunc {\n\tvar o sync.Once\n\tvar tpl *template.Template\n\n\t//steps on the page\n\tsteps := struct {\n\t\tStepDel string\n\t\tStepMarkPaid string\n\t}{\n\t\tStepDel: \"stepDel\",\n\t\tStepMarkPaid: \"stepMarkPaid\",\n\t}\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tctx, logger := GetLogger(s.getCtx(r))\n\t\to.Do(func() {\n\t\t\ttpl = s.loadWebTemplateDashboard(ctx, \"payment-view.html\")\n\t\t})\n\t\tctx, provider, data, errs, ok := s.createTemplateDataDashboard(w, r.WithContext(ctx), tpl, true)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tdata[TplParamActiveNav] = provider.GetURLPayments()\n\t\tdata[TplParamSteps] = steps\n\n\t\t//load the booking\n\t\tnow := data[TplParamCurrentTime].(time.Time)\n\t\tvar paymentUI *paymentUI\n\t\tbookIDStr := r.FormValue(URLParams.BookID)\n\t\tif bookIDStr != \"\" {\n\t\t\tctx, book, ok := s.loadTemplateBook(w, r.WithContext(ctx), tpl, data, errs, bookIDStr, false, false)\n\t\t\tif !ok {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tdata[TplParamFormAction] = book.GetURLPaymentView()\n\n\t\t\t//load the service\n\t\t\tctx, _, ok = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, book.Service.ID, now)\n\t\t\tif !ok {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t//probe for a payment\n\t\t\tctx, payment, err := LoadPaymentByProviderIDAndSecondaryIDAndType(ctx, s.getDB(), provider.ID, book.ID, PaymentTypeBooking)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"load payment\", \"error\", err, \"id\", book.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif payment == nil {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLBookings(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpaymentUI = s.createPaymentUI(payment)\n\t\t} else {\n\t\t\t//load the payment directly\n\t\t\tidStr := r.FormValue(URLParams.PaymentID)\n\t\t\tif idStr == \"\" {\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tid := uuid.FromStringOrNil(idStr)\n\t\t\tif id == uuid.Nil {\n\t\t\t\tlogger.Errorw(\"invalid uuid\", \"id\", idStr)\n\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tctx, payment, err := LoadPaymentByID(ctx, s.getDB(), &id)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"load payment\", \"error\", err, \"id\", id)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpaymentUI = s.createPaymentUI(payment)\n\t\t\tdata[TplParamFormAction] = paymentUI.GetURLView()\n\n\t\t\t//probe for a booking\n\t\t\tctx, book, ok := s.loadTemplateBook(w, r.WithContext(ctx), tpl, data, errs, payment.SecondaryID.String(), false, false)\n\t\t\tif ok {\n\t\t\t\tctx, _, _ = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, book.Service.ID, now)\n\t\t\t} else if paymentUI.ServiceID != \"\" {\n\t\t\t\tsvcID := uuid.FromStringOrNil(paymentUI.ServiceID)\n\t\t\t\tif svcID == uuid.Nil {\n\t\t\t\t\tlogger.Errorw(\"invalid uuid\", \"id\", paymentUI.ServiceID)\n\t\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tctx, _, _ = s.loadTemplateService(w, r.WithContext(ctx), tpl, data, provider, &svcID, now)\n\t\t\t}\n\t\t}\n\t\tdata[TplParamPayment] = paymentUI\n\n\t\t//set-up the confirmation\n\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgPaymentMarkPaid)\n\t\tdata[TplParamConfirmSubmitName] = URLParams.Step\n\t\tdata[TplParamConfirmSubmitValue] = steps.StepMarkPaid\n\n\t\t//check the method\n\t\tif r.Method == http.MethodGet {\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\n\t\t//process the step\n\t\tstep := r.FormValue(URLParams.Step)\n\t\tswitch step {\n\t\tcase steps.StepDel:\n\t\t\tctx, err := DeletePayment(ctx, s.getDB(), paymentUI.ID)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"delete payment\", \"error\", err, \"id\", paymentUI.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase steps.StepMarkPaid:\n\t\t\tctx, err := UpdatePaymentDirectCapture(ctx, s.getDB(), paymentUI.ID, &now)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"update payment captured\", \"error\", err, \"id\", paymentUI.ID)\n\t\t\t\ts.SetCookieErr(w, Err)\n\t\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\t\treturn\n\t\t\t}\n\t\tdefault:\n\t\t\tlogger.Errorw(\"invalid step\", \"id\", paymentUI.ID, \"step\", step)\n\t\t\ts.SetCookieErr(w, Err)\n\t\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\t\ts.SetCookieMsg(w, MsgUpdateSuccess)\n\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPayments(), http.StatusSeeOther)\n\t}\n}", "func (mw loggingMiddleware) PostPayment(ctx context.Context, p Payment) (err error) {\n\tdefer func(begin time.Time) {\n\t\tmw.logger.Log(\"method\", \"PostPayment\", \"id\", p.Account, \"took\", time.Since(begin), \"err\", err)\n\t}(time.Now())\n\treturn mw.next.PostPayment(ctx, p)\n}", "func MakePaymentEndpoints(service PaymentService, logger kitlog.Logger) http.Handler {\n\trouter := chi.NewRouter()\n\trouter.Method(http.MethodPost, \"/\", kithttp.NewServer(\n\t\ttransferMoney(service), decodeTransferMoneyRequest, encodeTransferMoneyResponse,\n\t\t[]kithttp.ServerOption{\n\t\t\tkithttp.ServerErrorLogger(logger),\n\t\t\tkithttp.ServerErrorEncoder(encodePaymentError),\n\t\t}...))\n\n\trouter.Method(http.MethodGet, \"/\", kithttp.NewServer(\n\t\tlistPayments(service), decodeListPaymentsRequest, encodeListPaymentsResponse,\n\t\t[]kithttp.ServerOption{\n\t\t\tkithttp.ServerErrorLogger(logger),\n\t\t\tkithttp.ServerErrorEncoder(encodePaymentError),\n\t\t}...))\n\n\treturn router\n}", "func paymentDelete(service payment.UseCase) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tvars := mux.Vars(r)\n\t\tpaymentID, ok := vars[\"paymentID\"]\n\t\tif !ok {\n\t\t\trespondWithError(w, http.StatusNotFound, \"Missing route parameter 'paymentID'\")\n\t\t\treturn\n\t\t}\n\t\tif entity.IsValidID(paymentID) {\n\t\t\terr := service.Delete(entity.StringToID(paymentID))\n\t\t\tif err != nil {\n\t\t\t\trespondWithError(w, http.StatusNotFound, \"Payment ID does not exist\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\trespondWithJSON(w, http.StatusNoContent, nil)\n\t\t} else {\n\t\t\trespondWithError(w, http.StatusBadRequest, \"Invalid Payment ID\")\n\t\t\treturn\n\t\t}\n\t})\n}", "func NewHandler(s Service, v *validator.Validate, c *cache.Cache) FundHandler {\n\treturn FundHandler{service: s, validate: v, cache: c}\n}", "func MakeHandler(svc notifiers.Service, tracer opentracing.Tracer, logger logger.Logger) http.Handler {\n\topts := []kithttp.ServerOption{\n\t\tkithttp.ServerErrorEncoder(apiutil.LoggingErrorEncoder(logger, encodeError)),\n\t}\n\n\tmux := bone.New()\n\n\tmux.Post(\"/subscriptions\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"create_subscription\")(createSubscriptionEndpoint(svc)),\n\t\tdecodeCreate,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Get(\"/subscriptions/:id\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"view_subscription\")(viewSubscriptionEndpint(svc)),\n\t\tdecodeSubscription,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Get(\"/subscriptions\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"list_subscriptions\")(listSubscriptionsEndpoint(svc)),\n\t\tdecodeList,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Delete(\"/subscriptions/:id\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"delete_subscription\")(deleteSubscriptionEndpint(svc)),\n\t\tdecodeSubscription,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.GetFunc(\"/health\", mainflux.Health(\"notifier\"))\n\tmux.Handle(\"/metrics\", promhttp.Handler())\n\n\treturn mux\n}", "func (s *Server) handleDashboardPayments() http.HandlerFunc {\n\tvar o sync.Once\n\tvar tpl *template.Template\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tctx, logger := GetLogger(s.getCtx(r))\n\t\to.Do(func() {\n\t\t\ttpl = s.loadWebTemplateDashboard(ctx, \"payments.html\")\n\t\t})\n\t\tctx, provider, data, _, ok := s.createTemplateDataDashboard(w, r.WithContext(ctx), tpl, true)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\n\t\t//setup the breadcrumbs\n\t\tbreadcrumbs := []breadcrumb{\n\t\t\t{\"Invoices\", \"\"},\n\t\t}\n\t\tdata[TplParamBreadcrumbs] = breadcrumbs\n\t\tdata[TplParamActiveNav] = provider.GetURLPayments()\n\t\tdata[TplParamFormAction] = provider.GetURLPayments()\n\n\t\t//read the form\n\t\tfilterStr := r.FormValue(URLParams.Filter)\n\n\t\t//prepare the data\n\t\tdata[TplParamFilter] = filterStr\n\n\t\t//validate the filter\n\t\tvar err error\n\t\tfilter := PaymentFilterAll\n\t\tif filterStr != \"\" {\n\t\t\tfilter, err = ParsePaymentFilter(filterStr)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"parse filter\", \"error\", err, \"filter\", filterStr)\n\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t}\n\t\t}\n\n\t\t//load the payments\n\t\tctx, payments, err := ListPaymentsByProviderIDAndFilter(ctx, s.getDB(), provider.ID, filter)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(\"load payments\", \"error\", err, \"id\", provider.ID)\n\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\t\tdata[TplParamPayments] = s.createPaymentUIs(payments)\n\n\t\t//load the count\n\t\tctx, countUnPaid, err := CountPaymentsByProviderIDAndFilter(ctx, s.getDB(), provider.ID, PaymentFilterUnPaid)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(\"count payments unpaid\", \"error\", err, \"id\", provider.ID)\n\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\t\tdata[TplParamCountUnPaid] = countUnPaid\n\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t}\n}", "func GetPayment(repo repository.Repository) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tpaymentID := mux.Vars(r)[\"paymentID\"]\n\t\tval, err := repo.Get(paymentID)\n\t\tif err != nil {\n\t\t\tif err == repository.ErrNotFound {\n\t\t\t\tSendErrorResponse(w, r, http.StatusNotFound, errors.Errorf(\"paymentID:%s not found\", paymentID))\n\t\t\t\treturn\n\t\t\t}\n\t\t\tSendErrorResponse(w, r, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\t\tSendResponse(w, r, http.StatusOK, val)\n\t\treturn\n\t}\n}", "func CreatePayment(repo repository.Repository) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\tvar t *model.Payment\n\t\tif err := json.NewDecoder(r.Body).Decode(&t); err != nil {\n\t\t\tSendErrorResponse(w, r, http.StatusBadRequest, err)\n\t\t\treturn\n\t\t}\n\n\t\terr := validate(t)\n\t\tif err != nil {\n\t\t\tSendErrorResponse(w, r, http.StatusBadRequest, err)\n\t\t\treturn\n\t\t}\n\n\t\tdup, err := repo.Get(t.ID)\n\t\tif err == nil {\n\t\t\tif cmp.Equal(*t, *dup) {\n\t\t\t\tw.WriteHeader(http.StatusCreated)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tSendErrorResponse(w, r, http.StatusConflict, errors.New(\"already exists\"))\n\t\t\treturn\n\t\t}\n\n\t\terr = repo.Create(t)\n\t\tif err != nil {\n\t\t\tSendErrorResponse(w, r, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\t\tw.WriteHeader(http.StatusCreated)\n\t})\n}", "func CreatePayment(c *soso.Context) {\n\tif c.Token == nil {\n\t\tc.ErrorResponse(403, soso.LevelError, errors.New(\"User not authorized\"))\n\t\treturn\n\t}\n\treq := c.RequestMap\n\n\tpayID, _ := req[\"id\"].(float64)\n\tleadID, _ := req[\"lead_id\"].(float64)\n\n\tif leadID < 0 || payID <= 0 {\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, errors.New(\"Incorrect parameter\"))\n\t\treturn\n\t}\n\n\tif leadID != 0 {\n\t\t_, role, err := getConversationID(c.Token.UID, uint64(leadID))\n\t\tif err != nil {\n\t\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\t\treturn\n\t\t}\n\n\t\torderData, paymentData, err := retrieveOrder(uint64(payID))\n\t\tif err != nil {\n\t\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\t\treturn\n\t\t}\n\n\t\tif !canBuy(paymentData.Direction, role) {\n\t\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, fmt.Errorf(\"This side of order can not pay it\"))\n\t\t\treturn\n\t\t}\n\t\tif orderData.LeadId != uint64(leadID) {\n\t\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, fmt.Errorf(\"Parameters mangled\"))\n\t\t\treturn\n\t\t}\n\n\t}\n\n\t// now -- create the order\n\tctx, cancel := rpc.DefaultContext()\n\tdefer cancel()\n\tresp, err := paymentServiceClient.BuyOrder(ctx, &payment.BuyOrderRequest{\n\t\tPayId: uint64(payID),\n\t\tUser: &payment.UserInfo{\n\t\t\tIp: c.RemoteIP,\n\t\t\tUserId: c.Token.UID,\n\t\t\t// phone not needed here\n\t\t},\n\t})\n\n\tif err != nil { // RPC errors\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\treturn\n\t}\n\tif resp.Error > 0 { // service errors\n\t\tc.Response.ResponseMap = map[string]interface{}{\n\t\t\t\"ErrorCode\": resp.Error,\n\t\t\t\"ErrorMessage\": resp.ErrorMessage,\n\t\t}\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, errors.New(resp.ErrorMessage))\n\t\treturn\n\t}\n\n\tc.SuccessResponse(map[string]interface{}{\n\t\t\"redirect_url\": resp.RedirectUrl,\n\t})\n\n}", "func (env *Env) CreatePayment(w http.ResponseWriter, r *http.Request) {\n\tlog.Printf(\"POST payment\")\n\n\ttx := &model.TX{}\n\terr := render.DecodeJSON(r.Body, &tx)\n\n\tif err != nil {\n\t\tlog.Printf(\"Error deserialising payment: %v\\n\", err)\n\t\trender.Status(r, http.StatusBadRequest)\n\t\trender.JSON(w, r, err)\n\t\treturn\n\t}\n\n\ttx, err = env.db.CreateTX(*tx)\n\tif err != nil {\n\t\tlog.Printf(\"Error creating payment: %v\\n\", err)\n\t\trender.Status(r, http.StatusMethodNotAllowed)\n\t\trender.JSON(w, r, err)\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusCreated)\n}", "func MakeHandler(csvc clients.Service, psvc policies.Service, mux *bone.Mux, logger logger.Logger) http.Handler {\n\topts := []kithttp.ServerOption{\n\t\tkithttp.ServerErrorEncoder(apiutil.LoggingErrorEncoder(logger, api.EncodeError)),\n\t}\n\tmux.Post(\"/connect\", kithttp.NewServer(\n\t\totelkit.EndpointMiddleware(otelkit.WithOperation(\"connect\"))(connectThingsEndpoint(psvc)),\n\t\tdecodeConnectList,\n\t\tapi.EncodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Post(\"/disconnect\", kithttp.NewServer(\n\t\totelkit.EndpointMiddleware(otelkit.WithOperation(\"disconnect\"))(disconnectThingsEndpoint(psvc)),\n\t\tdecodeConnectList,\n\t\tapi.EncodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Post(\"/channels/:chanID/things/:thingID\", kithttp.NewServer(\n\t\totelkit.EndpointMiddleware(otelkit.WithOperation(\"connect_thing\"))(connectEndpoint(psvc)),\n\t\tdecodeConnectThing,\n\t\tapi.EncodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Delete(\"/channels/:chanID/things/:thingID\", kithttp.NewServer(\n\t\totelkit.EndpointMiddleware(otelkit.WithOperation(\"disconnect_thing\"))(disconnectEndpoint(psvc)),\n\t\tdecodeDisconnectThing,\n\t\tapi.EncodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Post(\"/identify\", kithttp.NewServer(\n\t\totelkit.EndpointMiddleware(otelkit.WithOperation(\"identify\"))(identifyEndpoint(csvc)),\n\t\tdecodeIdentify,\n\t\tapi.EncodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Put(\"/things/policies\", kithttp.NewServer(\n\t\totelkit.EndpointMiddleware(otelkit.WithOperation(\"update_policy\"))(updatePolicyEndpoint(psvc)),\n\t\tdecodeUpdatePolicy,\n\t\tapi.EncodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Get(\"/things/policies\", kithttp.NewServer(\n\t\totelkit.EndpointMiddleware(otelkit.WithOperation(\"list_policies\"))(listPoliciesEndpoint(psvc)),\n\t\tdecodeListPolicies,\n\t\tapi.EncodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Post(\"/channels/:chanID/access\", kithttp.NewServer(\n\t\totelkit.EndpointMiddleware(otelkit.WithOperation(\"authorize\"))(authorizeEndpoint(psvc)),\n\t\tdecodeCanAccess,\n\t\tapi.EncodeResponse,\n\t\topts...,\n\t))\n\treturn mux\n\n}", "func (s *Server) handleDashboardPaymentSettings() http.HandlerFunc {\n\tvar o sync.Once\n\tvar tpl *template.Template\n\n\t//steps on the page\n\tsteps := struct {\n\t\tStepDel string\n\t\tStepUpd string\n\t}{\n\t\tStepDel: \"stepDel\",\n\t\tStepUpd: \"stepUpd\",\n\t}\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tctx, logger := GetLogger(s.getCtx(r))\n\t\to.Do(func() {\n\t\t\ttpl = s.loadWebTemplateDashboard(ctx, \"payment-settings.html\")\n\t\t})\n\t\tctx, provider, data, errs, ok := s.createTemplateDataDashboard(w, r.WithContext(ctx), tpl, true)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\n\t\t//setup the breadcrumbs\n\t\tbreadcrumbs := []breadcrumb{\n\t\t\t{\"Payment Settings\", \"\"},\n\t\t}\n\t\tdata[TplParamBreadcrumbs] = breadcrumbs\n\t\tdata[TplParamActiveNav] = provider.GetURLPaymentSettings()\n\t\tdata[TplParamFormAction] = provider.GetURLPaymentSettings()\n\t\tdata[TplParamSteps] = steps\n\t\tdata[TplParamTypes] = PaymentTypes\n\n\t\t//handle the input\n\t\temail := r.FormValue(URLParams.Email)\n\t\tid := r.FormValue(URLParams.ID)\n\t\tstep := r.FormValue(URLParams.Step)\n\t\tpaymentType := r.FormValue(URLParams.Type)\n\n\t\t//prepare the data\n\t\tdata[TplParamEmail] = email\n\t\tdata[TplParamID] = id\n\t\tdata[TplParamType] = paymentType\n\n\t\t//prepare the confirmation modal\n\t\tswitch paymentType {\n\t\tcase PaymentTypes.TypePayPal:\n\t\t\tif provider.PayPalEmail != nil {\n\t\t\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgPayPalRemove)\n\t\t\t\tdata[TplParamConfirmSubmitValue] = steps.StepDel\n\t\t\t} else {\n\t\t\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgPayPalActivate)\n\t\t\t\tdata[TplParamConfirmSubmitValue] = steps.StepUpd\n\t\t\t}\n\t\t\tdata[TplParamConfirmSubmitName] = URLParams.Step\n\t\tcase PaymentTypes.TypeStripe:\n\t\t\tif provider.StripeToken != nil {\n\t\t\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgStripeRemove)\n\t\t\t\tdata[TplParamConfirmSubmitValue] = steps.StepDel\n\t\t\t} else {\n\t\t\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgStripeActivate)\n\t\t\t\tdata[TplParamConfirmSubmitValue] = steps.StepUpd\n\t\t\t}\n\t\t\tdata[TplParamConfirmSubmitName] = URLParams.Step\n\t\tcase PaymentTypes.TypeZelle:\n\t\t\tif provider.ZelleID != nil {\n\t\t\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgZelleRemove)\n\t\t\t\tdata[TplParamConfirmSubmitValue] = steps.StepDel\n\t\t\t} else {\n\t\t\t\tdata[TplParamConfirmMsg] = GetMsgText(MsgZelleActivate)\n\t\t\t\tdata[TplParamConfirmSubmitValue] = steps.StepUpd\n\t\t\t}\n\t\t\tdata[TplParamConfirmSubmitName] = URLParams.Step\n\t\t}\n\n\t\t//check the method\n\t\tif r.Method == http.MethodGet {\n\t\t\t//default the data\n\t\t\tif provider.PayPalEmail != nil {\n\t\t\t\tdata[TplParamEmail] = *provider.PayPalEmail\n\t\t\t}\n\t\t\tif provider.ZelleID != nil {\n\t\t\t\tdata[TplParamID] = *provider.ZelleID\n\t\t\t}\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\n\t\t//execute the correct operation\n\t\tswitch step {\n\t\tcase steps.StepDel:\n\t\t\tswitch paymentType {\n\t\t\tcase PaymentTypes.TypePayPal:\n\t\t\t\tprovider.PayPalEmail = nil\n\t\t\tcase PaymentTypes.TypeStripe:\n\t\t\t\t//revoke access\n\t\t\t\terr := RevokeOAuthTokenStripe(ctx, provider.StripeToken)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.Errorw(\"revoke stripe\", \"error\", err)\n\t\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tprovider.StripeToken = nil\n\t\t\tcase PaymentTypes.TypeZelle:\n\t\t\t\tprovider.ZelleID = nil\n\t\t\t}\n\n\t\t\t//save the provider\n\t\t\tctx, err := SaveProvider(ctx, s.getDB(), provider.Provider)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorw(\"save provider\", \"error\", err, \"provider\", provider)\n\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase steps.StepUpd:\n\t\t\tswitch paymentType {\n\t\t\tcase PaymentTypes.TypePayPal:\n\t\t\t\t//validate the data\n\t\t\t\tform := EmailForm{\n\t\t\t\t\tEmail: strings.TrimSpace(email),\n\t\t\t\t}\n\t\t\t\tok = s.validateForm(w, r.WithContext(ctx), tpl, data, errs, form, true)\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t//populate from the form\n\t\t\t\tprovider.PayPalEmail = &form.Email\n\n\t\t\t\t//save the provider\n\t\t\t\tctx, err := SaveProvider(ctx, s.getDB(), provider.Provider)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.Errorw(\"save provider\", \"error\", err, \"provider\", provider)\n\t\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\tcase PaymentTypes.TypeStripe:\n\t\t\t\ts.invokeHdlrGet(s.handleStripeLogin(), w, r.WithContext(ctx))\n\t\t\t\treturn\n\t\t\tcase PaymentTypes.TypeZelle:\n\t\t\t\t//validate the data\n\t\t\t\tform := ZelleIDForm{\n\t\t\t\t\tZelleID: id,\n\t\t\t\t}\n\t\t\t\tok = s.validateForm(w, r.WithContext(ctx), tpl, data, errs, form, true)\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t//populate from the form\n\t\t\t\tprovider.ZelleID = &form.ZelleID\n\n\t\t\t\t//save the provider\n\t\t\t\tctx, err := SaveProvider(ctx, s.getDB(), provider.Provider)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.Errorw(\"save provider\", \"error\", err, \"provider\", provider)\n\t\t\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\tdefault:\n\t\t\tlogger.Errorw(\"invalid step\", \"step\", step)\n\t\t\tdata[TplParamErr] = GetErrText(Err)\n\t\t\ts.renderWebTemplate(w, r.WithContext(ctx), tpl, data)\n\t\t\treturn\n\t\t}\n\n\t\t//success\n\t\ts.SetCookieMsg(w, MsgUpdateSuccess)\n\t\thttp.Redirect(w, r.WithContext(ctx), provider.GetURLPaymentSettings(), http.StatusSeeOther)\n\t}\n}", "func MakeHandler(svc manager.Service) http.Handler {\n\topts := []kithttp.ServerOption{\n\t\tkithttp.ServerErrorEncoder(encodeError),\n\t}\n\n\tregistration := kithttp.NewServer(\n\t\tregistrationEndpoint(svc),\n\t\tdecodeCredentials,\n\t\tencodeResponse,\n\t\topts...,\n\t)\n\n\tlogin := kithttp.NewServer(\n\t\tloginEndpoint(svc),\n\t\tdecodeCredentials,\n\t\tencodeResponse,\n\t\topts...,\n\t)\n\n\taddClient := kithttp.NewServer(\n\t\taddClientEndpoint(svc),\n\t\tdecodeAddClient,\n\t\tencodeResponse,\n\t\topts...,\n\t)\n\n\tviewClient := kithttp.NewServer(\n\t\tviewClientEndpoint(svc),\n\t\tdecodeViewClient,\n\t\tencodeResponse,\n\t\topts...,\n\t)\n\n\tremoveClient := kithttp.NewServer(\n\t\tremoveClientEndpoint(svc),\n\t\tdecodeViewClient,\n\t\tencodeResponse,\n\t\topts...,\n\t)\n\n\tr := bone.New()\n\n\tr.Post(\"/users\", registration)\n\tr.Post(\"/tokens\", login)\n\tr.Post(\"/clients\", addClient)\n\tr.Get(\"/clients/:id\", viewClient)\n\tr.Delete(\"/clients/:id\", removeClient)\n\tr.Handle(\"/metrics\", promhttp.Handler())\n\n\treturn r\n}", "func Handler(arg events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\n\tsession := amazon.NewContext(&arg)\n\tconn, err := amazon.GetDBConnection()\n\tif err != nil {\n\t\thttp.InternalServerException(session)\n\t\treturn session.Response.Respond(), nil\n\t}\n\n\tdefer conn.Close()\n\n\tdeps := accounts.CreateRegistrationControllerDeps(conn)\n\tregister := accounts.NewRegistrationController(deps)\n\n\tregister.WithEmail(session)\n\treturn session.Response.Respond(), nil\n}", "func makeHandler(server *ServerContext, privs handlerPrivs, method handlerMethod) http.Handler {\n\treturn http.HandlerFunc(func(r http.ResponseWriter, rq *http.Request) {\n\t\th := newHandler(server, privs, r, rq)\n\t\terr := h.invoke(method)\n\t\th.writeError(err)\n\t\th.logDuration(true) \n\t})\n}", "func (p *payment) CreatePayment(ctx context.Context, args gopayd.CreatePaymentArgs, req gopayd.CreatePayment) (*gopayd.PaymentACK, error) {\n\tif err := validator.New().Validate(\"paymentID\", validator.NotEmpty(args.PaymentID)); err.Err() != nil {\n\t\treturn nil, err\n\t}\n\tpa := &gopayd.PaymentACK{\n\t\tPayment: &req,\n\t\tSuccess: paymentSuccess,\n\t}\n\t// get and attempt to store transaction before processing payment.\n\ttx, err := bt.NewTxFromString(req.Transaction)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to parse transaction for paymentID %s\", args.PaymentID)\n\t}\n\t// TODO: validate the transaction inputs\n\toutputTotal := uint64(0)\n\ttxos := make([]gopayd.CreateTxo, 0)\n\t// iterate outputs and gather the total satoshis for our known outputs\n\tfor i, o := range tx.GetOutputs() {\n\t\tsk, err := p.script.ScriptKey(ctx, gopayd.ScriptKeyArgs{LockingScript: o.LockingScript.ToString()})\n\t\tif err != nil {\n\t\t\t// script isn't known to us, could be a change utxo, skip and carry on\n\t\t\tif lathos.IsNotFound(err) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\treturn nil, errors.Wrapf(err, \"failed to get store output for paymentID %s\", args.PaymentID)\n\t\t}\n\t\t// push new txo onto list for persistence later\n\t\ttxos = append(txos, gopayd.CreateTxo{\n\t\t\tOutpoint: fmt.Sprintf(\"%s%d\", tx.GetTxID(), i),\n\t\t\tTxID: tx.GetTxID(),\n\t\t\tVout: i,\n\t\t\tKeyName: null.StringFrom(keyname),\n\t\t\tDerivationPath: sk.DerivationPath,\n\t\t\tLockingScript: sk.LockingScript,\n\t\t\tSatoshis: o.Satoshis,\n\t\t})\n\t\toutputTotal += o.Satoshis\n\t}\n\t// get the invoice for the paymentID to check total satoshis required.\n\tinv, err := p.invStore.Invoice(ctx, gopayd.InvoiceArgs{PaymentID: args.PaymentID})\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to get invoice to validate output total for paymentID %s.\", args.PaymentID)\n\t}\n\t// if it doesn't fully pay the invoice, reject it\n\tif outputTotal < inv.Satoshis {\n\t\tlog.Info(\"satoshis are less than expt outputs\")\n\t\tpa.Error = 1\n\t\tpa.Success = paymentFailed\n\t\tpa.Memo = \"Outputs do not fully pay invoice for paymentID \" + args.PaymentID\n\t\treturn pa, nil\n\t}\n\tctx = p.txrunner.WithTx(ctx)\n\t// Store utxos and set invoice to paid.\n\tif _, err = p.store.StoreUtxos(ctx, gopayd.CreateTransaction{\n\t\tPaymentID: inv.PaymentID,\n\t\tTxID: tx.GetTxID(),\n\t\tTxHex: req.Transaction,\n\t\tOutputs: txos,\n\t}); err != nil {\n\t\tlog.Error(err)\n\t\tpa.Error = 1\n\t\tpa.Success = paymentFailed\n\t\tpa.Memo = err.Error()\n\t\treturn nil, errors.Wrapf(err, \"failed to complete payment for paymentID %s\", args.PaymentID)\n\t}\n\tif _, err := p.invStore.Update(ctx, gopayd.InvoiceUpdateArgs{PaymentID: args.PaymentID}, gopayd.InvoiceUpdate{\n\t\tRefundTo: req.RefundTo,\n\t}); err != nil {\n\t\tlog.Error(err)\n\t\tpa.Error = 1\n\t\tpa.Success = paymentFailed\n\t\tpa.Memo = err.Error()\n\t\treturn nil, errors.Wrapf(err, \"failed to update invoice payment for paymentID %s\", args.PaymentID)\n\t}\n\t// Broadcast the transaction.\n\tif err := p.sender.Send(ctx, args, req); err != nil {\n\t\tlog.Error(err)\n\t\tpa.Error = 1\n\t\tpa.Success = paymentFailed\n\t\tpa.Memo = err.Error()\n\t\treturn pa, errors.Wrapf(err, \"failed to send payment for paymentID %s\", args.PaymentID)\n\t}\n\treturn pa, errors.WithStack(p.txrunner.Commit(ctx))\n}", "func paymentUpdate(service payment.UseCase) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tvars := mux.Vars(r)\n\t\tpaymentID, ok := vars[\"paymentID\"]\n\t\tif !ok {\n\t\t\trespondWithError(w, http.StatusNotFound, \"Missing route parameter 'paymentID'\")\n\t\t\treturn\n\t\t}\n\t\tdefer r.Body.Close()\n\t\tvar payment *entity.Payment\n\t\tif err := json.NewDecoder(r.Body).Decode(&payment); err != nil {\n\t\t\trespondWithError(w, http.StatusBadRequest, \"Invalid request payload\")\n\t\t\treturn\n\t\t}\n\t\tif err := service.Update(entity.StringToID(paymentID), payment); err != nil {\n\t\t\trespondWithError(w, http.StatusInternalServerError, err.Error())\n\t\t\treturn\n\t\t}\n\t\trespondWithJSON(w, http.StatusOK, payment)\n\t})\n}", "func paymentRequired(rw http.ResponseWriter, r *http.Request) {\n\n}", "func makeCreateOrderHandler(m *mux.Router, endpoints endpoint.Endpoints, options []kithttp.ServerOption) {\n\tm.Methods(\"POST\").Path(\"/orders\").Handler(\n\t\tkithttp.NewServer(\n\t\t\tendpoints.CreateOrderEndpoint,\n\t\t\tdecodeCreateOrderRequest,\n\t\t\tencodeHTTPGenericResponse,\n\t\t\toptions...,\n\t\t))\n}", "func NewPayment(\n\tID int64,\n\tAmount value.Amount,\n\tDescription string,\n\tCreatedAt types.Time,\n\tUser UserInterface,\n\n\tfileUploader file.Uploader,\n\tdispatcher queue.Publisher,\n) *Payment {\n\n\tp := Payment{\n\t\tid: ID,\n\t\tamount: Amount,\n\t\tdescription: Description,\n\t\tcreatedAt: CreatedAt,\n\t\tuser: User,\n\n\t\tfileUploader: fileUploader,\n\t\tdispatcher: dispatcher,\n\t}\n\n\tp.dispatcher = rabbitmq.NewQueue()\n\n\treturn &p\n}", "func NewPayment() *Payment {\n\treturn &Payment{}\n}", "func (t *ManagePayment) createPayment(stub shim.ChaincodeStubInterface, args []string) ([]byte, error) {\n\tvar err error\n\tif len(args) != 6 {\n\t\terrMsg := \"{ \\\"message\\\" : \\\"Incorrect number of arguments. Expecting 6 arguments.\\\", \\\"code\\\" : \\\"503\\\"}\"\n\t\terr = stub.SetEvent(\"errEvent\", []byte(errMsg))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t} \n\t\treturn nil, nil\n\t}\n\tfmt.Println(\"creating a new Payment\")\n\t//input sanitation\n\tif len(args[0]) <= 0 {\n\t\treturn nil, errors.New(\"Agreement Id cannot be empty.\")\n\t}else if len(args[1]) <= 0 {\n\t\treturn nil, errors.New(\"Payment Type cannot be empty.\")\n\t}else if len(args[2]) <= 0 {\n\t\treturn nil, errors.New(\"Customer Payment cannot be empty.\")\n\t}else if len(args[3]) <= 0 {\n\t\treturn nil, errors.New(\"Receiver Payment cannot be empty.\")\n\t}else if len(args[4]) <= 0 {\n\t\treturn nil, errors.New(\"Amount Paid cannot be empty.\")\n\t}else if len(args[5]) <= 0 {\n\t\treturn nil, errors.New(\"Last Updated By cannot be empty.\")\n\t}\n\n\t// setting attributes\n\tpaymentId := \"PA\"+ strconv.FormatInt(time.Now().Unix(), 10) // check https://play.golang.org/p/8Du2FrDk2eH\n\tagreementId := args[0]\n\tpaymentType := args[1]\n\tcustomerAccount := args[2]\n\treceiverAccount := args[3]\n\tamountPaid,_ := strconv.ParseFloat(args[4],64);\n\tlastUpdatedBy := args[5]\n\tlastUpdateDate := time.Now().Unix() // current unix timestamp\n\t\n\tfmt.Println(paymentId);\n\tfmt.Println(agreementId);\n\tfmt.Println(paymentType);\n\tfmt.Println(customerAccount);\n\tfmt.Println(receiverAccount);\n\tfmt.Println(amountPaid);\n\tfmt.Println(lastUpdatedBy);\n\tfmt.Println(lastUpdateDate);\n\n\t// Fetching Payment details by Payment Id\n\tPaymentAsBytes, err := stub.GetState(paymentId)\n\tif err != nil {\n\t\treturn nil, errors.New(\"Failed to get Payment Id\")\n\t}\n\tres := Payment{}\n\tjson.Unmarshal(PaymentAsBytes, &res)\n\tfmt.Print(\" Payment Details: \")\n\tfmt.Println(res)\n\tif res.PaymentId == paymentId{\n\t\tfmt.Println(\"This Payment already exists: \" + paymentId)\n\t\terrMsg := \"{ \\\"message\\\" : \\\"This Payment already exists.\\\", \\\"code\\\" : \\\"503\\\"}\"\n\t\terr := stub.SetEvent(\"errEvent\", []byte(errMsg))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t} \n\t\treturn nil, errors.New(errMsg)\t\t\t\t//stop creating a new Payment if Payment exists already\n\t}\n\n\t// create a pointer/json to the struct 'Payment'\n\tPaymentJson := &Payment{paymentId, agreementId, paymentType, customerAccount, receiverAccount, amountPaid, lastUpdatedBy, lastUpdateDate};\n\tfmt.Printf(\"PaymentJson: %v \\n\", PaymentJson)\n\t// convert *Payment to []byte\n\tPaymentJsonasBytes, err := json.Marshal(PaymentJson)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t//store paymentId as key\n\terr = stub.PutState(paymentId, PaymentJsonasBytes)\t\t\t\t\t\t\t\t\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t//get the Payment index\n\tPaymentIndexStrAsBytes, err := stub.GetState(PaymentIndexStr)\n\tif err != nil {\n\t\treturn nil, errors.New(\"Failed to get Payment index\")\n\t}\n\tvar PaymentIndex []string\n\n\tjson.Unmarshal(PaymentIndexStrAsBytes, &PaymentIndex)\t\t\t\t\t\t\t//un stringify it aka JSON.parse()\n\tfmt.Print(\"PaymentIndex after unmarshal..before append: \")\n\tfmt.Println(PaymentIndex)\n\n\t//append\n\tPaymentIndex = append(PaymentIndex, paymentId)\t\t\t\t\t\t\t\t\t//add PaymentId to index list\n\tfmt.Println(\"! Payment index: \", PaymentIndex)\n\tjsonAsBytes, _ := json.Marshal(PaymentIndex)\n\terr = stub.PutState(PaymentIndexStr, jsonAsBytes)\t\t\t\t\t\t//store PaymentId as an index\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// event message to set on successful Payment creation\n\ttosend := \"{ \\\" Payment Id\\\" : \\\"\"+paymentId+\"\\\", \\\"message\\\" : \\\" Payment created succcessfully\\\", \\\"code\\\" : \\\"200\\\"}\"\n\terr = stub.SetEvent(\"evtsender\", []byte(tosend))\n\tif err != nil {\n\t\treturn nil, err\n\t} \t\n\tfmt.Println(\" Payment created succcessfully.\")\n\treturn nil, nil\n}", "func handler(request events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// Initiialize a connection to Sentry to capture errors and traces\n\tsentry.Init(sentry.ClientOptions{\n\t\tDsn: os.Getenv(\"SENTRY_DSN\"),\n\t\tTransport: &sentry.HTTPSyncTransport{\n\t\t\tTimeout: time.Second * 3,\n\t\t},\n\t\tServerName: os.Getenv(\"FUNCTION_NAME\"),\n\t\tRelease: os.Getenv(\"VERSION\"),\n\t\tEnvironment: os.Getenv(\"STAGE\"),\n\t})\n\n\t// Create headers if they don't exist and add\n\t// the CORS required headers, otherwise the response\n\t// will not be accepted by browsers.\n\theaders := request.Headers\n\tif headers == nil {\n\t\theaders = make(map[string]string)\n\t}\n\theaders[\"Access-Control-Allow-Origin\"] = \"*\"\n\n\t// Update the order with an OrderID\n\tord, err := acmeserverless.UnmarshalOrder(request.Body)\n\tif err != nil {\n\t\treturn handleError(\"unmarshal\", headers, err)\n\t}\n\tord.OrderID = uuid.Must(uuid.NewV4()).String()\n\n\tdynamoStore := dynamodb.New()\n\tord, err = dynamoStore.AddOrder(ord)\n\tif err != nil {\n\t\treturn handleError(\"store\", headers, err)\n\t}\n\n\tprEvent := acmeserverless.PaymentRequestedEvent{\n\t\tMetadata: acmeserverless.Metadata{\n\t\t\tDomain: acmeserverless.OrderDomain,\n\t\t\tSource: \"AddOrder\",\n\t\t\tType: acmeserverless.PaymentRequestedEventName,\n\t\t\tStatus: acmeserverless.DefaultSuccessStatus,\n\t\t},\n\t\tData: acmeserverless.PaymentRequestDetails{\n\t\t\tOrderID: ord.OrderID,\n\t\t\tCard: ord.Card,\n\t\t\tTotal: ord.Total,\n\t\t},\n\t}\n\n\t// Send a breadcrumb to Sentry with the payment request\n\tsentry.AddBreadcrumb(&sentry.Breadcrumb{\n\t\tCategory: acmeserverless.PaymentRequestedEventName,\n\t\tTimestamp: time.Now(),\n\t\tLevel: sentry.LevelInfo,\n\t\tData: acmeserverless.ToSentryMap(prEvent.Data),\n\t})\n\n\tem := sqs.New()\n\terr = em.SendPaymentRequestedEvent(prEvent)\n\tif err != nil {\n\t\treturn handleError(\"request payment\", headers, err)\n\t}\n\n\tstatus := acmeserverless.OrderStatus{\n\t\tOrderID: ord.OrderID,\n\t\tUserID: ord.UserID,\n\t\tPayment: acmeserverless.CreditCardValidationDetails{\n\t\t\tMessage: \"pending payment\",\n\t\t\tSuccess: false,\n\t\t},\n\t}\n\n\t// Send a breadcrumb to Sentry with the shipment request\n\tsentry.AddBreadcrumb(&sentry.Breadcrumb{\n\t\tCategory: acmeserverless.PaymentRequestedEventName,\n\t\tTimestamp: time.Now(),\n\t\tLevel: sentry.LevelInfo,\n\t\tData: acmeserverless.ToSentryMap(status.Payment),\n\t})\n\n\tpayload, err := status.Marshal()\n\tif err != nil {\n\t\treturn handleError(\"response\", headers, err)\n\t}\n\n\tresponse := events.APIGatewayProxyResponse{\n\t\tStatusCode: http.StatusOK,\n\t\tBody: string(payload),\n\t\tHeaders: headers,\n\t}\n\n\treturn response, nil\n}", "func MakeHandler(svc Service, opts []kithttp.ServerOption, responseEncoder kithttp.EncodeResponseFunc) http.Handler {\n\tauthMid := middleware.AuthMiddleware(svc.(*service).authenticator, \"\", \"\")\n\tlistHandler := kithttp.NewServer(\n\t\tauthMid(makeListEndpoint(svc)),\n\t\tdecodeListRequest,\n\t\tresponseEncoder, opts...)\n\tauthMid = middleware.AuthMiddleware(svc.(*service).authenticator, \"\", permission.StartCampaign)\n\tstartHandler := kithttp.NewServer(\n\t\tauthMid(makeStartEndpoint(svc)),\n\t\tdecodeStartRequest,\n\t\tresponseEncoder, opts...)\n\tprogressHandler := kithttp.NewServer(\n\t\tauthMid(makeProgressEndpoint(svc)),\n\t\tdecodeProgressRequest,\n\t\tresponseEncoder, opts...)\n\treportHandler := kithttp.NewServer(\n\t\tauthMid(makeReportEndpoint(svc)),\n\t\tdecodeReportRequest,\n\t\tresponseEncoder, opts...)\n\tauthMid = middleware.AuthMiddleware(svc.(*service).authenticator, \"\", permission.StopCampaign)\n\tstopHandler := kithttp.NewServer(\n\t\tauthMid(makeStopEndpoint(svc)),\n\t\tdecodeStopRequest,\n\t\tresponseEncoder, opts...)\n\tr := mux.NewRouter()\n\n\tr.Handle(\"/campaign/v1/list\", listHandler).Methods(\"GET\", \"POST\")\n\tr.Handle(\"/campaign/v1/start\", startHandler).Methods(\"POST\")\n\tr.Handle(\"/campaign/v1/progress\", progressHandler).Methods(\"GET\", \"POST\")\n\tr.Handle(\"/campaign/v1/stop\", stopHandler).Methods(\"POST\")\n\tr.Handle(\"/campaign/v1/report\", reportHandler).Methods(\"GET\", \"POST\")\n\treturn r\n}", "func (h RequestPPMPaymentHandler) Handle(params ppmop.RequestPPMPaymentParams) middleware.Responder {\n\treturn h.AuditableAppContextFromRequestWithErrors(params.HTTPRequest,\n\t\tfunc(appCtx appcontext.AppContext) (middleware.Responder, error) {\n\t\t\tppmID, err := uuid.FromString(params.PersonallyProcuredMoveID.String())\n\t\t\tif err != nil {\n\t\t\t\treturn handlers.ResponseForError(appCtx.Logger(), err), err\n\t\t\t}\n\n\t\t\tppm, err := models.FetchPersonallyProcuredMove(appCtx.DB(), appCtx.Session(), ppmID)\n\t\t\tif err != nil {\n\t\t\t\treturn handlers.ResponseForError(appCtx.Logger(), err), err\n\t\t\t}\n\n\t\t\terr = ppm.RequestPayment()\n\t\t\tif err != nil {\n\t\t\t\treturn handlers.ResponseForError(appCtx.Logger(), err), err\n\t\t\t}\n\n\t\t\tverrs, err := models.SavePersonallyProcuredMove(appCtx.DB(), ppm)\n\t\t\tif err != nil || verrs.HasAny() {\n\t\t\t\treturn handlers.ResponseForVErrors(appCtx.Logger(), verrs, err), err\n\t\t\t}\n\n\t\t\tppmPayload, err := payloadForPPMModel(h.FileStorer(), *ppm)\n\t\t\tif err != nil {\n\t\t\t\treturn handlers.ResponseForError(appCtx.Logger(), err), err\n\t\t\t}\n\t\t\treturn ppmop.NewRequestPPMPaymentOK().WithPayload(ppmPayload), nil\n\t\t})\n}", "func MakeHandler(svc users.Service, tracer opentracing.Tracer, l log.Logger) http.Handler {\n\tlogger = l\n\n\topts := []kithttp.ServerOption{\n\t\tkithttp.ServerErrorEncoder(encodeError),\n\t}\n\n\tmux := bone.New()\n\n\tmux.Post(\"/users\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"register\")(registrationEndpoint(svc)),\n\t\tdecodeCredentials,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Get(\"/users\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"user_info\")(userInfoEndpoint(svc)),\n\t\tdecodeViewInfo,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Put(\"/users\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"update_user\")(updateUserEndpoint(svc)),\n\t\tdecodeUpdateUser,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Post(\"/password/reset-request\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"res-req\")(passwordResetRequestEndpoint(svc)),\n\t\tdecodePasswordResetRequest,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Put(\"/password/reset\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"reset\")(passwordResetEndpoint(svc)),\n\t\tdecodePasswordReset,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Patch(\"/password\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"reset\")(passwordChangeEndpoint(svc)),\n\t\tdecodePasswordChange,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.Post(\"/tokens\", kithttp.NewServer(\n\t\tkitot.TraceServer(tracer, \"login\")(loginEndpoint(svc)),\n\t\tdecodeCredentials,\n\t\tencodeResponse,\n\t\topts...,\n\t))\n\n\tmux.GetFunc(\"/version\", mainflux.Version(\"users\"))\n\tmux.Handle(\"/metrics\", promhttp.Handler())\n\n\treturn mux\n}", "func Handler(cs mongo.CustomerStorage) http.Handler {\n\tr := mux.NewRouter()\n\n\tcustomerService := CustomerService.New(cs)\n\n\tr.HandleFunc(\"/\", indexHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/signup\", genSignUpHandler(customerService)).Methods(\"POST\")\n\n\treturn r\n}", "func Handler(ctx context.Context, event Request) (Response, error) {\n\tsecretKeyAccessToken := os.Getenv(\"SECRET_ACCESS_TOKEN\")\n\n\ttokenString := event.AuthorizationToken\n\tsecretAccessToken := []byte(secretKeyAccessToken)\n\n\terr := verifyToken(tokenString, secretAccessToken)\n\n\tif err != nil {\n\t\treturn Response{}, errors.New(\"Unauthorized\")\n\t}\n\n\treturn generatePolicy(\"customer\", \"Allow\", event.MethodArn), nil\n}", "func PayTicketHandler(w http.ResponseWriter, r *http.Request) {\n\n\tvars := mux.Vars(r)\n\n\tid, err := strconv.ParseInt(vars[\"ticketid\"], 10, 64)\n\tcheckError(err)\n\n\tif r.Body == nil {\n\t\thttp.Error(w, \"No request body sent.\", 400)\n\t\treturn\n\t}\n\n\tcardInfo := struct {\n\t\tCreditCard string `json:\"credit_card\"`\n\t}{}\n\n\terr = json.NewDecoder(r.Body).Decode(&cardInfo)\n\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), 400)\n\t\treturn\n\t}\n\n\tticket, err := tickets.Pay(id, cardInfo.CreditCard)\n\n\tif err != nil {\n\t\trenderer.JSON(w, http.StatusNotFound, map[string]string{\"error\": err.Error()})\n\t\treturn\n\t}\n\n\trenderer.JSON(w, http.StatusOK, ticket)\n}", "func Handler(ctx context.Context, request events.APIGatewayProxyRequest,\n\tdynamoDB *dynamodb.DynamoDB, cfg config.Configuration) (\n\tevents.APIGatewayProxyResponse, error) {\n\n\t//Instantiate item API Handler\n\tih, err := item.New(dynamoDB, cfg.AWS.DynamoDB.Table.Store)\n\tif err != nil {\n\t\treturn web.GetResponse(ctx, err.Error(), http.StatusInternalServerError)\n\t}\n\n\tlog.Debug().Msgf(\"Executing method %s for path: %s with body: %v\",\n\t\trequest.HTTPMethod, request.Path, request.Body)\n\n\tswitch request.HTTPMethod {\n\tcase http.MethodGet:\n\n\t\treturn getItems(ctx, request, ih)\n\n\t}\n\n\t//APIGateway would not allow the function to get to this point\n\t//Since all the supported http methods are in the switch\n\treturn web.GetResponse(ctx, struct{}{}, http.StatusMethodNotAllowed)\n\n}", "func MakeHandler() *http.Handler {\n\tapi := rest.NewApi()\n\n\tauthMiddleware, err := CreateAuthMiddleware()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tapi.Use(&rest.IfMiddleware{\n\t\t// Only authenticate non login or register requests\n\t\tCondition: func(request *rest.Request) bool {\n\t\t\treturn (request.URL.Path != variables.APIPathLoginUserServer) && (request.URL.Path != variables.APIPathRegisterUserServer)\n\t\t},\n\t\tIfTrue: authMiddleware,\n\t})\n\tapi.Use(rest.DefaultProdStack...)\n\trouter, err := rest.MakeRouter(\n\t\trest.Post(variables.APIPathLoginUserServer, authMiddleware.LoginHandler),\n\t\trest.Get(variables.APIPathRefreshUserServer, authMiddleware.RefreshHandler),\n\t\trest.Post(variables.APIPathRegisterUserServer, PostRegister),\n\t\trest.Get(variables.APIPathUserServer, GetUser),\n\t\trest.Post(variables.APIPathUserServer, PostUser),\n\t)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tapi.SetApp(router)\n\thandler := api.MakeHandler()\n\treturn &handler\n}", "func DeletePayment(repo repository.Repository) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tpaymentID := mux.Vars(r)[\"paymentID\"]\n\n\t\terr := repo.Delete(paymentID)\n\t\tif err != nil && err == repository.ErrNotFound {\n\t\t\tSendErrorResponse(w, r, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\t\tw.WriteHeader(http.StatusNoContent)\n\t\treturn\n\t}\n}", "func (p *Payment) Do() (*YooKassaResponse, error) {\n\tpayloadBytes, err := json.Marshal(p)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbody := bytes.NewReader(payloadBytes)\n\n\treq, err := http.NewRequest(\"POST\", consts.Endpoint+\"payments\", body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.SetBasicAuth(p.Kassa.ShopID, p.Kassa.SecretKey)\n\treq.Header.Set(consts.IdempotentHeader, p.IdempotenceKey)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\tresp, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\n\tstuff, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tyooKassaError := &YooKassaErrorResponse{}\n\terr = json.Unmarshal(stuff, yooKassaError)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif yooKassaError.Type == \"error\" {\n\t\treturn nil, yooKassaError\n\t}\n\n\trespKassa := &YooKassaResponse{}\n\terr = json.Unmarshal(stuff, respKassa)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn respKassa, nil\n}", "func makeHandler() (http.Handler, *serializerLog, *mapStorage) {\n\treturn makeHandlerSubPath(\"\")\n}", "func NewHTTPHandler(endpoints endpoint.Endpoints, options map[string][]kithttp.ServerOption) http.Handler {\n\tm := mux.NewRouter()\n\tm = m.PathPrefix(\"/v1/paymentsvc\").Subrouter()\n\n\tmakeRechargeWalletHandler(m, endpoints, options[\"RechargeWallet\"])\n\tmakeListTransactionsHandler(m, endpoints, options[\"ListTransactions\"])\n\n\treturn m\n}", "func AddPayment(c *gin.Context) {\n\tfmt.Println(\"Adding Payment\")\n\tl := loan.GetLoanInstance(false)\n\tif l.Principal == 0 {\n\t\tsendErrorResponse(c, http.StatusNotFound, \"AddPayment :: Loan is not started\")\n\t\treturn\n\t}\n\tinstallment := loan.NewInstallment()\n\te := c.BindJSON(&installment)\n\tif e != nil {\n\t\tmsgs := customErrMsg(e)\n\t\tsendErrorResponse(c, http.StatusBadRequest, \"AddPayment :: \"+msgs)\n\t\treturn\n\t}\n\n\te = l.AddPayment(installment)\n\tif e != nil {\n\t\tsendErrorResponse(c, http.StatusBadRequest, \"AddPayment :: \"+e.Error())\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"status\": http.StatusOK, \"message\": \"Payment Accepted\"})\n}", "func paymentFind(service payment.UseCase) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tvars := mux.Vars(r)\n\t\tpaymentID, ok := vars[\"paymentID\"]\n\t\tif !ok {\n\t\t\trespondWithError(w, http.StatusNotFound, \"Missing route parameter 'paymentID'\")\n\t\t\treturn\n\t\t}\n\t\tif entity.IsValidID(paymentID) {\n\t\t\tpayment, err := service.Find(entity.StringToID(paymentID))\n\t\t\tif err != nil {\n\t\t\t\trespondWithError(w, http.StatusNotFound, \"Payment ID does not exist\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\trespondWithJSON(w, http.StatusOK, payment)\n\t\t} else {\n\t\t\trespondWithError(w, http.StatusBadRequest, \"Invalid Payment ID\")\n\t\t\treturn\n\t\t}\n\t})\n}", "func (env *Env) GetPayment(w http.ResponseWriter, r *http.Request) {\n\tID := chi.URLParam(r, \"id\")\n\tlog.Printf(\"GET payment id %s\\n\", ID)\n\n\ttx, err := env.db.GetTX(ID)\n\n\tif err != nil {\n\t\tlog.Printf(\"Error retrieving payment %s: %v\\n\", ID, err)\n\t\trender.Status(r, http.StatusNotFound)\n\t\trender.JSON(w, r, err)\n\t\treturn\n\t}\n\trender.JSON(w, r, tx)\n}", "func MakePurchase(c *gin.Context) {\n\n\tform := new(forms.Purchase)\n\tif err := c.Bind(form); err != nil {\n\t\tc.JSON(http.StatusBadRequest, forms.Response{\"bind\", \"can't bind params\"})\n\t\treturn\n\t}\n\n\t// validate\n\tif errs := form.Validate(); errs != nil {\n\t\tc.JSON(http.StatusBadRequest, errs)\n\t\treturn\n\t}\n\n\t// do\n\tcreds, err := form.Do()\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, forms.Response{\"do\", err.Error()})\n\t\treturn\n\t}\n\n\t// ok\n\tc.JSON(http.StatusOK, forms.Response{\"credits_remain\", creds})\n\treturn\n}", "func Handler(cs creating.Service, ls listing.Service, ds drawing.Service) http.Handler {\n\trouter := httprouter.New()\n\n\trouter.GET(\"/health\", health())\n\trouter.POST(\"/decks\", createDeck(cs))\n\trouter.GET(\"/decks/:id\", getDeck(ls))\n\trouter.PATCH(\"/decks/:id/draw/:amount\", drawCards(ds))\n\treturn router\n}", "func Handler(request events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\tvar (\n\t\tdatas []model.Todo\n\t\terr error\n\t\tb []byte\n\t)\n\n\ttodo := model.NewTodo()\n\tif datas, err = todo.List(); err != nil {\n\t\treturn response.Error(err)\n\t}\n\tif b, err = json.Marshal(datas); err != nil {\n\t\treturn response.Error(err)\n\t}\n\treturn response.Success(string(b))\n}", "func (pc *PolicyConfigurator) Handler() common.HTTPRequestHandler {\n\treturn pc.handle\n}", "func Handler(ctx context.Context, payload events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\tvar resp []byte\n\n\tswitch path := payload.Path; path {\n\tcase PRODUCT_PATH:\n\t\tswitch method := payload.HTTPMethod; method {\n\t\tcase GET:\n\t\t\tfmt.Printf(\"GET method for products.\\n\")\n\t\tcase POST:\n\t\t\tnewStock, err := event.CreateStock(payload.Body)\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\n\t\t\tresp, _ = json.Marshal(newStock)\n\t\tcase PUT:\n\t\t\tfmt.Printf(\"PUT method for products.\\n\")\n\t\tcase DELETE:\n\t\t\tfmt.Printf(\"DELETE method for products.\\n\")\n\t\t}\n\n\tcase STORE_PATH:\n\t\tswitch method := payload.HTTPMethod; method {\n\t\tcase GET:\n\t\t\tfmt.Printf(\"GET method for stocks.\\n\")\n\t\tcase POST:\n\t\t\tnewStockLoc, err := event.CreateStockLocation(payload.Body)\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\n\t\t\tresp, _ = json.Marshal(newStockLoc)\n\t\tcase PUT:\n\t\t\tfmt.Printf(\"PUT method for stocks.\\n\")\n\t\tcase DELETE:\n\t\t\tfmt.Printf(\"DELETE method for stocks.\\n\")\n\t\t}\n\tdefault:\n\t\tfmt.Printf(\"panik: %s.\\n\", path)\n\t}\n\n\treturn events.APIGatewayProxyResponse{\n\t\tBody: string(resp),\n\t\tStatusCode: 200,\n\t}, nil\n}", "func (r *Responder) PaymentRequired() { r.write(http.StatusPaymentRequired) }", "func NewHandler(k Keeper) sdk.Handler {\n\treturn func(ctx sdk.Context, msg sdk.Msg) (*sdk.Result, error) {\n\t\tctx = ctx.WithEventManager(sdk.NewEventManager())\n\t\tswitch msg := msg.(type) {\n\t\tcase MsgCreateCDP:\n\t\t\treturn handleMsgCreateCDP(ctx, k, msg)\n\t\tcase MsgDeposit:\n\t\t\treturn handleMsgDeposit(ctx, k, msg)\n\t\tcase MsgWithdraw:\n\t\t\treturn handleMsgWithdraw(ctx, k, msg)\n\t\tcase MsgDrawDebt:\n\t\t\treturn handleMsgDrawDebt(ctx, k, msg)\n\t\tcase MsgRepayDebt:\n\t\t\treturn handleMsgRepayDebt(ctx, k, msg)\n\t\tdefault:\n\t\t\treturn nil, sdkerrors.Wrapf(sdkerrors.ErrUnknownRequest, \"unrecognized %s message type: %T\", ModuleName, msg)\n\t\t}\n\t}\n}", "func Handler(service e2e.Service, hooks *twirp.ServerHooks) *handler.Server {\n\tes := NewExecutableSchema(Config{Resolvers: &Resolver{service}})\n\tsrv := handler.New(es)\n\tsrv.AddTransport(transport.POST{})\n\tsrv.Use(extension.Introspection{})\n\tif hooks == nil {\n\t\treturn srv\n\t}\n\tsrv.AroundFields(func(ctx context.Context, next graphql.Resolver) (res interface{}, err error) {\n\t\tf := graphql.GetFieldContext(ctx)\n\t\tparent := f.Parent.Path().String()\n\t\tif parent != \"\" {\n\t\t\treturn next(ctx)\n\t\t}\n\t\tctx = ctxsetters.WithMethodName(ctx, f.Field.Name)\n\t\tif hooks.RequestRouted != nil {\n\t\t\tctx, err = hooks.RequestRouted(ctx)\n\t\t\tif err != nil {\n\t\t\t\tif terr, ok := err.(twirp.Error); ok && hooks.Error != nil {\n\t\t\t\t\tctx = hooks.Error(ctx, terr)\n\t\t\t\t}\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t\tres, err = next(ctx)\n\t\tif terr, ok := err.(twirp.Error); ok && hooks.Error != nil {\n\t\t\tctx = hooks.Error(ctx, terr)\n\t\t}\n\t\treturn res, err\n\t})\n\treturn srv\n}", "func MakeHandler(s IdentityService, logger kitlog.Logger) http.Handler {\n\topts := []kithttp.ServerOption{\n\t\tkithttp.ServerErrorHandler(transport.NewLogErrorHandler(logger)),\n\t\tkithttp.ServerErrorEncoder(encodeError),\n\t}\n\n\tcreateUserHandler := kithttp.NewServer(\n\t\tmakeCreateUserEndpoint(s),\n\t\tdecodeCreateUserRequest,\n\t\tencodeResponse,\n\t\topts...,\n\t)\n\n\tr := mux.NewRouter()\n\n\tr.Handle(\"/v1/users\", createUserHandler).Methods(\"POST\")\n\n\treturn r\n}", "func Handler(req events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// Log body and pass to the DAO\n\tfmt.Printf(\"Received body: %v\\n\", req)\n\n\trequest := new(vm.GeneralRequest)\n\tresponse := request.Validate(req.Body)\n\tif response.Code != 0 {\n\t\treturn events.APIGatewayProxyResponse{Body: response.Marshal(), StatusCode: 500}, nil\n\t}\n\n\trequest.Date = time.Now().Unix()\n\n\tvar mainTable = \"main\"\n\tif value, ok := os.LookupEnv(\"dynamodb_table_main\"); ok {\n\t\tmainTable = value\n\t}\n\n\t// insert data into the DB\n\tdal.Insert(mainTable, request)\n\n\t// Log and return result\n\tfmt.Println(\"Wrote item: \", request)\n\treturn events.APIGatewayProxyResponse{Body: response.Marshal(), StatusCode: 200}, nil\n}", "func (h *Handler) GetProofHandler(w http.ResponseWriter, r *http.Request) {\n\tcredential := chi.URLParam(r, \"credential\")\n\tproof, err := h.service.GenerateProof(credential)\n\tif err != nil {\n\t\t// TODO(walfly): make this more nuanced in the way it chooses a status code\n\t\thttp.Error(w, err.Error(), http.StatusNotFound)\n\t\treturn\n\t}\n\n\tjs, err := json.Marshal(proof)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t_, _ = w.Write(js)\n}", "func NewPaymentPage(signatureHandler SignatureHandler) *PaymentPage {\n\tpaymentPage := PaymentPage{\"https://paymentpage.ecommpay.com/payment\", signatureHandler}\n\n\treturn &paymentPage\n}", "func NewHandler() iop.DataHandler {\n\treturn &Handler{NewLedger(DefaultAuthorizer())}\n}", "func (_BaseContent *BaseContentTransactor) ProcessRequestPayment(opts *bind.TransactOpts, request_ID *big.Int, payee common.Address, label string, amount *big.Int) (*types.Transaction, error) {\n\treturn _BaseContent.contract.Transact(opts, \"processRequestPayment\", request_ID, payee, label, amount)\n}", "func MakeHandler(service Service, logger kitlog.Logger, ml machineLearning.Service) http.Handler {\n\topts := gokit.GetServerOpts(logger, ml)\n\n\talertHandler := kithttp.NewServer(makeAlertEndpoint(service), gokit.DecodeString, gokit.EncodeResponse, opts...)\n\timageAlertHandler := kithttp.NewServer(makeImageAlertEndpoint(service), gokit.DecodeFromBase64, gokit.EncodeResponse, opts...)\n\n\theartbeatAlertHandler := kithttp.NewServer(makeHeartbeatMessageEncpoint(service), gokit.DecodeString, gokit.EncodeResponse, opts...)\n\theartbeatImageHandler := kithttp.NewServer(makeImageHeartbeatEndpoint(service), gokit.DecodeFromBase64, gokit.EncodeResponse, opts...)\n\n\tbusienssAlertHandler := kithttp.NewServer(makeBusinessAlertEndpoint(service), gokit.DecodeString, gokit.EncodeResponse, opts...)\n\n\talertErrorHandler := kithttp.NewServer(makeAlertErrorHandler(service), gokit.DecodeString, gokit.EncodeResponse, opts...)\n\n\tr := mux.NewRouter()\n\n\tr.Handle(\"/alert/\", alertHandler).Methods(\"POST\")\n\tr.Handle(\"/alert/image\", imageAlertHandler).Methods(\"POST\")\n\n\tr.Handle(\"/alert/heartbeat\", heartbeatAlertHandler).Methods(\"POST\")\n\tr.Handle(\"/alert/heartbeat/image\", heartbeatImageHandler).Methods(\"POST\")\n\n\tr.Handle(\"/alert/error\", alertErrorHandler).Methods(\"POST\")\n\n\tr.Handle(\"/alert/business\", busienssAlertHandler).Methods(\"POST\")\n\n\treturn r\n}", "func NewTransactionHandler(addr string, r *mux.Router) TransactionHandler {\n\ttransactionClient := transactionPb.NewTransactionServiceProtobufClient(addr, &http.Client{})\n\ttransactionHandler := TransactionHandler{\n\t\tClient: transactionClient,\n\t\tRouter: r,\n\t}\n\ttransactionHandler.route()\n\treturn transactionHandler\n}", "func NewHandler(k Keeper) sdk.Handler {\n\treturn func(ctx sdk.Context, msg sdk.Msg) (*sdk.Result, error) {\n\t\tctx = ctx.WithEventManager(sdk.NewEventManager())\n\t\tswitch msg := msg.(type) {\n\t\tcase MsgCreateAtomicSwap:\n\t\t\treturn handleMsgCreateAtomicSwap(ctx, k, msg)\n\t\tcase MsgClaimAtomicSwap:\n\t\t\treturn handleMsgClaimAtomicSwap(ctx, k, msg)\n\t\tcase MsgRefundAtomicSwap:\n\t\t\treturn handleMsgRefundAtomicSwap(ctx, k, msg)\n\t\tdefault:\n\t\t\treturn nil, sdkerrors.Wrapf(sdkerrors.ErrUnknownRequest, \"unrecognized %s message type: %T\", ModuleName, msg)\n\t\t}\n\t}\n}", "func Handler(w http.ResponseWriter, r *http.Request) {\n\thelper.SetupResponse(&w, r)\n\ti := invoice.Invoice{}\n\tif (*r).Method == \"OPTIONS\" {\n\t\treturn\n\t}\n\tif (*r).Method == \"GET\" {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tuserID := r.FormValue(\"userID\")\n\t\tinvoiceID := r.FormValue(\"invoiceID\")\n\t\tlessonID := r.FormValue(\"lessonID\")\n\t\tmode := r.FormValue(\"mode\")\n\n\t\tif mode == \"1\" {\n\t\t\tlogs := i.Read(invoiceID)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else if mode == \"2\" {\n\t\t\tlogs := i.ReadItemLineItem(invoiceID, lessonID)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else if mode == \"3\" {\n\t\t\tlogs := i.GetUnpaidInvoice(userID)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else if mode == \"4\" {\n\t\t\tlogs := i.GetInvoiceLineItemByInvoiceID(invoiceID)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else {\n\t\t\tlogs := i.GetAllInvoice()\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t}\n\t} else if (*r).Method == \"POST\" {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\t// Invoice\n\t\tuserID := r.FormValue(\"userID\")\n\t\tcreateDate := r.FormValue(\"createDate\")\n\t\ttotal := r.FormValue(\"total\")\n\t\tdetail := r.FormValue(\"detail\")\n\t\tstatus := r.FormValue(\"status\")\n\n\t\t// Line item\n\t\tinvoiceID := r.FormValue(\"invoiceID\")\n\t\tlessonID := r.FormValue(\"lessonID\")\n\t\tquantityDay := r.FormValue(\"quantityDay\")\n\t\tamountTotal := r.FormValue(\"amountTotal\")\n\n\t\tmode := r.FormValue(\"mode\")\n\n\t\tif mode == \"1\" {\n\t\t\tlogs := i.AddItemToLineItem(invoiceID, lessonID, quantityDay, amountTotal)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else {\n\t\t\tlogs := i.Create(userID, createDate, total, detail, status)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t}\n\n\t} else if (*r).Method == \"PUT\" {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\t// Invoice\n\t\tinvoiceID := r.FormValue(\"invoiceID\")\n\t\tuserID := r.FormValue(\"userID\")\n\t\tcreateDate := r.FormValue(\"createDate\")\n\t\ttotal := r.FormValue(\"total\")\n\t\tdetail := r.FormValue(\"detail\")\n\t\tstatus := r.FormValue(\"status\")\n\n\t\t// Line item\n\t\tlessonID := r.FormValue(\"lessonID\")\n\t\tquantityDay := r.FormValue(\"quantityDay\")\n\t\tamountTotal := r.FormValue(\"amountTotal\")\n\n\t\tmode := r.FormValue(\"mode\")\n\n\t\tif mode == \"1\" {\n\t\t\tlogs := i.UpdateItemLineItem(invoiceID, lessonID, quantityDay, amountTotal)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else if mode == \"2\" {\n\t\t\tlogs := i.UpdateStatusInvoice(invoiceID, status)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else {\n\t\t\tlogs := i.Update(invoiceID, userID, createDate, total, detail, status)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t}\n\n\t} else if (*r).Method == \"DELETE\" {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\tinvoiceID := r.FormValue(\"invoiceID\")\n\t\tlessonID := r.FormValue(\"lessonID\")\n\n\t\tmode := r.FormValue(\"mode\")\n\t\tif mode == \"1\" {\n\t\t\tlogs := i.DeleteItemLineItem(invoiceID, lessonID)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else if mode == \"2\" {\n\t\t\tlogs := i.CancelInvoice(invoiceID)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t} else {\n\t\t\tlogs := i.Delete(invoiceID)\n\t\t\tjson.NewEncoder(w).Encode(logs)\n\t\t}\n\n\t} else {\n\t\tfmt.Fprintf(w, \"Please use get medthod\")\n\t}\n}", "func Handler(ctx context.Context, evt events.APIGatewayV2HTTPRequest) (Response, error) {\n\tif strings.HasPrefix(evt.RawPath, \"/api\") {\n\t\tswitch evt.RawPath {\n\t\tcase \"/api/pdf\":\n\t\t\titem := []ItemDetail{\n\t\t\t\t{\n\t\t\t\t\tName: \"Front End Consultation\",\n\t\t\t\t\tDesc: \"Experience Review\",\n\t\t\t\t\tAmount: 150,\n\t\t\t\t\tQty: 4,\n\t\t\t\t\tTotal: 600,\n\t\t\t\t},\n\t\t\t}\n\t\t\tdata := Invoice{\n\t\t\t\tInvoiceNo: \"Invoice\",\n\t\t\t\tInvoiceDate: \"January 1, 2019\",\n\t\t\t\tCurrency: \"AUD\",\n\t\t\t\tAmountDue: 600,\n\t\t\t\tItems: item,\n\t\t\t\tTotal: 600,\n\t\t\t}\n\t\t\tinvoice, err := CreateInvoice(\"\", data)\n\t\t\tif err != nil {\n\t\t\t\tprintln(\"Error generating pdf %s\", err)\n\t\t\t\treturn Response{StatusCode: 404}, errors.New(\"Failed to generate the pdf\")\n\t\t\t} else {\n\t\t\t\treturn Response{\n\t\t\t\t\tStatusCode: 200,\n\t\t\t\t\tHeaders: map[string]string{\n\t\t\t\t\t\t\"content-type\": \"application/pdf\",\n\t\t\t\t\t},\n\t\t\t\t\tBody: base64.StdEncoding.EncodeToString(invoice),\n\t\t\t\t\tIsBase64Encoded: true,\n\t\t\t\t}, nil\n\t\t\t}\n\t\tdefault:\n\t\t\treturn Response{StatusCode: 404}, errors.New(\"Api not defined: \" + evt.RawPath)\n\t\t}\n\t} else {\n\t\tcontent, contentType, err := GetStaticContent(rootPath, evt.RawPath)\n\t\tif err != nil {\n\t\t\treturn Response{StatusCode: 404}, err\n\t\t} else {\n\t\t\treturn Response{\n\t\t\t\tStatusCode: 200,\n\t\t\t\tHeaders: map[string]string{\n\t\t\t\t\t\"content-type\": contentType,\n\t\t\t\t},\n\t\t\t\tBody: content,\n\t\t\t}, nil\n\t\t}\n\t}\n}", "func NewHandler(keeper Keeper) sdk.Handler {\n\treturn func(ctx sdk.Context, msg sdk.Msg) (*sdk.Result, error) {\n\t\tctx = ctx.WithEventManager(sdk.NewEventManager())\n\t\tswitch msg := msg.(type) {\n\t\tcase types.MsgFund:\n\t\t\treturn handleMsgFund(ctx, keeper, msg)\n\t\tdefault:\n\t\t\treturn nil, sdkerrors.Wrapf(sdkerrors.ErrUnknownRequest, \"unrecognized %s message type: %T\", ModuleName, msg)\n\t\t}\n\t}\n}", "func NewHandler(\n\tcustomerService objects.CustomerService,\n\tbookService objects.BookService,\n) *Handler {\n\th := &Handler{\n\t\tcustomerService: customerService,\n\t\tbookService: bookService,\n\t}\n\treturn h\n}", "func Handler(request sign.Request) (sign.Response, error) {\n\n\tcert, err := slsslSign.Sign(signer.SignRequest{\n\t\tRequest: string(request.CertificateRequest),\n\t\tProfile: profile,\n\t})\n\n\tif err != nil {\n\t\treturn sign.Response{}, err\n\t}\n\n\treturn sign.Response{\n\t\tCertificate: cert,\n\t}, nil\n}", "func (h *Host) ProcessPayment(stream siamux.Stream, bh types.BlockHeight) (modules.PaymentDetails, error) {\n\t// read the PaymentRequest\n\tvar pr modules.PaymentRequest\n\tif err := modules.RPCRead(stream, &pr); err != nil {\n\t\treturn nil, errors.AddContext(err, \"Could not read payment request\")\n\t}\n\n\t// process payment depending on the payment method\n\tif pr.Type == modules.PayByEphemeralAccount {\n\t\treturn h.staticPayByEphemeralAccount(stream, bh)\n\t}\n\tif pr.Type == modules.PayByContract {\n\t\treturn h.managedPayByContract(stream, bh)\n\t}\n\n\treturn nil, errors.Compose(fmt.Errorf(\"Could not handle payment method %v\", pr.Type), modules.ErrUnknownPaymentMethod)\n}", "func NewDeliveryHandler(out chan *DeliveryReceipt, verifyIPs bool) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, req *http.Request) {\n\t\tif verifyIPs {\n\t\t\t// Check if the request came from Nexmo\n\t\t\thost, _, err := net.SplitHostPort(req.RemoteAddr)\n\t\t\tif !IsTrustedIP(host) || err != nil {\n\t\t\t\thttp.Error(w, \"\", http.StatusInternalServerError)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tlog.Println(\"merf\")\n\n\t\tvar err error\n\t\t// Check if the query and body are empty. If it is, it's just Nexmo\n\t\t// making sure our service is up, so we don't want to return\n\t\t// an error.\n\t\tif req.URL.RawQuery == \"\" && req.ContentLength <= 0 {\n\t\t\tw.WriteHeader(http.StatusOK)\n\t\t\treturn\n\t\t}\n\n\t\tcontentType, ok := req.Header[\"Content-Type\"]\n\t\tif !ok {\n\t\t\tlog.Println(\"foo\")\n\t\t\thttp.Error(w, \"Content-Type not set\", http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\t// nexmo claims the response is going to be of type application/json\n\t\t// add support for application/json and maintain www-form-urlencoded support\n\t\tvar rm *rawDeliveryReceipt\n\t\tif contentType[0] == \"application/json\" {\n\t\t\trm, err = parseJSON(w, req)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t} else if contentType[0] == \"application/x-www-form-urlencoded\" {\n\t\t\trm, err = parseForm(w, req)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t} else {\n\t\t\thttp.Error(w, \"Content-Type \"+contentType[0]+\" not supported.\", http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tm, err := convertTimestamps(rm, w)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\t// Pass it out on the chan\n\t\tout <- m\n\t}\n\n}", "func createPaymentResponse(classOfPayment string) PaymentResponse {\n\tcost := Cost{\n\t\tClassOfPayment: []string{classOfPayment},\n\t}\n\tpaymentResponse := PaymentResponse{\n\t\tCosts: []Cost{cost},\n\t}\n\treturn paymentResponse\n}", "func NewHandler(transporteService domain.TransporteService) http.Handler {\n\thandler := &handler{\n\t\ttransporteService: transporteService,\n\t}\n\n\tgin.SetMode(gin.ReleaseMode)\n\n\trouter := gin.New()\n\trouter.Use(handler.recovery())\n\trouter.GET(\"/\", func(c *gin.Context) {\n\t\tc.HTML(http.StatusOK, \"index.html\", gin.H{\n\t\t\t\"title\": \"Home Page\",\n\t\t})\n\t})\n\n\tv1 := router.Group(\"/v1\")\n\n\tv1.GET(\"/calcularTransporte\", handler.PostCalcularTransporte)\n\tv1.GET(\"/veiculos\", handler.GetVeiculos)\n\n\treturn router\n}", "func addPayment(c *gin.Context) {\n\tpaymentsDB, err := setup(paymentsStorage)\n\n\t//connect to db\n\tif err != nil {\n\t\tlogHandler.Error(\"problem connecting to database\", log.Fields{\"dbname\": paymentsStorage.Cfg.Db, \"func\": \"addPayment\"})\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\"status\": \"error\", \"message\": \"Problem connecting to db\"})\n\t\treturn\n\t}\n\tdefer paymentsDB.Close()\n\n\tvar p storage.Payments\n\terr = c.BindJSON(&p)\n\n\terr = paymentsDB.CreatePayment(&p)\n\tif err != nil {\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\"status\": \"error\", \"message\": \"Could not add a payment\"})\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, gin.H{\"status\": \"success\", \"message\": \"Payment created\"})\n}", "func NewPayment(a payment.Repository, timeout time.Duration) payment.Usecase {\n\treturn &paymentUsecase{\n\t\trepo: a,\n\t\tcontextTimeout: timeout,\n\t}\n}", "func (pp *PayPalService) CreatePaymentAndGenerateNextURL(req *http.Request, paymentResource *models.PaymentResourceRest) (string, ResponseType, error) {\n\n\tlog.TraceR(req, \"performing PayPal request\", log.Data{\"company_number\": paymentResource.CompanyNumber})\n\n\tid := paymentResource.MetaData.ID\n\n\tredirectURL := fmt.Sprintf(\"%s/callback/payments/paypal/orders/%s\",\n\t\tpp.PaymentService.Config.PaymentsAPIURL, paymentResource.MetaData.ID)\n\n\tmoney := &paypal.Money{\n\t\tCurrency: gbp,\n\t\tValue: paymentResource.Amount,\n\t}\n\n\torder, err := pp.Client.CreateOrder(\n\t\tcontext.Background(),\n\t\tpaypal.OrderIntentCapture,\n\t\t[]paypal.PurchaseUnitRequest{\n\t\t\t{\n\t\t\t\tInvoiceID: id,\n\t\t\t\tAmount: &paypal.PurchaseUnitAmount{\n\t\t\t\t\tValue: paymentResource.Amount,\n\t\t\t\t\tCurrency: gbp,\n\t\t\t\t\tBreakdown: &paypal.PurchaseUnitAmountBreakdown{\n\t\t\t\t\t\tItemTotal: money,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tItems: []paypal.Item{\n\t\t\t\t\t{\n\t\t\t\t\t\tName: paymentResource.Reference,\n\t\t\t\t\t\tSKU: paymentResource.CompanyNumber, // SKU = Stock Keeping Unit\n\t\t\t\t\t\tQuantity: \"1\",\n\t\t\t\t\t\tUnitAmount: money,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tnil,\n\t\t&paypal.ApplicationContext{\n\t\t\tReturnURL: redirectURL,\n\t\t\tCancelURL: redirectURL,\n\t\t},\n\t)\n\tif err != nil {\n\t\treturn \"\", Error, fmt.Errorf(\"error creating order: [%v]\", err)\n\t}\n\n\tif order.Status != paypal.OrderStatusCreated {\n\t\tlog.Debug(fmt.Sprintf(\"paypal order response status: %s\", order.Status))\n\t\treturn \"\", Error, fmt.Errorf(\"failed to correctly create paypal order - status is not CREATED\")\n\t}\n\n\tvar nextURL string\n\tvar externalStatusURI string\n\n\tfor _, link := range order.Links {\n\t\tif link.Rel == \"approve\" {\n\t\t\tnextURL = link.Href\n\t\t}\n\t\tif link.Rel == \"self\" {\n\t\t\texternalStatusURI = link.Href\n\t\t}\n\t}\n\n\terr = pp.PaymentService.StoreExternalPaymentStatusDetails(paymentResource.MetaData.ID, externalStatusURI, order.ID)\n\tif err != nil {\n\t\treturn \"\", Error, fmt.Errorf(\"error storing PayPal external payment details for payment session: [%s]\", err)\n\t}\n\n\treturn nextURL, Success, nil\n}", "func NewHandler(logger *zap.Logger, destination string) *Handler {\n\t handler := &Handler{\n\t\t logger: logger,\n\t\t dispatcher: provisioners.NewMessageDispatcher(logger.Sugar()),\n\t\t destination: destination,\n\t }\n\t // The receiver function needs to point back at the handler itself, so set it up after\n\t // initialization.\n\t handler.receiver = provisioners.NewMessageReceiver(createReceiverFunction(handler), logger.Sugar())\n \n\t return handler\n }", "func NewHandler(dm DeveloperManager, am acc.AccountManager, gm *global.GlobalManager) sdk.Handler {\n\treturn func(ctx sdk.Context, msg sdk.Msg) sdk.Result {\n\t\tswitch msg := msg.(type) {\n\t\tcase DeveloperRegisterMsg:\n\t\t\treturn handleDeveloperRegisterMsg(ctx, dm, am, msg)\n\t\tcase DeveloperUpdateMsg:\n\t\t\treturn handleDeveloperUpdateMsg(ctx, dm, am, msg)\n\t\tcase GrantPermissionMsg:\n\t\t\treturn handleGrantPermissionMsg(ctx, dm, am, msg)\n\t\tcase PreAuthorizationMsg:\n\t\t\treturn handlePreAuthorizationMsg(ctx, dm, am, msg)\n\t\tcase DeveloperRevokeMsg:\n\t\t\treturn handleDeveloperRevokeMsg(ctx, dm, am, gm, msg)\n\t\tcase RevokePermissionMsg:\n\t\t\treturn handleRevokePermissionMsg(ctx, dm, am, msg)\n\t\tdefault:\n\t\t\terrMsg := fmt.Sprintf(\"Unrecognized developer msg type: %v\", reflect.TypeOf(msg).Name())\n\t\t\treturn sdk.ErrUnknownRequest(errMsg).Result()\n\t\t}\n\t}\n}", "func MakeHandler(ts Service, logger kitlog.Logger) http.Handler {\n\tr := mux.NewRouter()\n\n\topts := []kithttp.ServerOption{\n\t\tkithttp.ServerErrorLogger(logger),\n\t\tkithttp.ServerErrorEncoder(encodeError),\n\t}\n\n\ttrackCargoHandler := kithttp.NewServer(\n\t\tmakeTrackCargoEndpoint(ts),\n\t\tdecodeTrackCargoRequest,\n\t\tencodeResponse,\n\t\topts...,\n\t)\n\n\tr.Handle(\"/tracking/v1/cargos/{id}\", trackCargoHandler).Methods(\"GET\")\n\n\treturn r\n}", "func NewHandler(ck CoinKeeper) sdk.Handler {\n\treturn func(ctx sdk.Context, msg sdk.Msg) sdk.Result {\n\t\tswitch msg := msg.(type) {\n\t\tcase SendMsg:\n\t\t\treturn handleSendMsg(ctx, ck, msg)\n\t\tcase IssueMsg:\n\t\t\treturn handleIssueMsg(ctx, ck, msg)\n\t\tdefault:\n\t\t\terrMsg := \"Unrecognized bank Msg type: \" + reflect.TypeOf(msg).Name()\n\t\t\treturn sdk.ErrUnknownRequest(errMsg).Result()\n\t\t}\n\t}\n}", "func NewHandler(cfg Config) http.Handler {\n\tr := resolver.New(\n\t\tresolver.Config{\n\t\t\tLogger: cfg.Logger,\n\t\t\tReceiverFactory: cfg.ReceiverFactory,\n\t\t},\n\t)\n\tu := &jobs{cfg.Logger, r}\n\trouter := mux.NewRouter()\n\trouter.Use(\n\t\tfunc(next http.Handler) http.Handler {\n\t\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\t\tctx := newServicesContext(r.Context())\n\t\t\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t\t\t})\n\t\t})\n\troutes := []struct {\n\t\tname string\n\t\thandler http.HandlerFunc\n\t}{\n\t\t{\"sync_services\", u.syncServices},\n\t\t{\"gc\", u.garbageCollector},\n\t}\n\tfor _, route := range routes {\n\t\trouter.Path(\"/\" + route.name).\n\t\t\tMethods(http.MethodPost).\n\t\t\tHandlerFunc(route.handler).\n\t\t\tName(route.name)\n\t}\n\treturn router\n}", "func (s *CreatePaymentService) Execute(request CreatePaymentRequest) (response CreatePaymentResponse, err error) {\n\t// Figure out if the request implies creating a new user or user is already known\n\t// Does request has an auth token?\n\tuser, err := s.getUser(request)\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\t// Get product information from db\n\tproduct, _, err := s.ProductRepository.FindByCode(request.ProductCode)\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\t// Get the discount if any\n\tdiscount, count, err := s.DiscountRepository.FindByCode(request.DiscountCode)\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\t// Calculate the price and apply discount if any\n\tprice := payment.NewPrice(product.Price, product.Currency)\n\tif count != 0 && discount.IsValid() == true {\n\t\tprice.ApplyDiscount(discount.Percentage)\n\t}\n\n\tprice.ApplyDiscount(discount.Percentage)\n\n\t// Make payment with payment provider\n\tpaymentService := payment.Service{}\n\tparams := payment.Params{\n\t\tUserID: user.ID,\n\t\tProductID: product.ID,\n\t\tPrice: *price,\n\t\tNonce: request.Nonce,\n\t}\n\tpay, err := paymentService.Pay(\n\t\ts.PaymentRepository,\n\t\tparams,\n\t)\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\t// Generate 3 activation tokens in user account\n\ttokens, err := t.NewCount(user.ID, user.Email, 3)\n\tif err != nil {\n\t\treturn response, err\n\t}\n\tfor _, token := range *tokens {\n\t\t_, err = s.TokenRepository.Add(&token)\n\t\tif err != nil {\n\t\t\treturn response, err\n\t\t}\n\t}\n\n\t// Send an invoice email to the user\n\terr = s.EmailService.Send(\"payment\", s.createEmailPayload(\n\t\tuser.Name,\n\t\tuser.Email,\n\t\tproduct.Name,\n\t\tpay.Created,\n\t\tprice.GetAmount(),\n\t\tprice.WithTVA(),\n\t))\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\treturn CreatePaymentResponse{\n\t\tPaymentID: pay.ID,\n\t\tUserID: user.ID,\n\t\tProductID: product.ID,\n\t\tPrice: price.WithTVA(),\n\t\tCurrency: product.Currency,\n\t\tProductName: product.Name,\n\t\tUserName: user.Name,\n\t\tToken: user.JWT,\n\t\tSuccess: true,\n\t}, nil\n}", "func UpdatePayment(repo repository.Repository) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tpaymentID := mux.Vars(r)[\"paymentID\"]\n\n\t\tvar t *model.Payment\n\t\tif err := json.NewDecoder(r.Body).Decode(&t); err != nil {\n\t\t\tSendErrorResponse(w, r, http.StatusBadRequest, err)\n\t\t\treturn\n\t\t}\n\t\terr := validate(t)\n\t\tif err != nil {\n\t\t\tSendErrorResponse(w, r, http.StatusBadRequest, err)\n\t\t\treturn\n\t\t}\n\t\t//to ensure that the users does not try to modify a different payment\n\t\tt.ID = paymentID\n\t\terr = repo.Update(t)\n\t\tif err != nil {\n\t\t\tSendErrorResponse(w, r, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\t\tw.WriteHeader(http.StatusNoContent)\n\t\treturn\n\t})\n}", "func (h *Handler) Create(ctx *gin.Context) {\n\tvar payment models.PaymentInput\n\terr := json.NewDecoder(ctx.Request.Body).Decode(&payment)\n\tif err != nil {\n\t\thandleError(ctx, err, http.StatusBadRequest)\n\t} else {\n\t\tid, err := h.PaymentResource.Create(ctx, payment)\n\t\tif err != nil {\n\t\t\thandleError(ctx, err, 0)\n\t\t} else {\n\t\t\tctx.JSON(http.StatusCreated, models.JSONAPISuccessObject{Status: http.StatusCreated, ID: id})\n\t\t}\n\t}\n}", "func PolicyServiceHandler() http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\t// If the method is GET tell the client the server exists with a 200 response. Else the request is a POST return the server configuration.\n\t\tif r.Method == http.MethodGet {\n\t\t\tw.WriteHeader(http.StatusOK)\n\t\t\treturn\n\t\t}\n\n\t\t// Decode The HTTP Request Body From The Client To The cmd varible using the struct defined above.\n\t\tvar cmd windowstype.MdePolicyServiceRequest\n\t\tif err := cmd.Decode(r.Body); err != nil {\n\t\t\tpanic(err) // TODO: Error Handling\n\t\t}\n\n\t\t// // Use the verify function provided by the type to check it is correct\n\t\tif err := cmd.Verify(); err != nil {\n\t\t\tpanic(err) // TODO: Error Handling\n\t\t}\n\n\t\t// Verify the server is the one the device thinks it is talking to. This is to prevent someone implementing a proxy with a different domain.\n\t\t// TODO: log.Println(cmd.Header.To.Host)\n\n\t\t// Verify the users email and password against login directory\n\t\t// TODO: log.Println(cmd.Header.Security.UsernameToken.Username + \" \" + cmd.Header.Security.UsernameToken.Password.Text)\n\t\t// TODO: Can a gracefull error message be sent to the client cause if not just check the logijn in the enrollmentservice\n\n\t\t// Log Action\n\t\tlog.Println(\"Username: \" + cmd.Header.Security.UsernameToken.Username + \" Password: \" + cmd.Header.Security.UsernameToken.Password.Text + \" Action: PolicyService\")\n\n\t\t// Create Response\n\t\t// res := windowstype.Envelope{\n\t\t// \tS: \"http://www.w3.org/2003/05/soap-envelope\",\n\t\t// \tA: \"http://www.w3.org/2005/08/addressing\",\n\t\t// \tHeader: windowstype.NewHeader(\"http://schemas.microsoft.com/windows/pki/2009/01/enrollmentpolicy/IPolicy/GetPoliciesResponse\", cmd.Header.MessageID),\n\t\t// \tBody: windowstype.Body{\n\t\t// \t\tMdePolicyServiceResponse: windowstype.MdePolicyServiceResponse{\n\t\t// \t\t\t// \t\t\t\tAuthPolicy: \"OnPremise\",\n\t\t// \t\t\t// \t\t\t\tEnrollmentVersion: \"4.0\",\n\t\t// \t\t\t// \t\t\t\tEnrollmentPolicyServiceURL: \"https://\" + ConfigDomain + \"/EnrollmentServer/PolicyService.svc\", // TODO: Use propper url package for generation\n\t\t// \t\t\t// \t\t\t\tEnrollmentServiceURL: \"https://\" + ConfigDomain + \"/EnrollmentServer/EnrollmentService.svc\", // TODO: Use propper url package for generation\n\t\t// \t\t},\n\t\t// \t},\n\t\t// }\n\n\t\t// // Send Response To The Client\n\t\t// if err := res.Encode(w); err != nil {\n\t\t// \tpanic(err) // TODO: Error Handling\n\t\t// }\n\n\t\t// TODO:\n\t\t// \t- nextUpdateHours - What is a reccomneded value. I just made that one up for testing\n\t\t// \t- policiesNotChanged - Implement this functionality\n\n\t\tresponse := []byte(`<s:Envelope xmlns:s=\"http://www.w3.org/2003/05/soap-envelope\" xmlns:a=\"http://www.w3.org/2005/08/addressing\" xmlns:u=\"http://docs.oasis-open.org/wss/2004/01/oasis-200401-wss-wssecurity-utility-1.0.xsd\">\n\t\t <s:Header>\n\t\t\t <a:Action s:mustUnderstand=\"1\">http://schemas.microsoft.com/windows/pki/2009/01/enrollmentpolicy/IPolicy/GetPoliciesResponse</a:Action>\n\t\t\t <a:RelatesTo>` + cmd.Header.MessageID + `</a:RelatesTo>\n\t\t </s:Header>\n\t\t <s:Body xmlns:xsd=\"http://www.w3.org/2001/XMLSchema\" xmlns:xsi=\"http://www.w3.org/2001/XMLSchema-instance\">\n\t\t\t <GetPoliciesResponse xmlns=\"http://schemas.microsoft.com/windows/pki/2009/01/enrollmentpolicy\">\n\t\t\t\t<xcep:response>\n\t\t\t\t\t<policyID>` + \"d16de3a0-a087-4308-9344-e169fb528c0b\" + `</policyID>\n\t\t\t\t\t<policyFriendlyName>` + \"Mattrax Identity\" + `</policyFriendlyName>\n\t\t\t\t\t<nextUpdateHours>` + \"1\" + `</nextUpdateHours>\n\t\t\t\t\t<policiesNotChanged>` + \"false\" + `</policiesNotChanged>\n\t\t\t\t\t<policies>\n\t\t\t\t\t\t<policy>\n\t\t\t\t\t\t\t<policyOIDReference>0</policyOIDReference>\n\t\t\t\t\t\t\t<cAs xsi:nil=\"true\" />\n\t\t\t\t\t\t\t<attributes>\n\t\t\t\t\t\t\t\t<!--<commonName>CEPUnitTest</commonName>-->\n\t\t\t\t\t\t\t\t<policySchema>3</policySchema>\n\t\t\t\t\t\t\t\t<certificateValidity>\n\t\t\t\t\t\t\t\t<validityPeriodSeconds>1209600</validityPeriodSeconds>\n\t\t\t\t\t\t\t\t<renewalPeriodSeconds>172800</renewalPeriodSeconds>\n\t\t\t\t\t\t\t\t</certificateValidity>\n\t\t\t\t\t\t\t\t<permission>\n\t\t\t\t\t\t\t\t<enroll>true</enroll>\n\t\t\t\t\t\t\t\t<autoEnroll>false</autoEnroll>\n\t\t\t\t\t\t\t\t</permission>\n\t\t\t\t\t\t\t\t<privateKeyAttributes>\n\t\t\t\t\t\t\t\t<minimalKeyLength>2048</minimalKeyLength>\n\t\t\t\t\t\t\t\t<keySpec xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<keyUsageProperty xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<permissions xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<algorithmOIDReference xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<cryptoProviders xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t</privateKeyAttributes>\n\t\t\t\t\t\t\t\t<revision>\n\t\t\t\t\t\t\t\t<majorRevision>101</majorRevision>\n\t\t\t\t\t\t\t\t<minorRevision>0</minorRevision>\n\t\t\t\t\t\t\t\t</revision>\n\t\t\t\t\t\t\t\t<supersededPolicies xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<privateKeyFlags xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<subjectNameFlags xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<enrollmentFlags xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<generalFlags xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<hashAlgorithmOIDReference>0</hashAlgorithmOIDReference>\n\t\t\t\t\t\t\t\t<rARequirements xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<keyArchivalAttributes xsi:nil=\"true\" />\n\t\t\t\t\t\t\t\t<extensions xsi:nil=\"true\" />\n\t\t\t\t\t\t\t</attributes>\n\t\t\t\t\t\t</policy>\n\t\t\t\t\t</policies>\n\t\t\t\t</response>\n\t\t\t\t<xcep:cAs></xcep:cAs>\n\t\t\t\t<xcep:oIDs></xcep:oIDs>\n\t\t\t </GetPoliciesResponse>\n\t\t </s:Body>\n\t\t</s:Envelope>`)\n\n\t\tw.Header().Set(\"Content-Length\", strconv.Itoa(len(response)))\n\t\tw.Header().Set(\"Transfer-Encoding\", \"identity\")\n\t\tw.Write(response)\n\t}\n}", "func (s *Server) handleCustomerMakePurchase(writer http.ResponseWriter, request *http.Request) {\n\n\tvar item *types.Sales\n\n\tif err := json.NewDecoder(request.Body).Decode(&item); err != nil {\n\t\tlog.Println(err)\n\t\thttp.Error(writer, http.StatusText(http.StatusBadRequest), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tpurchase, err := s.customersSvc.MakePurchase(request.Context(), item)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\thttp.Error(writer, http.StatusText(http.StatusInternalServerError), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\trespondJSON(writer, purchase)\n\n\n\n\n\n\n\t\n}", "func NewHandler(keeper Keeper) sdk.Handler {\n\treturn func(ctx sdk.Context, msg sdk.Msg) sdk.Result {\n\t\tswitch msg := msg.(type) {\n\t\tcase MsgLockCoins:\n\t\t\treturn handleMsgLockCoins(ctx, keeper, msg)\n\t\tcase MsgUnlockCoins:\n\t\t\treturn handleMsgUnlockCoins(ctx, keeper, msg)\n\n\t\tdefault:\n\t\t\terrMsg := fmt.Sprintf(\"unrecognized %s message type: %T\", RouterKey, msg)\n\t\t\treturn sdk.ErrUnknownRequest(errMsg).Result()\n\t\t}\n\t}\n}", "func New(secret string, h Func) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tb, err := ioutil.ReadAll(r.Body)\n\t\tif err != nil {\n\t\t\tlog.WithError(err).Error(\"reading body\")\n\t\t\tresponse.InternalServerError(w)\n\t\t\treturn\n\t\t}\n\n\t\tsignature := r.Header.Get(\"Stripe-Signature\")\n\t\te, err := webhook.ConstructEvent(b, signature, secret)\n\t\tif err != nil {\n\t\t\tlog.WithError(err).Error(\"constructing event\")\n\t\t\tresponse.InternalServerError(w)\n\t\t\treturn\n\t\t}\n\n\t\tctx := log.WithFields(log.Fields{\n\t\t\t\"event_id\": e.ID,\n\t\t\t\"event_type\": e.Type,\n\t\t})\n\n\t\tctx.Info(\"handling stripe event\")\n\t\tif err := h(&e); err != nil {\n\t\t\tctx.WithError(err).Error(\"handling stripe event\")\n\t\t\tresponse.InternalServerError(w)\n\t\t\treturn\n\t\t}\n\n\t\tctx.Info(\"handled stripe event\")\n\t\tresponse.OK(w)\n\t})\n}", "func CreateOrder(c *soso.Context) {\n\tif c.Token == nil {\n\t\tc.ErrorResponse(403, soso.LevelError, errors.New(\"User not authorized\"))\n\t\treturn\n\t}\n\treq := c.RequestMap\n\n\tredirectKey, _ := req[\"redirect\"].(string)\n\tredirect, ok := conf.GetSettings().PaymentsRedirects[redirectKey]\n\tif !ok {\n\t\tc.ErrorResponse(http.StatusBadRequest, soso.LevelError, errors.New(\"unknown redirect key\"))\n\t\treturn\n\t}\n\n\tamount, _ := req[\"amount\"].(float64)\n\tleadID, _ := req[\"lead_id\"].(float64)\n\n\tcurrency, _ := req[\"currency\"].(float64)\n\tcurrencyName, currencyOK := payment.Currency_name[int32(currency)]\n\n\t// retrieve card number from payments service\n\tshopCardID, _ := req[\"card\"].(float64)\n\tshopCardNumber, err := getCardNumber(c.Token.UID, uint64(shopCardID))\n\n\tif err != nil {\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\treturn\n\t}\n\n\tif amount <= 0 || leadID <= 0 || !currencyOK || shopCardNumber == \"\" {\n\t\tc.ErrorResponse(http.StatusBadRequest, soso.LevelError, errors.New(\"Incorrect parameter\"))\n\t\treturn\n\t}\n\n\tleadInfo, err := getLeadInfo(c.Token.UID, uint64(leadID))\n\tif err != nil {\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\treturn\n\t}\n\n\tdirection, err := paymentDirection(leadInfo.UserRole, true)\n\tif err != nil {\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\treturn\n\t}\n\n\tif direction != payment.Direction_CLIENT_RECV && leadInfo.Shop.Suspended {\n\t\tc.ErrorResponse(http.StatusForbidden, soso.LevelError, errors.New(\"shop is suspended\"))\n\t\treturn\n\t}\n\n\tdata, err := json.Marshal(&payment.UsualData{\n\t\tDirection: direction,\n\t\tConversationId: leadInfo.ConversationId,\n\t})\n\tif err != nil {\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\treturn\n\t}\n\n\trequest := &payment.CreateOrderRequest{\n\t\tData: &payment.OrderData{\n\t\t\tAmount: uint64(amount),\n\t\t\tCurrency: payment.Currency(currency),\n\n\t\t\tLeadId: uint64(leadID),\n\t\t\tShopCardNumber: shopCardNumber,\n\n\t\t\tGateway: \"payture\",\n\t\t\tServiceName: \"api\",\n\t\t\tServiceData: string(data),\n\t\t\tRedirect: redirect,\n\t\t},\n\t\tInfo: &payment.UserInfo{\n\t\t\tUserId: c.Token.UID,\n\t\t},\n\t}\n\n\tif direction == payment.Direction_CLIENT_PAYS {\n\t\tplan, err := getMonetizationPlan(leadInfo.Shop.PlanId)\n\t\tif err != nil {\n\t\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\t\treturn\n\t\t}\n\t\tif plan.TransactionCommission != 0 && plan.CoinsExchangeRate != 0 {\n\t\t\tif plan.PrimaryCurrency != currencyName {\n\t\t\t\tc.ErrorResponse(http.StatusBadRequest, soso.LevelError, errors.New(\"Unexpected currency\"))\n\t\t\t}\n\t\t\trequest.Data.CommissionSource = uint64(leadInfo.Shop.SupplierId)\n\t\t\tfee := uint64(amount*plan.TransactionCommission*plan.CoinsExchangeRate + 0.5)\n\t\t\tif fee == 0 {\n\t\t\t\tfee = 1\n\t\t\t}\n\t\t\trequest.Data.CommissionFee = fee\n\t\t}\n\t}\n\n\t// now -- create the order\n\tctx, cancel := rpc.DefaultContext()\n\tdefer cancel()\n\n\tresp, err := paymentServiceClient.CreateOrder(ctx, request)\n\n\tif err != nil { // RPC errors\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, err)\n\t\treturn\n\t}\n\tif resp.Error > 0 { // service errors\n\t\tc.Response.ResponseMap = map[string]interface{}{\n\t\t\t\"ErrorCode\": resp.Error,\n\t\t\t\"ErrorMessage\": resp.ErrorMessage,\n\t\t}\n\t\tc.ErrorResponse(http.StatusInternalServerError, soso.LevelError, errors.New(resp.ErrorMessage))\n\t\treturn\n\t}\n\n\tc.SuccessResponse(map[string]interface{}{\n\t\t\"id\": resp.Id,\n\t})\n}", "func v4handleDBPostProcesspayment(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tdb.Connection.Close(nil)\n\t}()\n var errorGeneral string\n var errorGeneralNbr string\n var requestData modelito.RequestPayment\n \n errorGeneral=\"\"\nrequestData,errorGeneral =obtainPostParmsProcessPayment(r,errorGeneral) //logicrequest_post.go\n\n\t////////////////////////////////////////////////validate parms\n\t/// START\n\t////////////////////////////////////////////////validate parms\n\t/// START\n \n if errorGeneral==\"\" {\n\n\t\terrorGeneral,errorGeneralNbr= v4ProcessProcessPayment(w , requestData) //logicbusiness.go \n\t}\n\n if errorGeneral!=\"\"{\n \t//send error response if any\n \t//prepare an error JSON Response, if any\n\t\tlog.Print(\"CZ STEP Get the ERROR response JSON ready\")\n\t\t\n\t\t\t/// START\n\t\tfieldDataBytesJson,err := getJsonResponseError(errorGeneral, errorGeneralNbr)\n\t\t////////// write the response (ERROR)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(fieldDataBytesJson)\t\n\t\tif(err!=nil){\n\t\t\t\n\t\t}\n\t\n } \n\t\t\t\t\t\n}", "func Handler(ctx context.Context, req events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\tCurrentAccountID = &req.RequestContext.AccountID\n\n\t// Set baseRequest information lost by integration with gorilla mux\n\tbaseRequest = url.URL{}\n\tbaseRequest.Scheme = req.Headers[\"X-Forwarded-Proto\"]\n\tbaseRequest.Host = req.Headers[\"Host\"]\n\tbaseRequest.Path = req.RequestContext.Stage\n\n\t// If no name is provided in the HTTP request body, throw an error\n\treturn muxLambda.ProxyWithContext(ctx, req)\n}", "func NewPayment(store gopayd.PaymentWriter, script gopayd.ScriptKeyReader, invStore gopayd.InvoiceReaderWriter, sender gopayd.PaymentSender, txrunner gopayd.Transacter) *payment {\n\treturn &payment{\n\t\tstore: store,\n\t\tscript: script,\n\t\tinvStore: invStore,\n\t\tsender: sender,\n\t\ttxrunner: txrunner,\n\t}\n}", "func Handler(w http.ResponseWriter, r *http.Request) {\n\thandlerKeySecret := KeySecret{}\n\tdefer r.Body.Close()\n\tif err := json.NewDecoder(r.Body).Decode(&handlerKeySecret); err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\ttokens := []KeySecret{}\n\tquery := \"SELECT key, secret, rules FROM tokens WHERE key=$1 and secret=$2 LIMIT 1\"\n\tcq := config.PrestConf.Adapter.Query(query, handlerKeySecret.Key, handlerKeySecret.Secret)\n\terr := json.Unmarshal(cq.Bytes(), &tokens)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\tif len(tokens) == 0 {\n\t\thttp.Error(w, \"Key/Secret not found\", http.StatusBadRequest)\n\t\treturn\n\t}\n\ttokenJson, err := json.Marshal(tokens[0])\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t}\n\ttokenString, err := token.Generate(fmt.Sprintf(string(tokenJson)))\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t}\n\tauthPF := Auth{\n\t\tData: tokens[0],\n\t\tToken: tokenString,\n\t}\n\tw.WriteHeader(http.StatusOK)\n\tret, _ := json.Marshal(authPF)\n\tw.Write(ret)\n}", "func (client *Client) PaymentCreate(draft *PaymentDraft) (result *Payment, err error) {\n\terr = client.Create(PaymentURLPath, nil, draft, &result)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result, nil\n}", "func MakeHTTPHandler(endpoints Endpoints, tracer stdopentracing.Tracer,\n\tlogger log.Logger) http.Handler {\n\tr := mux.NewRouter()\n\toptions := []httptransport.ServerOption{\n\t\thttptransport.ServerErrorEncoder(errorEncoder),\n\t}\n\n\tCreateSoireeHTTPHandler(endpoints, tracer, logger, r, options)\n\tDeliverOrderHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetOrderHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetOrdersBySoireeHTTPHandler(endpoints, tracer, logger, r, options)\n\tSearchOrdersHTTPHandler(endpoints, tracer, logger, r, options)\n\tPutOrderHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetSoireesHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetStatHTTPHandler(endpoints, tracer, logger, r, options)\n\tLoginProHTTPHandler(endpoints, tracer, logger, r, options)\n\tRegisterProHTTPHandler(endpoints, tracer, logger, r, options)\n\tCreateEstabHTTPHandler(endpoints, tracer, logger, r, options)\n\tUpdateEstabHTTPHandler(endpoints, tracer, logger, r, options)\n\tUpdateProHTTPHandler(endpoints, tracer, logger, r, options)\n\tCreateConsoHTTPHandler(endpoints, tracer, logger, r, options)\n\tCreateMenuHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetConsoHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetConsoByOrderIDHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetMenuHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetEstablishmentTypeHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetSoireeOrdersHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetProEstablishmentsHTTPHandler(endpoints, tracer, logger, r, options)\n\tGetAnalysePHTTPHandler(endpoints, tracer, logger, r, options)\n\tDeleteEstabHTTPHandler(endpoints, tracer, logger, r, options)\n\tDeleteSoireeHTTPHandler(endpoints, tracer, logger, r, options)\n\n\treturn r\n}", "func Handler(request events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\trepo := secondary.NewDynamoRepository()\n\tservice := votes.NewService(repo)\n\tprimary := primary.NewAPIGatewayPrimaryAdapter(service)\n\n\treturn primary.HandleVote(request)\n}", "func WithPaymentCtx(repo repository.Repository, next func(repository.Repository) http.HandlerFunc) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tpaymentID := mux.Vars(r)[\"paymentID\"]\n\t\tif paymentID == \"\" {\n\t\t\tSendErrorResponse(w, r, http.StatusNotFound, errors.Errorf(\"bad request paymentID:%s\", paymentID))\n\t\t\treturn\n\t\t}\n\t\t_, err := repo.Get(paymentID)\n\t\tif err != nil {\n\t\t\tif err == repository.ErrNotFound {\n\t\t\t\tSendErrorResponse(w, r, http.StatusNotFound, errors.Errorf(\"paymentID:%s not found\", paymentID))\n\t\t\t\treturn\n\t\t\t}\n\t\t\tSendErrorResponse(w, r, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\t\tnext(repo).ServeHTTP(w, r)\n\t})\n}", "func (c BuildTxHandler) Handler(ctx context.Context, request *commonProtoc.BuildTxRequest) (\n\t*commonProtoc.BuildTxResponse, error) {\n\n\tbuildTxVO := c.buildRequest(request)\n\tresVO, err := buildTxService.BuildTx(buildTxVO)\n\n\tif err.IsNotNull() {\n\t\treturn nil, BuildException(err)\n\t}\n\treturn c.buildResponse(resVO), nil\n}", "func makeGetDealByStateHandler(m *mux.Router, endpoints endpoint.Endpoints, options []http.ServerOption) {\n\tm.Methods(\"GET\").Path(\"/api/deals\").Handler(handlers.CORS(handlers.AllowedMethods([]string{\"GET\"}), handlers.AllowedOrigins([]string{\"*\"}))(http.NewServer(endpoints.GetDealByStateEndpoint, decodeGetDealByStateRequest, encodeGetDealByStateResponse, options...)))\n}", "func (a *App) SendPayment(dest dapp.Identity, amount string) (dapp.TX, error) {\n\t// watch for payments to developer address. When one is seen, return\n\n\treturn dapp.TX(\"\"), errors.New(\"not implemented\")\n}", "func Handler() http.Handler {\n\t// Allocate a new router. Gorilla router with O(n) complexity is used as there\n\t// is just one route. Replace it by a trie based multiplexer if the number\n\t// of routes is growing.\n\tr := mux.NewRouter()\n\tapi := r.Path(\"/api/\").Subrouter()\n\n\t// TODO: for type safety use http.Method{Name} constants instead if manually writing\n\t// method names when Go 1.7 is stable and no support of other versions is required.\n\tapi.HandleFunc(\n\t\t\"/recent_purchases/{username:[A-Za-z0-9_.-]+}\", wrap(handlers.PopularPurchases),\n\t).Methods(\"GET\")\n\n\treturn api\n}", "func NewHandler(logger *zap.Logger, destination string) *Handler {\n\thandler := &Handler{\n\t\tlogger: logger,\n\t\tdispatcher: provisioners.NewMessageDispatcher(logger.Sugar()),\n\t\tdestination: fmt.Sprintf(\"http://%s\", destination),\n\t}\n\t// The receiver function needs to point back at the handler itself, so set it up after\n\t// initialization.\n\thandler.receiver = provisioners.NewMessageReceiver(createReceiverFunction(handler), logger.Sugar())\n\n\treturn handler\n}", "func Handler(ctx context.Context, req events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// If no name is provided in the HTTP request body, throw an error\n\t// requestUser := userDetails.GetUser(&req)\n\t// ctxWithUser := context.WithValue(ctx, api.DceCtxKey, *requestUser)\n\t// return muxLambda.ProxyWithContext(ctxWithUser, req)\n\treturn muxLambda.ProxyWithContext(ctx, req)\n}" ]
[ "0.69572175", "0.65829253", "0.6421519", "0.640149", "0.6159726", "0.6038519", "0.5934669", "0.5926997", "0.59172374", "0.5886439", "0.58437103", "0.5839155", "0.5768513", "0.57279027", "0.5699862", "0.56855047", "0.56638527", "0.5659348", "0.56425124", "0.5614429", "0.5570015", "0.55591613", "0.5558756", "0.5544448", "0.5504645", "0.5503517", "0.54432786", "0.54308504", "0.54222697", "0.54051965", "0.53812057", "0.5377582", "0.5341609", "0.53366655", "0.53166765", "0.53119427", "0.53113645", "0.5295358", "0.5293327", "0.52779335", "0.5262428", "0.526083", "0.5258481", "0.5255581", "0.52313817", "0.5227671", "0.5218075", "0.520425", "0.5200852", "0.51932687", "0.5186238", "0.5161846", "0.5160437", "0.51518697", "0.51465285", "0.5136794", "0.51255125", "0.51220286", "0.5118389", "0.51174605", "0.51128423", "0.51001775", "0.5092055", "0.50903684", "0.50862056", "0.5082498", "0.5080944", "0.50752217", "0.5069738", "0.50688416", "0.50636744", "0.50630057", "0.50574654", "0.5054065", "0.5051676", "0.50495565", "0.50474286", "0.50418943", "0.50411177", "0.5040678", "0.5038095", "0.5021378", "0.50202274", "0.50128824", "0.50125927", "0.5012273", "0.5003767", "0.500194", "0.49908292", "0.49894702", "0.49878415", "0.49824423", "0.49703276", "0.496965", "0.4969241", "0.49669254", "0.49646613", "0.49626094", "0.4958538", "0.49527135" ]
0.77579606
0
////////////////////////////////////////////////////////////////////////////// CALLBACKS // ////////////
func (mc *MouseCannon) StateEntry() { for i := 0; i < 20; i++ { mc.bullets = append(mc.bullets, &Bullet{ color: sdl.Color{255, 0, 0, 255}, posRect: &sdl.Rect{0, 0, 2, 2}, }) engine.Register(mc.bullets[i]) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func callback(\n\tservice models.DeviceService,\n\tid string,\n\taction string,\n\tactionType models.ActionType,\n\tlc logger.LoggingClient) error {\n\n\tclient := &http.Client{}\n\turl := service.Addressable.GetCallbackURL()\n\tif len(url) > 0 {\n\t\tbody, err := getBody(id, actionType)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treq, err := http.NewRequest(string(action), url, bytes.NewReader(body))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treq.Header.Add(clients.ContentType, clients.ContentTypeJSON)\n\n\t\tgo makeRequest(client, req, lc)\n\t} else {\n\t\tlc.Info(\"callback::no addressable for \" + service.Name)\n\t}\n\treturn nil\n}", "func (client *Client) ServiceStatusWithCallback(request *ServiceStatusRequest, callback func(response *ServiceStatusResponse, err error)) (<-chan int) {\nresult := make(chan int, 1)\nerr := client.AddAsyncTask(func() {\nvar response *ServiceStatusResponse\nvar err error\ndefer close(result)\nresponse, err = client.ServiceStatus(request)\ncallback(response, err)\nresult <- 1\n})\nif err != nil {\ndefer close(result)\ncallback(nil, err)\nresult <- 0\n}\nreturn result\n}", "func (n *Sub) parseCallback() (err error) {\n\tvar notifyURL *model.NotifyURL\n\tcbm := make(map[string]int8)\n\tlog.Info(\"callback(%v), topic(%s), group(%s)\", n.w.Callback, n.w.Topic, n.w.Group)\n\terr = json.Unmarshal([]byte(n.w.Callback), &cbm)\n\tif err != nil {\n\t\tlog.Error(\" Notify.parseCallback sub parse callback err %v, topic(%s), group(%s), callback(%s)\",\n\t\t\terr, n.w.Topic, n.w.Group, n.w.Callback)\n\t\treturn\n\t}\n\tcbs := make([]*model.Callback, 0, len(cbm))\n\tfor u, p := range cbm {\n\t\tnotifyURL, err = parseNotifyURL(u)\n\t\tif err != nil {\n\t\t\tlog.Error(\"Notify.parseCallback url parse error(%v), url(%s), topic(%s), group(%s)\",\n\t\t\t\terr, u, n.w.Topic, n.w.Group)\n\t\t\treturn\n\t\t}\n\t\tcbs = append(cbs, &model.Callback{URL: notifyURL, Priority: p})\n\t}\n\tsort.Slice(cbs, func(i, j int) bool { return cbs[i].Priority > cbs[j].Priority })\n\tn.w.Callbacks = cbs\n\treturn\n}", "func runCallback(receivedMessage *Message, consumerMessage *sarama.ConsumerMessage) {\n\tcallback := subscribeMap[consumerMessage.Topic][receivedMessage.MessageType]\n\n\tif callback == nil {\n\t\tlogrus.Error(fmt.Sprintf(\"callback not found for topic : %s, message type : %s\", consumerMessage.Topic,\n\t\t\treceivedMessage.MessageType))\n\t\treturn\n\t}\n\n\tgo callback(&Message{\n\t\tTopic: consumerMessage.Topic,\n\t\tMessage: receivedMessage.Message,\n\t\tMessageType: receivedMessage.MessageType,\n\t\tService: receivedMessage.Service,\n\t\tTraceId: receivedMessage.TraceId,\n\t\tMessageId: receivedMessage.MessageId,\n\t}, nil)\n}", "func (app *testbot) Callback(w http.ResponseWriter, r *http.Request) {\n\tevents, err := app.bot.ParseRequest(r)\n\tif err != nil {\n\t\tif err == linebot.ErrInvalidSignature {\n\t\t\tw.WriteHeader(400)\n\t\t} else {\n\t\t\tw.WriteHeader(500)\n\t\t}\n\t\treturn\n\t}\n\tfor _, event := range events {\n\t\tlog.Printf(\"Got event %v\", event)\n\t\tswitch event.Type {\n\t\tcase linebot.EventTypeMessage:\n\t\t\tswitch message := event.Message.(type) {\n\t\t\tcase *linebot.TextMessage:\n\t\t\t\tif err := app.handleText(message, event.ReplyToken, event.Source); err != nil {\n\t\t\t\t\tlog.Print(err)\n\t\t\t\t}\n\t\t\tcase *linebot.ImageMessage:\n\t\t\t\tif err := app.handleImage(message, event.ReplyToken); err != nil {\n\t\t\t\t\tlog.Print(err)\n\t\t\t\t}\n\t\t\tcase *linebot.VideoMessage:\n\t\t\t\tif err := app.handleVideo(message, event.ReplyToken); err != nil {\n\t\t\t\t\tlog.Print(err)\n\t\t\t\t}\n\t\t\tcase *linebot.AudioMessage:\n\t\t\t\tif err := app.handleAudio(message, event.ReplyToken); err != nil {\n\t\t\t\t\tlog.Print(err)\n\t\t\t\t}\n\t\t\tcase *linebot.LocationMessage:\n\t\t\t\tif err := app.handleLocation(message, event.ReplyToken); err != nil {\n\t\t\t\t\tlog.Print(err)\n\t\t\t\t}\n\t\t\tcase *linebot.StickerMessage:\n\t\t\t\tif err := app.handleSticker(message, event.ReplyToken); err != nil {\n\t\t\t\t\tlog.Print(err)\n\t\t\t\t}\n\t\t\tdefault:\n\t\t\t\tlog.Printf(\"Unknown message: %v\", message)\n\t\t\t}\n\t\tcase linebot.EventTypeFollow:\n\t\t\tif err := app.replyText(event.ReplyToken, \"Got followed event\"); err != nil {\n\t\t\t\tlog.Print(err)\n\t\t\t}\n\t\tcase linebot.EventTypeUnfollow:\n\t\t\tlog.Printf(\"Unfollowed this bot: %v\", event)\n\t\tcase linebot.EventTypeJoin:\n\t\t\tif err := app.replyText(event.ReplyToken, \"Joined \"+string(event.Source.Type)); err != nil {\n\t\t\t\tlog.Print(err)\n\t\t\t}\n\t\tcase linebot.EventTypeLeave:\n\t\t\tlog.Printf(\"Left: %v\", event)\n\n\t\tcase linebot.EventTypeBeacon:\n\t\t\tif err := app.replyText(event.ReplyToken, \"Got beacon: \"+event.Beacon.Hwid); err != nil {\n\t\t\t\tlog.Print(err)\n\t\t\t}\n\t\tdefault:\n\t\t\tlog.Printf(\"Unknown event: %v\", event)\n\t\t}\n\t}\n}", "func notifyReady() {\n}", "func callback() {\n\tlog.Println(\"shutdown requested\")\n}", "func onEventCallback(e event.Event, ctx interface{}) {\n\tservice := ctx.(*qutoService)\n\tservice.eventChan <- e\n}", "func onEventCallback(e event.Event, ctx interface{}) {\n\tservice := ctx.(*metadataService)\n\tservice.eventChan <- e\n}", "func callback(nlm *C.struct_nl_msg, nla unsafe.Pointer) C.int {\n\tcbID := uintptr(nla)\n\tcallbacksLock.RLock()\n\tcbArg := callbacks[cbID]\n\tcallbacksLock.RUnlock()\n\n\tif cbArg == nil {\n\t\tpanic(fmt.Sprintf(\"No netlink callback with ID %d\", cbID))\n\t}\n\n\tcbMsg := &Message{nlm: nlm}\n\tif err := cbArg.fn(cbMsg, cbArg.arg); err != nil {\n\t\tcbArg.err = err\n\t\treturn C.NL_STOP\n\t}\n\treturn C.NL_OK\n}", "func (s *Service) TapdCallBack(c context.Context, body io.ReadCloser) (err error) {\n\n\tvar (\n\t\tURLs []string\n\t\tjsonByte []byte\n\t\teventRequest *model.EventRequest\n\t\teventInterface = make(map[string]interface{})\n\t\tcreatedTime time.Time\n\t\tworkspaceID int\n\t\teventID int\n\t)\n\n\tif jsonByte, err = ioutil.ReadAll(body); err != nil {\n\t\treturn\n\t}\n\n\t//get event\n\tif err = json.Unmarshal(jsonByte, &eventRequest); err != nil {\n\t\treturn\n\t}\n\n\tif eventRequest.Secret != s.c.Tapd.CallbackToken {\n\t\terr = ecode.Unauthorized\n\t\treturn\n\t}\n\n\tif err = json.Unmarshal(jsonByte, &eventInterface); err != nil {\n\t\treturn\n\t}\n\n\t// add log\n\tworkspaceID, _ = strconv.Atoi(eventRequest.WorkspaceID)\n\teventID, _ = strconv.Atoi(eventRequest.EventID)\n\n\teventLog := &model.EventLog{\n\t\tEvent: string(eventRequest.Event),\n\t\tWorkspaceID: workspaceID,\n\t\tEventID: eventID,\n\t}\n\tif err = s.dao.AddEventLog(eventLog); err != nil {\n\t\treturn\n\t}\n\n\t//handle special param\n\tif createdTime, err = time.Parse(\"2006-01-02 15:04:05\", eventRequest.Created); err != nil {\n\t\treturn\n\t}\n\teventInterface[\"id\"] = eventRequest.EventID\n\teventInterface[\"created\"] = createdTime.Unix()\n\n\tif URLs, err = s.GetEnableHookURL(c, eventRequest.Event, workspaceID); err != nil {\n\t\treturn\n\t}\n\n\tfor _, URL := range URLs {\n\t\ts.transferChan.Do(context.Background(), func(c context.Context) {\n\t\t\ts.dao.CallHookUrlAsForm(context.Background(), URL, eventInterface)\n\t\t})\n\t}\n\treturn\n}", "func makeCallback(handle *syscall.Handle) uintptr {\n\tcb := syscall.NewCallback(func(h syscall.Handle, p uintptr) uintptr {\n\t\tb := make([]uint16, 200)\n\t\t_, err := getWindowText(h, &b[0], int32(len(b)))\n\t\tif err != nil {\n\t\t\t// ignore the error\n\t\t\treturn 1 // continue enumeration\n\t\t}\n\t\twindowTitle := syscall.UTF16ToString(b)\n\t\tif strings.HasPrefix(windowTitle, \"Spotify\") {\n\t\t\t// note the window\n\t\t\t*handle = h\n\t\t\treturn 0 // stop enumeration\n\t\t}\n\t\treturn 1 // continue enumeration\n\t})\n\treturn cb\n}", "func executeCallback(pipeline *sdk.Pipeline, operation *sdk.Operation, data []byte) error {\n\tvar err error\n\n\tcburl := operation.CallbackUrl\n\tparams := operation.GetParams()\n\theaders := operation.GetHeaders()\n\n\tmethod := os.Getenv(\"default-method\")\n\tif method == \"\" {\n\t\tmethod = \"POST\"\n\t}\n\n\tif m, ok := headers[\"method\"]; ok {\n\t\tmethod = m\n\t}\n\n\thttpreq, err := buildHttpRequest(cburl, method, data, params, headers)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot connect to Function on URL: %s\", cburl)\n\t}\n\n\tif operation.Requesthandler != nil {\n\t\toperation.Requesthandler(httpreq)\n\t}\n\n\tclient := &http.Client{}\n\tresp, err := client.Do(httpreq)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer resp.Body.Close()\n\tif operation.OnResphandler != nil {\n\t\t_, err = operation.OnResphandler(resp)\n\t} else {\n\t\tif resp.StatusCode < 200 || resp.StatusCode > 299 {\n\t\t\tcbresult, _ := ioutil.ReadAll(resp.Body)\n\t\t\terr := fmt.Errorf(\"%v:%s\", err, string(cbresult))\n\t\t\treturn err\n\t\t}\n\t}\n\treturn err\n\n}", "func (self *SinglePad) OnUpCallback() interface{}{\n return self.Object.Get(\"onUpCallback\")\n}", "func newCallBackHandler() (raw.OnewayHandler, <-chan map[string]string) {\n\tserverCalledBack := make(chan map[string]string)\n\treturn func(ctx context.Context, body []byte) error {\n\t\tserverCalledBack <- extractBaggage(ctx)\n\t\treturn nil\n\t}, serverCalledBack\n}", "func (b *Bot) implementCallback(update *tgbotapi.Update) error {\n\tcallback := Callback{}\n\terr := json.Unmarshal([]byte(update.CallbackQuery.Data), &callback)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tgo func() {\n\t\terr = b.sendTextMessage(Message{\n\t\t\tchatID: callback.ChatID,\n\t\t\ttext: downloadStarted})\n\t\tif err != nil {\n\t\t\tlogrus.Error(err)\n\t\t}\n\t}()\n\n\tworkDir, err := os.Getwd()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdownloadsDir := filepath.Join(workDir, b.conf.DownloadsDir)\n\tchatIDDir := filepath.Join(downloadsDir, fmt.Sprint(update.CallbackQuery.Message.Chat.ID))\n\n\tfilename, err := downloadAudioByLink(callback.URL, chatIDDir)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfullFilePath := filepath.Join(chatIDDir, filename)\n\tdefer func() {\n\t\terr = removeContents(chatIDDir)\n\t\tif err != nil {\n\t\t\tlogrus.Errorf(\"remove file %s after upload error: %v\", fullFilePath, err)\n\t\t}\n\t}()\n\n\t_, err = b.api.Send(tgbotapi.NewAudioUpload(callback.ChatID, fullFilePath))\n\n\treturn err\n}", "func parseCallback(r *http.Request, cb *callback) (int, error) {\n\tdefer func() {\n\t\terr := r.Body.Close()\n\t\tif err != nil {\n\t\t\tlog.Printf(\"body close: %v\", err)\n\t\t}\n\t}()\n\n\tcontentType := r.Header.Get(\"Content-Type\")\n\n\tswitch r.Method {\n\tcase \"POST\":\n\t\tswitch contentType {\n\t\tcase \"application/json\":\n\t\t\t// unmarshall JSON to callback struct\n\t\t\tdecoder := json.NewDecoder(r.Body)\n\t\t\tif err := decoder.Decode(cb); err != nil {\n\t\t\t\treturn http.StatusBadRequest, err\n\t\t\t}\n\n\t\tdefault:\n\t\t\treturn http.StatusBadRequest, fmt.Errorf(\"Unsupported content type: %s\", contentType)\n\t\t}\n\n\tdefault:\n\t\treturn http.StatusMethodNotAllowed, fmt.Errorf(\"Unsupported request method: %s\", r.Method)\n\t}\n\n\treturn 0, nil\n}", "func (c *Operation) callback(w http.ResponseWriter, r *http.Request) { //nolint: funlen,gocyclo\n\tif len(r.URL.Query()[\"error\"]) != 0 {\n\t\tif r.URL.Query()[\"error\"][0] == \"access_denied\" {\n\t\t\thttp.Redirect(w, r, c.homePage, http.StatusTemporaryRedirect)\n\t\t}\n\t}\n\n\ttk, err := c.tokenIssuer.Exchange(r)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to exchange code for token: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to exchange code for token: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\t// user info from token will be used for to retrieve data from cms\n\tinfo, err := c.tokenResolver.Resolve(tk.AccessToken)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to get token info: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get token info: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tuserID, subject, err := c.getCMSData(tk, \"email=\"+info.Subject, info.Scope)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to get cms data: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get cms data: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tcallbackURLCookie, err := r.Cookie(callbackURLCookie)\n\tif err != nil && !errors.Is(err, http.ErrNoCookie) {\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get authMode cookie: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tif callbackURLCookie != nil && callbackURLCookie.Value != \"\" {\n\t\ttxnID := uuid.NewString()\n\t\tdata := txnData{\n\t\t\tUserID: userID,\n\t\t\tScope: info.Scope,\n\t\t\tToken: tk.AccessToken,\n\t\t}\n\n\t\tdataBytes, mErr := json.Marshal(data)\n\t\tif mErr != nil {\n\t\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\t\tfmt.Sprintf(\"failed to marshal txn data: %s\", mErr.Error()))\n\t\t\treturn\n\t\t}\n\n\t\terr = c.store.Put(txnID, dataBytes)\n\t\tif err != nil {\n\t\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\t\tfmt.Sprintf(\"failed to save txn data: %s\", err.Error()))\n\n\t\t\treturn\n\t\t}\n\n\t\thttp.Redirect(w, r, callbackURLCookie.Value+\"?txnID=\"+txnID, http.StatusTemporaryRedirect)\n\n\t\treturn\n\t}\n\n\tvcsProfileCookie, err := r.Cookie(vcsProfileCookie)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to get cookie: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get cookie: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tcred, err := c.prepareCredential(subject, info.Scope, vcsProfileCookie.Value)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to create credential: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\tfmt.Sprintf(\"failed to create credential: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\n\tt, err := template.ParseFiles(c.didAuthHTML)\n\tif err != nil {\n\t\tlogger.Errorf(err.Error())\n\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\tfmt.Sprintf(\"unable to load html: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tif err := t.Execute(w, map[string]interface{}{\n\t\t\"Path\": generate + \"?\" + \"profile=\" + vcsProfileCookie.Value,\n\t\t\"Cred\": string(cred),\n\t}); err != nil {\n\t\tlogger.Errorf(fmt.Sprintf(\"failed execute qr html template: %s\", err.Error()))\n\t}\n}", "func MQCALLBACK_Go(hConn C.MQHCONN, mqmd *C.MQMD, mqgmo *C.MQGMO, mqBuffer C.PMQVOID, mqcbc *C.MQCBC) {\n\n\tvar cbHObj *MQObject\n\n\t// Find the real callback function and invoke it\n\t// Invoked function should match signature of the MQCB_FUNCTION type\n\tgogmo := NewMQGMO()\n\tgomd := NewMQMD()\n\tgocbc := NewMQCBC()\n\n\t// For EVENT callbacks, the GMO and MD may be NULL\n\tif mqgmo != (C.PMQGMO)(C.NULL) {\n\t\tcopyGMOfromC(mqgmo, gogmo)\n\t}\n\n\tif mqmd != (C.PMQMD)(C.NULL) {\n\t\tcopyMDfromC(mqmd, gomd)\n\t}\n\n\t// This should never be NULL\n\tcopyCBCfromC(mqcbc, gocbc)\n\n\tmqreturn := &MQReturn{MQCC: int32(mqcbc.CompCode),\n\t\tMQRC: int32(mqcbc.Reason),\n\t\tverb: \"MQCALLBACK\",\n\t}\n\n\tkey := makeKey(hConn, mqcbc.Hobj)\n\tmapLock()\n\tinfo, ok := cbMap[key]\n\tmapUnlock()\n\n\t// The MQ Client libraries seem to sometimes call us with an EVENT\n\t// even if it's not been registered. And therefore the cbMap does not\n\t// contain a matching callback function with the hObj. It has\n\t// been seen with a 2033 return (see issue #75).\n\t//\n\t// This feels like wrong behaviour from the client, but we need to find a\n\t// way to deal with it even if it gets fixed in future.\n\t// The way I've chosen is to find the first entry in\n\t// the map associated with the hConn and call its registered function with\n\t// a dummy hObj.\n\tif !ok {\n\t\tif gocbc.CallType == MQCBCT_EVENT_CALL && mqcbc.Hobj == C.MQHO_NONE {\n\t\t\tkey = makePartialKey(hConn)\n\t\t\tmapLock()\n\t\t\tfor k, i := range cbMap {\n\t\t\t\tif strings.HasPrefix(k, key) {\n\t\t\t\t\tok = true\n\t\t\t\t\tinfo = i\n\t\t\t\t\tcbHObj = &MQObject{qMgr: info.hObj.qMgr, Name: \"\"}\n\t\t\t\t\t// Only care about finding one match in the table\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tmapUnlock()\n\t\t}\n\t} else {\n\t\tcbHObj = info.hObj\n\t}\n\n\tif ok {\n\t\tif gogmo.MsgHandle.hMsg != C.MQHM_NONE {\n\t\t\tgogmo.MsgHandle.qMgr = cbHObj.qMgr\n\t\t}\n\n\t\tgocbc.CallbackArea = info.callbackArea\n\t\tgocbc.ConnectionArea = info.connectionArea\n\n\t\t// Get the data\n\t\tb := C.GoBytes(unsafe.Pointer(mqBuffer), C.int(mqcbc.DataLength))\n\t\t// And finally call the user function\n\t\tinfo.callbackFunction(cbHObj.qMgr, cbHObj, gomd, gogmo, b, gocbc, mqreturn)\n\t}\n}", "func Callback(c *gin.Context) {\n\tprovider := c.Param(\"provider\")\n\n\tvar logincode vo.LoginReq\n\tif err := c.ShouldBindQuery(&logincode); err != nil {\n\t\tfmt.Println(\"xxxx\", err)\n\t}\n\n\tfmt.Println(\"provider\", provider, logincode)\n\n\tuserInfo := vo.GetUserInfoFromOauth(provider, logincode.Code, logincode.State)\n\tfmt.Println(\"get user info\", userInfo)\n\n\tif userInfo == nil {\n\t\tc.JSON(http.StatusOK, sailor.HTTPAirdbResponse{\n\t\t\tCode: enum.AirdbSuccess,\n\t\t\tSuccess: true,\n\t\t\tData: vo.LoginResp{\n\t\t\t\tNickname: \"xxx\",\n\t\t\t\tHeadimgurl: \"xxx.png\",\n\t\t\t},\n\t\t})\n\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, sailor.HTTPAirdbResponse{\n\t\tCode: enum.AirdbSuccess,\n\t\tSuccess: true,\n\t\tData: vo.LoginResp{\n\t\t\tNickname: userInfo.Login,\n\t\t\tHeadimgurl: userInfo.AvatarURL,\n\t\t},\n\t})\n}", "func SpotifyCallback(w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(\"Received Request: /SpotifyCallback\")\n\turlParams := r.URL.Query()\n\tAuthCode = urlParams.Get(\"code\")\n\n\tgetSpotToken()\n\tgetCurrentUser()\n\tcreatePlaylist()\n}", "func (self *SinglePad) OnConnectCallback() interface{}{\n return self.Object.Get(\"onConnectCallback\")\n}", "func (*listener) OnConnect() {}", "func (client *Client) QueryCustomerAddressListWithCallback(request *QueryCustomerAddressListRequest, callback func(response *QueryCustomerAddressListResponse, err error)) (<-chan int) {\nresult := make(chan int, 1)\nerr := client.AddAsyncTask(func() {\nvar response *QueryCustomerAddressListResponse\nvar err error\ndefer close(result)\nresponse, err = client.QueryCustomerAddressList(request)\ncallback(response, err)\nresult <- 1\n})\nif err != nil {\ndefer close(result)\ncallback(nil, err)\nresult <- 0\n}\nreturn result\n}", "func (m *Monitor) notifyCallback(input, output []byte) {\n\tif m.Callback != nil {\n\t\tm.Callback(m, input, output)\n\t}\n}", "func callbackHandler(res http.ResponseWriter, req *http.Request) {\n\n\t// Complete the authentication process and fetch all of the\n\t// basic information about the user from the provider.\n\tuser, err := gothic.CompleteUserAuth(res, req)\n\tif err != nil {\n\t\tfmt.Fprintln(res, err)\n\t\treturn\n\t}\n\n\t// Execute the template for this user and respond with\n\t// the user page.\n\ttemplates.ExecuteTemplate(res, \"user\", user)\n}", "func preNotifyReady() {\n}", "func Callback(w http.ResponseWriter, r *http.Request) {\n\tcode := ParseResponse(w, r)\n\taccess := AccessToken(code, w, r)\n\tfmt.Fprintf(w, access.Token)\n\tGetData(access.Token, w, r)\n}", "func (r *RabbitMq) registerCallback() {\r\n\tr.RabbitMqChannel.QueueDeclare(\r\n\t\t\"callback-mq-producer\",\r\n\t\ttrue,\r\n\t\tfalse,\r\n\t\tfalse,\r\n\t\tfalse,\r\n\t\tnil,\r\n\t)\r\n\tr.StartConsumeCallback()\r\n\tlog.Println(\"Create Queue callback\")\r\n}", "func main() {\n\n\twait := make(chan struct{}, 0)\n\tRegisterCallback(\"ping\", pingCB())\n\tRegisterCallback(\"add\", addCB())\n\tRegisterErrorCallback(\"raiseError\", err)\n\tRegisterValue(\"wasmVal\", \"Hello World\")\n\n\t<-wait\n}", "func handleCallbackEvent(w http.ResponseWriter, api *slack.Client, event slackevents.EventsAPIEvent) {\n\tswitch ev := event.InnerEvent.Data.(type) {\n\tcase *slackevents.AppMentionEvent:\n\n\t\t// Remove user mentions\n\t\tquestion := helpers.RemoveUserMention(ev.Text)\n\n\t\t// Sanitize input\n\t\tquestion = helpers.ExtractQuestionFromMessage(question)\n\t\t// Exit if there is no question\n\t\tif question == \"\" {\n\t\t\tresponse := helpers.GetRandomStringFromSlice(constants.GreetingMessages)\n\t\t\tapi.PostMessage(ev.Channel, slack.MsgOptionText(response, false))\n\t\t\treturn\n\t\t}\n\n\t\t// Any question that has less than 10 characters is probably jumbled\n\t\t// TODO: Change this to count words or do better\n\t\tif len(question) < 10 {\n\t\t\tresponse := \"I don't know how to answer that :thinking_face:\"\n\t\t\tapi.PostMessage(ev.Channel, slack.MsgOptionText(response, false))\n\t\t\treturn\n\t\t}\n\n\t\t// Get the answer from the database\n\t\tresult, err := questionAnswerDAO.GetByQuestion(question, event.TeamID)\n\n\t\t// If no answers were found, prompt channel to help\n\t\tif err != nil {\n\t\t\tquestion = fmt.Sprintf(\"%s\", question)\n\t\t\tresponseAttachment := slack.MsgOptionAttachments(getAnswerNotFoundAttachment(question))\n\t\t\tresponse := helpers.GetRandomStringFromSlice(constants.AnswerNotFoundMessages)\n\t\t\tapi.PostMessage(ev.Channel, slack.MsgOptionText(response, false), responseAttachment)\n\t\t\treturn\n\n\t\t}\n\n\t\t// Respond with answer found in the database\n\t\tresponse := helpers.GetRandomStringFromSlice(constants.AnswerFoundMessages)\n\t\tresultLastUpdated := time.Now().UTC().Sub(result.LastUpdated).Round(time.Hour)\n\t\tattachmentMessage := fmt.Sprintf(\"%s \\n _%s_ \\n \\n Last updated by %s %s ago\", question, result.Answer, result.LastUpdatedBy, resultLastUpdated)\n\t\tresponseAttachment := slack.MsgOptionAttachments(getAnswerFoundAttachment(attachmentMessage))\n\t\tapi.PostMessage(ev.Channel, slack.MsgOptionText(response, false), responseAttachment)\n\t\treturn\n\tcase *slackevents.MessageEvent:\n\t\tif ev.SubType == \"bot_add\" {\n\t\t\t_, _, channelID, err := api.OpenIMChannel(ev.User)\n\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"%s\\n\", err)\n\t\t\t}\n\n\t\t\tapi.PostMessage(channelID, slack.MsgOptionText(\"Thank you for adding me to the workspace!\", false))\n\t\t\treturn\n\t\t}\n\tdefault:\n\t\tlog.Print(fmt.Sprintf(\"Uncaught Event: %+v\\n\", ev))\n\t}\n}", "func (handler *HTTPCallBackHanlder) Callback(c echo.Context) error {\n\n\tctx := c.Request().Context()\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\tevents, err := handler.Bot.ParseRequest(c.Request())\n\tif err != nil {\n\t\tif err == linebot.ErrInvalidSignature {\n\t\t\tc.String(400, linebot.ErrInvalidSignature.Error())\n\t\t} else {\n\t\t\tc.String(500, \"internal\")\n\t\t}\n\t}\n\n\tfor _, event := range events {\n\t\tif event.Type == linebot.EventTypeMessage {\n\t\t\tswitch message := event.Message.(type) {\n\t\t\tcase *linebot.TextMessage:\n\t\t\t\tmessageFromPing := servicemanagement.PingService(message.Text, handler.ServicesInfo, time.Second*5)\n\t\t\t\tif _, err = handler.Bot.ReplyMessage(event.ReplyToken, linebot.NewTextMessage(messageFromPing)).Do(); err != nil {\n\t\t\t\t\tlog.Print(err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn c.JSON(200, \"\")\n}", "func callback(w http.ResponseWriter, r *http.Request) {\n\tsession, err := store.Get(r, sessionStoreKey)\n\tif err != nil {\n\t\tfmt.Fprintln(w, \"aborted\")\n\t\treturn\n\t}\n\n\tif r.URL.Query().Get(\"state\") != session.Values[\"state\"] {\n\t\tfmt.Fprintln(w, \"no state match; possible csrf OR cookies not enabled\")\n\t\treturn\n\t}\n\n\ttoken, err := oauthCfg.Exchange(oauth2.NoContext, r.URL.Query().Get(\"code\"))\n\tif err != nil {\n\t\tfmt.Fprintln(w, \"there was an issue getting your token\")\n\t\treturn\n\t}\n\n\tif !token.Valid() {\n\t\tfmt.Fprintln(w, \"retreived invalid token\")\n\t\treturn\n\t}\n\n\tclient := github.NewClient(oauthCfg.Client(oauth2.NoContext, token))\n\n\tuser, _, err := client.Users.Get(context.Background(), \"\")\n\tif err != nil {\n\t\tfmt.Println(w, \"error getting name\")\n\t\treturn\n\t}\n\n\tsession.Values[\"githubUserName\"] = user.Name\n\tsession.Values[\"githubAccessToken\"] = token\n\tsession.Save(r, w)\n\n\thttp.Redirect(w, r, \"/\", 302)\n\n}", "func GoCallbackWrapper(ptr_q *unsafe.Pointer, ptr_nfad *unsafe.Pointer) int {\n q := (*Queue)(unsafe.Pointer(ptr_q))\n payload := build_payload(q.c_gh, ptr_nfad)\n return q.cb(payload)\n}", "func CallbackHandler(w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(\"In CallbackHandler!\")\n\tfmt.Println(r.URL.Query().Get(\"hub.verify_token\"))\n\tif r.URL.Query().Get(\"hub.verify_token\") == cfg.Instagram.Verify {\n\t\tfmt.Println(\"verify hihi OK!!\")\n\t\tw.Write([]byte(r.URL.Query().Get(\"hub.challenge\")))\n\t} else {\n\t\thttp.Error(w, http.StatusText(http.StatusInternalServerError), http.StatusInternalServerError)\n\t}\n}", "func SyncCallback(e *sync.SyncEvent) error {\n\tvar id string\n\tvar e2 sync.SyncEvent\n\n\t// Change notifications must usually be routed depending on the type of change.\n\t// This is done by using Field (to go down a struct) and Value (to access a map) methods.\n\t// Those functions can be chained. Checking for error later will tell if\n\t// any of the steps failed.\n\tif e2 = e.Field(\"Nodes\").Value(&id); e2.Error() == nil {\n\t\t// Here we know the change is a Node object in the Nodes map.\n\t\t// The key is stored in 'id'.\n\n\t\t// Current gets us the Node object\n\t\tc, _ := e2.Current()\n\n\t\tfmt.Printf(\"Modified Node with key %s: %v\\n\", id, c)\n\n\t} else if e2 = e.Field(\"Edges\").Value(&id); e2.Error() == nil {\n\t\t// Here we know the change is an Edge object in the Nodes map.\n\t\t// The key is stored in 'id'.\n\n\t\t// Current gets us the Node object\n\t\tc, _ := e2.Current()\n\n\t\t// Note that, since the Edge object is stored as 2 different keys,\n\t\t// The callback will be called twice.\n\n\t\tfmt.Printf(\"Modified Edge with key %s: %v\\n\", id, c)\n\n\t} else if b, err := e.Field(\"QuitDemo\").Bool(); err == nil {\n\n\t\t// Since QuitDemo is a boolean, we can use Bool() method to get the value\n\t\t// directly.\n\t\tstopTimeWheel = b\n\n\t}\n\treturn nil\n}", "func (self *PhysicsP2) CallbackContext() interface{}{\n return self.Object.Get(\"callbackContext\")\n}", "func RVExtensionRegisterCallback(cbptr unsafe.Pointer) {\n\tcb = C.callbackProc(cbptr)\n\n\tlog.Println(\"Calling callback function ……\")\n\tfunction := C.CString(\"registered\")\n\tdefer C.free(unsafe.Pointer(function))\n\tC.bridge_cb(cb, name, function, function)\n}", "func (ac *asyncCallbacksHandler) push(f func()) {\n\tac.cbQueue <- f\n}", "func (p *MyThriftClient) CallBack(callTime int64, name string, paramMap map[string]string) (r []string, err error) {\n if err = p.sendCallBack(callTime, name, paramMap); err != nil { return }\n return p.recvCallBack()\n}", "func (s *Server) HandlerCallback(w http.ResponseWriter, r *http.Request) {\n\tsession, err := s.SessionStore.Get(r, s.Config.SessionName)\n\tif err != nil {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": err.Error(),\n\t\t}).Error(\"failed to get session\")\n\t\thttp.Error(w, http.StatusText(http.StatusInternalServerError), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tvar val interface{}\n\tvar ok bool\n\n\tvar provider string\n\tval = session.Values[\"provider\"]\n\tif provider, ok = val.(string); !ok {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": \"provider is not found\",\n\t\t}).Info(\"session is broken.\")\n\t\thttp.Error(w, http.StatusText(http.StatusBadRequest), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tvar callback string\n\tval = session.Values[\"callback\"]\n\tif callback, ok = val.(string); !ok {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": \"callback is not found\",\n\t\t}).Info(\"session is broken.\")\n\t\thttp.Error(w, http.StatusText(http.StatusBadRequest), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tvar next string\n\tval = session.Values[\"next\"]\n\tif next, ok = val.(string); !ok {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": \"next is not found\",\n\t\t}).Info(\"session is broken.\")\n\t\thttp.Error(w, http.StatusText(http.StatusBadRequest), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tvar state string\n\tval = session.Values[\"state\"]\n\tif state, ok = val.(string); !ok {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": \"state is not found\",\n\t\t}).Info(\"session is broken.\")\n\t\thttp.Error(w, http.StatusText(http.StatusBadRequest), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tproviderConfig, ok := s.ProviderConfigs[provider]\n\tif !ok {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": \"provider is not found\",\n\t\t}).Info(\"session is broken.\")\n\t\thttp.Error(w, http.StatusText(http.StatusBadRequest), http.StatusBadRequest)\n\t\treturn\n\t}\n\tconf := providerConfig.Config()\n\tconf.RedirectURL = callback\n\n\tquery := r.URL.Query()\n\n\tif state != query.Get(\"state\") {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": \"state is not correct\",\n\t\t}).Info(\"session is broken.\")\n\t\thttp.Error(w, http.StatusText(http.StatusBadRequest), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tcode := query.Get(\"code\")\n\tt, err := conf.Exchange(r.Context(), code)\n\tif err != nil {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": err.Error(),\n\t\t}).Info(\"oauth verification failed\")\n\t\thttp.Error(w, http.StatusText(http.StatusForbidden), http.StatusForbidden)\n\t\treturn\n\t}\n\n\tvar uid string\n\tvar info map[string]interface{}\n\tif infoctx, ok := providerConfig.(ProviderInfoContext); ok {\n\t\tuid, info, err = infoctx.InfoContext(r.Context(), &conf, t)\n\t} else {\n\t\tuid, info, err = providerConfig.Info(&conf, t)\n\t}\n\tif err != nil {\n\t\tif err == ErrForbidden {\n\t\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\t\"err\": err.Error(),\n\t\t\t}).Warn(\"access forbidden\")\n\t\t\thttp.Error(w, http.StatusText(http.StatusForbidden), http.StatusForbidden)\n\t\t} else {\n\t\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\t\"err\": err.Error(),\n\t\t\t}).Warn(\"user info cannot get\")\n\t\t\thttp.Error(w, http.StatusText(http.StatusInternalServerError), http.StatusInternalServerError)\n\t\t}\n\t\treturn\n\t}\n\tjsonInfo, encodedInfo, err := encodeInfo(info)\n\tif err != nil {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": err.Error(),\n\t\t}).Info(\"failed encoding info\")\n\t}\n\n\tsession.Values[\"uid\"] = uid\n\tsession.Values[\"info\"] = encodedInfo\n\tsession.Values[\"logged_in_at\"] = time.Now()\n\n\tif err := session.Save(r, w); err != nil {\n\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\"err\": err.Error(),\n\t\t}).Error(\"failed to save session\")\n\t\thttp.Error(w, http.StatusText(http.StatusInternalServerError), http.StatusInternalServerError)\n\t}\n\n\tlogrus.WithFields(logrus.Fields{\n\t\t\"user\": uid,\n\t\t\"info\": jsonInfo,\n\t}).Info(\"user login\")\n\n\thttp.Redirect(w, r, next, http.StatusFound)\n}", "func registeredCB(\n ptr unsafe.Pointer,\n frameworkMessage *C.ProtobufObj,\n masterMessage *C.ProtobufObj) {\n if (ptr != nil) {\n var driver *SchedulerDriver = (*SchedulerDriver)(ptr)\n\n if (driver.Scheduler.Registered == nil) {\n return\n }\n\n frameworkData := C.GoBytes(\n frameworkMessage.data,\n C.int(frameworkMessage.size))\n\n var frameworkId FrameworkID\n err := proto.Unmarshal(frameworkData, &frameworkId); if err != nil {\n return\n }\n\n masterData := C.GoBytes(masterMessage.data, C.int(masterMessage.size))\n var masterInfo MasterInfo\n err = proto.Unmarshal(masterData, &masterInfo); if err != nil {\n return\n }\n\n driver.Scheduler.Registered(driver, frameworkId, masterInfo)\n }\n}", "func handleCallback(w http.ResponseWriter, r *http.Request) {\n\t// in the real world you should check the state query parameter, but this is omitted for brevity reasons.\n\n\t// Exchange the access code for an access (and optionally) a refresh token\n\ttoken, err := client.GetOAuth2Config().Exchange(context.Background(), r.URL.Query().Get(\"code\"))\n\tif err != nil {\n\t\thttp.Error(w, errors.Wrap(err, \"Could not exhange token\").Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\t// Render the output\n\trenderTemplate(w, \"callback.html\", struct {\n\t\t*oauth2.Token\n\t\tIDToken interface{}\n\t}{\n\t\tToken: token,\n\t\tIDToken: token.Extra(\"id_token\"),\n\t})\n}", "func onFileOpen(filename string) {\n\t//\n}", "func CallbackHandler(callbackChan chan Callback) http.HandlerFunc {\n\treturn http.HandlerFunc(func(resp http.ResponseWriter, req *http.Request) {\n\t\tvar cb Callback\n\t\tcb.Parse(req)\n\t\tresp.WriteHeader(http.StatusOK)\n\t\tgo func() {\n\t\t\tcallbackChan <- cb\n\t\t}()\n\t})\n}", "func (self *PhysicsP2) PostBroadphaseCallback() interface{}{\n return self.Object.Get(\"postBroadphaseCallback\")\n}", "func (data KeepAliveData) Complete() {\n}", "func (self *SinglePad) PollStatus() {\n self.Object.Call(\"pollStatus\")\n}", "func (l *Libvirt) DomainEventCallbackBlockJob() (err error) {\n\tvar buf []byte\n\n\n\t_, err = l.requestStream(326, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (self *SinglePad) CallbackContext() interface{}{\n return self.Object.Get(\"callbackContext\")\n}", "func onConnect(c *gnet.Connection, solicited bool) {\n\tfmt.Printf(\"Event Callback: connnect event \\n\")\n}", "func (r *Request) getCallback() mq.Response {\n\tif r.cb == nil {\n\t\tpanic(\"test: request already responded to\")\n\t}\n\n\tcb := r.cb\n\tr.cb = nil\n\treturn cb\n}", "func callback(srv *rospy_tutorials.AddTwoInts) error {\n\tsrv.Response.Sum = srv.Request.A + srv.Request.B\n\treturn nil\n}", "func handleResponseNotification(task *task.MessageTask, response *libcoap.Pdu, env *task.Env){\n handleNotification(env, task, response)\n}", "func (self *SinglePad) AddCallbacks(context interface{}, callbacks interface{}) {\n self.Object.Call(\"addCallbacks\", context, callbacks)\n}", "func (pb *Pubsub) run() {\n\tfor {\n\t\tselect {\n\t\tcase t := <-pb.updateCh.Get():\n\t\t\tpb.updateCh.Load()\n\t\t\tif pb.done.HasFired() {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpb.callCallback(t.(*watcherInfoWithUpdate))\n\t\tcase <-pb.done.Done():\n\t\t\treturn\n\t\t}\n\t}\n}", "func goCallback(l, p uintptr) int {\n\tL := &State{S: (*C.lua_State)(unsafe.Pointer(l))}\n\tf := reg[p].(func(*State) int)\n\t//fmt.Println(f,L)\n\n\treturn f(L)\n}", "func licenseCallback(status int) {\n\tif status == lexactivator.LA_OK {\n\t\tfmt.Println(\"License is genuinely activated.\")\n\t} else if status == lexactivator.LA_EXPIRED {\n\t\tfmt.Println(\"License is genuinely activated, but has expired\")\n\t} else if status == lexactivator.LA_SUSPENDED {\n\t\tfmt.Println(\"License is genuinely activated, but has been suspended\")\n\t} else {\n\t\tfmt.Println(\"License error status:\", status)\n\t}\n}", "func (o *CanvasItem) X_UpdateCallback() {\n\t//log.Println(\"Calling CanvasItem.X_UpdateCallback()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"CanvasItem\", \"_update_callback\")\n\n\t// Call the parent method.\n\t// void\n\tretPtr := gdnative.NewEmptyVoid()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n}", "func (self *SinglePad) OnDownCallback() interface{}{\n return self.Object.Get(\"onDownCallback\")\n}", "func CallbackHandler(c *gin.Context) {\n\n\t// Retrieve query params for state and code\n\tstate := c.Query(\"state\")\n\tcode := c.Query(\"code\")\n\t//2次進攻redirectURL\n\tresp, err := http.Post(\"https://app.asana.com/-/oauth_token\",\n\t\t\"application/x-www-form-urlencoded\",\n\t\tstrings.NewReader(\"grant_type=authorization_code&client_id=\"+clientID+\"&client_secret=\"+clientSecret+\"&redirect_uri=\"+redirectURL+\"&state=\"+state+\"&code=\"+code))\n\tif err != nil {\n\t\tutil.Error(err.Error())\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tutil.Error(\"resp error\")\n\t}\n\n\tServerslice1 := UserType{}\n\te := json.Unmarshal([]byte(body), &Serverslice1)\n\tif e != nil {\n\t\tutil.Error(e.Error())\n\t}\n\n\t// Save the username in the session\n\t//session.Set(userkey, Serverslice1.Data.Name)\n\n\t//fmt.Println(body)\n\n\t//rsp回來的資料\n\tutil.Info(string(\" > User \"+Serverslice1.Data.Name) + \" login ! \")\n\tc.Writer.Write([]byte(\"Hi, \" + string(Serverslice1.Data.Name)))\n}", "func (t *SelfTester) SetOnNewPoliciesReadyCb(cb func()) {\n}", "func CallbackHandler(callbackChan chan Callback) http.HandlerFunc {\n\treturn http.HandlerFunc(func(resp http.ResponseWriter, req *http.Request) {\n\t\tvar cb Callback\n\t\terr := cb.Parse(req)\n\t\tif err != nil {\n\t\t\tresp.WriteHeader(http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t\tresp.WriteHeader(http.StatusOK)\n\t\t// start seperate goroutine to allow http request to return.\n\t\tgo func() {\n\t\t\tcallbackChan <- cb\n\t\t}()\n\t})\n}", "func Callback(cbReq *CallbackRequest, opts *CallbackOptions) error {\n\tclient := opts.Client\n\tif client == nil {\n\t\tclient = http.DefaultClient\n\t}\n\tbuf := bytes.NewBuffer(nil)\n\terr := json.NewEncoder(buf).Encode(cbReq)\n\tif err != nil {\n\t\treturn err\n\t}\n\tsignature, err := opts.Signer.Sign(buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\treq, err := http.NewRequest(\"POST\", cbReq.StatusCallbackUrl, buf)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq.Header.Set(\"X-OpenGDPR-Processor-Domain\", opts.ProcessorDomain)\n\treq.Header.Set(\"X-OpenGDPR-Signature\", signature)\n\t// Attempt to make callback\n\tfor i := 0; i < opts.MaxAttempts; i++ {\n\t\tresp, err := client.Do(req)\n\t\tif err != nil || resp.StatusCode != 200 {\n\t\t\ttime.Sleep(opts.Backoff)\n\t\t\tcontinue\n\t\t}\n\t\t// Success\n\t\treturn nil\n\t}\n\treturn fmt.Errorf(\"callback timed out for %s\", cbReq.StatusCallbackUrl)\n}", "func (l *Libvirt) DomainEventCallbackJobCompleted() (err error) {\n\tvar buf []byte\n\n\n\t_, err = l.requestStream(363, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (m *mod) handleCallBackEvent(ctx context.Context, event *slackbotv1.Event) error {\n\tswitch slackevents.EventsAPIType(event.Type) {\n\tcase slackevents.AppMention:\n\t\treturn m.handleAppMentionEvent(ctx, event)\n\tcase slackevents.Message:\n\t\treturn m.handleMessageEvent(ctx, event)\n\tdefault:\n\t\treturn fmt.Errorf(\"received unsuported event type: %s\", event.Type)\n\t}\n}", "func (l *PoolListener) Done() {}", "func SetupCloseHandler() {\n\n}", "func CallbackHandler(c *gin.Context) {\n\t// Retrieve query params for state and code\n\tstate := c.Query(\"state\")\n\tcode := c.Query(\"code\")\n\tprovider := c.Param(\"provider\")\n\n\t// Handle callback and check for errors\n\tuser, _, err := config.Gocial.Handle(state, code)\n\tif err != nil {\n\t\tc.Writer.Write([]byte(\"Error: \" + err.Error()))\n\t\treturn\n\t}\n\n\tvar newUser = getOrRegisterUser(provider, user)\n\tvar jtwToken = createToken(&newUser)\n\n\tc.JSON(200, gin.H{\n\t\t\"data\": newUser,\n\t\t\"token\": jtwToken,\n\t\t\"message\": \"berhasil login\",\n\t})\n}", "func (app *appImpl) PollEventsOnMain() {\n\n}", "func submitCallback(t *kernel.Task, id uint64, cb *linux.IOCallback, cbAddr usermem.Addr) error {\n\tif cb.Reserved2 != 0 {\n\t\treturn syserror.EINVAL\n\t}\n\n\tfd := t.GetFileVFS2(cb.FD)\n\tif fd == nil {\n\t\treturn syserror.EBADF\n\t}\n\tdefer fd.DecRef(t)\n\n\t// Was there an eventFD? Extract it.\n\tvar eventFD *vfs.FileDescription\n\tif cb.Flags&linux.IOCB_FLAG_RESFD != 0 {\n\t\teventFD = t.GetFileVFS2(cb.ResFD)\n\t\tif eventFD == nil {\n\t\t\treturn syserror.EBADF\n\t\t}\n\t\tdefer eventFD.DecRef(t)\n\n\t\t// Check that it is an eventfd.\n\t\tif _, ok := eventFD.Impl().(*eventfd.EventFileDescription); !ok {\n\t\t\treturn syserror.EINVAL\n\t\t}\n\t}\n\n\tioseq, err := memoryFor(t, cb)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Check offset for reads/writes.\n\tswitch cb.OpCode {\n\tcase linux.IOCB_CMD_PREAD, linux.IOCB_CMD_PREADV, linux.IOCB_CMD_PWRITE, linux.IOCB_CMD_PWRITEV:\n\t\tif cb.Offset < 0 {\n\t\t\treturn syserror.EINVAL\n\t\t}\n\t}\n\n\t// Prepare the request.\n\taioCtx, ok := t.MemoryManager().LookupAIOContext(t, id)\n\tif !ok {\n\t\treturn syserror.EINVAL\n\t}\n\tif ready := aioCtx.Prepare(); !ready {\n\t\t// Context is busy.\n\t\treturn syserror.EAGAIN\n\t}\n\n\tif eventFD != nil {\n\t\t// The request is set. Make sure there's a ref on the file.\n\t\t//\n\t\t// This is necessary when the callback executes on completion,\n\t\t// which is also what will release this reference.\n\t\teventFD.IncRef()\n\t}\n\n\t// Perform the request asynchronously.\n\tfd.IncRef()\n\tt.QueueAIO(getAIOCallback(t, fd, eventFD, cbAddr, cb, ioseq, aioCtx))\n\treturn nil\n}", "func onStateCallback(userdata interface{}, state libutp.State) {\n\tc := userdata.(*Conn)\n\tswitch state {\n\tcase libutp.StateConnect, libutp.StateWritable:\n\t\tc.onConnectOrWritable(state)\n\tcase libutp.StateEOF:\n\t\tc.onConnectionFailure(io.EOF)\n\tcase libutp.StateDestroying:\n\t\tclose(c.baseConnDestroyed)\n\t}\n}", "func (this *ModuleManager) CallWithCallback(topic string, f, cb interface{}, cbParams, params []interface{}) (err error) {\n\tif m := this.GetModule(topic); m != nil {\n\t\terr = m.CallWithCallback(f, cb, cbParams, params)\n\t} else {\n\t\t// fmt.Println(this)\n\t\terr = Post.PutQueueWithCallback(f, cb, cbParams, params...)\n\t}\n\treturn\n}", "func (object *MQObject) CB(goOperation int32, gocbd *MQCBD, gomd *MQMD, gogmo *MQGMO) error {\n\tvar mqrc C.MQLONG\n\tvar mqcc C.MQLONG\n\tvar mqOperation C.MQLONG\n\tvar mqcbd C.MQCBD\n\tvar mqmd C.MQMD\n\tvar mqgmo C.MQGMO\n\n\terr := checkMD(gomd, \"MQCB\")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = checkGMO(gogmo, \"MQCB\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmqOperation = C.MQLONG(goOperation)\n\tcopyCBDtoC(&mqcbd, gocbd)\n\tcopyMDtoC(&mqmd, gomd)\n\tcopyGMOtoC(&mqgmo, gogmo)\n\n\tkey := makeKey(object.qMgr.hConn, object.hObj)\n\n\t// The callback function is a C function that is a proxy for the MQCALLBACK_Go function\n\t// defined here. And that in turn will call the user's callback function\n\tmqcbd.CallbackFunction = (C.MQPTR)(unsafe.Pointer(C.MQCALLBACK_C))\n\n\tC.MQCB(object.qMgr.hConn, mqOperation, (C.PMQVOID)(unsafe.Pointer(&mqcbd)),\n\t\tobject.hObj,\n\t\t(C.PMQVOID)(unsafe.Pointer(&mqmd)), (C.PMQVOID)(unsafe.Pointer(&mqgmo)),\n\t\t&mqcc, &mqrc)\n\n\tmqreturn := MQReturn{MQCC: int32(mqcc),\n\t\tMQRC: int32(mqrc),\n\t\tverb: \"MQCB\",\n\t}\n\n\tif mqcc != C.MQCC_OK {\n\t\treturn &mqreturn\n\t}\n\n\t// Add or remove the control information in the map used by the callback routines\n\tswitch mqOperation {\n\tcase C.MQOP_DEREGISTER:\n\t\tmapLock()\n\t\tdelete(cbMap, key)\n\t\tmapUnlock()\n\tcase C.MQOP_REGISTER:\n\t\t// Stash the hObj and real function to be called\n\t\tinfo := &cbInfo{hObj: object,\n\t\t\tcallbackFunction: gocbd.CallbackFunction,\n\t\t\tconnectionArea: nil,\n\t\t\tcallbackArea: gocbd.CallbackArea}\n\t\tmapLock()\n\t\tcbMap[key] = info\n\t\tmapUnlock()\n\tdefault: // Other values leave the map alone\n\t}\n\n\treturn nil\n}", "func Callback(state string, codeVerifier string, redirectURL string, oauth2config oauth2.Config, targetLabel string, finish chan bool) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tdefer func() {\n\t\t\tfinish <- true\n\t\t}()\n\t\tvar msg string\n\t\tvar page string\n\n\t\t// checking state\n\t\tif state != r.URL.Query().Get(\"state\") {\n\t\t\tmsg = fmt.Sprintf(errorMarkup, \"Invalid state\")\n\t\t\tpage = fmt.Sprintf(callbackPage, msg)\n\t\t} else {\n\t\t\t// State OK, continue OpenID Connect Flow\n\t\t\tcode := r.URL.Query().Get(\"code\")\n\t\t\tctx := context.Background()\n\t\t\toauth2Token, err := oauth2config.Exchange(ctx, code, oauth2.SetAuthURLParam(\"code_verifier\", codeVerifier))\n\t\t\tif err != nil {\n\t\t\t\t// Exchange error\n\t\t\t\tmsg = fmt.Sprintf(errorMarkup, err.Error())\n\t\t\t\tpage = fmt.Sprintf(callbackPage, msg)\n\t\t\t} else {\n\t\t\t\t// Exchange success\n\t\t\t\tpage = fmt.Sprintf(callbackPage, successMarkup)\n\n\t\t\t\t// Storing tokens on current target\n\t\t\t\toauth2Token.AccessToken = oauth2Token.Extra(\"id_token\").(string)\n\t\t\t\terr = StorageTokens(targetLabel, *oauth2Token)\n\t\t\t\tif err != nil {\n\t\t\t\t\t// Exchange error\n\t\t\t\t\tmsg = fmt.Sprintf(errorMarkup, err.Error())\n\t\t\t\t\tpage = fmt.Sprintf(callbackPage, msg)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tw.Header().Add(\"Content-Type\", \"text/html\")\n\t\t_, err := w.Write([]byte(page))\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"Client error writing callback page: (%s)\\n\", err.Error())\n\t\t}\n\t}\n}", "func callbackHandler(c *gin.Context) {\n\t// Retrieve query params for state and code\n\tstate := c.Query(\"state\")\n\tcode := c.Query(\"code\")\n\t//provider := c.Param(\"provider\")\n\n\t// Handle callback and check for errors\n\tuser, token, err := gocial.Handle(state, code)\n\tif err != nil {\n\t\tc.Writer.Write([]byte(\"Error: \" + err.Error()))\n\t\treturn\n\t}\n\tfmt.Printf(\"informacion del usuario\")\n\tfmt.Printf(\"%#v\", token)\n\tfmt.Printf(\"%#v\", user)\n\tfmt.Printf(\"fin de la informacion de usuario\")\n\n\t// If no errors, show user\n\t// c.Writer.Write([]byte(\"FullName: \" + user.FullName + \"\\n\"))\n\t// c.Writer.Write([]byte(\"Email: \" + user.Email + \"\\n\"))\n\t// c.Writer.Write([]byte(\"Username: \" + user.Username + \"\\n\"))\n\t// c.Writer.Write([]byte(\"Avatar: \" + user.Avatar + \"\\n\"))\n\n\t//buscamos el usuario\n\tuserModel, err := FindOneUser(&UserModel{Username: user.Username})\n\n\tif err != nil {\n\t\t//no se ha encontrado el usuario, no existe, hay que registrarlo\n\n\t\t//rellenamos los campos con los datos que nos llegan de github\n\t\tuserModel.Username = user.Username\n\t\tuserModel.Email = user.Email\n\t\tuserModel.Bio = user.FullName\n\t\tuserModel.Image = nil\n\t\t//asignamos una contraseña por defecto, que nos servirá para acceder al usuario también haciendo login normal\n\t\tuserModel.PasswordHash = \"12345678\"\n\n\t\t//lo guardamos en la base de datos\n\t\tif err := SaveOne(&userModel); err != nil {\n\t\t\tc.JSON(http.StatusUnprocessableEntity, common.NewError(\"database\", err))\n\t\t\treturn\n\t\t}\n\t\t//guardamos en el contexto de gin\n\t\tc.Set(\"my_user_model\", userModel)\n\t\t//hacemos el redirect\n\t\tc.Redirect(http.StatusFound, \"http://localhost:8081/social/\"+userModel.Username)\n\n\t} else {\n\t\t//el usuario existe, hace login y nos genera un token\n\n\t\t//guardamos en el contexto de gin\n\t\tc.Set(\"my_user_model\", userModel)\n\t\t//hacemos el redirect\n\t\tc.Redirect(http.StatusFound, \"http://localhost:8081/social/\"+userModel.Username)\n\t}\n}", "func DebugMessageCallback(callback DebugProc, userParam unsafe.Pointer) {\n userDebugCallback = callback\n C.glowDebugMessageCallback(gpDebugMessageCallback, (C.GLDEBUGPROC)(unsafe.Pointer(&callback)), userParam)\n}", "func (s *Basememcached_protocolListener) EnterRetrieval_response(ctx *Retrieval_responseContext) {}", "func (client *Client) onReconnectSuccess(callback func()) {\n\tclient.onReconnectSuccessCallBack = callback\n}", "func processMessage(consumerMessage *sarama.ConsumerMessage) {\n\treceivedMessage := unmarshal(consumerMessage)\n\trunCallback(receivedMessage, consumerMessage)\n}", "func callbackFunction(s *scope, name string, obj pyObject, requiredArguments int, arguments string) *pyFunc {\n\tif obj != nil && obj != None {\n\t\tf := obj.(*pyFunc)\n\t\ts.Assert(len(f.args) == requiredArguments, \"%s callbacks must take exactly %d %s (%s takes %d)\", name, requiredArguments, arguments, f.name, len(f.args))\n\t\treturn f\n\t}\n\treturn nil\n}", "func (l *Libvirt) DomainEventCallbackMetadataChange() (err error) {\n\tvar buf []byte\n\n\n\t_, err = l.requestStream(379, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (uem *UyuniEventMapper) call(function string, args ...interface{}) (interface{}, error) {\n\tvar res interface{}\n\treturn res, uem.GetRpc().Call(function, args, &res)\n}", "func (self *Tween) OnUpdateCallback(callback interface{}, callbackContext interface{}) *Tween{\n return &Tween{self.Object.Call(\"onUpdateCallback\", callback, callbackContext)}\n}", "func callbacks(signal os.Signal) []func() {\n\tlock.RLock()\n\tcalls := handlerMap[signal]\n\tlock.RUnlock()\n\treturn calls\n}", "func (ac *asyncCallbacksHandler) run() {\n\tfor {\n\t\tf := <-ac.cbQueue\n\t\tif f == nil {\n\t\t\treturn\n\t\t}\n\t\tf()\n\t}\n}", "func (l *Libvirt) DomainEventCallbackBalloonChange() (err error) {\n\tvar buf []byte\n\n\n\t_, err = l.requestStream(331, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func On(callback func(string)) error {\n stdinScanner := bufio.NewScanner(os.Stdin)\n for stdinScanner.Scan() {\n text := stdinScanner.Text()\n callback(text)\n }\n\n return stdinScanner.Err()\n}", "func callback(resp http.ResponseWriter, request *http.Request) {\n\tbody, err := ioutil.ReadAll(request.Body)\n\tif err != nil {\n\t\tlog.Println(\"Failed reading body\")\n\t\tresp.WriteHeader(401)\n\t\tresp.Write([]byte(fmt.Sprintf(`{\"success\": false}`)))\n\t\treturn\n\t}\n\n\tlog.Println(string(body))\n\n\tresp.WriteHeader(200)\n\tresp.Write([]byte(fmt.Sprintf(`{\"success\": true}`)))\n\treturn\n}", "func (s *scanCoordinator) serverCallback(protoReq interface{}, ctx interface{}, conn net.Conn,\n\tcancelCh <-chan bool) {\n\n\tif protoReq == queryport.Ping {\n\t\tif ctx != nil {\n\t\t\tif conCtx := ctx.(*ConnectionContext); conCtx != nil {\n\t\t\t\tconCtx.ResetCache()\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\n\tttime := time.Now()\n\n\treq, err := NewScanRequest(protoReq, ctx, cancelCh, s)\n\tatime := time.Now()\n\tw := NewProtoWriter(req.ScanType, conn)\n\tdefer func() {\n\t\ts.handleError(req.LogPrefix, w.Done())\n\t\treq.Done()\n\t}()\n\n\tif req.ScanType == HeloReq {\n\t\ts.handleHeloRequest(req, w)\n\t\treturn\n\t}\n\n\tlogging.LazyVerbose(func() string {\n\t\treturn fmt.Sprintf(\"%s REQUEST %s\", req.LogPrefix, logging.TagStrUD(req))\n\t})\n\n\tif req.Consistency != nil {\n\t\tlogging.LazyVerbose(func() string {\n\t\t\treturn fmt.Sprintf(\"%s requested timestamp: %s => %s Crc64 => %v\", req.LogPrefix,\n\t\t\t\tstrings.ToLower(req.Consistency.String()), ScanTStoString(req.Ts), req.Ts.GetCrc64())\n\t\t})\n\t}\n\n\tif req.hasRollback != nil && req.hasRollback.Load() == true {\n\t\ts.handleError(req.LogPrefix, w.Error(ErrIndexRollback))\n\t\treturn\n\t}\n\n\tif s.tryRespondWithError(w, req, err) {\n\t\treturn\n\t}\n\n\tif req.Stats != nil {\n\t\treq.Stats.scanReqAllocDuration.Add(time.Now().Sub(atime).Nanoseconds())\n\t}\n\n\tif err := s.isScanAllowed(*req.Consistency, req); err != nil {\n\t\ts.tryRespondWithError(w, req, err)\n\t\treturn\n\t}\n\n\tif req.Stats != nil {\n\t\telapsed := time.Now().Sub(ttime).Nanoseconds()\n\t\treq.Stats.scanReqInitDuration.Add(elapsed)\n\t\treq.Stats.scanReqInitLatDist.Add(elapsed)\n\n\t\tnow := time.Now().UnixNano()\n\t\treq.Stats.numRequests.Add(1)\n\t\treq.Stats.lastScanTime.Set(now)\n\t\tif req.GroupAggr != nil {\n\t\t\treq.Stats.numRequestsAggr.Add(1)\n\t\t} else {\n\t\t\treq.Stats.numRequestsRange.Add(1)\n\t\t}\n\n\t\tfor _, partitionId := range req.PartitionIds {\n\t\t\treq.Stats.updatePartitionStats(partitionId,\n\t\t\t\tfunc(stats *IndexStats) {\n\t\t\t\t\tstats.numRequests.Add(1)\n\t\t\t\t\tstats.lastScanTime.Set(now)\n\t\t\t\t})\n\t\t}\n\t}\n\n\tt0 := time.Now()\n\tis, err := s.getRequestedIndexSnapshot(req)\n\tif err != nil {\n\t\tlogging.Infof(\"%s Error in getRequestedIndexSnapshot %v\", req.LogPrefix, err)\n\t\tif s.tryRespondWithError(w, req, err) {\n\t\t\treturn\n\t\t}\n\t}\n\n\tdefer DestroyIndexSnapshot(is)\n\n\tlogging.LazyVerbose(func() string {\n\t\treturn fmt.Sprintf(\"%s snapshot timestamp: %s\",\n\t\t\treq.LogPrefix, ScanTStoString(is.Timestamp()))\n\t})\n\n\tdefer func() {\n\t\tif req.Stats != nil {\n\t\t\telapsed := time.Now().Sub(ttime).Nanoseconds()\n\t\t\treq.Stats.scanReqDuration.Add(elapsed)\n\t\t\treq.Stats.scanReqLatDist.Add(elapsed)\n\t\t}\n\t}()\n\n\tif len(req.Ctxs) != 0 {\n\t\tvar err error\n\t\tdonech := make(chan bool)\n\t\tvar mutex sync.Mutex\n\n\t\tgo func() {\n\t\t\tselect {\n\t\t\tcase <-req.getTimeoutCh():\n\t\t\t\tmutex.Lock()\n\t\t\t\tdefer mutex.Unlock()\n\n\t\t\t\tselect {\n\t\t\t\tcase <-donech:\n\t\t\t\tdefault:\n\t\t\t\t\terr = common.ErrScanTimedOut\n\t\t\t\t\tclose(donech)\n\t\t\t\t}\n\t\t\tcase <-req.CancelCh:\n\t\t\t\tmutex.Lock()\n\t\t\t\tdefer mutex.Unlock()\n\n\t\t\t\tselect {\n\t\t\t\tcase <-donech:\n\t\t\t\tdefault:\n\t\t\t\t\terr = common.ErrClientCancel\n\t\t\t\t\tclose(donech)\n\t\t\t\t}\n\t\t\tcase <-donech:\n\t\t\t}\n\t\t}()\n\n\t\tnumCtxs := 0\n\t\tfor _, ctx := range req.Ctxs {\n\t\t\tif !ctx.Init(donech) {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tnumCtxs++\n\t\t}\n\n\t\tcont := func() bool {\n\t\t\tmutex.Lock()\n\t\t\tdefer mutex.Unlock()\n\n\t\t\tif s.tryRespondWithError(w, req, err) {\n\t\t\t\tfor i := 0; i < numCtxs; i++ {\n\t\t\t\t\treq.Ctxs[i].Done()\n\t\t\t\t}\n\t\t\t\treturn false\n\t\t\t}\n\n\t\t\tclose(donech)\n\t\t\treturn true\n\t\t}()\n\n\t\tif !cont {\n\t\t\treturn\n\t\t}\n\t}\n\n\ts.processRequest(req, w, is, t0)\n\n\tif len(req.Ctxs) != 0 {\n\t\tfor _, ctx := range req.Ctxs {\n\t\t\tctx.Done()\n\t\t}\n\t}\n}", "func pendSVHandler()", "func githubCallbackHandler(rw http.ResponseWriter, req *http.Request) {\n\tcode := req.URL.Query().Get(\"code\")\n\taccessToken, err := githubService.GetGithubAccessToken(code)\n\tif err != nil {\n\t\tlogger.ErrorLogger.Printf(\"COULD NOT GET ACCESS TOKEN: %v\\n\", err)\n\t\thttp.Error(rw, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tgitOpErr := githubService.HandleGithubOperations(accessToken)\n\tif gitOpErr != nil {\n\t\tlogger.ErrorLogger.Printf(\"ERROR OCCURED IN PROCESSING GIT OPERATIONS: %v\\n\", err)\n\t\thttp.Error(rw, gitOpErr.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\trw.WriteHeader(http.StatusOK)\n\tio.WriteString(rw, `<b>Git operator has performed some changes in your repo...</b>`)\n}", "func TestCallbackInvokedWhenSetLate(t *testing.T) {\n\twg := &sync.WaitGroup{}\n\twg.Add(1)\n\tapp := blockedABCIApplication{\n\t\twg: wg,\n\t}\n\t_, c := setupClientServer(t, app)\n\treqRes := c.CheckTxAsync(types.RequestCheckTx{})\n\n\tdone := make(chan struct{})\n\tcb := func(_ *types.Response) {\n\t\tclose(done)\n\t}\n\treqRes.SetCallback(cb)\n\tapp.wg.Done()\n\t<-done\n\n\tvar called bool\n\tcb = func(_ *types.Response) {\n\t\tcalled = true\n\t}\n\treqRes.SetCallback(cb)\n\trequire.True(t, called)\n}", "func RunAfterCallbacks() {\n\tfor _, cb := range afterCallbacks {\n\t\tif err := cb(); err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t}\n}", "func (_AnchorChain *AnchorChainTransactor) Callback(opts *bind.TransactOpts, state bool, _result []string) (*types.Transaction, error) {\n\treturn _AnchorChain.contract.Transact(opts, \"callback\", state, _result)\n}", "func (c *controller) Callback(ctx context.Context, request *web.Request) web.Result {\n\tif resp := c.service.callback(ctx, request); resp != nil {\n\t\treturn resp\n\t}\n\treturn c.responder.NotFound(errors.New(\"broker for callback not found\"))\n}", "func onMessage(c *gnet.Connection, channel uint16,\n\tmsg []byte) error {\n\tfmt.Printf(\"Event Callback: message event: addr= %s, channel %v, msg= %s \\n\", c.Addr(), channel, msg)\n\treturn nil\n}", "func go_message_callback(jid *C.char, msg_type *C.char, from *C.char,\n\tmessage *C.char) {\n\n\tvar jid_i = C.GoString(jid)\n\n\tif client, ok := clients[jid_i]; ok {\n\t\tvar msg_type_i = C.GoString(msg_type)\n\t\tvar from_i = C.GoString(from)\n\t\tvar message_i = C.GoString(message)\n\n\t\tmsg := &Message{\n\t\t\tMessageType: msg_type_i,\n\t\t\tFrom: from_i,\n\t\t\tText: message_i,\n\t\t}\n\t\tclient.Channel <- msg\n\t}\n}", "func (c *Controller) Jsonp(callback string, data interface{}, status int, msg ...string) {\n ctx := c.Context()\n message := App().Status().Text(status, ctx.Header(\"Accept-Language\", \"\"), msg...)\n r := render.NewJsonp(callback, map[string]interface{}{\n \"status\": status,\n \"message\": message,\n \"data\": data,\n })\n\n ctx.PushLog(\"status\", status)\n ctx.SetHeader(\"Content-Type\", r.ContentType())\n ctx.End(r.HttpCode(), r.Content())\n}", "func updateAsync(updateStruct *Update, callback chan *Callback) {\n\trecords, err := update(updateStruct)\n\tcb := new(Callback)\n\tcb.Data = records\n\tcb.Error = err\n\tcallback <- cb\n}", "func (ece *ExecutionCoordinatorEntry) ExecuteCallbacks() {\n\tcr := server.consensus.Get(ece.Id)\n\tfor _, cb := range cr.Callbacks {\n\t\tgo cb(cr)\n\t}\n}" ]
[ "0.59013164", "0.5869913", "0.5864536", "0.57989216", "0.57925755", "0.5727821", "0.57164985", "0.57053274", "0.5696967", "0.5694784", "0.56891954", "0.5655518", "0.5637238", "0.56134903", "0.56065613", "0.55933195", "0.5590466", "0.55772394", "0.55689883", "0.55651474", "0.55583954", "0.5557103", "0.5552084", "0.55511546", "0.5526272", "0.54355264", "0.54287434", "0.5415898", "0.5399122", "0.5396066", "0.53778815", "0.53776604", "0.5375687", "0.5363289", "0.53178596", "0.5313901", "0.5309193", "0.5293608", "0.5285302", "0.52733874", "0.52714914", "0.52528965", "0.5247551", "0.52360266", "0.5218297", "0.5218207", "0.52172613", "0.52114177", "0.5202585", "0.5200366", "0.519683", "0.519608", "0.51942855", "0.5181388", "0.5181225", "0.51729834", "0.5172652", "0.5171395", "0.51656127", "0.5165049", "0.51600045", "0.515055", "0.5148573", "0.5146776", "0.5142373", "0.5131242", "0.5131063", "0.51305634", "0.5128361", "0.5124909", "0.5118729", "0.51087743", "0.5101513", "0.51002765", "0.50925434", "0.50869787", "0.50828195", "0.50810814", "0.50769454", "0.5074219", "0.5073879", "0.5071689", "0.50714415", "0.5059376", "0.505662", "0.50561553", "0.5038862", "0.5023733", "0.50176126", "0.5016351", "0.5014167", "0.50057673", "0.5005357", "0.50037473", "0.50014675", "0.49980694", "0.4993457", "0.49824622", "0.49742576", "0.4973868", "0.49735433" ]
0.0
-1
////////////////////////////////////////////////////////////////////////////// FUNCTIONS // ////////////
func (mc *MouseCannon) isArmed() bool { return mc.armed }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func curate(file *pe.File) []byte {\n\n\tm := make(map[string]interface{})\n\tfields := make([]string, 0)\n\n\tif file.HasDOSHdr {\n\t\tm[\"dos_header\"] = file.DOSHeader\n\t\tfields = append(fields, \"dos_header\")\n\t}\n\n\tif file.HasRichHdr {\n\t\tm[\"rich_header\"] = file.RichHeader\n\t\tfields = append(fields, \"rich_header\")\n\t}\n\n\tif file.HasCOFF {\n\t\tm[\"coff\"] = file.COFF\n\t\tfields = append(fields, \"coff\")\n\t}\n\n\tif file.HasNTHdr {\n\t\tm[\"nt_header\"] = file.NtHeader\n\t\tfields = append(fields, \"nt_header\")\n\t}\n\n\tif file.HasSections {\n\t\tm[\"sections\"] = file.Sections\n\t\tfields = append(fields, \"sections\")\n\t}\n\n\tif file.HasExport {\n\t\tm[\"export\"] = file.Export\n\t\tfields = append(fields, \"export\")\n\t}\n\n\tif file.HasImport {\n\t\tm[\"import\"] = file.Imports\n\t\tfields = append(fields, \"import\")\n\t}\n\n\tif file.HasResource {\n\t\tm[\"resource\"] = file.Resources\n\t\tfields = append(fields, \"resource\")\n\t}\n\n\tif file.HasException {\n\t\tm[\"exception\"] = file.Exceptions\n\t\tfields = append(fields, \"exception\")\n\t}\n\n\tif file.HasReloc {\n\t\tm[\"reloc\"] = file.Relocations\n\t\tfields = append(fields, \"reloc\")\n\t}\n\n\tif file.HasDebug {\n\t\tm[\"debug\"] = file.Debugs\n\t\tfields = append(fields, \"debug\")\n\t}\n\n\tif file.HasGlobalPtr {\n\t\tm[\"global_ptr\"] = file.GlobalPtr\n\t\tfields = append(fields, \"global_ptr\")\n\t}\n\n\tif file.HasTLS {\n\t\tm[\"tls\"] = file.TLS\n\t\tfields = append(fields, \"tls\")\n\t}\n\n\tif file.HasLoadCFG {\n\t\tm[\"load_config\"] = file.LoadConfig\n\t\tfields = append(fields, \"load_config\")\n\t}\n\n\tif file.HasBoundImp {\n\t\tm[\"bound_import\"] = file.BoundImports\n\t\tfields = append(fields, \"bound_import\")\n\t}\n\n\tif file.HasIAT {\n\t\tm[\"iat\"] = file.IAT\n\t\tfields = append(fields, \"iat\")\n\t}\n\n\tif file.HasDelayImp {\n\t\tm[\"delay_import\"] = file.DelayImports\n\t\tfields = append(fields, \"delay_import\")\n\t}\n\n\tif file.HasCLR {\n\t\tm[\"clr\"] = file.CLR\n\t\tfields = append(fields, \"clr\")\n\t}\n\n\tif file.HasSecurity {\n\t\tm[\"security\"] = file.Certificates\n\t\tif file.IsSigned {\n\t\t\tif file.Certificates.Verified {\n\t\t\t\tm[\"signature\"] = \"Signed file, valid signature\"\n\t\t\t} else {\n\t\t\t\tm[\"signature\"] = \"Signed file, invalid signature\"\n\t\t\t}\n\t\t}\n\t\tfields = append(fields, \"security\")\n\t} else {\n\t\tm[\"signature\"] = \"File is not signed\"\n\t}\n\n\tm[\"meta\"] = fields\n\treturn toJSON(m)\n}", "func (t *SimpleChaincode) getLoc(stub *shim.ChaincodeStub , args []string) ([]byte,error) {\n \n \t \n \ts := []string{args[0], \"requester\"};\n s1 := strings.Join(s, \"_\");\n \t \n \n requester_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n \t//------------------------------------------------------------\n \ts = []string{args[0], \"beneficiary\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n beneficiary_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n \ts = []string{args[0], \"amount\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n amount_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"expiry_date\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n expiry_date_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"status\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n status_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"advising_bank\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n advising_bank_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"document_hash\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n document_hash_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"loc_filename\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n loc_filename_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n \ts = []string{args[0], \"contract_hash\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n contract_hash_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"bol_hash\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n bol_hash_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n \t\n \ts = []string{string(requester_string),string(beneficiary_string),string(amount_string),string(expiry_date_string),string(status_string),string(advising_bank_string),string(document_hash_string),string(loc_filename_string),string(contract_hash_string),string(bol_hash_string)};\n \n // s=[]string{string(contract_hash_string),string(bol_hash_string)};\n final_string := strings.Join(s, \"|\");\n \t\n \t\n \t\n //s := strconv.Itoa(counter) ;\n //ret_s := []byte(s);\n return []byte(final_string), nil;\n \n }", "func sampleFunc() {\n\t// TODO[TK]\n\t// todo\n\t// FIXME\n}", "func pyrs(play *legends) { //\"u\" , pointer getinto that address and get the value of the address\n\tfmt.Println(play.dhoni)\n\tfmt.Println(play.virat)\n\tfmt.Println(play.rohit)\n\tfmt.Println(play.no1)\n\tfmt.Println(play.no2)\n\tfmt.Println(play.no3)\n}", "func printErr (err error){ // this function is used for printing error\r\n if err != nil{\r\n fmt.Println(err)\r\n }\r\n }", "func RetriveFirstLine(str string,resp *Value){\n str=str[2:len(str)-1]\n var temp string\n for i:=0;i!=len(str);i++{\n if int(str[i])!=32{\n temp=temp+str[i:i+1]\n }\n }\n mode:=strings.Split(temp,\":\")[0]\n infor:=strings.Split(temp,\":\")[1]\n //fmt.Printf(\"mode is %s,infor is %s\\n\",mode,infor)\n attributes:=strings.Split(infor,\",\")\n io:=strings.Split(attributes[0],\"=\")[1]\n bw:=strings.Split(attributes[1],\"=\")[1]\n iops:=strings.Split(attributes[2],\"=\")[1]\n runt:=strings.Split(attributes[3],\"=\")[1]\n //fmt.Printf(\"io=%s,bw=%s,iops=%s,runt=%s\\n\",io,bw,iops,runt)\n if mode==\"read\"{\n resp.Read_io=io\n resp.Read_bw=bw\n resp.Read_iops=iops\n resp.Read_runt=runt\n }else if mode==\"write\"{\n resp.Write_io=io\n resp.Write_bw=bw\n resp.Write_iops=iops\n resp.Write_runt=runt\n }\n}", "func Storage_basic( function string ,flexible_in1 interface{} ,flexible_in2 interface{} ,w http.ResponseWriter, r *http.Request ) (flexible_out1 interface{} ,flexible_out2 interface{}) {\n\n// IN function : trans ,check ,initialize ,sort etc           \n// IN flexible_in1  : see attachment\n// IN flexible_in2  : see attachment\n// IN w : response-writer\n// IN r : request- paramete\n\n// out1 flexible_out1 : see attachment\n// out2 flexible_out2 : see attachment\n\n// fmt.Fprintf( w, \"storage_basic start \\n\" )\n// fmt.Fprintf( w, \"storage_basic function \\n\" ,function )\n\n///\n/// select process by function\n///\n\n\tswitch function {\n\n case \"open\" :\n value1, _ := flexible_in1.(string)\n value2, _ := flexible_in2.(string)\n\n flexible_out1 = File_Open ( w ,r ,value1 ,value2 )\n\n break;\n\n case \"create\" :\n\n value1, _ := flexible_in1.(string)\n value2, _ := flexible_in2.(string)\n\n flexible_out1 = File_Create ( w ,r ,value1 ,value2 )\n\n break;\n\n case \"delete\" :\n\n value1, _ := flexible_in1.(string)\n value2, _ := flexible_in2.(string)\n\n File_Delete ( w , r ,value1 ,value2 )\n\n break;\n\n case \"copy\" :\n\n value1, _ := flexible_in1.(string)\n value2, _ := flexible_in2.([]type5.General_Work)\n\n File_Copy ( w , r ,value1 ,value2[0].String_Work ,value2[1].String_Work )\n\n break;\n\n case \"rename\" :\n\n value1, _ := flexible_in1.(string)\n value2, _ := flexible_in2.([]type5.General_Work)\n\n File_Rename ( w , r ,value1 ,value2[0].String_Work ,value2[1].String_Work )\n\n break;\n\n case \"write\" : // write line-data\n\n value1, _ := flexible_in1.(*storage.Writer)\n value2, _ := flexible_in2.([]string)\n\n File_write ( w ,value1 ,value2 )\n\n break;\n\n case \"write2\" : // write struct-data\n\n value1, _ := flexible_in1.([]type5.General_Work)\n\n File_Write_Struct ( w ,value1[0].Sw_Work ,value1[0].Int64_Work ,flexible_in2 )\n\n break;\n\n case \"list\" : // get bucket-list\n\n value1, _ := flexible_in1.(string)\n\n flexible_out1 = Bucket_List ( w ,r , value1 )\n\n break;\n\n case \"list2\" : // get object-list\n\n value1, _ := flexible_in1.(string)\n\n flexible_out1 = Object_List ( w ,r , value1 )\n\n break;\n\n case \"list3\" : // get object-list which is detail\n\n value1, _ := flexible_in1.(string)\n\n flexible_out1 = Object_List_Detail ( w ,r , value1 )\n\n break;\n\n case \"show1\" : // show graf on web ( type1 )\n\n value1, _ := flexible_in1.(string)\n value2, _ := flexible_in2.(string)\n\n get.Image_file_show ( w ,r , value1 ,value2 )\n\n break;\n\n case \"show2\" : // show graf on web ( type2 )\n\n value1, _ := flexible_in1.(type5.Image_Show)\n\n get.Image_file_show2 ( w ,r , value1 )\n\n break;\n\n\n }\n\n\treturn flexible_out1 ,flexible_out2\n\n}", "func (self *SinglePad) Index() int{\n return self.Object.Get(\"index\").Int()\n}", "func CQO() { ctx.CQO() }", "func Water_line_re_num( wname string ,w http.ResponseWriter, r *http.Request ) (record_number int64) {\n\n// IN wname : water-name    \n// IN w : レスポンスライター\n// IN r : リクエストパラメーター\n\n// OUT    : water-number\n\n// fmt.Fprintf( w, \"trans4.water_line_re_num start \\n\" )\n\n bucket := \"sample-7777\"\n filename1 := \"Water_Line.txt\"\n\n///\n/// open Water_Line file\n///\n\n reader_minor , _ := storage2.Storage_basic( \"open\" ,bucket ,filename1 , w , r )\n\n reader, _ := reader_minor.(io.ReadCloser)\n\n// reader := storage2.File_Open(w ,r ,bucket ,filename1)\n\n defer reader.Close()\n\n// get file-reader\n\n sreader := bufio.NewReaderSize(reader, 4096)\n\n index := 0\n\n record_number = 0\n\n for {\n\n index ++\n// fmt.Fprintf(w, \"trans4.water_line_re_num : lndex %v\\n\", index )\n\n// read one-record\n\n line ,_ := sreader.ReadString('\\n')\n\n num := len(line)\n\n// fmt.Fprintf(w, \"trans4.water_line_re_num : num %v\\n\", num )\n\n if num > 1 {\n\n// fmt.Fprintf(w, \"trans4.water_line_re_num : line %s\\n\", line )\n\n///\n/// change format which is struct\n///\n\n water_line_struct := struct_set.Water_line( w , line )\n\n if water_line_struct.Name == wname {\n\n record_number ++ // add water-line number\n\n }\n\n } else if num == 0 {\n\n// io.WriteString(w, \"\\n trans4.water_line_re_num : data end \\n\")\n\n break\n\n }\n }\n\n return\trecord_number\n\n}", "func (p *callgrindProfiler) getFunctionParameters(function *lisp.LVal) (string, int, string) {\n\tvar source string\n\tline := 0\n\tif function.Source == nil {\n\t\tif cell := function.Cells[0]; cell != nil && cell.Source != nil {\n\t\t\tsource = cell.Source.File\n\t\t\tline = cell.Source.Line\n\t\t} else {\n\t\t\tsource = \"no-source\"\n\t\t}\n\t} else {\n\t\tsource = function.Source.File\n\t\tline = function.Source.Line\n\t}\n\tfName := fmt.Sprintf(\"%s:%s\", function.FunData().Package, getFunNameFromFID(p.runtime, function.FunData().FID))\n\treturn source, line, fName\n}", "func pudding_stir(prime *pudding) {\n\n}", "func inspectPn(request datatype.GetObjRequest) {\n\n\tvar (\n\t\tresult *s3.GetObjectOutput\n\t\tnerrors int = 0\n\t\tusermd string\n\t\tdocument *documentpb.Document\n\t\tstart2 = time.Now()\n\n\t)\n\tif result, err = api.GetObject(request); err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase s3.ErrCodeNoSuchKey:\n\t\t\t\tgLog.Warning.Printf(\"Error: [%v] Error: [%v]\", s3.ErrCodeNoSuchKey, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tgLog.Error.Printf(\"Error: %v\", aerr.Error())\n\t\t\t\tnerrors += 1\n\t\t\t}\n\t\t} else {\n\t\t\tgLog.Error.Printf(\"Error:%v\", err.Error())\n\t\t\tnerrors += 1\n\t\t}\n\t} else {\n\t\tdefer result.Body.Close()\n\t\tif usermd, err = utils.GetUserMeta(result.Metadata); err == nil {\n\t\t\tuserm := meta.UserMd{}\n\t\t\tjson.Unmarshal([]byte(usermd), &userm)\n\t\t} else {\n\t\t\tgLog.Error.Printf(\"Error %v - The user metadata %s is invalid\", err, result.Metadata)\n\t\t\treturn\n\t\t}\n\t\tgLog.Info.Printf(\"Get Object key %s - Elapsed time %v \", request.Key, time.Since(start2))\n\n\t\t/*\n\t\t\tretrieve the backup document\n\t\t*/\n\n\t\tif body, err := utils.ReadObjectv(result.Body, CHUNKSIZE); err == nil {\n\t\t\tdefer result.Body.Close()\n\t\t\tdocument, err = mosesbc.GetDocument(body.Bytes())\n\t\t\tfmt.Printf(\"\\tDocument id: %s - Version Id: %s - Number of pages: %d - Document size: %d\\n\",document.DocId, document.VersionId, document.NumberOfPages, document.Size)\n\t\t\tusermd,_ := base64.StdEncoding.DecodeString(document.Metadata)\n\t\t\tfmt.Printf(\"\\tDocument user metadada: %s\\n\",string(usermd))\n\t\t\ts3meta,_ := base64.StdEncoding.DecodeString(document.S3Meta)\n\t\t\tfmt.Printf(\"\\tDocument s3 metadata: %s\\n\",string(s3meta))\n\t\t\tPdf:= document.Pdf\n\t\t\tif len(Pdf.Pdf) > 0 {\n\t\t\t\tfmt.Printf(\"\\tDocument PDFid: %s - PDF size %d\\n\",Pdf.PdfId,Pdf.Size)\n\t\t\t}\n\t\t\tif document.Clip {\n\t\t\t\tfmt.Printf(\"\\tDocument %s has a clipping page (page 0)\\n\",document.DocId)\n\t\t\t}\n\t\t\tmosesbc.InspectBlobs(document, maxPage,verbose)\n\t\t} else {\n\t\t\tgLog.Error.Printf(\"Error %v when retrieving the document %s\\n\", err, request.Key)\n\t\t\tnerrors = 1\n\t\t}\n\t}\n}", "func (c *Create ) MagicNumberMap() {\n\n}", "func(this*Window)ReadAddr()(begin int,end int,err error){\nf,err:=this.File(\"addr\")\nif err!=nil{\nreturn\n}\nif _,err= f.Seek(0,0);err!=nil{\nreturn\n}\n_,err= fmt.Fscanf(f,\"%d %d\",&begin,&end)\nreturn\n}", "func(this*Window)Seek(offset int64,whence int)(ret int64,err error){\nf,err:=this.File(\"body\")\nif err!=nil{\nreturn 0,err\n}\nreturn f.Seek(offset,whence)\n}", "func getJsObjectByUuid(uuid string) Uint32_t {\n\n\t// So here I will try to create a local Js representation of the object.\n\tobjInfos, err := GoJavaScript.CallGoFunction(\"Client\", \"GetGoObjectInfos\", uuid)\n\n\tif err == nil {\n\t\t// So here I got an object map info.\n\t\t// Create the object JS object.\n\t\tobj := Jerry_create_object()\n\t\tif !Jerry_value_is_object(obj) {\n\t\t\tlog.Panicln(\"---> fail to create a new object! \", uuid)\n\t\t}\n\n\t\t// I will keep the reference in the js cache to be able to remove release\n\t\t// the c pointer reference latter.\n\t\tGoJavaScript.GetCache().SetJsObject(uuid, obj)\n\n\t\t// Set the uuid property.\n\t\tJerry_set_object_property(obj, \"uuid_\", uuid)\n\n\t\t// Set native object to the object.\n\t\tC.create_native_object(C.CString(uuid), uint32_t_To_Jerry_value_t(obj))\n\n\t\t// Now I will set the object method.\n\t\tmethods := objInfos.(map[string]interface{})[\"Methods\"].(map[string]interface{})\n\t\tfor name, src := range methods {\n\t\t\tif len(src.(string)) == 0 {\n\t\t\t\t// Set the go function here.\n\t\t\t\tcstr := C.CString(name)\n\t\t\t\tdefer C.free(unsafe.Pointer(cstr))\n\t\t\t\tC.setGoMethod(C.CString(name), uint32_t_To_Jerry_value_t(obj))\n\t\t\t} else {\n\t\t\t\t// append the object function here.\n\t\t\t\tappendJsFunction(obj, name, src.(string))\n\t\t\t}\n\t\t}\n\n\t\t// I can remove the methods from the infos.\n\t\tdelete(objInfos.(map[string]interface{}), \"Methods\")\n\n\t\t// Now the object properties.\n\t\tfor name, value := range objInfos.(map[string]interface{}) {\n\t\t\tif reflect.TypeOf(value).Kind() == reflect.Slice {\n\t\t\t\tslice := reflect.ValueOf(value)\n\t\t\t\tvalues := jerry_value_t_To_uint32_t(C.create_array(C.uint32_t(slice.Len())))\n\t\t\t\tfor i := 0; i < slice.Len(); i++ {\n\t\t\t\t\te := slice.Index(i).Interface()\n\t\t\t\t\tif reflect.TypeOf(e).Kind() == reflect.Map {\n\t\t\t\t\t\t// Here The value contain a map... so I will append\n\t\t\t\t\t\tif e.(map[string]interface{})[\"TYPENAME\"] != nil {\n\t\t\t\t\t\t\tif e.(map[string]interface{})[\"TYPENAME\"].(string) == \"GoJavaScript.ObjectRef\" {\n\t\t\t\t\t\t\t\tvalue_ := getJsObjectByUuid(e.(map[string]interface{})[\"UUID\"].(string))\n\t\t\t\t\t\t\t\tr := C.set_property_by_index(uint32_t_To_Jerry_value_t(values), C.uint32_t(uint32(i)), uint32_t_To_Jerry_value_t(goToJs(value_)))\n\t\t\t\t\t\t\t\t// Release the result\n\t\t\t\t\t\t\t\tJerry_release_value(jerry_value_t_To_uint32_t(r))\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tlog.Println(\"---> unknow object propertie type 231\")\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tr := C.set_property_by_index(uint32_t_To_Jerry_value_t(values), C.uint32_t(uint32(i)), uint32_t_To_Jerry_value_t(goToJs(e)))\n\t\t\t\t\t\t// Release the result\n\t\t\t\t\t\tJerry_release_value(jerry_value_t_To_uint32_t(r))\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tJerry_set_object_property(obj, name, values)\n\n\t\t\t} else if reflect.TypeOf(value).Kind() == reflect.Map {\n\t\t\t\tif value.(map[string]interface{})[\"TYPENAME\"] != nil {\n\t\t\t\t\tif value.(map[string]interface{})[\"TYPENAME\"].(string) == \"GoJavaScript.ObjectRef\" {\n\t\t\t\t\t\tvalue_ := getJsObjectByUuid(value.(map[string]interface{})[\"UUID\"].(string))\n\t\t\t\t\t\tJerry_set_object_property(obj, name, value_)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tlog.Println(\"---> unknow object propertie type 245\")\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// Standard object property, int, string, float...\n\t\t\t\tJerry_set_object_property(obj, name, value)\n\t\t\t}\n\t\t}\n\t\treturn obj\n\t}\n\n\t// The property is undefined.\n\treturn nil\n}", "func readEvent(r io.Reader)(*Event,error){\no,t,b,e,f,s,err:=readFields(r)\nif err!=nil{\nreturn nil,err\n}\nvar ev Event\n\n\n/*43:*/\n\n\n//line goacme.w:522\n\nswitch o{\ncase'E':ev.Origin= Edit\ncase'F':ev.Origin= File\ncase'K':ev.Origin= Keyboard\ncase'M':ev.Origin= Mouse\ndefault:return nil,ErrInvalidOrigin\n}\n\n\n\n/*:43*/\n\n\n//line goacme.w:482\n\n\n\n/*48:*/\n\n\n//line goacme.w:560\n\nswitch t{\ncase'D':ev.Type= Delete\ncase'd':ev.Type= Delete|Tag\ncase'I':ev.Type= Insert\ncase'i':ev.Type= Insert|Tag\ncase'L':ev.Type= Look\ncase'l':ev.Type= Look|Tag\ncase'X':ev.Type= Execute\ncase'x':ev.Type= Execute|Tag\ndefault:return nil,ErrInvalidType\n}\n\n\n\n/*:48*/\n\n\n//line goacme.w:483\n\n\n\n/*50:*/\n\n\n//line goacme.w:586\n\nev.begin= b\nev.Begin= b\nev.end= e\nev.End= e\n\n\n\n/*:50*/\n\n\n//line goacme.w:484\n\n\n\n/*52:*/\n\n\n//line goacme.w:608\n\nev.flag= f\n\nif ev.Type&Execute==Execute{\nev.IsBuiltin= (ev.flag&1)==1\n}else if ev.Type&Look==Look{\nev.NoLoad= (ev.flag&1)==1\nev.IsFile= (ev.flag&4)==4\n}\n\nev.Text= s\n\n// if there is an expansion\nif(ev.flag&2)==2{\n_,_,ev.Begin,ev.End,_,ev.Text,err= readFields(r)\nif err!=nil{\nreturn nil,err\n}\n}\n// if there is a chording\nif(ev.flag&8)==8{\n_,_,_,_,_,ev.Arg,err= readFields(r)\nif err!=nil{\nreturn nil,err\n}\n_,_,_,_,_,_,err= readFields(r)\nif err!=nil{\nreturn nil,err\n}\n}\n\n\n/*54:*/\n\n\n//line goacme.w:645\n\nif len(ev.Text)> 0{\nf:=strings.Fields(ev.Text)\nif len(f)> 1{\nev.Text= f[0]\ns:=ev.Arg\nif len(s)> 0{\ns= \" \"+ev.Arg\n}\nev.Arg= strings.Join(f[1:],\" \")+s\n}\n}\n\n\n\n\n/*:54*/\n\n\n//line goacme.w:638\n\n\n\n\n/*:52*/\n\n\n//line goacme.w:485\n\nreturn&ev,nil\n}", "func main() {\n\t// 1. uncomment the code below\n\tpng, header := []byte{'P', 'N', 'G'}, []byte{}\n\n\tfor i := range png {\n\t\theader = append(header, png[i])\n\t}\n\tfmt.Printf(\"%T %q %q\", header, header, png)\n\t// 2. append elements to header to make it equal with the png slice\n\n\t// 3. compare the slices using the bytes.Equal function\n\tif bytes.Equal(png, header) {\n\t\tfmt.Println(\"They are equal.\")\n\t} else {\n\t\tfmt.Println(\"They are not equal.\")\n\t}\n\n\t// 4. print whether they're equal or not\n}", "func(this*Window)Write(p[]byte)(int,error){\nf,err:=this.File(\"body\")\nif err!=nil{\nreturn 0,err\n}\n\n\n/*71:*/\n\n\n//line goacme.w:1016\n\nf= &wrapper{f:f}\n\n\n\n\n\n/*:71*/\n\n\n//line goacme.w:244\n\nreturn f.Write(p)\n}", "func PicUsefulLog() {\n\n}", "func ByteOrder() binary.ByteOrder { return byteOrder }", "func Location() {}", "func init() {}", "func init() {}", "func init() {}", "func init() {}", "func (mtr *Mxmx1intmacMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"Lane0Sbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane0Sbe.Size()\n\n\tif fldName == \"Lane0Dbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane0Dbe.Size()\n\n\tif fldName == \"Lane1Sbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane1Sbe.Size()\n\n\tif fldName == \"Lane1Dbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane1Dbe.Size()\n\n\tif fldName == \"Lane2Sbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane2Sbe.Size()\n\n\tif fldName == \"Lane2Dbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane2Dbe.Size()\n\n\tif fldName == \"Lane3Sbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane3Sbe.Size()\n\n\tif fldName == \"Lane3Dbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane3Dbe.Size()\n\n\tif fldName == \"M0PbPbusDrdy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.M0PbPbusDrdy.Size()\n\n\tif fldName == \"M1PbPbusDrdy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.M1PbPbusDrdy.Size()\n\n\tif fldName == \"M2PbPbusDrdy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.M2PbPbusDrdy.Size()\n\n\tif fldName == \"M3PbPbusDrdy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.M3PbPbusDrdy.Size()\n\n\treturn offset\n}", "func getstring(org string) string {\n\tresult := \"\"\n\t// it is a little status machine( IN ESPCAE: '\"...\\', NOT IN ESPCE: '\"...'(\n\t// mormal string that do not have char '\\')\n\tIN_ESPCAE_STATUS := false\n\tfor _, v := range org[1 : len(org)-1] {\n\t\tif IN_ESPCAE_STATUS {\n\t\t\t// if now in espcae status, it will espcae char '\\', '\\\"', and \"\\'\"\n\t\t\tif v == '\\\\' || v == '\"' || v == '\\'' {\n\t\t\t\tresult += string(v)\n\t\t\t\t// back to normal status\n\t\t\t\tIN_ESPCAE_STATUS = false\n\t\t\t} else {\n\t\t\t\t// it do not matter, we do not need to espcae it, so we back the\n\t\t\t\t// normal string( e.g. \"\\a\" -> \"\\a\" )\n\t\t\t\tresult += string('\\\\')\n\t\t\t\tresult += string(v)\n\t\t\t}\n\t\t} else if v == '\\\\' {\n\t\t\t// if match the char '\\', then change self's status\n\t\t\tIN_ESPCAE_STATUS = true\n\t\t} else {\n\t\t\t// in normal status\n\t\t\tresult += string(v)\n\t\t}\n\t}\n\treturn result\n}", "func (mtr *Dprdpr1intsramseccMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"DprPktinFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprPktinFifoUncorrectable.Size()\n\n\tif fldName == \"DprPktinFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprPktinFifoCorrectable.Size()\n\n\tif fldName == \"DprCsumFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprCsumFifoUncorrectable.Size()\n\n\tif fldName == \"DprCsumFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprCsumFifoCorrectable.Size()\n\n\tif fldName == \"DprPhvFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprPhvFifoUncorrectable.Size()\n\n\tif fldName == \"DprPhvFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprPhvFifoCorrectable.Size()\n\n\tif fldName == \"DprOhiFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprOhiFifoUncorrectable.Size()\n\n\tif fldName == \"DprOhiFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprOhiFifoCorrectable.Size()\n\n\tif fldName == \"DprPtrFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprPtrFifoUncorrectable.Size()\n\n\tif fldName == \"DprPtrFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprPtrFifoCorrectable.Size()\n\n\tif fldName == \"DprPktoutFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprPktoutFifoUncorrectable.Size()\n\n\tif fldName == \"DprPktoutFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DprPktoutFifoCorrectable.Size()\n\n\treturn offset\n}", "func filenodehash(filename string,ch string,nodenum int64,readbitlen int64) []byte{\n\t//H(ch)is parsed into k indexes.\n\t//Calculate the hash value HCH of ch\n\tvar Hch string = GetSHA256HashCodeString(ch)\n\tvar Hchbyte, _ = hex.DecodeString(Hch)\n\t//Hch,_ := hex.DecodeString(ch)\n\tfmt.Println(\"Hch is \", Hch)\n\tfmt.Println(\"Hchbyte is \", Hchbyte)\n\t//Convert Hch to 01 string\n\tvar Hchstring string = biu.ToBinaryString(Hchbyte)\n\t//remove all \"[\"\n\tHchstring = strings.Replace(Hchstring, \"[\", \"\", -1)\n\t//remove all \"]\"\n\tHchstring = strings.Replace(Hchstring, \"]\", \"\", -1)\n\t//remove all space\n\tHchstring = strings.Replace(Hchstring, \" \", \"\", -1)\n\tfmt.Println(\"Hchstring is \", Hchstring)\n\t//convert nodenum to 01\n\tvar bittosting string = biu.ToBinaryString(nodenum)\n\n\tbittosting = strings.Replace(bittosting, \"[\", \"\", -1)\n\tbittosting = strings.Replace(bittosting, \"]\", \"\", -1)\n\tbittosting = strings.Replace(bittosting, \" \", \"\", -1)\n\tvar stringlen = len(bittosting)\n\n\tfmt.Println(\"nodenum is \", bittosting)\n\tfmt.Println(\"stringlen is \", stringlen)\n\n\tvar stringiter int = 0\n\tvar zerolen int = 0\n\tfor stringiter = 0; stringiter < stringlen; stringiter++ {\n\t\tif '0' != bittosting[stringiter] {\n\t\t\t//zerolen = stringiter + 1\n\t\t\tzerolen = stringiter\n\t\t\tbreak\n\t\t}\n\t}\n\n\tfmt.Println(\"zerolen is \", zerolen)\n\n\n\n\t//The calculation requires eachlen bits to represent the total number of leaf nodes.\n\t//For example, if the number of leaf nodes is 245441, 17 bits are needed to represent it\n\tvar eachlen uintptr = ((unsafe.Sizeof(nodenum) * 8) - uintptr(zerolen))\n\tfmt.Println(\"eachlen is \", eachlen)\n\n\n\n\t//由Hchstring切割得到原文件序号\n\tvar fileposition []int64\n\t//将Hchstring的bit字符串按每eachlen一份进行切割,生成[]string\n\tvar Hcharray []string = ChunkString(Hchstring, int(eachlen))\n\t//fmt.Println(\"chunkarray is \", chunkarray)\n\tvar filebititer int = 0\n\tfor filebititer = 0; filebititer < len(Hcharray); filebititer++ {\n\t\tvar tmpint int64 = 0\n\t\tvar partiter int = 0\n\t\tfor partiter = 0; partiter < len(Hcharray[filebititer]); partiter++ {\n\t\t\ttmpint = (tmpint << 1)\n\t\t\tif '1' == Hcharray[filebititer][partiter] {\n\t\t\t\ttmpint = (tmpint) ^ 1\n\t\t\t}\n\t\t\tif tmpint >= nodenum {\n\t\t\t\ttmpint = tmpint % nodenum\n\t\t\t}\n\n\t\t}\n\t\tfileposition = append(fileposition, tmpint)\n\t}\n\n\tfmt.Println(\"fileposition is \", fileposition)\n\tfileposition = RemoveRepeatedElement(fileposition)\n\tfmt.Println(\"fileposition is \", fileposition)\n\tvar fileretdata []byte\n\t//retdata, _ := ReadBlock(filename, readbitlen, 0*readbitlen)\n\t//fmt.Println(\"000000000000retdata is \", retdata)\n\tvar readiter int\n\tfor readiter = 0; readiter < len(fileposition); readiter++ {\n\t\t//fmt.Println(\"readiter is \", readiter)\n\t\t//fmt.Println(\"now fileposition is \", fileposition[readiter])\n\t\tretdata, _ := ReadBlock(filename, readbitlen, (fileposition[readiter])*readbitlen)\n\t\t//fmt.Println(\"retdata is \", retdata)\n\t\tfor _,nounceum := range retdata{\n\t\t\tfileretdata=append(fileretdata,nounceum)\n\t\t}\n\n\t}\n\tfmt.Println(\"fileretdata is \", fileretdata)\n\tfileretdata_hash := GetSHA256HashCode([]byte(fileretdata))\n\n\tvar filebyte_hash []byte\n\tfilebyte_hash, _ = hex.DecodeString(fileretdata_hash)\n\tfmt.Println(\"filebyte_hash is \", filebyte_hash)\n\treturn filebyte_hash\n\n}", "func (mtr *Dppdpp0intcreditMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"PtrCreditOvflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.PtrCreditOvflow.Size()\n\n\tif fldName == \"PtrCreditUndflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.PtrCreditUndflow.Size()\n\n\tif fldName == \"PktCreditOvflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.PktCreditOvflow.Size()\n\n\tif fldName == \"PktCreditUndflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.PktCreditUndflow.Size()\n\n\tif fldName == \"FramerCreditOvflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.FramerCreditOvflow.Size()\n\n\tif fldName == \"FramerCreditUndflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.FramerCreditUndflow.Size()\n\n\tif fldName == \"FramerHdrfldVldOvfl\" {\n\t\treturn offset\n\t}\n\toffset += mtr.FramerHdrfldVldOvfl.Size()\n\n\tif fldName == \"FramerHdrfldOffsetOvfl\" {\n\t\treturn offset\n\t}\n\toffset += mtr.FramerHdrfldOffsetOvfl.Size()\n\n\tif fldName == \"ErrFramerHdrsizeZeroOvfl\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrFramerHdrsizeZeroOvfl.Size()\n\n\treturn offset\n}", "func main()", "func main()", "func main()", "func (px *Paxos) status(Seq int) (Fate, interface{}) {\n\t// Your code here.\n\tpx.mu.Lock()\n\tdefer px.mu.Unlock()\n\tVal, ok := px.Stati[Seq]\n\n\tif !ok {\n\t\treturn Pending, nil\n\t}\n\n\tif Val == Decided {\n\t\t// return Decided, px.result[Seq]\n\t\treturn Decided, px.Val[Seq]\n\t}\n\treturn Val, nil\n}", "func get_line()bool{/* inputs the next line */\nrestart:\nif changing&&include_depth==change_depth{\n\n\n/*37:*/\n\n\n//line gocommon.w:534\n{\nchange_line++\nif input_ln(change_file)!=nil{\nerr_print(\"! Change file ended without @z\")\n\nbuffer= append(buffer,[]rune(\"@z\")...)\n}\nif len(buffer)> 0{/* check if the change has ended */\nif change_pending{\nif_section_start_make_pending(false)\nif change_pending{\nchanged_section[section_count]= true\nchange_pending= false\n}\n}\nif len(buffer)>=2&&buffer[0]=='@'{\nif unicode.IsUpper(buffer[1]){\nbuffer[1]= unicode.ToLower(buffer[1])\n}\nif buffer[1]=='x'||buffer[1]=='y'{\nloc= 2\nerr_print(\"! Where is the matching @z?\")\n\n}else if buffer[1]=='z'{\nprime_the_change_buffer()\nchanging= !changing\nprint_where= true\n}\n}\n}\n}\n\n\n\n/*:37*/\n\n\n//line gocommon.w:418\n\n}\nif!changing||include_depth> change_depth{\n\n\n/*36:*/\n\n\n//line gocommon.w:504\n{\nline[include_depth]++\nfor input_ln(file[include_depth])!=nil{/* pop the stack or quit */\nprint_where= true\nif include_depth==0{\ninput_has_ended= true\nbreak\n}else{\nfile[include_depth]= nil\nfile_name= file_name[:include_depth]\nfile= file[:include_depth]\nline= line[:include_depth]\ninclude_depth--\nif changing&&include_depth==change_depth{\nbreak\n}\nline[include_depth]++\n}\n}\nif!changing&&!input_has_ended{\nif len(buffer)==len(change_buffer){\nif buffer[0]==change_buffer[0]{\nif len(change_buffer)> 0{\ncheck_change()\n}\n}\n}\n}\n}\n\n\n\n/*:36*/\n\n\n//line gocommon.w:421\n\nif changing&&include_depth==change_depth{\ngoto restart\n}\n}\nif input_has_ended{\nreturn false\n}\nloc= 0\nif len(buffer)>=2&&buffer[0]=='@'&&(buffer[1]=='i'||buffer[1]=='I'){\nloc= 2\nfor loc<len(buffer)&&unicode.IsSpace(buffer[loc]){\nloc++\n}\nif loc>=len(buffer){\nerr_print(\"! Include file name not given\")\n\ngoto restart\n}\n\ninclude_depth++/* push input stack */\n\n\n/*35:*/\n\n\n//line gocommon.w:460\n{\nl:=loc\nif buffer[loc]=='\"'{\nloc++\nl++\nfor loc<len(buffer)&&buffer[loc]!='\"'{\nloc++\n}\n}else{\nfor loc<len(buffer)&&!unicode.IsSpace(buffer[loc]){\nloc++\n}\n}\n\nfile_name= append(file_name,string(buffer[l:loc]))\n\n\nif f,err:=os.Open(file_name[include_depth]);err==nil{\nfile= append(file,bufio.NewReader(f))\nline= append(line,0)\nprint_where= true\ngoto restart/* success */\n}\ntemp_file_name:=os.Getenv(\"GOWEBINPUTS\")\nif len(temp_file_name)!=0{\n\nfor _,fn:=range strings.Split(temp_file_name,\":\"){\nfile_name[include_depth]= fn+\"/\"+file_name[include_depth]\nif f,err:=os.Open(file_name[include_depth]);err==nil{\nfile= append(file,bufio.NewReader(f))\nline= append(line,0)\nprint_where= true\ngoto restart/* success */\n}\n}\n}\nfile_name= file_name[:include_depth]\nfile= file[:include_depth]\nline= line[:include_depth]\ninclude_depth--\nerr_print(\"! Cannot open include file\")\ngoto restart\n}\n\n\n\n/*:35*/\n\n\n//line gocommon.w:442\n\n}\nreturn true\n}", "func (mtr *Dppdpp1intcreditMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"PtrCreditOvflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.PtrCreditOvflow.Size()\n\n\tif fldName == \"PtrCreditUndflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.PtrCreditUndflow.Size()\n\n\tif fldName == \"PktCreditOvflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.PktCreditOvflow.Size()\n\n\tif fldName == \"PktCreditUndflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.PktCreditUndflow.Size()\n\n\tif fldName == \"FramerCreditOvflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.FramerCreditOvflow.Size()\n\n\tif fldName == \"FramerCreditUndflow\" {\n\t\treturn offset\n\t}\n\toffset += mtr.FramerCreditUndflow.Size()\n\n\tif fldName == \"FramerHdrfldVldOvfl\" {\n\t\treturn offset\n\t}\n\toffset += mtr.FramerHdrfldVldOvfl.Size()\n\n\tif fldName == \"FramerHdrfldOffsetOvfl\" {\n\t\treturn offset\n\t}\n\toffset += mtr.FramerHdrfldOffsetOvfl.Size()\n\n\tif fldName == \"ErrFramerHdrsizeZeroOvfl\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrFramerHdrsizeZeroOvfl.Size()\n\n\treturn offset\n}", "func getMarker(bloc,data []byte)Marker{\n\tmarker := hex.EncodeToString(bloc[0:2])\n\ttypeMarker := intReader.ReadInt16(bloc[2:4])\n\tlength := intReader.ReadInt32(bloc[4:8])\n\tvar formatData interface{}\n\tswitch typeMarker {\n\tcase 2 :\n\t\tif length > 4 {\n\t\t\t// Read data at offset\n\t\t\toffset := intReader.ReadInt32(bloc[8:])\n\t\t\tformatData = string(data[offset:offset+length])\n\t\t}else{\n\t\t\tformatData = string(bloc)\n\t\t}\n\tcase 3 : formatData = intReader.ReadInt16(bloc[8:])\n\tcase 4 : formatData = intReader.ReadInt32(bloc[8:])\n\t\tdefault:formatData = bloc[8:]\n\t}\n\n\treturn Marker{marker,typeMarker,length,formatData}\n}", "func id_lookup(\nid[]rune,/* string with id */\nt int32/* the ilk; used by \\.{GOWEAVE} only */)int32{\n\n\n/*46:*/\n\n\n//line gocommon.w:644\n\nh:=id[0]\nfor i:=1;i<len(id);i++{\nh= (h+h+id[i])%hash_size\n}\n\n\n\n/*:46*/\n\n\n//line gocommon.w:632\n\n\n\n/*47:*/\n\n\n//line gocommon.w:653\n\np:=hash[h]\nfor p!=-1&&!names_match(p,id,t){\np= name_dir[p].llink\n}\nif p==-1{\np:=int32(len(name_dir))/* the current identifier is new */\nname_dir= append(name_dir,name_info{})\nname_dir[p].llink= -1\ninit_node(p)\nname_dir[p].llink= hash[h]\nhash[h]= p/* insert p at beginning of hash list */\n}\n\n\n\n/*:47*/\n\n\n//line gocommon.w:633\n\nif p==-1{\n\n\n/*49:*/\n\n\n//line gocommon.w:671\n\np= int32(len(name_dir)-1)\nname_dir[p].name= append(name_dir[p].name,id...)\n\n\n/*97:*/\n\n\n//line gotangle.w:146\n\n\n\n\n\n/*:97*/\n\n\n//line gocommon.w:674\n\n\n\n\n/*:49*/\n\n\n//line gocommon.w:635\n\n}\nreturn p\n}", "func (o *queuePtr) getRelativeOffset(p *queuePtr) uint16 {\n\tif o.index == p.index {\n\t\tif o.offset > p.offset {\n\t\t\tpanic(\" getRelativeOffset(p *queuePtr) \")\n\t\t}\n\t\treturn p.offset - o.offset\n\t}\n\treturn p.offset\n}", "func (obj *object) getStructName(line string) {\n\t/*nested := false\n\tif !strings.Contains(line, \"type \") {\n\t\tnested = true\n\t}*/\n\n\tline = strings.TrimSpace(line)\n\tline = strings.TrimPrefix(strings.TrimSuffix(line, \"{\"), \"type\")\n\tline = strings.TrimSpace(line)\n\tobj.Name = strings.TrimSpace(strings.TrimSuffix(line, \"struct\"))\n\tif strings.Contains(obj.Name, \"[]\") {\n\t\tobj.Name = strings.TrimSpace(strings.TrimSuffix(obj.Name, \"[]\"))\n\t}\n\tobj.Tp = obj.Name\n\tobj.JsonKey = obj.Name\n\t/*if nested {\n\t\tobj.CommonFileds = append(obj.CommonFileds, &field{JsonKey: obj.Name, Tp: obj.Name})\n\t}*/\n}", "func (api *DeprecatedApiService) getTransactionIntro(params map[string]string) map[string]string {\n\tresult := make(map[string]string)\n\n\t_, is_only_check_exist := params[\"only_check_exist\"]\n\n\ttrsid, ok1 := params[\"id\"]\n\tif !ok1 {\n\t\tresult[\"err\"] = \"param id must.\"\n\t\treturn result\n\t}\n\n\tvar trshx []byte\n\tif txhx, e := hex.DecodeString(trsid); e == nil && len(txhx) == 32 {\n\t\ttrshx = txhx\n\t} else {\n\t\tresult[\"err\"] = \"transaction hash error.\"\n\t\treturn result\n\t}\n\n\t// Query transaction\n\tblkhei, trsresbytes, err := api.blockchain.GetChainEngineKernel().StateRead().ReadTransactionBytesByHash(trshx)\n\tif err != nil {\n\t\tresult[\"err\"] = err.Error()\n\t\treturn result\n\t}\n\tif trsresbytes == nil {\n\t\tresult[\"err\"] = \"transaction not fond.\"\n\t\treturn result\n\t}\n\n\t// Whether to just judge whether it exists\n\tif is_only_check_exist && len(trsresbytes) > 0 {\n\t\tresult[\"ret\"] = \"0\"\n\t\tresult[\"exist\"] = \"yes\"\n\t\treturn result\n\t}\n\n\ttrsres, _, err := transactions.ParseTransaction(trsresbytes, 0)\n\tif err != nil {\n\t\tresult[\"err\"] = err.Error()\n\t\treturn result\n\t}\n\n\t// Resolve actions\n\tvar allactions = trsres.GetActionList()\n\tvar actions_ary []string\n\tvar actions_strings = \"\"\n\tfor _, act := range allactions {\n\t\tvar kind = act.Kind()\n\t\tactstr := fmt.Sprintf(`{\"k\":%d`, kind)\n\t\tif kind == 1 {\n\t\t\tacc := act.(*actions.Action_1_SimpleToTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"to\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 13 {\n\t\t\tacc := act.(*actions.Action_13_FromTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 14 {\n\t\t\tacc := act.(*actions.Action_14_FromToTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"to\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 2 {\n\t\t\tacc := act.(*actions.Action_2_OpenPaymentChannel)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"left_addr\":\"%s\",\"left_amt\":\"%s\",\"right_addr\":\"%s\",\"right_amt\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t\tacc.LeftAddress.ToReadable(),\n\t\t\t\tacc.LeftAmount.ToFinString(),\n\t\t\t\tacc.RightAddress.ToReadable(),\n\t\t\t\tacc.RightAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 3 {\n\t\t\tacc := act.(*actions.Action_3_ClosePaymentChannel)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 4 {\n\t\t\tacc := act.(*actions.Action_4_DiamondCreate)\n\t\t\tactstr += fmt.Sprintf(`,\"number\":\"%d\",\"name\":\"%s\",\"address\":\"%s\"`,\n\t\t\t\tacc.Number,\n\t\t\t\tacc.Diamond,\n\t\t\t\tacc.Address.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 5 {\n\t\t\tacc := act.(*actions.Action_5_DiamondTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"count\":1,\"names\":\"%s\",\"from\":\"%s\",\"to\":\"%s\"`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Diamond,\n\t\t\t\ttrsres.GetAddress().ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 6 {\n\t\t\tacc := act.(*actions.Action_6_OutfeeQuantityDiamondTransfer)\n\t\t\tdmds := make([]string, len(acc.DiamondList.Diamonds))\n\t\t\tfor i, v := range acc.DiamondList.Diamonds {\n\t\t\t\tdmds[i] = string(v)\n\t\t\t}\n\t\t\tactstr += fmt.Sprintf(`,\"count\":%d,\"names\":\"%s\",\"from\":\"%s\",\"to\":\"%s\"`,\n\t\t\t\tacc.DiamondList.Count,\n\t\t\t\tstrings.Join(dmds, \",\"),\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 7 {\n\t\t\tacc := act.(*actions.Action_7_SatoshiGenesis)\n\t\t\tactstr += fmt.Sprintf(`,\"trs_no\":%d,\"btc_num\":%d,\"hac_subsidy\":%d,\"address\":\"%s\",\"lockbls_id\":\"%s\"`,\n\t\t\t\tacc.TransferNo,\n\t\t\t\tacc.BitcoinQuantity,\n\t\t\t\tacc.AdditionalTotalHacAmount,\n\t\t\t\tacc.OriginAddress.ToReadable(),\n\t\t\t\thex.EncodeToString(actions.GainLockblsIdByBtcMove(uint32(acc.TransferNo))),\n\t\t\t)\n\t\t} else if kind == 8 {\n\t\t\tacc := act.(*actions.Action_8_SimpleSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"to\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 9 {\n\t\t\tacc := act.(*actions.Action_9_LockblsCreate)\n\t\t\tactstr += fmt.Sprintf(`,\"lockbls_id\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.LockblsId),\n\t\t\t\tacc.TotalStockAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 10 {\n\t\t\tacc := act.(*actions.Action_10_LockblsRelease)\n\t\t\tactstr += fmt.Sprintf(`,\"lockbls_id\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.LockblsId),\n\t\t\t\tacc.ReleaseAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 11 {\n\t\t\tacc := act.(*actions.Action_11_FromToSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"to\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 28 {\n\t\t\tacc := act.(*actions.Action_28_FromSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 12 {\n\t\t\tacc := act.(*actions.Action_12_ClosePaymentChannelBySetupAmount)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 21 {\n\t\t\tacc := act.(*actions.Action_21_ClosePaymentChannelBySetupOnlyLeftAmount)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 22 {\n\t\t\tacc := act.(*actions.Action_22_UnilateralClosePaymentChannelByNothing)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\"\":\"%s\",\"bill_number\"\":0`,\n\t\t\t\thex.EncodeToString(acc.ChannelId), acc.AssertCloseAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 23 {\n\t\t\tacc := act.(*actions.Action_23_UnilateralCloseOrRespondChallengePaymentChannelByRealtimeReconciliation)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"%s\",\"bill_number\":%d`,\n\t\t\t\thex.EncodeToString(acc.Reconciliation.GetChannelId()), acc.AssertAddress.ToReadable(), acc.Reconciliation.GetAutoNumber(),\n\t\t\t)\n\t\t} else if kind == 24 {\n\t\t\tacc := act.(*actions.Action_24_UnilateralCloseOrRespondChallengePaymentChannelByChannelChainTransferBody)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"%s\",\"bill_number\":%d`,\n\t\t\t\thex.EncodeToString(acc.ChannelChainTransferTargetProveBody.GetChannelId()), acc.AssertAddress.ToReadable(), acc.ChannelChainTransferTargetProveBody.GetAutoNumber(),\n\t\t\t)\n\t\t} else if kind == 27 {\n\t\t\tacc := act.(*actions.Action_27_ClosePaymentChannelByClaimDistribution)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"any\",\"bill_number\"\":\"closed\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t}\n\t\tactstr += \"}\"\n\t\tactions_ary = append(actions_ary, actstr)\n\t}\n\tactions_strings = strings.Join(actions_ary, \",\")\n\n\t// Transaction return data\n\ttxaddr := fields.Address(trsres.GetAddress())\n\tvar txfee = trsres.GetFee()\n\tvar txfeeminergot = trsres.GetFeeOfMinerRealReceived()\n\tresult[\"jsondata\"] = fmt.Sprintf(\n\t\t`{\"block\":{\"height\":%d,\"timestamp\":%d},\"type\":%d,\"address\":\"%s\",\"fee\":\"%s\",\"feeminergot\":\"%s\",\"timestamp\":%d,\"actioncount\":%d,\"actions\":[%s]`,\n\t\tblkhei,\n\t\ttrsres.GetTimestamp(),\n\t\ttrsres.Type(),\n\t\ttxaddr.ToReadable(), // Primary address\n\t\ttxfee.ToFinString(),\n\t\ttxfeeminergot.ToFinString(),\n\t\ttrsres.GetTimestamp(),\n\t\tlen(allactions),\n\t\tactions_strings,\n\t)\n\n\tif _, ok := params[\"txbodyhex\"]; ok {\n\t\tresult[\"jsondata\"] += fmt.Sprintf(`,\"txbodyhex\":\"%s\"`,\n\t\t\thex.EncodeToString(trsresbytes))\n\t}\n\n\t// Wrap up and return\n\tresult[\"jsondata\"] += \"}\"\n\treturn result\n}", "func (self *Graphics) Tint() int{\n return self.Object.Get(\"tint\").Int()\n}", "func common_init(){\n\n\n/*44:*/\n\n\n//line gocommon.w:620\n\nfor i,_:=range hash{\nhash[i]= -1\n}\n\n\n\n/*:44*/\n\n\n\n/*51:*/\n\n\n//line gocommon.w:688\n\nname_root= -1/* the binary search tree starts out with nothing in it */\n\n\n\n/*:51*/\n\n\n//line gocommon.w:47\n\n\n\n/*82:*/\n\n\n//line gocommon.w:1230\n\nflags['b']= true\nflags['h']= true\nflags['p']= true\n\n\n\n/*:82*/\n\n\n//line gocommon.w:48\n\n\n\n/*89:*/\n\n\n//line gocommon.w:1368\n\nscan_args()\n\n\n/*165:*/\n\n\n//line gotangle.w:1532\n\nvar err error\nif go_file,err= os.OpenFile(go_file_name,os.O_WRONLY|os.O_CREATE|os.O_TRUNC,0666);err!=nil{\nfatal(\"! Cannot open output file \",go_file_name)\n\n}\n\n\n\n/*:165*/\n\n\n//line gocommon.w:1370\n\n\n\n\n/*:89*/\n\n\n//line gocommon.w:49\n\n}", "func Regift(AccessUnits []*ts.AccessUnit) []byte {\n\tfmt.Println(\"\\nRegift()\\n\\n\")\n\n\t//fmt.Println(\"AccessUnits[0]:\")\n\t//fmt.Println(AccessUnits[0])\n\t//fmt.Println(\"AccessUnits[0].PesMap:\")\n\t//fmt.Println(AccessUnits[0].PesMap)\n\n\taudioByte := make([]byte, 0)\n\tvideoByte := make([]byte, 0)\n\taudioSamples := make([]mp4box.Sample, 0)\n\tvideoSamples := make([]mp4box.Sample, 0)\n\ttrackID := 1\n\t// Need a array of boxes to hold the boxes\n\t// until they are ready to print\n\t// boxes = make([]mpeg4boxes, 0)\n\t// IMPORTANT NOTE: To have a array of the boxes they all have to\n\t// be in the same interface. I think this means all the box files\n\t// will need to be in the same package.\n\n\t// Dave you will need to print to file in reverse order from driver.go\n\t// Boxes is exported so that this is convenient\n\tBoxes := make([]mp4box.Box, 0)\n\n\tvar audioSize int = 0\n\tvar videoSize int = 0\n\tvar pcrDelta uint32 = 0\n\n\tfor _, AccessUnit := range AccessUnits {\n\t\t// fmt.Println( \"for _, AccessUnit := \" )\n\t\t// fmt.Println( AccessUnit )\n\n\t\tdelta := 0\n\n\t\t// fmt.Println(\"VIDEO_STREAM_TYPE = \", AccessUnit.PesMap[VIDEO_STREAM_TYPE])\n\n\t\t//fmt.Println(\"for_, pes := AUDIO_STREAM_TYPE\")\n\t\tfor _, pes := range AccessUnit.PesMap[AUDIO_STREAM_TYPE] {\n\t\t\t//fmt.Println(\"audio pes payload= \", pes.Payload)\n\t\t\taudioByte = append(audioByte, pes.Payload...)\n\t\t}\n\t\t//fmt.Println(\"AFTER for_, pes := AUDIO_STREAM_TYPE\")\n\n\t\tdelta = len(audioByte) - audioSize\n\n\t\taudioSize = len(audioByte)\n\n\t\taudioSamples = append(audioSamples, mp4box.Sample{uint32(AccessUnit.Pcr), uint32(delta), 0, 0})\n\n\t\t// fmt.Println(\"audioSamples = \", audioSamples)\n\n\t\tfor _, pes := range AccessUnit.PesMap[VIDEO_STREAM_TYPE] {\n\t\t\tvideoByte = append(videoByte, pes.Payload...)\n\t\t}\n\n\t\tdelta = len(videoByte) - videoSize\n\n\t\tvideoSize = len(videoByte)\n\n\t\tvideoSamples = append(videoSamples, mp4box.Sample{uint32(AccessUnit.Pcr), uint32(delta), 0, 0})\n\t}\n\n\tif len(videoSamples) < 2 {\n\t\tlog.Fatal(\"Not enough data to genertae pcr delta\")\n\t\tos.Exit(71)\n\n\t}\n\n\tpcrDelta = (videoSamples[len(videoSamples)-1].SampleDuration) - (videoSamples[len(videoSamples)-2].SampleDuration)\n\n\tif pcrDelta == 0 {\n\t\tlog.Fatal(\"pcrDelta is 0, cannot generate delta\")\n\t\tos.Exit(71)\n\n\t}\n\n\tfmt.Println(\"video pcrDelta\", pcrDelta)\n\n\tif (videoSamples[len(videoSamples)-1].SampleDuration % uint32(pcrDelta)) == 0 {\n\n\t\tfor i := 0; i < len(videoSamples); i++ {\n\n\t\t\tvideoSamples[i].SampleDuration = pcrDelta\n\n\t\t}\n\n\t}\n\n\tif len(audioSamples) < 2 {\n\t\tlog.Fatal(\"Not enough data to genertae pcr delta\")\n\t\tos.Exit(71)\n\n\t}\n\n\tpcrDelta = (audioSamples[len(audioSamples)-1].SampleDuration) - (audioSamples[len(audioSamples)-2].SampleDuration)\n\n\tif pcrDelta == 0 {\n\t\tlog.Fatal(\"pcrDelta is 0, cannot generate delta\")\n\t\tos.Exit(71)\n\n\t}\n\n\tfmt.Println(\"audio pcrDelta\", pcrDelta)\n\n\tif (audioSamples[len(audioSamples)-1].SampleDuration % uint32(pcrDelta)) == 0 {\n\n\t\tfor i := 0; i < len(audioSamples); i++ {\n\n\t\t\taudioSamples[i].SampleDuration = pcrDelta\n\n\t\t}\n\n\t}\n\n\tfmt.Println(\"\\nvideoSamples = \", videoSamples)\n\n\tfmt.Println(\"\\naudioSamples = \", audioSamples)\n\n\t// Create mdat and add it to boxes array\n\tpayload := append(audioByte, videoByte...)\n\tmdat := mp4box.NewMdat(uint32(audioSize+videoSize+8), payload)\n\n\tBoxes = append(Boxes, mdat)\n\t// Setting Flags for the trun should be done programatically from the\n\t// PES data but that can come later\n\taudioTrunFlags := make([]byte, 0, 3)\n\taudioTrunFlags = append(audioTrunFlags, 0x00)\n\taudioTrunFlags = append(audioTrunFlags, 0x0B)\n\taudioTrunFlags = append(audioTrunFlags, 0x00)\n\t// Add audio Samples to boxes array. Appended to rear of boxes array\n\tfmt.Println(uint32(len(audioSamples)))\n\taudioTrun := mp4box.NewTrun(\n\t\t0, //size is calculated later\n\t\t0, //version will be zero until we have a reason to do otherwise\n\t\taudioTrunFlags,\n\t\tuint32(len(audioSamples)),\n\t\t0, //dataoffset = MOOF.SIZE + 8, points to start of mdat, must be calculated later\n\t\t0, //no reason for first-sample-flags\n\t\taudioSamples)\n\taudioTrun.CalculateSize()\n\t// Add audio trun to boxes array. Appended to rear of boxes array\n\tBoxes = append(Boxes, audioTrun)\n\n\t// Add tfhd to boxes array. Appended to rear of boxes array\n\taudioTfhdFlags := make([]byte, 0, 3)\n\taudioTfhdFlags = append(audioTfhdFlags, 0x00)\n\taudioTfhdFlags = append(audioTfhdFlags, 0x00)\n\taudioTfhdFlags = append(audioTfhdFlags, 0x20)\n\taudioTfhd := mp4box.NewTfhd(\n\t\t0, //size is calculated later\n\t\t0, //version is typically 0\n\t\taudioTfhdFlags,\n\t\tuint32(trackID),\n\t\t0, //base-data-offset not obsevred in sample fragments\n\t\t0, //sample-description-index not observed in sample fragments\n\t\t0, //default-sample-duration not observed in sample fragments\n\t\t0, //default-sample-size not observed in sample fragments\n\t\t0) //default-sample-flags not observed in sample fragments\n\ttrackID++\n\taudioTfhd.CalculateSize()\n\tBoxes = append(Boxes, audioTfhd)\n\t// Add audio traf to boxes array. Appended to rear of boxes array\n\taudioTraf := mp4box.NewTraf(0) //Size = 8 + audioTfhd.size + audioTrun.Size, calculated later\n\tvar totalTrun uint32 = 0\n\tfor i := len(Boxes) - 1; i > 0; i-- {\n\t\tif Boxes[i].GetBoxType() == uint32(0x74726166) {\n\t\t\tbreak\n\t\t}\n\t\tif Boxes[i].GetBoxType() == uint32(0x7472756E) {\n\t\t\ttotalTrun += Boxes[i].GetSize()\n\t\t}\n\t}\n\taudioTraf.CalculateSize(totalTrun, audioTfhd.Size)\n\tBoxes = append(Boxes, audioTraf)\n\t// Add video samples to boxes array. Appended to rear of boxes array\n\t// Setting Flags for the trun should be done programatically from the\n\t// PES data but that can come later\n\tvideoTrunFlags := make([]byte, 0, 3)\n\tvideoTrunFlags = append(videoTrunFlags, 0x00)\n\tvideoTrunFlags = append(videoTrunFlags, 0x0B)\n\tvideoTrunFlags = append(videoTrunFlags, 0x00)\n\tfmt.Println(\"Num videoSamples is: \", len(videoSamples))\n\t// Add video Samples to boxes array. Appended to rear of boxes array\n\tvideoTrun := mp4box.NewTrun(\n\t\t0, //size is calculated later\n\t\t0, //version will be zero until we have a reason to do otherwise\n\t\tvideoTrunFlags,\n\t\tuint32(len(videoSamples)),\n\t\t0, //dataoffset = MOOF.SIZE + 8 + len(audioByte),\n\t\t//points to end of audio data in mdat, must be calculated later\n\t\t0, //no reason for first-sample-flags\n\t\tvideoSamples)\n\tvideoTrun.CalculateSize()\n\tfmt.Println(\"Num samplecount is: \", videoTrun.SampleCount)\n\t// Add video trun to boxes array.\n\tBoxes = append(Boxes, videoTrun)\n\n\t// Add tfhd to boxes array. Appended to rear of boxes array\n\tvideoTfhdFlags := make([]byte, 0, 3)\n\tvideoTfhdFlags = append(videoTfhdFlags, 0x00)\n\tvideoTfhdFlags = append(videoTfhdFlags, 0x00)\n\tvideoTfhdFlags = append(videoTfhdFlags, 0x20)\n\tvideoTfhd := mp4box.NewTfhd(\n\t\t0, //size is calculated later\n\t\t0, //version is typically 0\n\t\tvideoTfhdFlags,\n\t\tuint32(trackID),\n\t\t0, //base-data-offset not observed in sample fragments\n\t\t0, //sample-description-index not observed in sample fragments\n\t\t0, //default-sample-duration not observed in sample fragments\n\t\t0, //default-sample-size not observed in sample fragments\n\t\t0) //default-sample-flags not observed in sample fragments\n\ttrackID++\n\tvideoTfhd.CalculateSize()\n\tBoxes = append(Boxes, videoTfhd)\n\t// Add video traf to boxes array. Append to front of boxes array\n\tvideoTraf := mp4box.NewTraf(0) //Size = 8 + audioTfhd.size + audioTrun.Size, calculated later\n\ttotalTrun = 0\n\tfor i := len(Boxes) - 1; i > 0; i-- {\n\t\tif Boxes[i].GetBoxType() == uint32(0x74726166) {\n\t\t\tbreak\n\t\t}\n\t\tif Boxes[i].GetBoxType() == uint32(0x7472756E) {\n\t\t\ttotalTrun += Boxes[i].GetSize()\n\t\t}\n\t}\n\tvideoTraf.CalculateSize(totalTrun, videoTfhd.Size)\n\tBoxes = append(Boxes, videoTraf)\n\t// Add mfhd to boxes array. Appended to rear of boxes array\n\tmfhdFlags := make([]byte, 0, 3)\n\tmfhdFlags = append(mfhdFlags, 0x00)\n\tmfhdFlags = append(mfhdFlags, 0x00)\n\tmfhdFlags = append(mfhdFlags, 0x00)\n\tmfhd := mp4box.NewMfhd(\n\t\t16, //Size = 16 always\n\t\t0,\n\t\tmfhdFlags,\n\t\tsequenceNumber)\n\tsequenceNumber++ //advance sequenceNumber for the next moof-mdat fragment\n\tBoxes = append(Boxes, mfhd)\n\t// Add moof to boxes array. Appended to rear of boxes array\n\tmoof := mp4box.NewMoof(0) //Size is 8 + MFHD.SIZE + TRAFs.SIZE, calculated later\n\tvar totalTraf uint32 = 0\n\tfor i := len(Boxes) - 1; i > 0; i-- {\n\t\tif Boxes[i].GetBoxType() == uint32(0x74726166) {\n\t\t\ttotalTraf += Boxes[i].GetSize()\n\t\t}\n\t}\n\tmoof.CalculateSize(totalTraf, mfhd.Size)\n\tBoxes = append(Boxes, moof)\n\t\n\n\tboxesBytes := make([]byte, 0)\n\tfor i := (len(Boxes) - 1); i >= 0; i-- {\n\t\tboxesBytes = append(boxesBytes, Boxes[i].Write()...)\n\t}\n\n\treturn boxesBytes\n}", "func(this*Window)File(file string)(io.ReadWriteSeeker,error){\nfid,ok:=this.files[file]\nif!ok{\nvar err error\nif fid,err= fsys.Open(fmt.Sprintf(\"%d/%s\",this.id,file),plan9.ORDWR);err!=nil{\nif fid,err= fsys.Open(fmt.Sprintf(\"%d/%s\",this.id,file),plan9.OREAD);err!=nil{\nif fid,err= fsys.Open(fmt.Sprintf(\"%d/%s\",this.id,file),plan9.OWRITE);err!=nil{\nreturn nil,err\n}\n}\n}\nthis.files[file]= fid\n}\nvar f io.ReadWriteSeeker= fid\n\n\n/*71:*/\n\n\n//line goacme.w:1016\n\nf= &wrapper{f:f}\n\n\n\n\n\n/*:71*/\n\n\n//line goacme.w:334\n\nreturn f,nil\n}", "func(this*Window)Read(p[]byte)(int,error){\nf,err:=this.File(\"body\")\nif err!=nil{\nreturn 0,err\n}\nreturn f.Read(p)\n}", "func printErr (err error){ // this function is used for printing error\n if err != nil{\n fmt.Println(err)\n }\n}", "func (mtr *Mxmx0intmacMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"Lane0Sbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane0Sbe.Size()\n\n\tif fldName == \"Lane0Dbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane0Dbe.Size()\n\n\tif fldName == \"Lane1Sbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane1Sbe.Size()\n\n\tif fldName == \"Lane1Dbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane1Dbe.Size()\n\n\tif fldName == \"Lane2Sbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane2Sbe.Size()\n\n\tif fldName == \"Lane2Dbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane2Dbe.Size()\n\n\tif fldName == \"Lane3Sbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane3Sbe.Size()\n\n\tif fldName == \"Lane3Dbe\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Lane3Dbe.Size()\n\n\tif fldName == \"M0PbPbusDrdy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.M0PbPbusDrdy.Size()\n\n\tif fldName == \"M1PbPbusDrdy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.M1PbPbusDrdy.Size()\n\n\tif fldName == \"M2PbPbusDrdy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.M2PbPbusDrdy.Size()\n\n\tif fldName == \"M3PbPbusDrdy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.M3PbPbusDrdy.Size()\n\n\treturn offset\n}", "func check_complete(){\nif len(change_buffer)> 0{/* changing is false */\nbuffer= change_buffer\nchange_buffer= nil\nchanging= true\nchange_depth= include_depth\nloc= 0\nerr_print(\"! Change file entry did not match\")\n\n}\n}", "func GetResTransactionCost()int{\n return 1 // TODO: tune later\n}", "func (px *Paxos) Status(seq int) (bool, interface{}) {\n // Your code here.\n DPrintf(\"Paxos %v is getting status ...\\n\", px.me)\n if seq < px.Min() {\n DPrintf(\"Required Status less than min available...\\n\")\n return false, nil\n }\n ins := px.getProp(seq)\n DPrintf(\"Return status '%v' with value %v...\\n\", ins.status, ins.v_a)\n return ins.status, ins.v_a\n}", "func XGETBV() { ctx.XGETBV() }", "func (self *TileSprite) FrameName() string{\n return self.Object.Get(\"frameName\").String()\n}", "func (r *importReader) pos() {\n\tif r.int64() != deltaNewFile {\n\t} else if l := r.int64(); l == -1 {\n\t} else {\n\t\tr.string()\n\t}\n}", "func TestGetNone4A(t *testing.T) {\n}", "func (object MQObject) getInternal(gomd *MQMD,\n\tgogmo *MQGMO, buffer []byte, useCap bool) (int, error) {\n\n\tvar mqrc C.MQLONG\n\tvar mqcc C.MQLONG\n\tvar mqmd C.MQMD\n\tvar mqgmo C.MQGMO\n\tvar datalen C.MQLONG\n\tvar ptr C.PMQVOID\n\n\terr := checkMD(gomd, \"MQGET\")\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\terr = checkGMO(gogmo, \"MQGET\")\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tbufflen := 0\n\tif useCap {\n\t\tbufflen = cap(buffer)\n\t} else {\n\t\tbufflen = len(buffer)\n\t}\n\n\tcopyMDtoC(&mqmd, gomd)\n\tcopyGMOtoC(&mqgmo, gogmo)\n\n\tif bufflen > 0 {\n\t\t// There has to be something in the buffer for CGO to be able to\n\t\t// find its address. We know there's space backing the buffer so just\n\t\t// set the first byte to something.\n\t\tif useCap && len(buffer) == 0 {\n\t\t\tbuffer = append(buffer, 0)\n\t\t}\n\t\tptr = (C.PMQVOID)(unsafe.Pointer(&buffer[0]))\n\t} else {\n\t\tptr = nil\n\t}\n\n\tC.MQGET(object.qMgr.hConn, object.hObj, (C.PMQVOID)(unsafe.Pointer(&mqmd)),\n\t\t(C.PMQVOID)(unsafe.Pointer(&mqgmo)),\n\t\t(C.MQLONG)(bufflen),\n\t\tptr,\n\t\t&datalen,\n\t\t&mqcc, &mqrc)\n\n\tgodatalen := int(datalen)\n\tcopyMDfromC(&mqmd, gomd)\n\tcopyGMOfromC(&mqgmo, gogmo)\n\n\tmqreturn := MQReturn{MQCC: int32(mqcc),\n\t\tMQRC: int32(mqrc),\n\t\tverb: \"MQGET\",\n\t}\n\n\tif mqcc != C.MQCC_OK {\n\t\treturn godatalen, &mqreturn\n\t}\n\n\treturn godatalen, nil\n\n}", "func Supported() bool { return true }", "func init_node(node int32){\nname_dir[node].equiv= -1\n}", "func (self *Graphics) Name() string{\n return self.Object.Get(\"name\").String()\n}", "func getHelper(nodeRPC string, input string, fname string, cmd string) {\n\n\tcolorprint.Info(\">>>> Please enter the name of the file that you would like to obtain\")\n\tfmt.Scan(&fname)\n\tcolorprint.Debug(\"<<<< \" + fname)\n\tcolorprint.Info(\">>>> Please enter the address of the node you want to connect to\")\n\tfmt.Scan(&input)\n\tcolorprint.Debug(\"<<<< \" + input)\n\tnodeAddr := input\n\t// Connect to utility.Service via RPC // returns *Client, err\n\tavail, _, _ := CheckFileAvailability(fname, nodeAddr)\n\tif avail && (cmd == \"get\") {\n\t\tcolorprint.Info(\">>>> Would you like to get the file from the node[\" + nodeRPC + \"]?(y/n)\")\n\t\tfmt.Scan(&input)\n\t\tcolorprint.Debug(\"<<<< \" + input)\n\t\tif input == \"y\" {\n\t\t\t// TODO\n\t\t}\n\t}\n}", "func (mtr *Dppdpp0intreg1Metrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"ErrPhvSopNoEop\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPhvSopNoEop.Size()\n\n\tif fldName == \"ErrPhvEopNoSop\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPhvEopNoSop.Size()\n\n\tif fldName == \"ErrOhiSopNoEop\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrOhiSopNoEop.Size()\n\n\tif fldName == \"ErrOhiEopNoSop\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrOhiEopNoSop.Size()\n\n\tif fldName == \"ErrFramerCreditOverrun\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrFramerCreditOverrun.Size()\n\n\tif fldName == \"ErrPacketsInFlightCreditOverrun\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPacketsInFlightCreditOverrun.Size()\n\n\tif fldName == \"ErrNullHdrVld\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrNullHdrVld.Size()\n\n\tif fldName == \"ErrNullHdrfldVld\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrNullHdrfldVld.Size()\n\n\tif fldName == \"ErrMaxPktSize\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrMaxPktSize.Size()\n\n\tif fldName == \"ErrMaxActiveHdrs\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrMaxActiveHdrs.Size()\n\n\tif fldName == \"ErrPhvNoDataReferenceOhi\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPhvNoDataReferenceOhi.Size()\n\n\tif fldName == \"ErrCsumMultipleHdr\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrCsumMultipleHdr.Size()\n\n\tif fldName == \"ErrCsumMultipleHdrCopy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrCsumMultipleHdrCopy.Size()\n\n\tif fldName == \"ErrCrcMultipleHdr\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrCrcMultipleHdr.Size()\n\n\tif fldName == \"ErrPtrFifoCreditOverrun\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPtrFifoCreditOverrun.Size()\n\n\tif fldName == \"ErrClipMaxPktSize\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrClipMaxPktSize.Size()\n\n\tif fldName == \"ErrMinPktSize\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrMinPktSize.Size()\n\n\treturn offset\n}", "func treeGene( geneArray[] string, arrayPos int, shapeType string) int{\n\n//Function check to see if the there is any shapes in the in the listArray\n//if not it changes the structures as it moves to the end\n\tif listArray[0] != shapeType { \n\t\t\tlistArray[1]= \"<inst>\"\n\t\t\tlistArray[0] = shapeType\n\t\tarrayPos++\n\t\ttreeGene(geneArray[0:],arrayPos,shapeType)\t// Called the function so it can be processed with the valid format \t\n\t}else{ if listArray[1] == \"\" || listArray[1] == \"<inst>\"{ // after transforming it is place in a format that can be parsed \n\t\t\tif geneArray[arrayPos] == shapeType{ \t// Ensure we are not Validating a Shape\n\t\t\t\tarrayPos++\n\t\t\t}\n\t\t\t// Retrieve the Coordinated from the array\n\t\t\tvar curCoord string=geneArray[arrayPos] \n\t\t\tarrayPos++\n\t\t\t\n\t\t\t// Using Slices we get each Values \n\t\t\tx:=curCoord[0:1] \n\t\t\ty:=curCoord[1:2]\n\t\t\tfmt.Printf(\"\\n |\\n\")\n\t\t\t\n\t\t\t// The Only Difference with Circle is that it Has a Coordinated\n\t\t\tfmt.Printf(shapeType+\"\\n / \\\\\\n\")\n\t\t\tif shapeType == \"CIR\"{\n\t\t\t\tfmt.Printf(\"<coord>\\n\")\n\t\t\t}\t\t\t\n\t\t\tfmt.Printf(\"<x><y>\\n\"+x+\" \"+y)\n\t\t\t\n\t\t}\n\t\t }\n\n\treturn arrayPos\n}", "func (tw *Topwords) get_inserter() func(uint8) {\n\tcurrent := &tw.start\n\tinword := false\n\n\treturn func(b uint8) {\n\t\tb = filterToLower(b)\n\n\t\tif b == ' ' {\n\t\t\t//fmt.Printf(\"_\")\n\t\t\tif inword {\n\t\t\t\tif current.word_count == 0 {\n\t\t\t\t\ttw.uniquewordcount += 1\n\t\t\t\t}\n\t\t\t\tcurrent.word_count += 1\n\t\t\t\tcurrent = &tw.start // Start at the beginning again\n\t\t\t}\n\t\t\tinword = false\n\t\t\treturn\n\t\t}\n\t\t//fmt.Printf(\"%v\", string(b))\n\t\tinword = true\n\t\tchildren := &(current.children)\n\t\tfor e := range *children {\n\t\t\tif (*children)[e].char == b {\n\t\t\t\t//fmt.Printf(\"<\")\n\t\t\t\tcurrent = (*children)[e]\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\t//fmt.Printf(\"^\")\n\t\tnew_node := &Node{b, 0, nil}\n\t\tcurrent.children = append(current.children, new_node)\n\t\tcurrent = new_node\n\t\t// fmt.Printf(\"\\n+%v\\n\", current.children)\n\t\treturn\n\n\t}\n\n}", "func (mtr *Dppdpp1intsramseccMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"DppPhvFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DppPhvFifoUncorrectable.Size()\n\n\tif fldName == \"DppPhvFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DppPhvFifoCorrectable.Size()\n\n\tif fldName == \"DppOhiFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DppOhiFifoUncorrectable.Size()\n\n\tif fldName == \"DppOhiFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DppOhiFifoCorrectable.Size()\n\n\treturn offset\n}", "func main(){\n\t// make a request and then the request will return response in bytes\n\t// and error. An error might be empty, for now we define it as underscore\n\t// we can define a variable that you don't intend to use as _ (underscore)\n\n\tresp, _ := http.Get(\"https://www.washingtonpost.com/news-sitemaps/index.xml\")\n\t//unpack it\n\tbytes, _ := ioutil.ReadAll(resp.Body)\n\t// parse to string\n\t// string_body := string(bytes)\n\t// fmt.Println(string_body)\n\tresp.Body.Close()\n\n\tvar s SitemapIndex\n\txml.Unmarshal(bytes, &s) // memory address s\n\n\t// fmt.Println(s.Locations)\n\t// After we know how to loop, we can print all the URLs as multilines of strings\n\tfmt.Printf(\"Here %s some %s\", \"are\",\"variables\")\n\t// we use _, Location because in map in will contain both key and value.\n\t// then _ is the index of the url (key) which we don't want to print and know. \n\t// So, we use _ that we already declared (just to iterate like i or something)\n\t// what we really want is the next one which is the \"value\" or the real url data which uses Location as iterator\n\t// *Note that _, Location = 1, http://...\n\t// *and Location is just a variable which is not associated w/ Location struct.\n\tfor _, Location := range s.Locations {\n\t\tfmt.Printf(\"\\n%s\", Location)\n\t}\n}", "func (mtr *Dppdpp0intsramseccMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"DppPhvFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DppPhvFifoUncorrectable.Size()\n\n\tif fldName == \"DppPhvFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DppPhvFifoCorrectable.Size()\n\n\tif fldName == \"DppOhiFifoUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DppOhiFifoUncorrectable.Size()\n\n\tif fldName == \"DppOhiFifoCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.DppOhiFifoCorrectable.Size()\n\n\treturn offset\n}", "func (objectIDs *DbObjectIDs) getUniqueID() (string, error) {\n\tid := objectIDs.ownerControllerName + \":\" + string(objectIDs.idsType.ownerObjectType)\n\tfor _, key := range objectIDs.idsType.GetExternalIDKeys() {\n\t\tvalue, ok := objectIDs.objectIDs[key]\n\t\tif !ok {\n\t\t\treturn \"\", fmt.Errorf(\"key %v is required but not present\", key)\n\t\t}\n\t\tid += \":\" + value\n\t}\n\treturn id, nil\n}", "func (mtr *Dppdpp1intreg1Metrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"ErrPhvSopNoEop\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPhvSopNoEop.Size()\n\n\tif fldName == \"ErrPhvEopNoSop\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPhvEopNoSop.Size()\n\n\tif fldName == \"ErrOhiSopNoEop\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrOhiSopNoEop.Size()\n\n\tif fldName == \"ErrOhiEopNoSop\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrOhiEopNoSop.Size()\n\n\tif fldName == \"ErrFramerCreditOverrun\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrFramerCreditOverrun.Size()\n\n\tif fldName == \"ErrPacketsInFlightCreditOverrun\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPacketsInFlightCreditOverrun.Size()\n\n\tif fldName == \"ErrNullHdrVld\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrNullHdrVld.Size()\n\n\tif fldName == \"ErrNullHdrfldVld\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrNullHdrfldVld.Size()\n\n\tif fldName == \"ErrMaxPktSize\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrMaxPktSize.Size()\n\n\tif fldName == \"ErrMaxActiveHdrs\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrMaxActiveHdrs.Size()\n\n\tif fldName == \"ErrPhvNoDataReferenceOhi\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPhvNoDataReferenceOhi.Size()\n\n\tif fldName == \"ErrCsumMultipleHdr\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrCsumMultipleHdr.Size()\n\n\tif fldName == \"ErrCsumMultipleHdrCopy\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrCsumMultipleHdrCopy.Size()\n\n\tif fldName == \"ErrCrcMultipleHdr\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrCrcMultipleHdr.Size()\n\n\tif fldName == \"ErrPtrFifoCreditOverrun\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrPtrFifoCreditOverrun.Size()\n\n\tif fldName == \"ErrClipMaxPktSize\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrClipMaxPktSize.Size()\n\n\tif fldName == \"ErrMinPktSize\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ErrMinPktSize.Size()\n\n\treturn offset\n}", "func (self *Graphics) RenderOrderID() int{\n return self.Object.Get(\"renderOrderID\").Int()\n}", "func fidxdb0_1(sdb, ddb *bolt.DB) (err error) {\n\tfmt.Println(\"[IdxDB] fix 0 -> 1, #128\")\n\n\tvar feedsBucket = []byte(\"f\")\n\n\t// create necessary buckets\n\n\terr = ddb.Update(func(tx *bolt.Tx) (err error) {\n\n\t\t// feeds\n\t\tif _, err = tx.CreateBucketIfNotExists(feedsBucket); err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tvar info *bolt.Bucket\n\t\tif info, err = tx.CreateBucketIfNotExists(metaBucket); err != nil {\n\t\t\treturn\n\t\t}\n\n\t\t// 0 -> 1\n\t\treturn info.Put(versionKey, versionBytes(1))\n\t})\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// service functions\n\tvar decSeq = func(val []byte) (seq uint32) {\n\t\tseq = binary.LittleEndian.Uint32(val) // le\n\t\treturn\n\t}\n\n\t// returns new uint32\n\tvar encSeq = func(seq uint32) (val []byte) {\n\t\tval = make([]byte, 4)\n\t\tbinary.BigEndian.PutUint32(val, seq) // be\n\t\treturn\n\t}\n\n\t// range cover objects\n\n\treturn sdb.View(func(stx *bolt.Tx) (err error) {\n\n\t\tsfeeds := stx.Bucket(feedsBucket)\n\n\t\treturn ddb.Update(func(dtx *bolt.Tx) (err error) {\n\n\t\t\tdfeeds := dtx.Bucket(feedsBucket)\n\n\t\t\t// iterate over feeds\n\n\t\t\tvar sfc = sfeeds.Cursor() // source feeds cursor\n\n\t\t\tfor f, _ := sfc.First(); f != nil; f, _ = sfc.Next() {\n\n\t\t\t\tvar sf = sfeeds.Bucket(f) // source feed\n\n\t\t\t\tvar df *bolt.Bucket // destination feed\n\t\t\t\tif df, err = dfeeds.CreateBucket(f); err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\terr = sf.ForEach(func(seqb, val []byte) (err error) {\n\t\t\t\t\tseq := decSeq(seqb) // le\n\t\t\t\t\treturn df.Put(encSeq(seq), val) // be\n\t\t\t\t})\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\treturn\n\n\t\t})\n\t})\n\n\t// done\n\n}", "func getRelHash(r string) [8]int {\n\th1 := sha1.New()\n\th1.Write([]byte(r))\n\tb1 := h1.Sum(nil)\n\tdata1 := b1[0]\n\tid1 := data1 % 4\n\tidint := int(id1)\n\tvar nodelist [8]int\n\tfor k := 0; k < 8; k++ {\n\t\tnodelist[k] = (k * 4) + idint\n\t}\n\t//fmt.Println(\"Nodelist for given relation\", nodelist)\n\treturn nodelist\n}", "func (mtr *Msmsintprp1Metrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"Read\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Read.Size()\n\n\tif fldName == \"Security\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Security.Size()\n\n\tif fldName == \"Decode\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Decode.Size()\n\n\treturn offset\n}", "func (q *dnsQuestion) getQuestion(req []byte, offset uint16, head *dnsHeader) {\n\tost := offset\n\ttmp := ost\n\tost = q.getQName(req, ost)\n\tq.qType = binary.BigEndian.Uint16(req[ost : ost+twoByteSize])\n\tost += twoByteSize\n\tq.qClass = binary.BigEndian.Uint16(req[ost : ost+twoByteSize])\n\tost += twoByteSize\n\tq.head = head\n\tq.queByte = req[tmp:ost]\n}", "func getBiography(age int, name string, status string) (string, string){\n\tageNow := strconv.Itoa(age)\n\n\treturn name + \" adalah seorang \"+ status,\n\t\t \"umurnya \"+ ageNow\n\n\n}", "func check_change(){\nn:=0/* the number of discrepancies found */\nif compare_runes(buffer,change_buffer)!=0{\nreturn\n}\nchange_pending= false\nif!changed_section[section_count]{\nif_section_start_make_pending(true)\nif!change_pending{\nchanged_section[section_count]= true\n}\n}\nfor true{\nchanging= true\nprint_where= true\nchange_line++\nif err:=input_ln(change_file);err!=nil{\nerr_print(\"! Change file ended before @y\")\n\nchange_buffer= nil\nchanging= false\nreturn\n}\nif len(buffer)> 1&&buffer[0]=='@'{\nvar xyz_code rune\nif unicode.IsUpper(buffer[1]){\nxyz_code= unicode.ToLower(buffer[1])\n}else{\nxyz_code= buffer[1]\n}\n\n\n/*28:*/\n\n\n//line gocommon.w:333\n\nif xyz_code=='x'||xyz_code=='z'{\nloc= 2\nerr_print(\"! Where is the matching @y?\")\n\n}else if xyz_code=='y'{\nif n> 0{\nloc= 2\nerr_print(\"! Hmm... %d of the preceding lines failed to match\",n)\n\n}\nchange_depth= include_depth\nreturn\n}\n\n\n\n/*:28*/\n\n\n//line gocommon.w:309\n\n}\n\n\n/*23:*/\n\n\n//line gocommon.w:222\n\n{\nchange_buffer= buffer\nbuffer= nil\n}\n\n\n\n/*:23*/\n\n\n//line gocommon.w:311\n\nchanging= false\nline[include_depth]++\nfor input_ln(file[include_depth])!=nil{/* pop the stack or quit */\nif include_depth==0{\nerr_print(\"! GOWEB file ended during a change\")\n\ninput_has_ended= true\nreturn\n}\ninclude_depth--\nline[include_depth]++\n}\nif compare_runes(buffer,change_buffer)!=0{\nn++\n}\n}\n}", "func get_section_name(p int32)(dest[]rune,complete bool){\nq:=p+1\nfor p!=-1{\ndest= append(dest,name_dir[p].name[1:]...)\nif name_dir[p].ispref{\np= name_dir[q].llink\nq= p\n}else{\np= -1\nq= -2\n}\n}\ncomplete= true\nif q!=-2{\ncomplete= false/* complete name not yet known */\n}\nreturn\n}", "func (rpcServer * RPCServer)makeInsertsToSuccessor(){\n\t//open a read transaction\n\trpcServer.boltDB.View(func(tx *bolt.Tx) error {\n\t\tvar cursor *bolt.Cursor\n\t\tcursor = tx.Cursor()\n\t\t\n\t\tvar bucket *bolt.Bucket\n\t\t\n\n\t\t//traverse through all keys\n\t\tfor k, _ := cursor.First(); k != nil; k, _ = cursor.Next() {\n\t\t\tbucket = tx.Bucket(k)\n\t\t\t\n\t\t\t//traverse through all relation and value pairs\n\t\t\tbucket.ForEach(func(relation, value []byte) error {\n\t\t\t\t//create paramter - successor\n\t\t\t\n\t\t\t\t//add to array of interface\n\t\t\t\t\n\t\t\t\tparameterArray := make([]interface{},3)\n\t\t\t\tparameterArray[0] = string(k)\n\t\t\t\tparameterArray[1] = string(relation)\n\t\t\t\tparameterArray[2] = string(value)\n\t\t\t\t\n\n\t\t\t\t//create json message\n\t\t\t\tjsonMessage := rpcclient.RequestParameters{}\n\t\t\t\tjsonMessage.Method = \"Insert\";\n\t\t\t\tjsonMessage.Params = parameterArray\n\t\t\t\t\n\t\t\t\tjsonBytes,err :=json.Marshal(jsonMessage)\n\t\t\t\tif err!=nil{\n\t\t\t\t\trpcServer.logger.Println(err)\n\t\t\t\t\treturn err\n\t\t\t\t} \n \n\t\t\t\trpcServer.logger.Println(string(jsonBytes))\n\n\t\t\t\tclientServerInfo,err := rpcServer.chordNode.PrepareClientServerInfo(rpcServer.chordNode.FingerTable[1])\n\t\t\t\tif err!=nil{\n\t\t\t\t\t\n\t\t\t\t\trpcServer.logger.Println(err)\n\t\t\t\t\treturn nil\n\t\t\t\t\t\n\t\t\t\t}\n\t\t\t\tclient := &rpcclient.RPCClient{}\n\t\t\t\terr, _ = client.RpcCall(clientServerInfo, string(jsonBytes))\n\t\t\t\t\n\t\t\t\tif err != nil {\n\t\t\t\t\trpcServer.logger.Println(err)\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\t\t\n\t\t\t\t\n\t\t\t\treturn nil\n\t\t\t})\n\t\t}\n\t\treturn nil\n\t})\n\n}", "func (mtr *MsmsintmiscMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"ElamCapEccUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ElamCapEccUncorrectable.Size()\n\n\tif fldName == \"ElamCapEccCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.ElamCapEccCorrectable.Size()\n\n\tif fldName == \"Bl2RamEccUncorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Bl2RamEccUncorrectable.Size()\n\n\tif fldName == \"Bl2RamEccCorrectable\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Bl2RamEccCorrectable.Size()\n\n\treturn offset\n}", "func getTemplateInfoByName(t *Template, name string, owner string, version string) {\n\ttheUrl := config.CdnUrl + \"/template?name=\" + name\n\n\tif owner != \"\" {\n\t\ttheUrl += \"&owner=\" + owner\n\t}\n\n\tif version == \"\" {\n\t\ttheUrl += \"&version=latest\"\n\t} else {\n\t\ttheUrl += \"&version=\" + version\n\t}\n\n\tclnt := util.GetClient(config.CDN.AllowInsecure, 30)\n\n\tresponse, err := util.RetryGet(theUrl, clnt, 3)\n\n\tlog.Check(log.ErrorLevel, \"Retrieving template info, get: \"+theUrl, err)\n\tdefer util.Close(response)\n\n\tif response.StatusCode == 404 {\n\t\tlog.Error(\"Template \" + t.Name + \" not found\")\n\t}\n\tif response.StatusCode != 200 {\n\t\tlog.Error(\"Failed to get template info: \" + response.Status)\n\t}\n\n\tbody, err := ioutil.ReadAll(response.Body)\n\tlog.Check(log.ErrorLevel, \"Reading template info\", err)\n\n\tvar templ Template\n\tif log.Check(log.WarnLevel, \"Parsing response body\", json.Unmarshal(body, &templ)) {\n\t\tlog.Error(\"Failed to parse template info\")\n\t}\n\n\tt.Name = templ.Name\n\tt.Owner = templ.Owner\n\tt.Version = templ.Version\n\tt.Id = templ.Id\n\tt.MD5 = templ.MD5\n\tt.Parent = templ.Parent\n\tt.Size = templ.Size\n\tt.DigestMethod = templ.DigestMethod\n\tt.DigestHash = templ.DigestHash\n\n\tlog.Debug(\"Template identified as \" + t.Name + \"@\" + t.Owner + \":\" + t.Version)\n}", "func TestSinglePrewrite4A(t *testing.T) {\n}", "func main(){\ncommon_init()\n\n\n/*99:*/\n\n\n//line gotangle.w:168\n\nlast_unnamed= 0\ntext_info= append(text_info,text{})\ntext_info[0].text_link= 0\n\n\n\n/*:99*/\n\n\n\n/*127:*/\n\n\n//line gotangle.w:665\n{\nfor c:=0;c<len(ccode);c++{\nccode[c]= ignore\n}\nccode[' ']= new_section\nccode['\\t']= new_section\nccode['\\n']= new_section\nccode['\\v']= new_section\nccode['\\r']= new_section\nccode['\\f']= new_section\nccode['*']= new_section\nccode['@']= '@'\nccode['=']= strs\nccode['d']= definition\nccode['D']= definition\nccode['f']= format_code\nccode['F']= format_code\nccode['s']= format_code\nccode['S']= format_code\nccode['c']= begin_code\nccode['C']= begin_code\nccode['p']= begin_code\nccode['P']= begin_code\nccode['^']= control_text\nccode[':']= control_text\nccode['.']= control_text\nccode['t']= control_text\nccode['T']= control_text\nccode['r']= control_text\nccode['R']= control_text\nccode['q']= control_text\nccode['Q']= control_text\nccode['&']= join\nccode['<']= section_name\nccode['(']= section_name\nccode['\\'']= ord\n}\n\n\n\n/*:127*/\n\n\n//line gotangle.w:82\n\nif show_banner(){\nfmt.Print(banner)/* print a ``banner line'' */\n}\nphase_one()/* read all the user's text and compress it into tok_mem */\nphase_two()/* output the contents of the compressed tables */\nos.Exit(wrap_up())/* and exit gracefully */\n}", "func (l *lseditor) exec(com *command) bool {\n\tlinum := false\n\ti1 := com.index1\n\ti2 := com.index2\n\tif com.index1 == CurLine {\n\t\ti1 = l.currow\n\t}\n\tif com.index2 == CurLine {\n\t\ti2 = l.currow\n\t}\n\tif i1 == LastLine || i1 > l.numrows {\n\t\ti1 = l.numrows\n\t}\n\tif i2 == LastLine || i2 > l.numrows {\n\t\ti2 = l.numrows\n\t}\n\tif i1 < 1 {\n\t\ti1 = 1\n\t}\n\tswitch com.name {\n\tcase \".abort\":\n\t\treturn true\n\tcase \".end\":\n\t\terr := l.save()\n\t\tif err != nil {\n\t\t\tfmt.Fprintf(os.Stderr, \"< %s >\\n\", err.Error())\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\tcase \".save\":\n\t\terr := l.save()\n\t\tif err != nil {\n\t\t\tfmt.Fprintln(os.Stderr, err.Error())\n\t\t}\n\tcase \".find\":\n\t\tif i1 > l.numrows {\n\t\t\ti1 = l.numrows\n\t\t}\n\t\tfor i, row := range l.rows[i1-1:] {\n\t\t\tif strings.Contains(row, com.equals) {\n\t\t\t\tprintLinum(i1 + i)\n\t\t\t\tfmt.Println(row)\n\t\t\t}\n\t\t}\n\tcase \".d\", \".del\":\n\t\tif com.index2 == NoIndex {\n\t\t\tl.delLine(i1 - 1)\n\t\t} else {\n\t\t\tl.delRegion(i1-1, i2)\n\t\t}\n\tcase \".copy\":\n\t\teq, err := strconv.Atoi(com.equals)\n\t\tif err != nil {\n\t\t\tfmt.Fprintf(os.Stderr, \"< %s >\\n\", err)\n\t\t\treturn false\n\t\t}\n\n\t\treg := l.createRegion(i1-1, i2)\n\t\tif reg == nil {\n\t\t\tfmt.Fprintln(os.Stderr, \"< Invalid region >\")\n\t\t\treturn false\n\t\t}\n\n\t\tl.insertRegion(reg, eq)\n\tcase \".h\", \".help\":\n\t\thelpScreen()\n\tcase \".i\", \".insert\":\n\t\tif com.index1 == NoIndex || com.index1 == CurLine {\n\t\t\t// Do nothing; just print the current row number\n\t\t} else if com.index1 == LastLine || com.index1 > l.numrows {\n\t\t\tl.currow = l.numrows\n\t\t} else {\n\t\t\tl.currow = com.index1 - 1\n\t\t}\n\t\tl.printCurRow()\n\tcase \".p\", \".print\":\n\t\tlinum = true\n\t\tfallthrough\n\tcase \".l\", \".list\":\n\t\tif com.index1 == NoIndex {\n\t\t\tfor i, row := range l.rows {\n\t\t\t\tif linum {\n\t\t\t\t\tprintLinum(i + 1)\n\t\t\t\t}\n\t\t\t\tfmt.Println(row)\n\t\t\t}\n\t\t\treturn false\n\t\t} else if com.index1 == LastLine || com.index1 == l.numrows {\n\t\t\tif linum {\n\t\t\t\tprintLinum(l.numrows)\n\t\t\t}\n\t\t\tfmt.Println(l.rows[l.numrows-1])\n\t\t\treturn false\n\t\t} else if com.index1 > l.numrows {\n\t\t\treturn false\n\t\t}\n\t\tif i2 == NoIndex || i2 <= com.index1 {\n\t\t\tif linum {\n\t\t\t\tprintLinum(i1)\n\t\t\t}\n\t\t\tfmt.Println(l.rows[i1-1])\n\t\t} else {\n\t\t\tfor i, row := range l.rows[i1-1 : i2] {\n\t\t\t\tif linum {\n\t\t\t\t\tprintLinum(i1 + i)\n\t\t\t\t}\n\t\t\t\tfmt.Println(row)\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func (mtr *Mcmc1mchintmcMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"Ecc_1BitThreshPs1\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Ecc_1BitThreshPs1.Size()\n\n\tif fldName == \"Ecc_1BitThreshPs0\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Ecc_1BitThreshPs0.Size()\n\n\treturn offset\n}", "func (self *Graphics) FillAlpha() int{\n return self.Object.Get(\"fillAlpha\").Int()\n}", "func Make_Tile_Geojson2(tileid m.TileID,layername string,features_str []*geojson.Feature) []byte {\n\t// getitng featuerstr\n\tvar bytevals []byte\n\tif len(features_str) > 100000000 {\n\t\t//fmt.Println(\"concurrent\")\n\t} else {\n\n\t\t// intializing shit for cursor\n\t\tbound := m.Bounds(tileid)\n\t\tdeltax := bound.E-bound.W\n\t\tdeltay := bound.N - bound.S\n\n\t\t// random intializatioin for property collection\n\t\tvar keys []string\n\t\tvar values []*vector_tile.Tile_Value\n\t\tkeysmap := map[string]uint32{}\n\t\tvaluesmap := map[*vector_tile.Tile_Value]uint32{}\n\n\t\t// iterating through each feature\n\t\tfeatures := []*vector_tile.Tile_Feature{}\n\t\t\n\t\t// setting and converting coordinate\t\n\t\tcur := Cursor{LastPoint:[]int32{0,0},Bounds:bound,DeltaX:deltax,DeltaY:deltay,Count:0}\n\t\tcur = Convert_Cursor(cur)\n\t\tvar bytevals []byte\n\n\t\t// creating new mapper\n\t\t\n\t\t//position := []int32{0, 0}\n\t\tfor _,i := range features_str {\n\t\t\t// appplying the soft boudning box filter\n\t\t\t\t\n\t\t\t\t// getitng the featuer\t\t\n\t\t\n\n\t\t\t\t// adding properties and getting correct tags\n\t\t\t\tvar tags, geometry []uint32\n\t\t\t\tvar feat vector_tile.Tile_Feature\n\t\t\t\ttags, keys, values, keysmap, valuesmap = Update_Properties(i.Properties, keys, values, keysmap, valuesmap)\n\t\t\t\t//fmt.Println(i.Geometry)\n\t\t\t\t// logic for point feature'\n\t\t\t\tif i.Geometry == nil {\n\n\t\t\t\t} else if i.Geometry.Type == \"Point\" {\n\t\t\t\t\tgeometry = cur.Make_Point_Float(i.Geometry.Point)\n\t\t\t\t\tfeat_type := vector_tile.Tile_POINT\n\t\t\t\t\tfeat = vector_tile.Tile_Feature{Tags: tags, Type: &feat_type, Geometry: geometry}\n\t\t\t\t\tfeatures = append(features, &feat)\n\n\t\t\t\t} else if i.Geometry.Type == \"LineString\" {\n\t\t\t\t\tif len(i.Geometry.LineString) >= 2 {\n\t\t\t\t\t\tgeometry = cur.Make_Line_Float(i.Geometry.LineString)\n\t\t\t\t\t\tif geometry[3] > 2 {\n\t\t\t\t\t\t\tfeat_type := vector_tile.Tile_LINESTRING\n\t\t\t\t\t\t\tfeat = vector_tile.Tile_Feature{Tags: tags, Type: &feat_type, Geometry: geometry}\n\t\t\t\t\t\t\tfeatures = append(features, &feat)\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\t\t\t\t} else if i.Geometry.Type == \"Polygon\" {\n\t\t\t\t\tgeometry = cur.Make_Polygon_Float(i.Geometry.Polygon)\n\t\t\t\t\tfeat_type := vector_tile.Tile_POLYGON\n\t\t\t\t\tfeat = vector_tile.Tile_Feature{Tags: tags, Type: &feat_type, Geometry: geometry}\n\t\t\t\t\tfeatures = append(features, &feat)\n\t\t\t\t}\n\t\t}\n\n\t\t//fmt.Println(len(features))\n\t\tif len(features) > 0 {\n\t\t\tlayerVersion := uint32(15)\n\t\t\textent := vector_tile.Default_Tile_Layer_Extent\n\t\t\t//var bound []Bounds\n\t\t\tlayer := vector_tile.Tile_Layer{\n\t\t\t\tVersion: &layerVersion,\n\t\t\t\tName: &layername,\n\t\t\t\tExtent: &extent,\n\t\t\t\tValues: values,\n\t\t\t\tKeys: keys,\n\t\t\t\tFeatures: features,\n\t\t\t}\n\n\t\t\ttile := vector_tile.Tile{}\n\t\t\ttile.Layers = append(tile.Layers, &layer)\n\t\t\tbytevals,_ = proto.Marshal(&tile)\n\t\t\t//if len(bytevals) > 0 {\n\t\t\t//\tmbtile.Add_Tile(tileid,bytevals)\n\t\t\t//}\n\t\t} else {\n\t\t\tbytevals = []byte{}\n\t\t}\n\t\tif len(bytevals) > 0 { \n\t\t}\n\t}\n\treturn bytevals\n}", "func PrintFingerPrint(src string) {\n fp, _ := getFingerPrint(src)\n for _, it := range fp.items {\n fmt.Println(it.str, it.cnt)\n }\n}", "func (self *TileSprite) Name() string{\n return self.Object.Get(\"name\").String()\n}", "func (self *Graphics) Width() int{\n return self.Object.Get(\"width\").Int()\n}", "func treeSQR( sqrArray[] string, arrayPos int) int{\n\n//Function check to see if the there is any shapes in the in the listArray\n//if not it changes the structures as it moves to the end\nif listArray[0] != \"SQR\" { \n\t\t\tlistArray[1]= \"<inst>\"\n\t\t\tlistArray[0] = \"SQR\"\n\t\tarrayPos++\n\t\t\n\t\t// Called the function so it can be processed with the valid format \n\t\ttreeSQR(sqrArray[0:],arrayPos)\t\n\t}else{ if listArray[1] == \"\" || listArray[1] == \"<inst>\"{ // after transforming it is place in a format that can be parsed \n\t\n\t\t\tif sqrArray[arrayPos] == \"SQR\"{ // Ensure we are not Validating a Shape\n\t\t\t\tarrayPos++\n\t\t\t}\n\t\t\t\n\t\t\t// Retrieve the Coordinated from the array\n\t\t\t// Proceeding to the next value\n\t\t\tvar curCoord string=sqrArray[arrayPos]\n\t\t\tarrayPos++\n\t\t\tvar secCoord string=sqrArray[arrayPos]\n\t\t\t//arrayPos++\t\t\n\t\t\t\n\t\t\t// Using Slices we get each Values \n\t\t\tvar test string \n\t\t\tx:=curCoord[0:1]\n\t\t\ty:=curCoord[1:2]\n\t\t\tyy:=secCoord[0:1]\n\t\t\txx:=secCoord[1:2]\n\t\t\tarrayPos=arrayPos-2\n\t\t\t\n\t\t\t//The Printing format for the lower part of the tree\n\t\t\tfmt.Printf(\"\\n |\\n\")\n\t\t\tfmt.Printf(\"SQR\\n/ \\\\\\n\")\n\t\t\tfmt.Printf(\"<coord>,<coord>\\n\")\n\t\t\tfmt.Printf(\" / \\\\ / \\\\\\n\")\n\t\t\tfmt.Printf(\"<x><y>,<x><y>\\n\"+x+\" \"+y+\" \"+yy+\" \"+xx)\n\t\t\n\t\t\tlistArray[0] = \"<inst>\"\n\t\t\tlistArray[1] = \"<inst_list>\"\n\t\t\t\t\t\t\n\t\t\ttempCount=tempCount-1\n\t\t\t\tif(tempCount >= 0){\t\t\t\t\n\t\t\t\t\tlistArray[tempCount]=\"\"\n\t\t\t\t\t\n\t\t\t\t}\n\t\t\treturn arrayPos\t\n\t\t}\n\t\treturn arrayPos\n\t\t }\n\t\n\treturn arrayPos\n}", "func getOffset(n string, offset uint32, fb byte) ([4]byte, bool) {\n\tfilesL.RLock()\n\tdefer filesL.RUnlock()\n\n\t/* Get hold of the file */\n\tf, ok := files[n]\n\tif !ok {\n\t\tlog.Panicf(\"no file %q for offset\", n)\n\t}\n\n\t/* Make sure we have enough file */\n\tvar a [4]byte\n\tif uint32(len(f.contents)-1) < offset {\n\t\treturn a, false\n\t}\n\ta[0] = fb\n\tcopy(a[1:], f.contents[offset:])\n\treturn a, true\n}", "func main() {\n\n\t\n}", "func printHeadTag(r *http.Response) {\n\tvar head string\n\tvar err error\n\tvar b = bytes.NewBuffer(nil)\n\tvar n int64\n\tvar openInd, closeInd int\n\t//if len(r.Body) == 0 {\n\t//\tfmt.Printf(\"getrequester.getHeadTag: no body\\n\")\n\t//\treturn\n\t//}\n\tif n, err = b.ReadFrom(r.Body); err != nil {\n\t\tfmt.Printf(\"getrequester.getHeadTag err=%v\\n\", err.Error())\n\t\thead = b.String()\n\t\tif head == \"\" {\n\t\t\tfmt.Printf(\"getrequester.getHeadTag: nothing was read from the body\\n\")\n\t\t} else {\n\t\t\tfmt.Printf(\"%v\\n\", head)\n\t\t}\n\t\treturn\n\t}\n\tif n == 0 {\n\t\tfmt.Printf(\"getrequester.getHeadTag: no body\\n\")\n\t\treturn\n\t}\n\tr.Body.Close()\n\topenInd = strings.Index(b.String(), \"<head>\")\n\tcloseInd = strings.Index(b.String(), \"</head>\")\n\tif openInd == -1 || closeInd == -1 {\n\t\tfmt.Printf(\"getrequester.getHeadTag no head tag (%v;%v)\", openInd, closeInd)\n\t\treturn\n\t}\n\thead = b.String()\n\thead = head[openInd : closeInd+6]\n\tfmt.Printf(\"%v\\n\", head)\n}", "func (self *Graphics) OffsetX() int{\n return self.Object.Get(\"offsetX\").Int()\n}", "func (mtr *Mcmc2mchintmcMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"Ecc_1BitThreshPs1\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Ecc_1BitThreshPs1.Size()\n\n\tif fldName == \"Ecc_1BitThreshPs0\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Ecc_1BitThreshPs0.Size()\n\n\treturn offset\n}", "func (mtr *Dppdpp1intspareMetrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"Spare_0\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_0.Size()\n\n\tif fldName == \"Spare_1\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_1.Size()\n\n\tif fldName == \"Spare_2\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_2.Size()\n\n\tif fldName == \"Spare_3\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_3.Size()\n\n\tif fldName == \"Spare_4\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_4.Size()\n\n\tif fldName == \"Spare_5\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_5.Size()\n\n\tif fldName == \"Spare_6\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_6.Size()\n\n\tif fldName == \"Spare_7\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_7.Size()\n\n\tif fldName == \"Spare_8\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_8.Size()\n\n\tif fldName == \"Spare_9\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_9.Size()\n\n\tif fldName == \"Spare_10\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_10.Size()\n\n\tif fldName == \"Spare_11\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_11.Size()\n\n\tif fldName == \"Spare_12\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_12.Size()\n\n\tif fldName == \"Spare_13\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_13.Size()\n\n\tif fldName == \"Spare_14\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_14.Size()\n\n\tif fldName == \"Spare_15\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_15.Size()\n\n\tif fldName == \"Spare_16\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_16.Size()\n\n\tif fldName == \"Spare_17\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_17.Size()\n\n\tif fldName == \"Spare_18\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_18.Size()\n\n\tif fldName == \"Spare_19\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_19.Size()\n\n\tif fldName == \"Spare_20\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_20.Size()\n\n\tif fldName == \"Spare_21\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_21.Size()\n\n\tif fldName == \"Spare_22\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_22.Size()\n\n\tif fldName == \"Spare_23\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_23.Size()\n\n\tif fldName == \"Spare_24\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_24.Size()\n\n\tif fldName == \"Spare_25\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_25.Size()\n\n\tif fldName == \"Spare_26\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_26.Size()\n\n\tif fldName == \"Spare_27\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_27.Size()\n\n\tif fldName == \"Spare_28\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_28.Size()\n\n\tif fldName == \"Spare_29\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_29.Size()\n\n\tif fldName == \"Spare_30\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_30.Size()\n\n\tif fldName == \"Spare_31\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Spare_31.Size()\n\n\treturn offset\n}", "func TestPrewriteMultiple4A(t *testing.T) {\n}", "func (mtr *Msmsintprp2Metrics) getOffset(fldName string) int {\n\tvar offset int\n\n\tif fldName == \"Read\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Read.Size()\n\n\tif fldName == \"Security\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Security.Size()\n\n\tif fldName == \"Decode\" {\n\t\treturn offset\n\t}\n\toffset += mtr.Decode.Size()\n\n\treturn offset\n}", "func All( w http.ResponseWriter, r *http.Request ,basic_name string ,copy_file string ,new_file string )( err error) {\n\n// IN w   : response-writer\n// IN r   : request-parameter\n// IN basic_name : d.s. name of basic\n// IN copy_file : d.s. name which is cpied\n// IN new_file : new d.s. name\n\n// OUT err : error inf.\n\n// fmt.Fprintf( w, \"copy3.all start \\n\" )\n// fmt.Fprintf( w, \"copy3.all basic_name %v\\n\" ,basic_name)\n\n c := appengine.NewContext(r)\n\n\tq := datastore.NewQuery(copy_file)\n\n\tcount, err := q.Count(c)\n\tif err != nil {\n//\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn err\n\n\t}\n // allocate work area for records\n ds_data := make([]type2.Deliver, 0, count)\n\n\tif _, err := q.GetAll(c, &ds_data); err != nil { // get d.s. inf.\n\n//\t http.Error(w, err.Error(), http.StatusInternalServerError)\n\t return err\n\n\t} else{\n for _, ds_dataw := range ds_data { // put copy inf. for d.s.\n\n\t if _, err := datastore.Put(c, datastore.NewIncompleteKey(c, new_file, nil), &ds_dataw); err != nil {\n\n//\t\t http.Error(w,err.Error(), http.StatusInternalServerError)\n\t\t return err\n\n\t }\n\n\t }\n\t}\n\n//\tfmt.Fprintf( w, \"copy3.all normal end \\n\" )\n\n return nil\n}", "func checkInstList(stringarray []string, outputArray []string, curPosition int, length int,sentence string){\n\t//declaration of variables for the counter control that follows \n\t\tvar buffer bytes.Buffer\n\t\tvar count int=0\n\t\tvar amtShape int=0\n\t\tvar tempCount int = 2\n\t\t//Check to see the amount of semi colon in the sentence and by that we can evaluate how many shapes\n\t\tfor _, r := range sentence {\n if string(r) == \";\"{\n\t\t\tamtShape++}\t\t\n\t\t}\n\t\t//static assignment for the 1st part of the inst\n\t\toutputArray[1]=\"<inst>\"\n\t\n\t\t//if there is no one inst then we can processed in making a format for one inst\n\t\tif amtShape == 0 {\n\t\t\toutputArray[1]=\"<inst>\"\n\t\t\tfmt.Println(strings.Trim(fmt.Sprint(outputArray), \"[] \")+\" finish\")\n\t\t\tcheckInstructions(outputArray[0:], stringarray[0:],curPosition,sentence)\n\t\t//If there is more that two can we can add <inst_list>\n\t\t// buffer doest that but as a single string.\n\t\t}else{ //\n\t\tfor ; amtShape >= 0; amtShape--{\n\t\t\t\t\tbuffer.WriteString(\"<inst_list>\")\n\t\t\t\t\tnewOutput:=buffer.String()\t\t\t\t\t\n\t\t\t\t\toutputArray[tempCount]=(newOutput)\n\t\t\t\t}\t\t\n\t\t//Get the value of how man y inst_list was place by using a token of \"_\"\n\t\tfor _, r := range outputArray[2] {\n\t\t\t\t\t\tif string(r) == \"_\"{\n\t\t\t\t\t\t\tcount++\t}\n\t\t\t\t\t}\n\t\tcount--\t\n\t\t//Proceed with placing it in its respective Array location \t\t\n\t\tfor ; count >= 1; {\t\n\t\t\tif outputArray[tempCount] == \"<inst>\"{\n\t\t\t\tcount--\n\t\t\t\t\t}else{outputArray[tempCount]= \"<inst_list>\"\n\t\t\t\t\tcount--\n\t\t\t\ttempCount++\n\t\t\t\t}\n\t\t\t\t\n\t\t}\n\t\t//Print the formatted derivation\n\t\tfmt.Println(strings.Trim(fmt.Sprint(outputArray), \"[] \")+\" finish\")\n\t\ttempCount--\n\t\t//Change the format of the derivation\n\t\tif outputArray[tempCount] == \"<inst_list>\"{\n\t\t\toutputArray[tempCount] = \"<inst>\"\n\t\t}\n\t\t//Print the formatted derivation and Proceed with calling the function to check each inst_list\n\t\tfmt.Println(strings.Trim(fmt.Sprint(outputArray), \"[] \")+\" finish\")\t\n\t\tcheckInstructions(outputArray[0:], stringarray[0:],curPosition,sentence)\n\t\t}\n\t\t\n\t\t}" ]
[ "0.5175247", "0.4751098", "0.47478223", "0.47188663", "0.46758133", "0.46617737", "0.45669132", "0.45663005", "0.4564074", "0.45483142", "0.45308706", "0.45184872", "0.450741", "0.45017928", "0.44985196", "0.44984955", "0.44978857", "0.44974223", "0.4484171", "0.44783118", "0.44674402", "0.44670317", "0.44588712", "0.44583097", "0.44583097", "0.44583097", "0.44583097", "0.4449753", "0.44429353", "0.44424537", "0.44422397", "0.44413793", "0.44412974", "0.44412974", "0.44412974", "0.44384363", "0.44358075", "0.44357595", "0.44330567", "0.44282466", "0.4426932", "0.44268033", "0.44237587", "0.44224334", "0.44220394", "0.4421066", "0.44203433", "0.4417567", "0.44156122", "0.44133073", "0.44029635", "0.44019842", "0.4401051", "0.43980503", "0.43920556", "0.4389929", "0.43867794", "0.43862683", "0.43843755", "0.4379152", "0.43758532", "0.43748495", "0.43736225", "0.43725154", "0.4371369", "0.43713644", "0.43699622", "0.4365456", "0.43622264", "0.43607378", "0.43506905", "0.43449685", "0.43434596", "0.43383086", "0.43372044", "0.43329886", "0.433193", "0.43304133", "0.43281412", "0.43241173", "0.43225074", "0.4320729", "0.43202066", "0.43192208", "0.4318242", "0.4316968", "0.4315887", "0.4313601", "0.4313201", "0.43111324", "0.43105358", "0.43072286", "0.43065637", "0.4303898", "0.43038484", "0.43033594", "0.43027574", "0.4302642", "0.4301456", "0.4298848", "0.42978513" ]
0.0
-1
Kinds is a list of known pkger kinds.
func Kinds() []Kind { var out []Kind for k := range kinds { out = append(out, k) } return out }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (*bzlLibraryLang) Kinds() map[string]rule.KindInfo {\n\treturn kinds\n}", "func (es ExternalServices) Kinds() (kinds []string) {\n\tset := make(map[string]bool, len(es))\n\tfor _, e := range es {\n\t\tif !set[e.Kind] {\n\t\t\tkinds = append(kinds, e.Kind)\n\t\t\tset[e.Kind] = true\n\t\t}\n\t}\n\treturn kinds\n}", "func getApplicableKindsForPolicy(p *v1alpha1.Policy) []string {\n\tkindsMap := map[string]interface{}{}\n\tkinds := []string{}\n\t// iterate over the rules an identify all kinds\n\tfor _, rule := range p.Spec.Rules {\n\t\tfor _, k := range rule.ResourceDescription.Kinds {\n\t\t\tkindsMap[k] = nil\n\t\t}\n\t}\n\n\t// get the kinds\n\tfor k := range kindsMap {\n\t\tkinds = append(kinds, k)\n\t}\n\treturn kinds\n}", "func (c *Chart) UnknownKinds(known []string) []string {\n\tlookup := make(map[string]bool, len(known))\n\tfor _, k := range known {\n\t\tlookup[k] = true\n\t}\n\n\tu := []string{}\n\tfor n := range c.Kind {\n\t\tif _, ok := lookup[n]; !ok {\n\t\t\tu = append(u, n)\n\t\t}\n\t}\n\n\treturn u\n}", "func (*visibilityExtension) Kinds() map[string]rule.KindInfo {\n\treturn map[string]rule.KindInfo{\n\t\t\"package\": {\n\t\t\tMatchAny: true,\n\t\t\tMergeableAttrs: map[string]bool{\n\t\t\t\t\"default_visibility\": true,\n\t\t\t},\n\t\t},\n\t}\n}", "func (rs Repos) Kinds() (kinds []string) {\n\tset := map[string]bool{}\n\tfor _, r := range rs {\n\t\tkind := strings.ToUpper(r.ExternalRepo.ServiceType)\n\t\tif !set[kind] {\n\t\t\tkinds = append(kinds, kind)\n\t\t\tset[kind] = true\n\t\t}\n\t}\n\treturn kinds\n}", "func (_m *Resolver) ListUsageKinds(ctx context.Context, first *int, offset *int) ([]*gqlschema.UsageKind, error) {\n\tvar r0 []*gqlschema.UsageKind\n\tvar r1 error\n\tr1 = _m.err\n\n\treturn r0, r1\n}", "func KindStrings() []string {\n\tstrs := make([]string, len(_KindNames))\n\tcopy(strs, _KindNames)\n\treturn strs\n}", "func KindValues() []Kind {\n\treturn _KindValues\n}", "func CriticalKinds() []string {\n\tck := make([]string, 0, 6)\n\tck = append(ck, constant.RulesKind)\n\tck = append(ck, constant.AttributeManifestKind)\n\tck = append(ck, constant.AdapterKind)\n\tck = append(ck, constant.TemplateKind)\n\tck = append(ck, constant.InstanceKind)\n\tck = append(ck, constant.HandlerKind)\n\treturn ck\n}", "func (o *ObjRelMapper) GetOrderedKinds() []string {\n\treturn o.orderedKind\n}", "func emptyKotsKinds() KotsKinds {\n\tkotsKinds := KotsKinds{\n\t\tInstallation: kotsv1beta1.Installation{\n\t\t\tTypeMeta: metav1.TypeMeta{\n\t\t\t\tAPIVersion: \"kots.io/v1beta1\",\n\t\t\t\tKind: \"Installation\",\n\t\t\t},\n\t\t},\n\t\tKotsApplication: kotsv1beta1.Application{\n\t\t\tTypeMeta: metav1.TypeMeta{\n\t\t\t\tAPIVersion: \"kots.io/v1beta1\",\n\t\t\t\tKind: \"Application\",\n\t\t\t},\n\t\t},\n\t}\n\n\treturn kotsKinds\n}", "func (m *RegeneratingDiscoveryRESTMapper) KindsFor(resource schema.GroupVersionResource) ([]schema.GroupVersionKind, error) {\n\tm.mu.RLock()\n\tdefer m.mu.RUnlock()\n\n\treturn m.delegate.KindsFor(resource)\n\n}", "func (e *ExternalServiceStore) DistinctKinds(ctx context.Context) ([]string, error) {\n\te.ensureStore()\n\n\tq := sqlf.Sprintf(`\nSELECT ARRAY_AGG(DISTINCT(kind)::TEXT)\nFROM external_services\nWHERE deleted_at IS NULL\n`)\n\n\tvar kinds []string\n\terr := e.QueryRow(ctx, q).Scan(pq.Array(&kinds))\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn []string{}, nil\n\t\t}\n\t\treturn nil, err\n\t}\n\n\treturn kinds, nil\n}", "func parseChannelModeKinds(address, always, onset, none string) (\n\tkinds map[rune]int) {\n\tsize := len(always) + len(onset) + len(address)\n\tkinds = make(map[rune]int, size)\n\n\tfor _, mode := range always {\n\t\tkinds[mode] = ARGS_ALWAYS\n\t}\n\tfor _, mode := range onset {\n\t\tkinds[mode] = ARGS_ONSET\n\t}\n\tfor _, mode := range address {\n\t\tkinds[mode] = ARGS_ADDRESS\n\t}\n\tfor _, mode := range none {\n\t\tkinds[mode] = ARGS_NONE\n\t}\n\n\treturn\n}", "func newModeKinds(prefix, chanModes string) (*modeKinds, error) {\n\tuserPrefixes, err := parsePrefixString(prefix)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tchannelModes, err := parseChannelModeKindsCSV(chanModes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &modeKinds{\n\t\tuserPrefixes: userPrefixes,\n\t\tchannelModes: channelModes,\n\t}, nil\n}", "func PossibleKindValues() []Kind {\n\treturn []Kind{BlobStorage, BlockBlobStorage, FileStorage, Storage, StorageV2}\n}", "func parseChannelModeKinds(address, always, onset, none string) (\n\tkinds map[rune]int) {\n\tsize := len(always) + len(onset) + len(address)\n\tkinds = make(map[rune]int, size)\n\n\tfor _, mode := range always {\n\t\tkinds[mode] = ARGS_ALWAYS\n\t}\n\tfor _, mode := range onset {\n\t\tkinds[mode] = ARGS_ONSET\n\t}\n\tfor _, mode := range address {\n\t\tkinds[mode] = ARGS_ADDRESS\n\t}\n\tfor _, mode := range none {\n\t\tkinds[mode] = ARGS_NONE\n\t}\n\n\treturn kinds\n}", "func PossibleKindValues() []Kind {\n\treturn []Kind{MicrosoftCompute}\n}", "func (k *Key) Kind() string { return k.toks[len(k.toks)-1].Kind }", "func (ck *clusterKinds) getAll() map[string]bool {\n\treturn ck.isNamespaced\n}", "func PossibleKindValues() []Kind {\n\treturn []Kind{KindApplication, KindApplications, KindApplicationTypeApplications, KindDeltaNodesCheck, KindDeployedApplication, KindDeployedApplications, KindDeployedServicePackage, KindDeployedServicePackages, KindEvent, KindHealthEvaluation, KindNode, KindNodes, KindPartition, KindPartitions, KindReplica, KindReplicas, KindService, KindServices, KindSystemApplication, KindUpgradeDomainDeltaNodesCheck, KindUpgradeDomainNodes}\n}", "func getLabelToMatchForKind(kind string) []string {\n\tswitch kind {\n\tcase \"apiservice\": // API Services are not namespaced\n\t\treturn []string{\"apiservice\"}\n\tcase \"customresourcedefinition\": // CRD are not namespaced\n\t\treturn []string{\"customresourcedefinition\"}\n\tcase \"job\": // job metrics use specific label\n\t\treturn []string{\"job_name\", \"namespace\"}\n\tcase \"node\": // persistent nodes are not namespaced\n\t\treturn []string{\"node\"}\n\tcase \"persistentvolume\": // persistent volumes are not namespaced\n\t\treturn []string{\"persistentvolume\"}\n\tdefault:\n\t\treturn []string{kind, \"namespace\"}\n\t}\n}", "func GetResourceMarshalerKinds() []string {\n\tmarshalerMutex.Lock()\n\tdefer marshalerMutex.Unlock()\n\tkinds := make([]string, 0, len(resourceMarshalers))\n\tfor kind := range resourceMarshalers {\n\t\tkinds = append(kinds, kind)\n\t}\n\treturn kinds\n}", "func (LogSigner_KeyTypes) EnumDescriptor() ([]byte, []int) {\n\treturn file_api_proto_rawDescGZIP(), []int{13, 0}\n}", "func KindFromSides(a, b, c float64) (k Kind) {\n\n\tif !isValid(a, b, c) {\n\t\treturn NaT\n\t}\n\n\tif a == b {\n\t\tk++\n\t}\n\tif a == c {\n\t\tk++\n\t}\n\tif b == c {\n\t\tk++\n\t}\n\treturn\n}", "func KindMap(adapterInfo map[string]*adapter.Info, templateInfo map[string]*template.Info) map[string]proto.Message {\n\tkindMap := make(map[string]proto.Message)\n\t// typed instances\n\tfor kind, info := range templateInfo {\n\t\tkindMap[kind] = info.CtrCfg\n\t\tlog.Debugf(\"template kind: %s, %v\", kind, info.CtrCfg)\n\t}\n\t// typed handlers\n\tfor kind, info := range adapterInfo {\n\t\tkindMap[kind] = info.DefaultConfig\n\t\tlog.Debugf(\"adapter kind: %s, %v\", kind, info.DefaultConfig)\n\t}\n\tkindMap[constant.RulesKind] = &configpb.Rule{}\n\tlog.Debugf(\"Rules config kind: %s\", constant.RulesKind)\n\tkindMap[constant.AttributeManifestKind] = &configpb.AttributeManifest{}\n\tlog.Debugf(\"Attribute manifest kind: %s\", constant.AttributeManifestKind)\n\tkindMap[constant.AdapterKind] = &v1beta1.Info{}\n\tlog.Debugf(\"Adapter info config kind: %s\", constant.AdapterKind)\n\tkindMap[constant.TemplateKind] = &v1beta1.Template{}\n\tlog.Debugf(\"Template config kind: %s\", constant.TemplateKind)\n\tkindMap[constant.InstanceKind] = &configpb.Instance{}\n\tlog.Debugf(\"Instance config kind: %s\", constant.InstanceKind)\n\tkindMap[constant.HandlerKind] = &configpb.Handler{}\n\tlog.Debugf(\"Handler config kind: %s\", constant.HandlerKind)\n\treturn kindMap\n}", "func PossibleKindValues() []Kind {\n\treturn []Kind{KindMicrosoftCompute}\n}", "func (d *Descriptor) Kind() Kind {\n\treturn d.kind\n}", "func decodeClusterV1Kinds(decoder *yaml.YAMLOrJSONDecoder, kind string) ([][]byte, error) {\n\touts := [][]byte{}\n\n\tfor {\n\t\tvar out unstructured.Unstructured\n\n\t\tif err := decoder.Decode(&out); err == io.EOF {\n\t\t\tbreak\n\t\t} else if err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif out.GetKind() == kind && out.GetAPIVersion() == clusterv1.SchemeGroupVersion.String() {\n\t\t\tmarshaled, err := out.MarshalJSON()\n\t\t\tif err != nil {\n\t\t\t\treturn outs, err\n\t\t\t}\n\t\t\touts = append(outs, marshaled)\n\t\t}\n\t}\n\n\treturn outs, nil\n}", "func ListKind(kind string, namespace string, args ...string) (List, error) {\n\tget := []string{\"get\", kind, \"-n\", namespace, \"-o\", \"json\"}\n\tout, err := kubectl(append(get, args...)...)\n\tif err != nil {\n\t\treturn List{}, err\n\t}\n\tvar unmarshalled unmarshalledList\n\terr = json.Unmarshal(out, &unmarshalled)\n\tif err != nil {\n\t\treturn List{}, err\n\t}\n\tvar l List\n\tfor _, resource := range unmarshalled.Items {\n\t\tmarshalled, marshalErr := json.Marshal(resource)\n\t\tif marshalErr != nil {\n\t\t\treturn List{}, marshalErr\n\t\t}\n\t\tl.Items = append(l.Items, marshalled)\n\t}\n\treturn l, nil\n}", "func NewChannelModeKinds(\n\taddress, always, onset, none string) *ChannelModeKinds {\n\n\treturn &ChannelModeKinds{\n\t\tparseChannelModeKinds(address, always, onset, none),\n\t}\n}", "func (k Kind) String() string {\n\tif k == 0 {\n\t\treturn \"any\"\n\t}\n\tvar ret []string\n\tfor i := uint(0); i < 64; i++ {\n\t\tif k&(1<<i) != 0 {\n\t\t\tif name, ok := kindNames[1<<i]; ok {\n\t\t\t\tret = append(ret, name)\n\t\t\t}\n\t\t}\n\t}\n\tif len(ret) == 0 {\n\t\treturn \"unknown\"\n\t}\n\treturn strings.Join(ret, \" or \")\n}", "func (k Kind) String() string {\n\treturn kindNames[k]\n}", "func PossibleKindValues() []Kind {\n\treturn []Kind{KindFhir, KindFhirR4, KindFhirStu3}\n}", "func PossibleKindValues() []Kind {\n\treturn []Kind{KindDatabase, KindReadOnlyFollowing, KindReadWrite}\n}", "func (*KindTag) Descriptor() ([]byte, []int) {\n\treturn file_kindTagService_proto_rawDescGZIP(), []int{1}\n}", "func (o EndpointsListOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *EndpointsList) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (pkgs Pkgs) Patterns() []string {\n\treturn pkgs.pats\n}", "func (o *objectSet) kind(kind string) *objectSet {\n\tret := &objectSet{}\n\tfor k, v := range o.objMap {\n\t\tobjKind, _, _ := object.FromHash(k)\n\t\tif objKind == kind {\n\t\t\tret.append(v)\n\t\t}\n\t}\n\treturn ret\n}", "func convKindStr(v interface{}) string { return reflectKindStr(convKind(v)) }", "func getKindFlows(m *api.Message, s *sessionManager) (Flow, bool) {\n\tk := m.GetKind()\n\tvalue, ok := s.telego.kindFlows[k]\n\treturn value, ok\n}", "func (*KindTagWhere) Descriptor() ([]byte, []int) {\n\treturn file_kindTagService_proto_rawDescGZIP(), []int{0}\n}", "func (*UrlCategorys) Descriptor() ([]byte, []int) {\n\treturn file_src_nap_nap_proto_rawDescGZIP(), []int{23}\n}", "func GetGVKsFromAddToScheme(addToSchemeFunc func(*runtime.Scheme) error) ([]schema.GroupVersionKind, error) {\n\ts := runtime.NewScheme()\n\terr := addToSchemeFunc(s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tschemeAllKnownTypes := s.AllKnownTypes()\n\townGVKs := []schema.GroupVersionKind{}\n\tfor gvk := range schemeAllKnownTypes {\n\t\tif !isKubeMetaKind(gvk.Kind) {\n\t\t\townGVKs = append(ownGVKs, gvk)\n\t\t}\n\t}\n\n\treturn ownGVKs, nil\n}", "func (p Primitive) Kind() Kind { return Kind(p) }", "func (s *protoTsLibrary) KindInfo() rule.KindInfo {\n\treturn rule.KindInfo{\n\t\tMergeableAttrs: map[string]bool{\n\t\t\t\"srcs\": true,\n\t\t\t\"tsc\": true,\n\t\t\t\"args\": true,\n\t\t\t\"data\": true,\n\t\t\t\"tsconfig\": true,\n\t\t\t\"out_dir\": true,\n\t\t},\n\t\tResolveAttrs: map[string]bool{\n\t\t\t\"deps\": true,\n\t\t},\n\t}\n\n}", "func (o ValidatingAdmissionPolicyBindingListTypeOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ValidatingAdmissionPolicyBindingListType) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (l *PendingDeleteClusterList) Kind() string {\n\tif l == nil {\n\t\treturn PendingDeleteClusterListNilKind\n\t}\n\tif l.link {\n\t\treturn PendingDeleteClusterListLinkKind\n\t}\n\treturn PendingDeleteClusterListKind\n}", "func (o PriorityLevelConfigurationPatchOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *PriorityLevelConfigurationPatch) pulumi.StringPtrOutput { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (o PersistentVolumeListOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *PersistentVolumeList) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (p *PCache) List(kind string) map[string]interface{} {\n\tp.RLock()\n\tkindMap := p.kinds[kind]\n\tp.RUnlock()\n\titems := map[string]interface{}{}\n\n\tif kindMap != nil {\n\t\tkindMap.Lock()\n\t\tfor key, entry := range kindMap.entries {\n\t\t\titems[key] = entry\n\t\t}\n\t\tkindMap.Unlock()\n\t}\n\n\treturn items\n}", "func (s *GenericStorage) List(gvk schema.GroupVersionKind) (result []runtime.Object, walkerr error) {\n\twalkerr = s.walkKind(gvk, func(content []byte) error {\n\t\tobj, err := s.decode(content, gvk)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tresult = append(result, obj)\n\t\treturn nil\n\t})\n\treturn\n}", "func (*ControllerPlanningSpecK8S) Descriptor() ([]byte, []int) {\n\treturn file_alameda_api_v1alpha1_datahub_plannings_types_proto_rawDescGZIP(), []int{1}\n}", "func (m *modeKinds) ToProto() *api.ModeKinds {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\tvar proto api.ModeKinds\n\n\tm.RLock()\n\tdefer m.RUnlock()\n\n\tif m.userPrefixes != nil {\n\t\tproto.UserPrefixes = make([]*api.ModeKinds_UserPrefix, len(m.userPrefixes))\n\t\tfor i, prefix := range m.userPrefixes {\n\t\t\tproto.UserPrefixes[i] = &api.ModeKinds_UserPrefix{\n\t\t\t\tSymbol: string(prefix[0]),\n\t\t\t\tChar: string(prefix[1]),\n\t\t\t}\n\t\t}\n\t}\n\tif m.channelModes != nil {\n\t\tproto.ChannelModes = make(map[string]int32, len(m.channelModes))\n\t\tfor k, v := range m.channelModes {\n\t\t\tproto.ChannelModes[string(k)] = int32(v)\n\t\t}\n\t}\n\n\treturn &proto\n}", "func (m *InformationProtection) GetSensitivityLabels()([]SensitivityLabelable) {\n val, err := m.GetBackingStore().Get(\"sensitivityLabels\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]SensitivityLabelable)\n }\n return nil\n}", "func (d Document) Kind() string { return d.kind }", "func (PublicKeyDetails) EnumDescriptor() ([]byte, []int) {\n\treturn file_sigstore_common_proto_rawDescGZIP(), []int{1}\n}", "func (d *DefaultLabelStrategy) LabelNames(p *protocol.Protocol) []string {\n\tres := []string{\"name\", \"proto\", \"ip_version\", \"import_filter\", \"export_filter\"}\n\tif d.descriptionLabels && p.Description != \"\" {\n\t\tres = append(res, labelKeysFromDescription(p.Description)...)\n\t}\n\n\treturn res\n}", "func (c *ConsulServiceRegistry) Kind() string {\n\treturn Kind\n}", "func PossibleCampaignKindsValues() []CampaignKinds {\n\treturn []CampaignKinds{Announcements, DataPushes, NativePushes, Polls}\n}", "func (PkixPublicKey_SignatureAlgorithm) EnumDescriptor() ([]byte, []int) {\n\treturn file_google_cloud_binaryauthorization_v1_resources_proto_rawDescGZIP(), []int{5, 0}\n}", "func (*DocumentSymbolClientCapabilities_SymbolKind) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{148, 0}\n}", "func (o ParamKindPatchOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ParamKindPatch) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (d *Discrete) Kind() Kind {\n\treturn d.DiscreteKind\n}", "func (*WithWellKnownTypes) Descriptor() ([]byte, []int) {\n\treturn file_testing_proto_rawDescGZIP(), []int{1}\n}", "func sliceKind(val reflect.Value) reflect.Kind {\n\tswitch val.Type().Elem().Kind() {\n\tcase reflect.Ptr:\n\t\treturn ptrKind(val.Type().Elem())\n\t}\n\treturn val.Type().Elem().Kind()\n}", "func (*TEKSignatureList) Descriptor() ([]byte, []int) {\n\treturn file_proto_covidshieldv1_proto_rawDescGZIP(), []int{8}\n}", "func (*ListSyntheticsRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_v1_synthetics_proto_rawDescGZIP(), []int{9}\n}", "func (*WorkspaceSymbolClientCapabilities_SymbolKind) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{80, 0}\n}", "func (t *Topic) GetGroupVersionKind() schema.GroupVersionKind {\n\treturn SchemeGroupVersion.WithKind(\"Topic\")\n}", "func buildKindKey(object runtime.Object) (string, error) {\n\tgvk, err := apiutil.GVKForObject(object, scheme.Scheme)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn fmt.Sprintf(\"%s/%s/%s\", gvk.Group, gvk.Version, gvk.Kind), nil\n}", "func (d *BackupDescriptor) List() []string {\n\tlst := make([]string, len(d.Classes))\n\tfor i, cls := range d.Classes {\n\t\tlst[i] = cls.Name\n\t}\n\treturn lst\n}", "func Kind(kind string) unversioned.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func ProtocolList() map[int]string {\n\n\tprotocols := make(map[int]string)\n\n\tprotocols[0] = \"HOPOPT\"\n\tprotocols[1] = \"ICMP\"\n\tprotocols[2] = \"IGMP\"\n\tprotocols[3] = \"GGP\"\n\tprotocols[4] = \"IP-in-IP\"\n\tprotocols[5] = \"ST\"\n\tprotocols[6] = \"TCP\"\n\tprotocols[7] = \"CBT\"\n\tprotocols[8] = \"EGP\"\n\tprotocols[9] = \"IGP\"\n\tprotocols[10] = \"BBN-RCC-MON\"\n\tprotocols[11] = \"NVP-II\"\n\tprotocols[12] = \"PUP\"\n\tprotocols[13] = \"ARGUS\"\n\tprotocols[14] = \"EMCON\"\n\tprotocols[15] = \"XNET\"\n\tprotocols[16] = \"CHAOS\"\n\tprotocols[17] = \"UDP\"\n\tprotocols[18] = \"MUX\"\n\tprotocols[19] = \"DCN-MEAS\"\n\tprotocols[20] = \"HMP\"\n\tprotocols[21] = \"PRM\"\n\tprotocols[22] = \"XNS-IDP\"\n\tprotocols[23] = \"TRUNK-1\"\n\tprotocols[24] = \"TRUNK-2\"\n\tprotocols[25] = \"LEAF-1\"\n\tprotocols[26] = \"LEAF-2\"\n\tprotocols[27] = \"RDP\"\n\tprotocols[28] = \"IRTP\"\n\tprotocols[29] = \"ISO-TP4\"\n\tprotocols[30] = \"NETBLT\"\n\tprotocols[31] = \"MFE-NSP\"\n\tprotocols[32] = \"MERIT-INP\"\n\tprotocols[33] = \"DCCP\"\n\tprotocols[34] = \"3PC\"\n\tprotocols[35] = \"IDPR\"\n\tprotocols[36] = \"XTP\"\n\tprotocols[37] = \"DDP\"\n\tprotocols[38] = \"IDPR-CMTP\"\n\tprotocols[39] = \"TP++\"\n\tprotocols[40] = \"IL\"\n\tprotocols[41] = \"IPv6\"\n\tprotocols[42] = \"SDRP\"\n\tprotocols[43] = \"IPv6-Route\"\n\tprotocols[44] = \"IPv6-Frag\"\n\tprotocols[45] = \"IDRP\"\n\tprotocols[46] = \"RSVP\"\n\tprotocols[47] = \"GRE\"\n\tprotocols[48] = \"DSR\"\n\tprotocols[49] = \"BNA\"\n\tprotocols[50] = \"ESP\"\n\tprotocols[51] = \"AH\"\n\tprotocols[52] = \"I-NLSP\"\n\tprotocols[53] = \"SWIPE\"\n\tprotocols[54] = \"NARP\"\n\tprotocols[55] = \"MOBILE\"\n\tprotocols[56] = \"TLSP\"\n\tprotocols[57] = \"SKIP\"\n\tprotocols[58] = \"IPv6-ICMP\"\n\tprotocols[59] = \"IPv6-NoNxt\"\n\tprotocols[60] = \"IPv6-Opts\"\n\tprotocols[61] = \"\"\n\tprotocols[62] = \"CFTP\"\n\tprotocols[63] = \"\"\n\tprotocols[64] = \"SAT-EXPAK\"\n\tprotocols[65] = \"KRYPTOLAN\"\n\tprotocols[66] = \"RVD\"\n\tprotocols[67] = \"IPPC\"\n\tprotocols[68] = \"\"\n\tprotocols[69] = \"SAT-MON\"\n\tprotocols[70] = \"VISA\"\n\tprotocols[71] = \"IPCU\"\n\tprotocols[72] = \"CPNX\"\n\tprotocols[73] = \"CPHB\"\n\tprotocols[74] = \"WSN\"\n\tprotocols[75] = \"PVP\"\n\tprotocols[76] = \"BR-SAT-MON\"\n\tprotocols[77] = \"SUN-ND\"\n\tprotocols[78] = \"WB-MON\"\n\tprotocols[79] = \"WB-EXPAK\"\n\tprotocols[80] = \"ISO-IP\"\n\tprotocols[81] = \"VMTP\"\n\tprotocols[82] = \"SECURE-VMTP\"\n\tprotocols[83] = \"VINES\"\n\tprotocols[84] = \"TTP\"\n\tprotocols[84] = \"IPTM\"\n\tprotocols[85] = \"NSFNET-IGP\"\n\tprotocols[86] = \"DGP\"\n\tprotocols[87] = \"TCF\"\n\tprotocols[88] = \"EIGRP\"\n\tprotocols[89] = \"OSPF\"\n\tprotocols[90] = \"Sprite-RPC\"\n\tprotocols[91] = \"LARP\"\n\tprotocols[92] = \"MTP\"\n\tprotocols[93] = \"AX.25\"\n\tprotocols[94] = \"OS\"\n\tprotocols[95] = \"MICP\"\n\tprotocols[96] = \"SCC-SP\"\n\tprotocols[97] = \"ETHERIP\"\n\tprotocols[98] = \"ENCAP\"\n\tprotocols[100] = \"GMTP\"\n\tprotocols[101] = \"IFMP\"\n\tprotocols[102] = \"PNNI\"\n\tprotocols[103] = \"PIM\"\n\tprotocols[104] = \"ARIS\"\n\tprotocols[105] = \"SCPS\"\n\tprotocols[106] = \"QNX\"\n\tprotocols[107] = \"A/N\"\n\tprotocols[108] = \"IPComp\"\n\tprotocols[109] = \"SNP\"\n\tprotocols[110] = \"Compaq-Peer\"\n\tprotocols[111] = \"IPX-in-IP\"\n\tprotocols[112] = \"VRRP\"\n\tprotocols[113] = \"PGM\"\n\tprotocols[115] = \"L2TP\"\n\tprotocols[116] = \"DDX\"\n\tprotocols[117] = \"IATP\"\n\tprotocols[118] = \"STP\"\n\tprotocols[119] = \"SRP\"\n\tprotocols[120] = \"UTI\"\n\tprotocols[121] = \"SMP\"\n\tprotocols[122] = \"SM\"\n\tprotocols[123] = \"PTP\"\n\tprotocols[124] = \"IS-IS over IPv4\"\n\tprotocols[125] = \"FIRE\"\n\tprotocols[126] = \"CRTP\"\n\tprotocols[127] = \"CRUDP\"\n\tprotocols[128] = \"SSCOPMCE\"\n\tprotocols[129] = \"IPLT\"\n\tprotocols[130] = \"SPS\"\n\tprotocols[131] = \"PIPE\"\n\tprotocols[132] = \"SCTP\"\n\tprotocols[133] = \"FC\"\n\tprotocols[134] = \"RSVP-E2E-IGNORE\"\n\tprotocols[135] = \"Mobility Header\"\n\tprotocols[136] = \"UDPLite\"\n\tprotocols[137] = \"MPLS-in-IP\"\n\tprotocols[138] = \"manet\"\n\tprotocols[139] = \"HIP\"\n\tprotocols[140] = \"Shim6\"\n\tprotocols[141] = \"WESP\"\n\tprotocols[142] = \"ROHC\"\n\n\treturn protocols\n}", "func (SignMode) EnumDescriptor() ([]byte, []int) {\n\treturn file_x_authn_v1alpha1_signing_proto_rawDescGZIP(), []int{0}\n}", "func wrapDescriptors(file *descriptor.FileDescriptorProto) []*Descriptor {\n\tsl := make([]*Descriptor, 0, len(file.MessageType)+10)\n\tfor i, desc := range file.MessageType {\n\t\tsl = wrapThisDescriptor(sl, desc, nil, file, i)\n\t}\n\treturn sl\n}", "func (*AllSupportedTypes) Descriptor() ([]byte, []int) {\n\treturn file_testing_proto_rawDescGZIP(), []int{0}\n}", "func (s *GitLabBinding) GetGroupVersionKind() schema.GroupVersionKind {\n\treturn SchemeGroupVersion.WithKind(\"GitLabBinding\")\n}", "func (o KubernetesClusterSpecClassRefOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterSpecClassRef) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (c *Inventory) TypesToPurge() []schema.GroupVersionKind {\n\tcurrentObjects := c.inventoryData.ObjectsToDelete\n\tgroupVersionKindDict := make(map[schema.GroupVersionKind]struct{})\n\n\tfor _, currentObject := range currentObjects {\n\t\tgvk := currentObject.GetObjectKind().GroupVersionKind()\n\n\t\tif gvk.Kind == CustomResourceDefinition || gvk.Kind == Namespace {\n\t\t\tcontinue\n\t\t}\n\n\t\tif objMeta, err := meta.Accessor(currentObject); err == nil {\n\t\t\tc.log.V(1).Info(\"mark object for deletion\", \"gvk\", gvk.String(), \"namespace\", objMeta.GetNamespace(), \"name\", objMeta.GetName())\n\t\t\tgroupVersionKindDict[gvk] = struct{}{}\n\t\t}\n\t}\n\n\tgroupVersionKindList := make([]schema.GroupVersionKind, 0, len(groupVersionKindDict))\n\tfor k := range groupVersionKindDict {\n\t\tgroupVersionKindList = append(groupVersionKindList, k)\n\t}\n\n\treturn groupVersionKindList\n}", "func (*KindTagResponse) Descriptor() ([]byte, []int) {\n\treturn file_kindTagService_proto_rawDescGZIP(), []int{2}\n}", "func (o IndexesOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Indexes) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (o HorizontalPodAutoscalerListTypeOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v HorizontalPodAutoscalerListType) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (o HorizontalPodAutoscalerListTypeOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v HorizontalPodAutoscalerListType) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (VariantAnnotation_ClinicalSignificance) EnumDescriptor() ([]byte, []int) {\n\treturn file_google_genomics_v1_annotations_proto_rawDescGZIP(), []int{2, 2}\n}", "func (*HBox) Kind() *base.Kind {\n\treturn &hboxKind\n}", "func (o ValidatingAdmissionPolicyBindingPatchTypeOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ValidatingAdmissionPolicyBindingPatchType) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (af *filtBase) GetKindName() string {\n\treturn af.kind\n}", "func (*Interests) Descriptor() ([]byte, []int) {\n\treturn file_API_proto_rawDescGZIP(), []int{6}\n}", "func (o MachineInstanceSpecClassRefOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MachineInstanceSpecClassRef) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (c *KafkaTopicChannel) GetGroupVersionKind() schema.GroupVersionKind {\n\treturn SchemeGroupVersion.WithKind(\"KafkaTopicChannel\")\n}" ]
[ "0.67730343", "0.6455275", "0.6337542", "0.63176453", "0.62905663", "0.62271625", "0.61758685", "0.6062551", "0.5999965", "0.5919107", "0.5728519", "0.56822056", "0.54708886", "0.52446043", "0.5142412", "0.511807", "0.50216275", "0.49991712", "0.49917233", "0.4989127", "0.49874306", "0.4967197", "0.4965158", "0.49565208", "0.49277693", "0.48921114", "0.48765093", "0.4871589", "0.48590547", "0.48560187", "0.48469263", "0.48342714", "0.4831805", "0.4809343", "0.47975114", "0.4788564", "0.47818786", "0.4762655", "0.47460273", "0.4738821", "0.47198984", "0.47189656", "0.47145912", "0.4712526", "0.4708339", "0.47067136", "0.46914372", "0.46817866", "0.46777228", "0.46690577", "0.46570006", "0.4653332", "0.4644787", "0.46376744", "0.4635976", "0.46298307", "0.462459", "0.46238992", "0.4597053", "0.45856318", "0.45773393", "0.4573746", "0.45482773", "0.45421082", "0.45303246", "0.45186642", "0.45131168", "0.4510863", "0.4495788", "0.4490684", "0.44852534", "0.44825602", "0.4471211", "0.4459475", "0.44592708", "0.44592708", "0.44592708", "0.44592708", "0.44592708", "0.44592708", "0.44592708", "0.44592708", "0.44560587", "0.44554907", "0.44537887", "0.4441416", "0.44400004", "0.44370842", "0.44284856", "0.44276804", "0.44264126", "0.4425004", "0.4425004", "0.4424173", "0.44197005", "0.4419127", "0.44124264", "0.44094187", "0.4409205", "0.44088405" ]
0.70997757
0
String provides the kind in human readable form.
func (k Kind) String() string { if kinds[k] { return string(k) } if k == KindUnknown { return "unknown" } return string(k) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (k Kind) String() string {\n\tif k == 0 {\n\t\treturn \"any\"\n\t}\n\tvar ret []string\n\tfor i := uint(0); i < 64; i++ {\n\t\tif k&(1<<i) != 0 {\n\t\t\tif name, ok := kindNames[1<<i]; ok {\n\t\t\t\tret = append(ret, name)\n\t\t\t}\n\t\t}\n\t}\n\tif len(ret) == 0 {\n\t\treturn \"unknown\"\n\t}\n\treturn strings.Join(ret, \" or \")\n}", "func (k Kind) String() string {\n\treturn kindNames[k]\n}", "func (k Kind) String() string {\n\tswitch k {\n\tcase Invalid:\n\t\treturn \"Invalid\"\n\tcase Comment:\n\t\treturn \"Comment\"\n\tcase Key:\n\t\treturn \"Key\"\n\tcase Table:\n\t\treturn \"Table\"\n\tcase ArrayTable:\n\t\treturn \"ArrayTable\"\n\tcase KeyValue:\n\t\treturn \"KeyValue\"\n\tcase Array:\n\t\treturn \"Array\"\n\tcase InlineTable:\n\t\treturn \"InlineTable\"\n\tcase String:\n\t\treturn \"String\"\n\tcase Bool:\n\t\treturn \"Bool\"\n\tcase Float:\n\t\treturn \"Float\"\n\tcase Integer:\n\t\treturn \"Integer\"\n\tcase LocalDate:\n\t\treturn \"LocalDate\"\n\tcase LocalTime:\n\t\treturn \"LocalTime\"\n\tcase LocalDateTime:\n\t\treturn \"LocalDateTime\"\n\tcase DateTime:\n\t\treturn \"DateTime\"\n\t}\n\tpanic(fmt.Errorf(\"Kind.String() not implemented for '%d'\", k))\n}", "func (x Kind) String() string {\n\tif str, ok := _KindMap[x]; ok {\n\t\treturn str\n\t}\n\treturn fmt.Sprintf(\"Kind(%d)\", x)\n}", "func (x DumpKind) String() string {\n\tswitch x {\n\tcase DumpKindVersion:\n\t\treturn \"Version\"\n\tcase DumpKindControl:\n\t\treturn \"Control\"\n\tcase DumpKindMsgId:\n\t\treturn \"MsgId\"\n\tcase DumpKindTypeMsg:\n\t\treturn \"TypeMsg\"\n\tcase DumpKindValueMsg:\n\t\treturn \"ValueMsg\"\n\tcase DumpKindMsgLen:\n\t\treturn \"MsgLen\"\n\tcase DumpKindAnyMsgLen:\n\t\treturn \"AnyMsgLen\"\n\tcase DumpKindAnyLensLen:\n\t\treturn \"AnyLensLen\"\n\tcase DumpKindTypeIdsLen:\n\t\treturn \"TypeIdsLen\"\n\tcase DumpKindTypeId:\n\t\treturn \"TypeId\"\n\tcase DumpKindPrimValue:\n\t\treturn \"PrimValue\"\n\tcase DumpKindByteLen:\n\t\treturn \"ByteLen\"\n\tcase DumpKindValueLen:\n\t\treturn \"ValueLen\"\n\tcase DumpKindIndex:\n\t\treturn \"Index\"\n\tcase DumpKindWireTypeIndex:\n\t\treturn \"WireTypeIndex\"\n\t}\n\treturn \"\"\n}", "func (k ThresholdKind) String() string {\n\tswitch k {\n\tcase KindEntity:\n\t\treturn \"entity\"\n\tcase KindNodeValidator:\n\t\treturn \"validator node\"\n\tcase KindNodeCompute:\n\t\treturn \"compute node\"\n\tcase KindNodeStorage:\n\t\treturn \"storage node\"\n\tcase KindNodeKeyManager:\n\t\treturn \"key manager node\"\n\tcase KindRuntimeCompute:\n\t\treturn \"compute runtime\"\n\tcase KindRuntimeKeyManager:\n\t\treturn \"key manager runtime\"\n\tdefault:\n\t\treturn \"[unknown threshold kind]\"\n\t}\n}", "func (k *Kind) String() string {\n\treturn string(*k)\n}", "func (k *Kind) String() string {\n\treturn string(*k)\n}", "func (ty ValKind) String() string {\n\tswitch ty {\n\tcase KindI32:\n\t\treturn \"i32\"\n\tcase KindI64:\n\t\treturn \"i64\"\n\tcase KindF32:\n\t\treturn \"f32\"\n\tcase KindF64:\n\t\treturn \"f64\"\n\tcase KindExternref:\n\t\treturn \"externref\"\n\tcase KindFuncref:\n\t\treturn \"funcref\"\n\t}\n\tpanic(\"unknown kind\")\n}", "func convKindStr(v interface{}) string { return reflectKindStr(convKind(v)) }", "func (s *Base) String() string {\n\treturn s.kind\n}", "func (t EventKind) String() string {\n\tswitch t {\n\tcase CreateEvent:\n\t\treturn \"create\"\n\tcase DeleteEvent:\n\t\treturn \"delete\"\n\tcase UpdateEvent:\n\t\treturn \"update\"\n\tdefault:\n\t\treturn fmt.Sprintf(\"EventKind(%d)\", t)\n\t}\n}", "func kindType(rse string) string {\n\tname := strings.ToLower(rse)\n\tif strings.Contains(name, \"_tape\") || strings.Contains(name, \"_mss\") || strings.Contains(name, \"_export\") {\n\t\treturn \"TAPE\"\n\t}\n\treturn \"DISK\"\n}", "func (self ValueKind) String() string {\n\tswitch self {\n\tcase I32:\n\t\treturn \"i32\"\n\tcase I64:\n\t\treturn \"i64\"\n\tcase F32:\n\t\treturn \"f32\"\n\tcase F64:\n\t\treturn \"f64\"\n\tcase AnyRef:\n\t\treturn \"anyref\"\n\tcase FuncRef:\n\t\treturn \"funcref\"\n\t}\n\tpanic(\"Unknown value kind\")\n}", "func (b BasicKind) String() string {\n\treturn string(b)\n}", "func (t Type) String() string {\n\tswitch t {\n\tcase Immutable:\n\t\treturn \"IMMUTABLE\"\n\tcase Temporal:\n\t\treturn \"TEMPORAL\"\n\tdefault:\n\t\treturn \"UNKNOWN\"\n\t}\n}", "func (t Type) String() string {\n\tswitch t {\n\tdefault:\n\t\treturn \"Unknown\"\n\tcase '+':\n\t\treturn \"SimpleString\"\n\tcase '-':\n\t\treturn \"Error\"\n\tcase ':':\n\t\treturn \"Integer\"\n\tcase '$':\n\t\treturn \"BulkString\"\n\tcase '*':\n\t\treturn \"Array\"\n\tcase 'R':\n\t\treturn \"RDB\"\n\t}\n}", "func (x ControlKind) String() string {\n\tswitch x {\n\tcase ControlKindNil:\n\t\treturn \"Nil\"\n\tcase ControlKindEnd:\n\t\treturn \"End\"\n\tcase ControlKindIncompleteType:\n\t\treturn \"IncompleteType\"\n\t}\n\treturn \"\"\n}", "func (st StatType) String() string {\n\tswitch st {\n\tcase Avg:\n\t\treturn ElemAvg\n\tcase Max:\n\t\treturn ElemMax\n\tcase Min:\n\t\treturn ElemMin\n\t}\n\treturn \"unknown\"\n}", "func (t Type) String() string {\r\n\tswitch t {\r\n\tcase CONNECT:\r\n\t\treturn \"Connect\"\r\n\tcase CONNACK:\r\n\t\treturn \"Connack\"\r\n\tcase PUBLISH:\r\n\t\treturn \"Publish\"\r\n\tcase PUBACK:\r\n\t\treturn \"Puback\"\r\n\tcase PUBREC:\r\n\t\treturn \"Pubrec\"\r\n\tcase PUBREL:\r\n\t\treturn \"Pubrel\"\r\n\tcase PUBCOMP:\r\n\t\treturn \"Pubcomp\"\r\n\tcase SUBSCRIBE:\r\n\t\treturn \"Subscribe\"\r\n\tcase SUBACK:\r\n\t\treturn \"Suback\"\r\n\tcase UNSUBSCRIBE:\r\n\t\treturn \"Unsubscribe\"\r\n\tcase UNSUBACK:\r\n\t\treturn \"Unsuback\"\r\n\tcase PINGREQ:\r\n\t\treturn \"Pingreq\"\r\n\tcase PINGRESP:\r\n\t\treturn \"Pingresp\"\r\n\tcase DISCONNECT:\r\n\t\treturn \"Disconnect\"\r\n\t}\r\n\r\n\treturn \"Unknown\"\r\n}", "func (o FioSpecVolumePersistentVolumeClaimSpecDataSourceOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumePersistentVolumeClaimSpecDataSource) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (t Type) String() string {\n\tswitch t {\n\tdefault:\n\t\treturn \"\"\n\tcase VPC:\n\t\treturn \"aws_vpc\"\n\tcase Subnet:\n\t\treturn \"aws_subnet\"\n\tcase SG:\n\t\treturn \"aws_security_group\"\n\tcase VPN_GW:\n\t\treturn \"aws_vpn_gateway\"\n\tcase VPN_CONNECTION:\n\t\treturn \"aws_vpn_connection\"\n\tcase CUSTOMER_GW:\n\t\treturn \"aws_customer_gateway\"\n\tcase ELB:\n\t\treturn \"aws_elb\"\n\tcase ASG:\n\t\treturn \"aws_autoscaling_group\"\n\tcase LAUNCH_CONFIG:\n\t\treturn \"aws_launch_configuration\"\n\tcase RDS:\n\t\treturn \"aws_db_instance\"\n\tcase IAM_POLICY:\n\t\treturn \"aws_iam_policy\"\n\tcase IAM_ROLE:\n\t\treturn \"aws_iam_role\"\n\tcase EC2_INSTANCE:\n\t\treturn \"aws_instance\"\n\tcase ROUTE53_RECORD:\n\t\treturn \"aws_route53_record\"\n\tcase CW_ALARM:\n\t\treturn \"aws_cloudwatch_metric_alarm\"\n\tcase LAMBDA:\n\t\treturn \"aws_lambda_function\"\n\tcase S3_BUCKET:\n\t\treturn \"aws_s3_bucket\"\n\t}\n}", "func (c *cachedKind) String() string {\n\treturn c.name\n}", "func (o IopingSpecVolumePersistentVolumeClaimSpecDataSourceOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumePersistentVolumeClaimSpecDataSource) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (t Type) String() string {\n\tswitch t {\n\tcase UNKNOWN:\n\t\treturn \"unknown\"\n\tcase STRING:\n\t\treturn \"string\"\n\tcase BOOLEAN:\n\t\treturn \"boolean\"\n\tcase NUMBER:\n\t\treturn \"number\"\n\tcase DATE:\n\t\treturn \"date\"\n\tcase TIME:\n\t\treturn \"time\"\n\tcase DATETIME:\n\t\treturn \"datetime\"\n\tcase LOCATION:\n\t\treturn \"location\"\n\tcase ADDRESS:\n\t\treturn \"address\"\n\t}\n\treturn \"unsupported\"\n}", "func (t Type) String() string {\n\tswitch t {\n\tdefault:\n\t\treturn \"Unknown\"\n\tcase Null:\n\t\treturn \"Null\"\n\tcase RESP:\n\t\treturn \"RESP\"\n\tcase Telnet:\n\t\treturn \"Telnet\"\n\tcase Native:\n\t\treturn \"Native\"\n\tcase HTTP:\n\t\treturn \"HTTP\"\n\tcase WebSocket:\n\t\treturn \"WebSocket\"\n\tcase JSON:\n\t\treturn \"JSON\"\n\t}\n}", "func (m Manifest) Kind() string {\n\treturn m[\"kind\"].(string)\n}", "func (o RegionAutoscalerOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *RegionAutoscaler) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (t Type) String() string {\n\tswitch t {\n\tcase Int32:\n\t\treturn \"int32\"\n\tcase Int64:\n\t\treturn \"int64\"\n\tcase Float64:\n\t\treturn \"float64\"\n\tcase String:\n\t\treturn \"string\"\n\tcase Bool:\n\t\treturn \"bool\"\n\tcase Timestamp:\n\t\treturn \"timestamp\"\n\tcase JSON:\n\t\treturn \"json\"\n\tdefault:\n\t\treturn \"unsupported\"\n\t}\n}", "func (o CrossVersionObjectReferenceOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CrossVersionObjectReference) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (o CrossVersionObjectReferenceOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CrossVersionObjectReference) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (t Mode) String() string {\n\tswitch t {\n\tcase Anime:\n\t\treturn \"anime\"\n\tcase Photo:\n\t\treturn \"photo\"\n\t}\n\treturn fmt.Sprintf(\"unknown type=%d\", t)\n}", "func (o MetadataResponseOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v MetadataResponse) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (c Column) Kind() string {\n\tif c.DataType.IsInt() {\n\t\treturn c.DataType.String()\n\t}\n\tvar a []string\n\tif c.DataSize > 0 {\n\t\ta = append(a, ds.Unit(c.DataSize).String())\n\t}\n\tif c.DataType.IsString() {\n\t\ta = append(a, c.Charset)\n\t}\n\tif s := strings.Join(a, \", \"); s != \"\" {\n\t\treturn fmt.Sprintf(\"%s(%s)\", c.DataType.String(), s)\n\t}\n\treturn c.DataType.String()\n}", "func (o PriorityClassOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *PriorityClass) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (o SavedAttachedDiskResponseOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v SavedAttachedDiskResponse) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (res ResultType) String() string {\n\treturn fmt.Sprintf(\"kind: %s success: %t\", res.kind, res.success)\n}", "func (t ProviderType) String() string {\n\treturn types[t]\n}", "func (o ArgoCDSpecServerAutoscaleHpaScaleTargetRefOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecServerAutoscaleHpaScaleTargetRef) string { return v.Kind }).(pulumi.StringOutput)\n}", "func TestKindString(t *testing.T) {\n\ttestcases := []struct {\n\t\tkind Kind\n\t\tstr string\n\t}{\n\t\t{KindUndefined, \"Undefined\"},\n\t\t{KindNativeError, \"NativeError\"},\n\t\t{KindRegExp, \"RegExp\"},\n\t\t{KindWebAssemblyCompiledModule, \"WebAssemblyCompiledModule\"},\n\n\t\t// Verify that we have N kinds and they are stringified reasonably.\n\t\t{kNumKinds, \"NoSuchKind:47\"},\n\t}\n\tfor _, test := range testcases {\n\t\tif test.kind.String() != test.str {\n\t\t\tt.Errorf(\"Expected kind %q (%d) to stringify to %q\",\n\t\t\t\ttest.kind, test.kind, test.str)\n\t\t}\n\t}\n}", "func (t Resource) String() string {\n\tswitch t.Type {\n\tdefault:\n\t\treturn \"\"\n\tcase VPC:\n\t\treturn t.aws_vpc_spec()\n\tcase Subnet:\n\t\treturn t.aws_subnet_spec()\n\tcase SG:\n\t\treturn t.aws_sg_spec()\n\tcase VPN_GW:\n\t\treturn t.aws_vpn_gw_spec()\n\tcase VPN_CONNECTION:\n\t\treturn t.aws_vpn_connection_spec()\n\tcase CUSTOMER_GW:\n\t\treturn t.aws_customer_gw_spec()\n\tcase ELB:\n\t\treturn t.aws_elb_spec()\n\tcase ASG:\n\t\treturn t.aws_autoscaling_group_spec()\n\tcase LAUNCH_CONFIG:\n\t\treturn t.aws_launch_configuration_spec()\n\tcase RDS:\n\t\treturn t.aws_rds_instance_spec()\n\tcase IAM_POLICY:\n\t\treturn t.aws_iam_policy_spec()\n\tcase IAM_ROLE:\n\t\treturn t.aws_iam_role_spec()\n\tcase EC2_INSTANCE:\n\t\treturn t.aws_ec2_instance_spec()\n\tcase ROUTE53_RECORD:\n\t\treturn t.aws_route53_zone_record_spec()\n\tcase CW_ALARM:\n\t\treturn t.aws_cloudwatch_metric_alarm_spec()\n\tcase LAMBDA:\n\t\treturn t.aws_lambda_function_spec()\n\tcase S3_BUCKET:\n\t\treturn t.aws_s3_bucket_spec()\n\t}\n}", "func (a APIType) String() string {\n\tswitch a {\n\tcase EGL:\n\t\treturn \"EGL\"\n\tcase GLES2:\n\t\treturn \"GLES2\"\n\tcase GLES3:\n\t\treturn \"GLES3\"\n\tcase GLES31:\n\t\treturn \"GLES31\"\n\tcase VK:\n\t\treturn \"VK\"\n\t}\n\treturn fmt.Sprintf(\"UNKNOWN (%d)\", a)\n}", "func (t EntityType) String() string {\n\tswitch t {\n\tcase Mention:\n\t\treturn \"Mention\"\n\tcase Hashtag:\n\t\treturn \"Hashtag\"\n\tcase URL:\n\t\treturn \"URL\"\n\t}\n\treturn \"Unknown\"\n}", "func (r Resource) Kind() string {\n\treturn r.kind\n}", "func (typ OperationType) String() string {\n\tswitch typ {\n\tcase Query:\n\t\treturn \"query\"\n\tcase Mutation:\n\t\treturn \"mutation\"\n\tcase Subscription:\n\t\treturn \"subscription\"\n\tdefault:\n\t\treturn fmt.Sprintf(\"OperationType(%d)\", int(typ))\n\t}\n}", "func (p PacketType) String() string {\n\tswitch p {\n\tcase PacketTypeOpen:\n\t\treturn \"open\"\n\tcase PacketTypeClose:\n\t\treturn \"close\"\n\tcase PacketTypePing:\n\t\treturn \"ping\"\n\tcase PacketTypePong:\n\t\treturn \"pong\"\n\tcase PacketTypeMessage:\n\t\treturn \"message\"\n\tcase PacketTypeUpgrade:\n\t\treturn \"upgrade\"\n\tcase PacketTypeNoop:\n\t\treturn \"noop\"\n\t}\n\treturn \"invalid\"\n}", "func (o BuildSpecStrategyOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BuildSpecStrategy) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (o LookupImageResultOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupImageResult) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (t ArtifactType) String() string {\n\treturn artifactTypeNames[t]\n}", "func (o CatalogOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Catalog) pulumi.StringPtrOutput { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (o SkuResponseOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v SkuResponse) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (e E_OpenconfigHerculesPlatform_PLATFORM_TYPE) String() string {\n\treturn ygot.EnumLogString(e, int64(e), \"E_OpenconfigHerculesPlatform_PLATFORM_TYPE\")\n}", "func (rt ReleaseType) String() string {\n\tswitch rt {\n\tcase ReleaseTypeError:\n\t\t// Developer error. Panic as we can't call Critical here.\n\t\tpanic(\"uninitialized release type\")\n\tcase Release:\n\t\treturn \"release\"\n\tcase Dev:\n\t\treturn \"dev\"\n\tcase Testing:\n\t\treturn \"testing\"\n\tdefault:\n\t\t// Developer error. Panic as we can't call Critical here.\n\t\tpanic(\"unknown release type\")\n\t}\n}", "func (o LinuxWebAppOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *LinuxWebApp) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (t Type) String() string {\n\tswitch t {\n\tcase TypeCustom:\n\t\treturn \"custom\"\n\tcase TypeASCII:\n\t\treturn \"ascii\"\n\tcase TypeBigInt:\n\t\treturn \"bigint\"\n\tcase TypeBlob:\n\t\treturn \"blob\"\n\tcase TypeBoolean:\n\t\treturn \"boolean\"\n\tcase TypeCounter:\n\t\treturn \"counter\"\n\tcase TypeDecimal:\n\t\treturn \"decimal\"\n\tcase TypeDouble:\n\t\treturn \"double\"\n\tcase TypeFloat:\n\t\treturn \"float\"\n\tcase TypeInt:\n\t\treturn \"int\"\n\tcase TypeText:\n\t\treturn \"text\"\n\tcase TypeTimestamp:\n\t\treturn \"timestamp\"\n\tcase TypeUUID:\n\t\treturn \"uuid\"\n\tcase TypeVarchar:\n\t\treturn \"varchar\"\n\tcase TypeTimeUUID:\n\t\treturn \"timeuuid\"\n\tcase TypeInet:\n\t\treturn \"inet\"\n\tcase TypeDate:\n\t\treturn \"date\"\n\tcase TypeTime:\n\t\treturn \"time\"\n\tcase TypeSmallInt:\n\t\treturn \"smallint\"\n\tcase TypeTinyInt:\n\t\treturn \"tinyint\"\n\tcase TypeList:\n\t\treturn \"list\"\n\tcase TypeMap:\n\t\treturn \"map\"\n\tcase TypeSet:\n\t\treturn \"set\"\n\tcase TypeVarint:\n\t\treturn \"varint\"\n\tcase TypeTuple:\n\t\treturn \"tuple\"\n\tdefault:\n\t\treturn fmt.Sprintf(\"unknown_type_%d\", t)\n\t}\n}", "func (c CurveType) String() string {\n\tswitch c {\n\tcase Hilbert:\n\t\treturn \"Hilbert\"\n\tcase Morton:\n\t\treturn \"Morton\"\n\t}\n\treturn \"\"\n}", "func (o NetworkAttachmentOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *NetworkAttachment) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func KindFromString(s string) (Kind, error) {\n\tswitch Kind(s) {\n\tcase Created:\n\t\treturn Created, nil\n\tcase Modified:\n\t\treturn Modified, nil\n\tdefault:\n\t\treturn \"\", fmt.Errorf(\"invalid kind\")\n\t}\n}", "func (e ExternalID) Kind() (string, error) {\n\ts := string(e)\n\tidx := strings.IndexRune(s, '/')\n\tif idx <= 0 {\n\t\treturn \"\", errors.Reason(\"invalid ExternalID: %q\", s).Err()\n\t}\n\treturn s[:idx], nil\n}", "func (av AttributeConnectionType) String() string {\n\tswitch av {\n\tcase AttributeConnectionTypeClient:\n\t\treturn \"client\"\n\tcase AttributeConnectionTypeFabric:\n\t\treturn \"fabric\"\n\tcase AttributeConnectionTypeHeartbeat:\n\t\treturn \"heartbeat\"\n\t}\n\treturn \"\"\n}", "func (o AttachedDiskResponseOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AttachedDiskResponse) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (p ProtocolVersion) String() string {\n\tswitch p {\n\tcase ProtocolVersionUnrecorded:\n\t\treturn \"Unrecorded\"\n\n\tcase ProtocolVersionLegacy:\n\t\treturn \"Legacy\"\n\n\tcase ProtocolVersionMultiLoopOut:\n\t\treturn \"Multi Loop Out\"\n\n\tcase ProtocolVersionSegwitLoopIn:\n\t\treturn \"Segwit Loop In\"\n\n\tcase ProtocolVersionPreimagePush:\n\t\treturn \"Preimage Push\"\n\n\tcase ProtocolVersionUserExpiryLoopOut:\n\t\treturn \"User Expiry Loop Out\"\n\n\tcase ProtocolVersionHtlcV2:\n\t\treturn \"HTLC V2\"\n\n\tcase ProtocolVersionLoopOutCancel:\n\t\treturn \"Loop Out Cancel\"\n\n\tcase ProtocolVersionProbe:\n\t\treturn \"Probe\"\n\n\tcase ProtocolVersionRoutingPlugin:\n\t\treturn \"Routing Plugin\"\n\n\tcase ProtocolVersionHtlcV3:\n\t\treturn \"HTLC V3\"\n\n\tcase ProtocolVersionMuSig2:\n\t\treturn \"MuSig2\"\n\n\tdefault:\n\t\treturn \"Unknown\"\n\t}\n}", "func (t *Type) String() string {\n\treturn TypeNames[*t]\n}", "func (mvt ValueType) String() string {\n\treturn strings.Title(strings.ToLower(mvt.ValueType.String())) // nolint SA1019\n}", "func (nt NodeType) String() string {\n\tswitch nt {\n\tcase DirType:\n\t\treturn \"dir type\"\n\tcase RegularType:\n\t\treturn \"file type\"\n\tcase BlockDeviceType:\n\t\treturn \"block device type\"\n\tcase CharDeviceType:\n\t\treturn \"char device type\"\n\tcase SocketType:\n\t\treturn \"socket type\"\n\tcase FIFOType:\n\t\treturn \"fifo type\"\n\tcase LinkType:\n\t\treturn \"link type\"\n\tdefault:\n\t\treturn \"unknown type\"\n\t}\n}", "func (tp Protocol) String() string {\n\tswitch tp {\n\tcase PurePayload:\n\t\treturn \"PurePayload\"\n\tcase TTHeader:\n\t\treturn \"TTHeader\"\n\tcase Framed:\n\t\treturn \"Framed\"\n\tcase HTTP:\n\t\treturn \"HTTP\"\n\tcase TTHeaderFramed:\n\t\treturn \"TTHeaderFramed\"\n\tcase GRPC:\n\t\treturn \"GRPC\"\n\t}\n\treturn Unknown\n}", "func (av AttributeOperationType) String() string {\n\tswitch av {\n\tcase AttributeOperationTypeRead:\n\t\treturn \"read\"\n\tcase AttributeOperationTypeWrite:\n\t\treturn \"write\"\n\tcase AttributeOperationTypeSearch:\n\t\treturn \"search\"\n\t}\n\treturn \"\"\n}", "func (o MachineInstanceSpecClassRefOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MachineInstanceSpecClassRef) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (o TargetPoolOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *TargetPool) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (h HotRegionType) String() string {\n\tswitch h {\n\tcase WriteType:\n\t\treturn \"write\"\n\tcase ReadType:\n\t\treturn \"read\"\n\t}\n\treturn \"unimplemented\"\n}", "func (class OsClass) String() string {\n\tswitch class {\n\tcase Generic:\n\t\treturn \"generic\"\n\tcase Device:\n\t\treturn \"device\"\n\tcase Host:\n\t\treturn \"host\"\n\tdefault:\n\t\tpanic(fmt.Errorf(\"unknown class %d\", class))\n\t}\n}", "func (e E_OpenconfigQos_Qos_Classifiers_Classifier_Config_Type) String() string {\n\treturn ygot.EnumLogString(e, int64(e), \"E_OpenconfigQos_Qos_Classifiers_Classifier_Config_Type\")\n}", "func (o *operator) Kind() string {\n\treturn o.kind\n}", "func (d Disposition) String() string {\n\tswitch d {\n\tcase Accepted:\n\t\treturn \"Accepted\"\n\tcase Rejected:\n\t\treturn \"Rejected\"\n\tcase Released:\n\t\treturn \"Released\"\n\tdefault:\n\t\treturn \"Unknown\"\n\t}\n}", "func (o NetworkInterfaceResponseOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v NetworkInterfaceResponse) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (t ProviderType) String() string {\n\tif int(t) >= 0 && int(t) < len(providerTypes) {\n\t\treturn providerTypes[t]\n\t} else {\n\t\treturn \"\"\n\t}\n}", "func (t *Type) Kind() string {\n\treturn t.kind\n}", "func (o MachineInstanceSpecResourceRefOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MachineInstanceSpecResourceRef) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (o NetworkPolicyOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *NetworkPolicy) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (typeType TagTypePrimitive) String() string {\n\treturn TypeNames[typeType]\n}", "func (o ArgoCDExportSpecStoragePvcDataSourceOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ArgoCDExportSpecStoragePvcDataSource) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (v SensorType) String() string {\n\tif v == DHT11 {\n\t\treturn \"DHT11\"\n\t} else if v == DHT12 {\n\t\treturn \"DHT12\"\n\t} else if v == DHT22 || v == AM2302 {\n\t\treturn \"DHT22|AM2302\"\n\t} else {\n\t\treturn \"!!! unknown !!!\"\n\t}\n}", "func (o LookupShareResultOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupShareResult) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (s InstType) String() string {\n\tif name, ok := instTypeToString[s]; ok {\n\t\treturn name\n\t}\n\treturn \"?\"\n}", "func (e EventType) String() string {\n\tswitch e {\n\tcase EventTypeGistCreated:\n\t\treturn \"gistCreated\"\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"unexpected EventType %s, in string()\", string(e)))\n\t}\n\n}", "func (s *String) Kind() int {\n\treturn 204\n}", "func (e EnumType) String() string {\n\treturn fmt.Sprintf(\"enum %v\", e.ident)\n}", "func (k Key) String() string {\n\treturn fmt.Sprintf(\"%s:%s\", k.Digest, k.Kind)\n}", "func (s *tsLibrary) Kind() string {\n\treturn s.KindName\n}", "func KindString(s string) (Kind, error) {\n\tif val, ok := _KindNameToValueMap[s]; ok {\n\t\treturn val, nil\n\t}\n\ts = strings.ToLower(s)\n\tif val, ok := _KindNameToValueMap[s]; ok {\n\t\treturn val, nil\n\t}\n\treturn 0, fmt.Errorf(\"%s does not belong to Kind values\", s)\n}", "func GetKind(node *yaml.RNode, path string) string {\n\treturn GetStringField(node, path, \"kind\")\n}", "func (level LevelType) String() string {\n\tif !level.valid() {\n\t\treturn UNKNOWN\n\t}\n\treturn LevelStrings[level]\n}", "func (o PriorityLevelConfigurationOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *PriorityLevelConfiguration) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (t *Type) String() string {\n\tif t.Precision != nil && t.Scale != nil {\n\t\treturn fmt.Sprintf(\"%s(%s,%s)\", t.Name.Name, t.Precision.String(), t.Scale.String())\n\t} else if t.Precision != nil {\n\t\treturn fmt.Sprintf(\"%s(%s)\", t.Name.Name, t.Precision.String())\n\t} else if t.Scale != nil {\n\t\t// I'm not sure how you're supposed to tell this from the t.Precision case.\n\t\treturn fmt.Sprintf(\"%s(%s)\", t.Name.Name, t.Scale.String())\n\t}\n\treturn t.Name.Name\n}", "func (af *filtBase) GetKindName() string {\n\treturn af.kind\n}", "func (t TherapistType) String() string {\n\tswitch t {\n\tcase UnknownTherapistType:\n\t\treturn \"unknown\"\n\tcase OccupationalTherapist:\n\t\treturn \"ergo\"\n\tcase Physiotherapist:\n\t\treturn \"physio\"\n\tcase SpeechTherapist:\n\t\treturn \"logo\"\n\tdefault:\n\t\treturn \"<unknown therapist type>\"\n\t}\n}", "func (o StatefulSetOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *StatefulSet) pulumi.StringOutput { return v.Kind }).(pulumi.StringOutput)\n}", "func (o ObjectAccessControlResponseOutput) Kind() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ObjectAccessControlResponse) string { return v.Kind }).(pulumi.StringOutput)\n}", "func (m MessageType) String() string {\n\tif s, ok := messageTypeToString[m]; ok {\n\t\treturn s\n\t}\n\treturn fmt.Sprintf(\"unknown (%d)\", m)\n}", "func (t ResourceType) String() string {\n\treturn C.GoString(C.rd_kafka_ResourceType_name(C.rd_kafka_ResourceType_t(t)))\n}" ]
[ "0.7882306", "0.78182817", "0.7631398", "0.74356145", "0.7273103", "0.7260341", "0.72308713", "0.72308713", "0.72005254", "0.7159091", "0.71464866", "0.70390964", "0.69976556", "0.6933344", "0.6829861", "0.67954165", "0.67454123", "0.674282", "0.67382026", "0.67114013", "0.67022324", "0.66523844", "0.66481346", "0.66352934", "0.6599222", "0.65896577", "0.6584011", "0.6581553", "0.6576051", "0.6550144", "0.6550144", "0.6541871", "0.65102136", "0.6481548", "0.6467704", "0.64613265", "0.64571863", "0.6441401", "0.6434986", "0.64129007", "0.6397653", "0.6397156", "0.6393585", "0.639354", "0.6393161", "0.6390268", "0.63789415", "0.6371498", "0.63693035", "0.6356304", "0.63436145", "0.6342716", "0.63376033", "0.63299465", "0.631433", "0.63118017", "0.629618", "0.62930477", "0.6288435", "0.6288099", "0.6277586", "0.62753344", "0.62647575", "0.62645286", "0.62608105", "0.62559706", "0.625015", "0.62500656", "0.62493604", "0.6240445", "0.6240324", "0.62400234", "0.6215941", "0.62113565", "0.6209437", "0.6199495", "0.61994475", "0.61992186", "0.6197311", "0.6193433", "0.6189797", "0.6184824", "0.617961", "0.6171086", "0.6161019", "0.6160757", "0.6157515", "0.61568385", "0.6152053", "0.6151149", "0.615105", "0.6146621", "0.6146003", "0.61455655", "0.6144861", "0.61447626", "0.61411303", "0.61400884", "0.61397576", "0.61385846" ]
0.7552743
3
OK validates the kind is valid.
func (k Kind) OK() error { if k == KindUnknown { return errors.New("invalid kind") } if !kinds[k] { return errors.New("unsupported kind provided") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (kind Kind) IsValid() bool {\n\treturn kind&Invalid == 0\n}", "func (m *Resource) IsOK() bool {\n\tswitch {\n\tcase len(m.name) == 0:\n\t\treturn false\n\tcase len(m.description) == 0:\n\t\treturn false\n\tcase m.schema == nil:\n\t\treturn false\n\tcase m.model == nil:\n\t\treturn false\n\tcase m.store == nil:\n\t\treturn false\n\tcase len(m.methods) == 0:\n\t\treturn false\n\tdefault:\n\t\treturn true\n\t}\n}", "func (c *Collection) invalidKind(kind string) bool {\n\tkinds := []string{\"ComponentStatus\"}\n\tfor _, item := range kinds {\n\t\tif kind == item {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (o *ManualDependency) GetKindOk() (*string, bool) {\n\tif o == nil || o.Kind == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Kind, true\n}", "func OK() ValidationError {\n\treturn ValidationError{}\n}", "func (d *Document) OK() error {\n\tif d.URL == \"\" {\n\t\treturn er.InvalidField(\"url\")\n\t}\n\tif d.DocType == \"\" {\n\t\treturn er.InvalidField(\"doc_type\")\n\t}\n\tif d.OwnerType == \"\" {\n\t\treturn er.InvalidField(\"owner_type\")\n\t}\n\treturn nil\n}", "func (o *ReconciliationTarget) GetKindOk() (*string, bool) {\n\tif o == nil || o.Kind == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Kind, true\n}", "func (r *BaseStandard) IsOk() bool {\n\tif r.AuditInfo.StatusCode < http.StatusOK || r.AuditInfo.StatusCode >= http.StatusMultipleChoices {\n\t\treturn false\n\t}\n\n\tif !r.HasItems() {\n\t\treturn false\n\t}\n\n\tif len(r.AuditInfo.Errors.Items) > 0 {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (o *TeamConfiguration) GetKindOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.Kind, true\n}", "func (ctx *GetByIDHostContext) OK(r *Hostwithdogs) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"hostwithdogs\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (s *RateLimitSuite) TestSpecIsOK(c *C) {\n\tc.Assert(plugin.NewRegistry().AddSpec(GetSpec()), IsNil)\n}", "func (s *RateLimitSuite) TestSpecIsOK(c *C) {\n\tc.Assert(plugin.NewRegistry().AddSpec(GetSpec()), IsNil)\n}", "func FormatOk(fmt string) bool {\n\tfmts := []string{\n\t\tFmt_ntriples,\n\t\tFmt_turtle,\n\t\tFmt_rdfxmlXmp,\n\t\tFmt_rdfxmlAbbrev,\n\t\tFmt_rdfxml,\n\t\tFmt_rss,\n\t\tFmt_atom,\n\t\tFmt_dot,\n\t\tFmt_jsonTriples,\n\t\tFmt_json,\n\t\tFmt_html,\n\t\tFmt_nquads,\n\t}\n\tfor _, f := range fmts {\n\t\tif fmt == f {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (ctl *Ctl) SpecIsValid() (bool, error) {\n\treturn true, nil\n}", "func (ctl *Ctl) SpecIsValid() (bool, error) {\n\treturn true, nil\n}", "func (resp *Response) OK() bool {\n\treturn resp.StatusCode < 400\n}", "func (a API) OK(data interface{}) (int, model.MessageResponse) {\n\treturn http.StatusOK, model.MessageResponse{\n\t\tData: data,\n\t\tMessages: model.Responses{{Code: Ok, Message: \"¡listo!\"}},\n\t}\n}", "func (req *Request) OK(body string) {\n\treq.Reply(http.StatusOK, body)\n}", "func (ctx *HealthHealthContext) OK(r *JSON) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/json\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (ctx *GetDogsByHostIDHostContext) OK(r *Dogs) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"dogs\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (r *Request) OK() error {\n\tif len(r.URL) == 0 {\n\t\treturn ErrRequired{Msg: \"url must be specified\"}\n\t}\n\treturn nil\n}", "func (c *SeaterController) OK(data interface{}) {\n\tc.Code(200)\n\tc.jsonResp(data)\n}", "func (s *APIStatusResponse) OK() bool {\n\treturn s.StatusCode == \"ok\"\n}", "func (a *assertEncapsulation) OK(condition bool, statusCode int, message string, args ...interface{}) {\n\tif err := ok(condition, statusCode, message, args...); err != nil {\n\t\ta.throw(err)\n\t}\n}", "func (pr *ResourceDescription) Validate() error {\n\tif len(pr.Kinds) == 0 {\n\t\treturn errors.New(\"The Kind is not specified\")\n\t}\n\n\tif pr.Selector != nil {\n\t\tselector, err := metav1.LabelSelectorAsSelector(pr.Selector)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\trequirements, _ := selector.Requirements()\n\t\tif len(requirements) == 0 {\n\t\t\treturn errors.New(\"The requirements are not specified in selector\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func OK(payload interface{}) Response {\n\treturn Response{\n\t\tStatusCode: http.StatusOK,\n\t\tPayload: payload,\n\t}\n}", "func (s Stored) Ok() bool {\n\treturn s.Type.Ok() && s.Encoding.Ok()\n}", "func (r *Responder) OK() { r.write(http.StatusOK) }", "func (h Health) Ok() bool {\n\treturn h.success\n}", "func (ctx *ShowVerificationContext) OK(r *Verification) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.verification+json\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (ctx *ShowStatusContext) OK(r *Status) error {\n\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.status+json\")\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (gvk GVK) Validate() error {\n\t// Check if the qualified group has a valid DNS1123 subdomain value\n\tif gvk.QualifiedGroup() == \"\" {\n\t\treturn fmt.Errorf(groupRequired)\n\t}\n\tif err := validation.IsDNS1123Subdomain(gvk.QualifiedGroup()); err != nil {\n\t\t// NOTE: IsDNS1123Subdomain returns a slice of strings instead of an error, so no wrapping\n\t\treturn fmt.Errorf(\"either Group or Domain is invalid: %s\", err)\n\t}\n\n\t// Check if the version follows the valid pattern\n\tif gvk.Version == \"\" {\n\t\treturn fmt.Errorf(versionRequired)\n\t}\n\tif !versionRegex.MatchString(gvk.Version) {\n\t\treturn fmt.Errorf(\"Version must match %s (was %s)\", versionPattern, gvk.Version)\n\t}\n\n\t// Check if kind has a valid DNS1035 label value\n\tif gvk.Kind == \"\" {\n\t\treturn fmt.Errorf(kindRequired)\n\t}\n\tif errors := validation.IsDNS1035Label(strings.ToLower(gvk.Kind)); len(errors) != 0 {\n\t\t// NOTE: IsDNS1035Label returns a slice of strings instead of an error, so no wrapping\n\t\treturn fmt.Errorf(\"invalid Kind: %#v\", errors)\n\t}\n\n\t// Require kind to start with an uppercase character\n\t// NOTE: previous validation already fails for empty strings, gvk.Kind[0] will not panic\n\tif string(gvk.Kind[0]) == strings.ToLower(string(gvk.Kind[0])) {\n\t\treturn fmt.Errorf(\"invalid Kind: must start with an uppercase character\")\n\t}\n\n\treturn nil\n}", "func (ctx *SpecsFilterContext) OK(r FilterSpecCollection) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.feedpushr.filter-spec.v1+json; type=collection\")\n\t}\n\tif r == nil {\n\t\tr = FilterSpecCollection{}\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (t Type) Ok() bool {\n\t_, ok := types[t]\n\treturn ok\n}", "func (self *Platform) ValidateSpec(spec *pb.ChaincodeSpec) error {\n\treturn nil\n}", "func (spec Spec) Validate() error {\n\tif !validModule.MatchString(spec.Module) {\n\t\treturn fmt.Errorf(\"invalid module: %q\", spec.Module)\n\t}\n\tif !validVersion.MatchString(spec.Version) {\n\t\treturn fmt.Errorf(\"invalid version: %q\", spec.Version)\n\t}\n\tif !validType.MatchString(spec.Type) {\n\t\treturn fmt.Errorf(\"invalid type: %q\", spec.Type)\n\t}\n\tif !validClass.MatchString(spec.Class) {\n\t\treturn fmt.Errorf(\"invalid class: %q\", spec.Class)\n\t}\n\treturn nil\n}", "func (c *Client) IsOK() bool {\n\turl := fmt.Sprintf(\"%s/v1/sys/health\", c.addr)\n\n\tr, _ := http.NewRequest(http.MethodGet, url, nil)\n\t//r.Header.Add(\"X-Vault-Token\", \"root\")\n\n\tresp, err := http.DefaultClient.Do(r)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func OK(condition bool, statusCode int, message string, args ...interface{}) {\n\tif err := ok(condition, statusCode, message, args...); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (ctx *GetSwaggerContext) OK(resp []byte) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/json\")\n\t}\n\tctx.ResponseData.WriteHeader(200)\n\t_, err := ctx.ResponseData.Write(resp)\n\treturn err\n}", "func kind(err error) errors.Kind {\n\tswitch {\n\tcase docker.IsErrUnauthorized(err):\n\t\treturn errors.NotAllowed\n\tdefault:\n\t\t// Liberally pick unavailable as the default error, so that lower\n\t\t// layers can retry errors that may be fruitfully retried.\n\t\t// This is always safe to do, but may cause extra work.\n\t\treturn errors.Unavailable\n\t}\n}", "func (mt *Vironauthtype) Validate() (err error) {\n\tif mt.Type == \"\" {\n\t\terr = goa.MergeErrors(err, goa.MissingAttributeError(`response`, \"type\"))\n\t}\n\tif mt.Provider == \"\" {\n\t\terr = goa.MergeErrors(err, goa.MissingAttributeError(`response`, \"provider\"))\n\t}\n\tif mt.URL == \"\" {\n\t\terr = goa.MergeErrors(err, goa.MissingAttributeError(`response`, \"url\"))\n\t}\n\tif mt.Method == \"\" {\n\t\terr = goa.MergeErrors(err, goa.MissingAttributeError(`response`, \"method\"))\n\t}\n\treturn\n}", "func TestParser_GetKind(t *testing.T) {\n\tp := &Parser{}\n\trequire.Equal(t, model.KindJSON, p.GetKind())\n}", "func (ctx *GetHealthContext) OK(resp []byte) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"text/plain\")\n\t}\n\tctx.ResponseData.WriteHeader(200)\n\t_, err := ctx.ResponseData.Write(resp)\n\treturn err\n}", "func (o *UtilTestOKBodyResponse) Validate(formats strfmt.Registry) error {\n\treturn nil\n}", "func (ctx *TagOrdinalValuesContext) OK(resp []byte) error {\n\tctx.ResponseData.Header().Set(\"Content-Type\", \"text/plain\")\n\tctx.ResponseData.WriteHeader(200)\n\t_, err := ctx.ResponseData.Write(resp)\n\treturn err\n}", "func (ind ErrInvalidNetworkDriver) BadRequest() {}", "func (vr *VersionResponse) IsOk() bool {\n\treturn len(vr.version) > 0\n}", "func OKWithMeta(w http.ResponseWriter, data interface{}, msg string, meta interface{}) error {\n\tsb := BuildSuccess(data, msg, meta)\n\treturn Write(w, sb, http.StatusOK)\n}", "func (s *String) Kind() int {\n\treturn 204\n}", "func (ct ContentType) IsValid() error {\n\tswitch ct {\n\tcase Collection, Credential, DIDResolutionResponse, Metadata, Connection, Key:\n\t\treturn nil\n\t}\n\n\treturn fmt.Errorf(\"invalid content type '%s', supported types are %s\", ct,\n\t\t[]ContentType{Collection, Credential, DIDResolutionResponse, Metadata, Connection, Key})\n}", "func (o *VersionedControllerService) GetTypeOk() (*string, bool) {\n\tif o == nil || o.Type == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Type, true\n}", "func (ctx *SpecsOutputContext) OK(r OutputSpecCollection) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.feedpushr.output-spec.v1+json; type=collection\")\n\t}\n\tif r == nil {\n\t\tr = OutputSpecCollection{}\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (m softwaresourcesummary) ValidateEnumValue() (bool, error) {\n\terrMessage := []string{}\n\tif _, ok := GetMappingAvailabilityEnum(string(m.Availability)); !ok && m.Availability != \"\" {\n\t\terrMessage = append(errMessage, fmt.Sprintf(\"unsupported enum value for Availability: %s. Supported values are: %s.\", m.Availability, strings.Join(GetAvailabilityEnumStringValues(), \",\")))\n\t}\n\tif _, ok := GetMappingOsFamilyEnum(string(m.OsFamily)); !ok && m.OsFamily != \"\" {\n\t\terrMessage = append(errMessage, fmt.Sprintf(\"unsupported enum value for OsFamily: %s. Supported values are: %s.\", m.OsFamily, strings.Join(GetOsFamilyEnumStringValues(), \",\")))\n\t}\n\tif _, ok := GetMappingArchTypeEnum(string(m.ArchType)); !ok && m.ArchType != \"\" {\n\t\terrMessage = append(errMessage, fmt.Sprintf(\"unsupported enum value for ArchType: %s. Supported values are: %s.\", m.ArchType, strings.Join(GetArchTypeEnumStringValues(), \",\")))\n\t}\n\n\tif _, ok := GetMappingSoftwareSourceLifecycleStateEnum(string(m.LifecycleState)); !ok && m.LifecycleState != \"\" {\n\t\terrMessage = append(errMessage, fmt.Sprintf(\"unsupported enum value for LifecycleState: %s. Supported values are: %s.\", m.LifecycleState, strings.Join(GetSoftwareSourceLifecycleStateEnumStringValues(), \",\")))\n\t}\n\tif len(errMessage) > 0 {\n\t\treturn true, fmt.Errorf(strings.Join(errMessage, \"\\n\"))\n\t}\n\treturn false, nil\n}", "func (ctx *DeleteDogContext) OK(resp []byte) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"text/plain\")\n\t}\n\tctx.ResponseData.WriteHeader(200)\n\t_, err := ctx.ResponseData.Write(resp)\n\treturn err\n}", "func (b *AuctionReq) Ok() error {\r\n\tswitch {\r\n\tcase strings.TrimSpace(b.AuctionID) == \"\":\r\n\t\treturn errors.IsRequiredErr(\"auction id\")\r\n\t}\r\n\r\n\treturn nil\r\n}", "func (u *userInput) OK() error {\n\t// check the white space\n\tu.Email = strings.TrimSpace(u.Email)\n\tu.Username = strings.TrimSpace(u.Username)\n\n\t// validate the regex\n\tif !user.RXEmail.MatchString(u.Email) {\n\t\treturn user.ErrInvalidEmail\n\t}\n\tif !user.RXUsername.MatchString(u.Username) {\n\t\treturn user.ErrInvalidUsername\n\t}\n\treturn nil\n}", "func (ctx *RegisterOrdinalValuesContext) OK(resp []byte) error {\n\tctx.ResponseData.Header().Set(\"Content-Type\", \"text/plain\")\n\tctx.ResponseData.WriteHeader(200)\n\t_, err := ctx.ResponseData.Write(resp)\n\treturn err\n}", "func (ctx *SignupAuthenticationContext) OK(r *Success) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.security.success\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func validate(status int, _ string) bool {\n\treturn status == 200\n}", "func (u NewAttestation) OK() error {\n\tif len(u.ClaimantID) == 0 {\n\t\treturn errMissingField(\"ClaimantID\")\n\t}\n\tif len(u.AttestorID) == 0 {\n\t\treturn errMissingField(\"AttestorID\")\n\t}\n\tif len(u.Claim) == 0 {\n\t\treturn errMissingField(\"Claim\")\n\t}\n\tif len(u.Claim) > 100 {\n\t\treturn &errMaxLengthExceeded{\"Claim\", 100}\n\t}\n\treturn nil\n}", "func (o *GetSlashingParametersOKBodyResult) Validate(formats strfmt.Registry) error {\n\treturn nil\n}", "func (o *GetHelloOKBody) Validate(formats strfmt.Registry) error {\n\treturn nil\n}", "func isMissingKind(err error) bool {\n\treturn strings.Contains(err.Error(), \"Object 'Kind' is missing in\")\n}", "func ParseKind(name string) (Kind, error) {\n\tif x, ok := _KindValue[name]; ok {\n\t\treturn x, nil\n\t}\n\treturn Kind(0), fmt.Errorf(\"%s is not a valid Kind\", name)\n}", "func (m createsoftwaresourcedetails) ValidateEnumValue() (bool, error) {\n\terrMessage := []string{}\n\n\tif len(errMessage) > 0 {\n\t\treturn true, fmt.Errorf(strings.Join(errMessage, \"\\n\"))\n\t}\n\treturn false, nil\n}", "func (ctx *DeleteHostContext) OK(resp []byte) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"text/plain\")\n\t}\n\tctx.ResponseData.WriteHeader(200)\n\t_, err := ctx.ResponseData.Write(resp)\n\treturn err\n}", "func (o *ControllerServiceAPI) GetTypeOk() (*string, bool) {\n\tif o == nil || o.Type == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Type, true\n}", "func OK(w http.ResponseWriter, data interface{}, message string) {\n\tsuccessResponse := BuildSuccess(data, message, MetaInfo{HTTPStatus: http.StatusOK})\n\tWrite(w, successResponse, http.StatusOK)\n}", "func IsOK(err error, message string) error {\n\tif err != nil {\n\t\tlog.Println(message)\n\t\t// if debug {\n\t\t// \tlog.Fatal(\"[ ERROR ] \", err)\n\t\t// }\n\t\treturn err\n\t}\n\treturn nil\n}", "func OK(data interface{}, w http.ResponseWriter) {\n\tjsonResponse(data, http.StatusOK, w)\n}", "func checkSpec(spec *pb.ChaincodeSpec) error {\n\t// Don't allow nil value\n\tif spec == nil {\n\t\tfmt.Printf(\"Expected chaincode specification, nil received\")\n\t}\n\n\tplatform, err := platforms.Find(spec.Type)\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to determine platform type: %s\", err)\n\t}\n\n\treturn platform.ValidateSpec(spec)\n}", "func (g Group) IsOk() bool {\n\tpaw.Logger.Debug(\"checking Group...\" + paw.Caller(1))\n\tif g > 0 && g < 4 {\n\t\treturn true\n\t}\n\treturn false\n\t// switch g {\n\t// case Grouped, GroupedR, GroupNone:\n\t// \treturn true\n\t// default:\n\t// \treturn false\n\t// }\n}", "func (ctx *GetFeedContext) OK(r *Feed) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/json\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func TestKindString(t *testing.T) {\n\ttestcases := []struct {\n\t\tkind Kind\n\t\tstr string\n\t}{\n\t\t{KindUndefined, \"Undefined\"},\n\t\t{KindNativeError, \"NativeError\"},\n\t\t{KindRegExp, \"RegExp\"},\n\t\t{KindWebAssemblyCompiledModule, \"WebAssemblyCompiledModule\"},\n\n\t\t// Verify that we have N kinds and they are stringified reasonably.\n\t\t{kNumKinds, \"NoSuchKind:47\"},\n\t}\n\tfor _, test := range testcases {\n\t\tif test.kind.String() != test.str {\n\t\t\tt.Errorf(\"Expected kind %q (%d) to stringify to %q\",\n\t\t\t\ttest.kind, test.kind, test.str)\n\t\t}\n\t}\n}", "func (ct ContentType) IsValid() error {\n\tswitch ct {\n\tcase Collection, Credential, DIDDocument, Metadata, Connection:\n\t\treturn nil\n\t}\n\n\treturn fmt.Errorf(\"invalid content type '%s', supported types are %s\", ct,\n\t\t[]ContentType{Collection, Credential, DIDDocument, Metadata, Connection})\n}", "func (v SyntheticsBasicAuthOauthROPType) IsValid() bool {\n\tfor _, existing := range allowedSyntheticsBasicAuthOauthROPTypeEnumValues {\n\t\tif existing == v {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Ok(msg string) *Response {\n\treturn &Response{\n\t\tCode: http.StatusOK,\n\t\tMessage: msg,\n\t\tData: nil,\n\t}\n}", "func (p *Parser) expect(kind ItemKind) bool {\n\tif p.next.Kind != kind {\n\t\tp.itemError(p.next, fmt.Errorf(\"expected %s\", kind))\n\t\treturn false\n\t}\n\tp.consume()\n\treturn true\n}", "func (ctx *GetFilterContext) OK(r *Filter) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/json\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (ii ErrInvalidID) BadRequest() {}", "func (in ErrInvalidName) BadRequest() {}", "func (o *ResourceInfoOKBody) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := o.validateResults(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *V1VirusDatasetRequest) GetFormatOk() (*V1TableFormat, bool) {\n\tif o == nil || o.Format == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Format, true\n}", "func (m Match) Ok() bool { return len(m.Errors) == 0 }", "func (m AnyThing) Validate() error {\n\treturn InKnownAnyThing.Validate(m)\n}", "func (o *ContainerTopOKBody) Validate(formats strfmt.Registry) error {\n\treturn nil\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func Kind(kind string) schema.GroupKind {\n\treturn SchemeGroupVersion.WithKind(kind).GroupKind()\n}", "func (ctx *PushOrdinalValuesContext) OK(resp []byte) error {\n\tctx.ResponseData.Header().Set(\"Content-Type\", \"text/plain\")\n\tctx.ResponseData.WriteHeader(200)\n\t_, err := ctx.ResponseData.Write(resp)\n\treturn err\n}", "func (h Howout) Valid() error {\n\tswitch h {\n\tcase Bowled, Catch, Runout, Notout, Didnotbat, Stumped:\n\t\treturn nil\n\t}\n\n\treturn errors.New(\"Invalid howout, it has to be one of [b,c,ro,no,dnb,st] \")\n}", "func (e Encoding) Ok() bool {\n\t_, ok := encodings[e]\n\treturn ok\n}", "func TestParser_GetKind(t *testing.T) {\n\tp := &Parser{}\n\trequire.Equal(t, model.KindTerraform, p.GetKind())\n}", "func (ctx *SiginAuthenticationContext) OK(r *Success) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.security.success\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (m *WireguardSpec) Validate(formats strfmt.Registry) error {\n\treturn nil\n}" ]
[ "0.62202156", "0.59624", "0.57328045", "0.57115155", "0.5710232", "0.56863034", "0.5651936", "0.5640751", "0.5633026", "0.55821824", "0.5531406", "0.5531406", "0.550324", "0.5461285", "0.5461285", "0.5432487", "0.53961223", "0.5378424", "0.53483063", "0.5335181", "0.53294885", "0.5309623", "0.5297801", "0.5296796", "0.52700686", "0.5251187", "0.524845", "0.5244753", "0.52227086", "0.521453", "0.5210973", "0.52071923", "0.517808", "0.5151254", "0.51318073", "0.51201826", "0.5117959", "0.5117187", "0.51095086", "0.5098496", "0.50695616", "0.50523716", "0.5048377", "0.50337386", "0.50149995", "0.5005527", "0.50045097", "0.49999452", "0.4998998", "0.49976158", "0.4995372", "0.49951476", "0.499053", "0.49831158", "0.4971886", "0.49709135", "0.4968343", "0.49670836", "0.49670687", "0.4949558", "0.49337977", "0.49295807", "0.49241355", "0.49168125", "0.49155474", "0.49141955", "0.4912834", "0.4912019", "0.48933685", "0.48921597", "0.488656", "0.48812884", "0.48744956", "0.4873446", "0.48681095", "0.48680368", "0.48668903", "0.4866063", "0.4865559", "0.48619425", "0.48554495", "0.48546404", "0.48474216", "0.48449168", "0.48446864", "0.48428106", "0.48390806", "0.48390806", "0.48390806", "0.48390806", "0.48390806", "0.48390806", "0.48390806", "0.48390806", "0.4837458", "0.48348176", "0.48333123", "0.48308674", "0.48291025", "0.48252758" ]
0.7533541
0
ResourceType converts a kind to a known resource type (if applicable).
func (k Kind) ResourceType() influxdb.ResourceType { switch k { case KindBucket: return influxdb.BucketsResourceType case KindCheck, KindCheckDeadman, KindCheckThreshold: return influxdb.ChecksResourceType case KindDashboard: return influxdb.DashboardsResourceType case KindLabel: return influxdb.LabelsResourceType case KindNotificationEndpoint, KindNotificationEndpointHTTP, KindNotificationEndpointPagerDuty, KindNotificationEndpointSlack: return influxdb.NotificationEndpointResourceType case KindNotificationRule: return influxdb.NotificationRuleResourceType case KindTask: return influxdb.TasksResourceType case KindTelegraf: return influxdb.TelegrafsResourceType case KindVariable: return influxdb.VariablesResourceType default: return "" } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func kindToResource(kind string, mixedCase bool) (plural, singular string) {\n\tif len(kind) == 0 {\n\t\treturn\n\t}\n\tif mixedCase {\n\t\t// Legacy support for mixed case names\n\t\tsingular = strings.ToLower(kind[:1]) + kind[1:]\n\t} else {\n\t\tsingular = strings.ToLower(kind)\n\t}\n\tswitch string(singular[len(singular)-1]) {\n\tcase \"s\":\n\t\tplural = singular\n\tcase \"y\":\n\t\tplural = strings.TrimSuffix(singular, \"y\") + \"ies\"\n\tdefault:\n\t\tplural = singular + \"s\"\n\t}\n\treturn\n}", "func MapResourceKindToListResourcesType(kind string) string {\n\tswitch kind {\n\tcase types.KindApp:\n\t\treturn types.KindAppServer\n\tcase types.KindDatabase:\n\t\treturn types.KindDatabaseServer\n\tcase types.KindKubernetesCluster:\n\t\treturn types.KindKubeServer\n\tdefault:\n\t\treturn kind\n\t}\n}", "func ResourceType() *spec.ResourceType {\n\treturn internal.ResourceType\n}", "func (r *DeliveryResource) ResourceType() string {\n\tleft := strings.Index(r.Kind, \"/\")\n\tright := strings.LastIndex(r.Kind, \"@\")\n\treturn r.Kind[left+1 : right]\n}", "func (r *CredentialReconciler) resourceForKind(kwg kindWithGroup) (metav1.APIResource, error) {\n\tresource, exists := r.resourceCache[kwg]\n\tif !exists {\n\t\t// TODO: this is definitely not the best way of dealing with a missing resource in the local\n\t\t// cache. While this is fine for a prototype, it can crash the controller pod or the node\n\t\t// (depending on the pod resources) if the kind with the specified group doesn't exist on\n\t\t// the API server (or if the API server goes off for a little walk), or crash the node.\n\t\tif err := r.updateResourceCache(); err != nil {\n\t\t\treturn metav1.APIResource{}, err\n\t\t}\n\t\treturn r.resourceForKind(kwg)\n\t}\n\treturn resource, nil\n}", "func (o CloudConfigurationRuleOutput) ResourceType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *CloudConfigurationRule) pulumi.StringOutput { return v.ResourceType }).(pulumi.StringOutput)\n}", "func (s *SyncResourceFilter) SetResourceType(v string) *SyncResourceFilter {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o DatasourceOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Datasource) *string { return v.ResourceType }).(pulumi.StringPtrOutput)\n}", "func (o *LabelProperties) SetResourceType(v string) {\n\n\to.ResourceType = &v\n\n}", "func ResourceTypeFromString(typeString string) (ResourceType, error) {\n\tswitch strings.ToUpper(typeString) {\n\tcase \"ANY\":\n\t\treturn ResourceAny, nil\n\tcase \"TOPIC\":\n\t\treturn ResourceTopic, nil\n\tcase \"GROUP\":\n\t\treturn ResourceGroup, nil\n\tcase \"BROKER\":\n\t\treturn ResourceBroker, nil\n\tdefault:\n\t\treturn ResourceUnknown, NewError(ErrInvalidArg, \"Unknown resource type\", false)\n\t}\n}", "func ParseResourceKind(name string) ResourceKind {\n\tk, ok := resourceNameToValue[name]\n\tif ok {\n\t\treturn k\n\t}\n\treturn UnKnownKind\n}", "func (o *QuotaSummaryAllOf) GetResourceTypeOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.ResourceType, true\n}", "func (s *SyncResourceSummary) SetResourceType(v string) *SyncResourceSummary {\n\ts.ResourceType = &v\n\treturn s\n}", "func (s *ResourceSharePermissionSummary) SetResourceType(v string) *ResourceSharePermissionSummary {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o GetVpcEndpointServiceResourcesResourceOutput) ResourceType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetVpcEndpointServiceResourcesResource) string { return v.ResourceType }).(pulumi.StringOutput)\n}", "func (f *Filter) AddResourceType(t string) {\n\tconst (\n\t\tresourceType = \"resource.type\"\n\t)\n\tif strings.Contains(f.String(), resourceType) {\n\t\tglog.Fatalf(\"Stackdriver filters may only contain one '%s'\", resourceType)\n\t}\n\tf.add(fmt.Sprintf(\"%s=\\\"%s\\\"\", resourceType, t))\n}", "func (o DatasourceResponseOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v DatasourceResponse) *string { return v.ResourceType }).(pulumi.StringPtrOutput)\n}", "func (o DatasourceSetResponseOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v DatasourceSetResponse) *string { return v.ResourceType }).(pulumi.StringPtrOutput)\n}", "func (s *ResourceSharePermissionDetail) SetResourceType(v string) *ResourceSharePermissionDetail {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o *LabelProperties) GetResourceTypeOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\n\treturn o.ResourceType, true\n}", "func (s *DisassociateInstanceStorageConfigInput) SetResourceType(v string) *DisassociateInstanceStorageConfigInput {\n\ts.ResourceType = &v\n\treturn s\n}", "func (s *Resource) SetResourceType(v string) *Resource {\n\ts.ResourceType = &v\n\treturn s\n}", "func (s *UpdateInstanceStorageConfigInput) SetResourceType(v string) *UpdateInstanceStorageConfigInput {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o DatasourceSetOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v DatasourceSet) *string { return v.ResourceType }).(pulumi.StringPtrOutput)\n}", "func (s *ListPermissionsInput) SetResourceType(v string) *ListPermissionsInput {\n\ts.ResourceType = &v\n\treturn s\n}", "func (s *ListResourcesInput) SetResourceType(v string) *ListResourcesInput {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o DatasourcePtrOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Datasource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ResourceType\n\t}).(pulumi.StringPtrOutput)\n}", "func (o *ListNodesParams) SetResourceType(resourceType *string) {\n\to.ResourceType = resourceType\n}", "func (o *LabelProperties) GetResourceType() *string {\n\tif o == nil {\n\t\treturn nil\n\t}\n\n\treturn o.ResourceType\n\n}", "func (o *BulkCreatePayloadAuthentications) SetResourceType(v string) {\n\to.ResourceType = &v\n}", "func ResourceTypeString(s string) (ResourceType, error) {\n\tif val, ok := _ResourceTypeNameToValueMap[s]; ok {\n\t\treturn val, nil\n\t}\n\treturn 0, fmt.Errorf(\"%s does not belong to ResourceType values\", s)\n}", "func ResourceTypeString(s string) (ResourceType, error) {\n\tif val, ok := _ResourceTypeNameToValueMap[s]; ok {\n\t\treturn val, nil\n\t}\n\treturn 0, fmt.Errorf(\"%s does not belong to ResourceType values\", s)\n}", "func (o *GetComplianceByResourceTypesParams) SetResourceType(resourceType []string) {\n\to.ResourceType = resourceType\n}", "func (s *ListInstanceStorageConfigsInput) SetResourceType(v string) *ListInstanceStorageConfigsInput {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o DatasourceSetResponsePtrOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *DatasourceSetResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ResourceType\n\t}).(pulumi.StringPtrOutput)\n}", "func (s *LFTagPolicyDetails) SetResourceType(v string) *LFTagPolicyDetails {\n\ts.ResourceType = &v\n\treturn s\n}", "func ResourceOfGroupKind(group string, kind string) func(r *Resource) bool {\n\treturn func(r *Resource) bool {\n\t\tkey := r.ResourceKey()\n\t\treturn key.Group == group && key.Kind == kind\n\t}\n}", "func StrToType(v string) ResourceType {\n\treturn typeMap[v]\n}", "func (s *ListPrincipalsInput) SetResourceType(v string) *ListPrincipalsInput {\n\ts.ResourceType = &v\n\treturn s\n}", "func validateFromResourceType(resourceType string) (string, error) {\n\tname, err := k8s.CanonicalResourceNameFromFriendlyName(resourceType)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif name == k8s.Authority {\n\t\treturn \"\", errors.New(\"cannot query traffic --from an authority\")\n\t}\n\treturn name, nil\n}", "func (o DatasourceResponsePtrOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *DatasourceResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ResourceType\n\t}).(pulumi.StringPtrOutput)\n}", "func (o *QuotaSummaryAllOf) SetResourceType(v string) {\n\to.ResourceType = v\n}", "func (s *AssociateInstanceStorageConfigInput) SetResourceType(v string) *AssociateInstanceStorageConfigInput {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o DatasourceSetPtrOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *DatasourceSet) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ResourceType\n\t}).(pulumi.StringPtrOutput)\n}", "func (s *JobError) SetResourceType(v string) *JobError {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o *QuotaSummaryAllOf) GetResourceType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.ResourceType\n}", "func (o *BulkCreatePayloadAuthentications) GetResourceTypeOk() (*string, bool) {\n\tif o == nil || o.ResourceType == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ResourceType, true\n}", "func (*Volume) ResourceType() string {\n\treturn \"Volume\"\n}", "func (s *DescribeInstanceStorageConfigInput) SetResourceType(v string) *DescribeInstanceStorageConfigInput {\n\ts.ResourceType = &v\n\treturn s\n}", "func NewResourceType(specType Type, statusType Type) *ResourceType {\n\tresult := &ResourceType{\n\t\tisStorageVersion: false,\n\t\towner: nil,\n\t\tfunctions: make(map[string]Function),\n\t\ttestcases: make(map[string]TestCase),\n\t\tscope: ResourceScopeResourceGroup,\n\t\tInterfaceImplementer: MakeInterfaceImplementer(),\n\t}\n\n\treturn result.WithSpec(specType).WithStatus(statusType)\n}", "func (s SupportedResourceType) GoString() string {\n\treturn s.String()\n}", "func (s *ErrorDetails) SetResourceType(v string) *ErrorDetails {\n\ts.ResourceType = &v\n\treturn s\n}", "func (s *SupportedResourceType) SetResourceType(v string) *SupportedResourceType {\n\ts.ResourceType = &v\n\treturn s\n}", "func (o ProtectionGroupOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ProtectionGroup) pulumi.StringPtrOutput { return v.ResourceType }).(pulumi.StringPtrOutput)\n}", "func (o SkuResponseOutput) ResourceType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v SkuResponse) string { return v.ResourceType }).(pulumi.StringOutput)\n}", "func (c *podConverter) ResourceType() runtime.Object {\n\treturn c.podResourceType\n}", "func (o SkuResponsePtrOutput) ResourceType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *SkuResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.ResourceType\n\t}).(pulumi.StringPtrOutput)\n}", "func (o VpcNetworkAclAttachmentOutput) ResourceType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *VpcNetworkAclAttachment) pulumi.StringOutput { return v.ResourceType }).(pulumi.StringOutput)\n}", "func (r Resource) Kind() string {\n\treturn r.kind\n}", "func (o *BulkCreatePayloadAuthentications) GetResourceType() string {\n\tif o == nil || o.ResourceType == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.ResourceType\n}", "func (b *AccessReviewRequestBuilder) ResourceType(value string) *AccessReviewRequestBuilder {\n\tb.resourceType = value\n\tb.bitmap_ |= 32\n\treturn b\n}", "func formatResourceName(kind schema.GroupKind, name string, withKind bool) string {\n\tif !withKind || kind.Empty() {\n\t\treturn name\n\t}\n\treturn strings.ToLower(kind.String()) + \"/\" + name\n}", "func (o ContainerServiceOutput) ResourceType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ContainerService) pulumi.StringOutput { return v.ResourceType }).(pulumi.StringOutput)\n}", "func decodeResource(reader io.Reader, kind string) (*TektonResource, error) {\n\n\t// create a duplicate for UniversalDeserializer and NewYAMLToJSONDecoder\n\t// to read from readers\n\tvar dup bytes.Buffer\n\tr := io.TeeReader(reader, &dup)\n\tcontents, err := io.ReadAll(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tobject, gvk, err := scheme.Codecs.UniversalDeserializer().Decode(contents, nil, nil)\n\tif err != nil || !isTektonKind(gvk) {\n\t\treturn nil, fmt.Errorf(\"parse error: invalid resource %+v:\\n%s\", err, contents)\n\t}\n\n\tdecoder := decoder.NewYAMLToJSONDecoder(&dup)\n\n\tvar res *unstructured.Unstructured\n\tfor {\n\t\tres = &unstructured.Unstructured{}\n\t\tif err := decoder.Decode(res); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to decode: %w\", err)\n\t\t}\n\n\t\tif len(res.Object) == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// break at the first object that has a kind and Catalog TEP expects\n\t\t// that the files have only one resource\n\t\tif res.GetKind() != \"\" {\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif k := res.GetKind(); k != kind {\n\t\treturn nil, fmt.Errorf(\"expected kind to be %s but got %s\", kind, k)\n\t}\n\n\tif _, err := convertToTyped(res); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &TektonResource{\n\t\tName: res.GetName(),\n\t\tKind: gvk.Kind,\n\t\tGVK: res.GroupVersionKind(),\n\t\tUnstructured: res,\n\t\tObject: object,\n\t}, nil\n}", "func (s *ServiceNameAndResourceType) SetResourceType(v string) *ServiceNameAndResourceType {\n\ts.ResourceType = &v\n\treturn s\n}", "func (c serverResources) findResourceFromResourceName(gvr schema.GroupVersionResource) (schema.GroupVersionKind, error) {\n\t_, serverGroupsAndResources, err := c.cachedClient.ServerGroupsAndResources()\n\tif err != nil && !strings.Contains(err.Error(), \"Got empty response for\") {\n\t\tif discovery.IsGroupDiscoveryFailedError(err) {\n\t\t\tlogDiscoveryErrors(err)\n\t\t} else if isServerCurrentlyUnableToHandleRequest(err) {\n\t\t\tlogger.V(3).Info(\"failed to find preferred resource version\", \"error\", err.Error())\n\t\t} else {\n\t\t\tlogger.Error(err, \"failed to find preferred resource version\")\n\t\t\treturn schema.GroupVersionKind{}, err\n\t\t}\n\t}\n\tapiResource, err := findResourceFromResourceName(gvr, serverGroupsAndResources)\n\tif err != nil {\n\t\treturn schema.GroupVersionKind{}, err\n\t}\n\treturn schema.GroupVersionKind{Group: apiResource.Group, Version: apiResource.Version, Kind: apiResource.Kind}, err\n}", "func (in *ResourceType) DeepCopy() *ResourceType {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(ResourceType)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewResourceType(regularExpression string, primaryResourceType string, ) *ResourceType {\n\tthis := ResourceType{}\n\tthis.RegularExpression = regularExpression\n\tthis.PrimaryResourceType = primaryResourceType\n\treturn &this\n}", "func (m *AmazonResourceEvidence) GetResourceType()(*string) {\n val, err := m.GetBackingStore().Get(\"resourceType\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (rs *Resource) GetType() *schema.ResourceType {\n\treturn rs.resType\n}", "func (t ResourceType) String() string {\n\treturn C.GoString(C.rd_kafka_ResourceType_name(C.rd_kafka_ResourceType_t(t)))\n}", "func (m *AmazonResourceEvidence) SetResourceType(value *string)() {\n err := m.GetBackingStore().Set(\"resourceType\", value)\n if err != nil {\n panic(err)\n }\n}", "func makeResource(mod string, res string) tokens.Type {\n\tfn := string(unicode.ToLower(rune(res[0]))) + res[1:]\n\treturn makeType(mod+\"/\"+fn, res)\n}", "func newResource(tm unversioned.TypeMetadata) (unversioned.Resource, error) {\n\trh, ok := resourceToType[tm]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"Unknown resource type (%s) and/or version (%s)\", tm.Kind, tm.APIVersion)\n\t}\n\tlog.Debugf(\"Found resource helper: %s\", rh)\n\n\t// Create new resource and fill in the type metadata.\n\tnew := reflect.New(rh)\n\telem := new.Elem()\n\telem.FieldByName(\"Kind\").SetString(tm.GetTypeMetadata().Kind)\n\telem.FieldByName(\"APIVersion\").SetString(tm.GetTypeMetadata().APIVersion)\n\n\treturn new.Interface().(unversioned.Resource), nil\n}", "func ToGodotResource(tscn *parser.TscnFile) (*godot.Resource, error) {\n\tif tscn.Key != TscnTypeGodotResource {\n\t\treturn nil, fmt.Errorf(\"can't convert %s to gd_resource\", tscn.Key)\n\t}\n\n\tres := &godot.Resource{\n\t\tExtResources: make(map[int64]*godot.ExtResource),\n\t\tSubResources: make(map[int64]*godot.SubResource),\n\t\tFields: make(map[string]interface{}),\n\t\tMetaData: godot.MetaData{\n\t\t\tLexerPosition: tscn.Pos,\n\t\t},\n\t}\n\n\tt, err := tscn.GetAttribute(\"type\")\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"gd_resource doesn't contain required attribute type\")\n\t}\n\tif t.String == nil {\n\t\treturn nil, errors.New(\"gd_resource attribute type must be a string\")\n\t}\n\n\tres.Type = *t.String\n\n\t// handle everything that isn't a node\n\tfor _, section := range tscn.Sections {\n\t\t// External resources\n\t\tif section.ResourceType == parser.ResourceTypeExtResource {\n\t\t\tr, err := convertSectionToExtResource(section)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tres.ExtResources[r.ID] = r\n\t\t\tcontinue\n\t\t}\n\n\t\t// Internal resources\n\t\tif section.ResourceType == parser.ResourceTypeSubResource {\n\t\t\tr, err := convertSectionToSubResource(section)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tres.SubResources[r.ID] = r\n\t\t\tcontinue\n\t\t}\n\n\t\t// resource section\n\t\tif section.ResourceType == parser.ResourceTypeResource {\n\t\t\tfor _, field := range section.Fields {\n\t\t\t\tres.Fields[field.Key] = convertGdValue(field.Value)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\t// something else found? Whoops, throw error\n\t\treturn nil, fmt.Errorf(\"invalid resource type found: %s [%s]\", section.ResourceType, section.Pos)\n\t}\n\n\treturn res, nil\n}", "func (d *portworx) GetAlertsUsingResourceTypeBySeverity(resourceType api.ResourceType, severity api.SeverityType) (*api.SdkAlertsEnumerateWithFiltersResponse, error) {\n\t/*\n\t\tresourceType : RESOURCE_TYPE_NONE | RESOURCE_TYPE_VOLUME | RESOURCE_TYPE_NODE | RESOURCE_TYPE_CLUSTER | RESOURCE_TYPE_DRIVE | RESOURCE_TYPE_POOL\n\t\tSeverityType : SEVERITY_TYPE_NONE | SEVERITY_TYPE_ALARM | SEVERITY_TYPE_WARNING | SEVERITY_TYPE_NOTIFY\n\t\te.x :\n\t\t\tvar resourceType api.ResourceType\n\t\t\tvar severity api.SeverityType\n\t\t\tresourceType = api.ResourceType_RESOURCE_TYPE_POOL\n\t\t\tseverity = api.SeverityType_SEVERITY_TYPE_ALARM\n\t\t\talerts, err := Inst().V.GetAlertsUsingResourceTypeBySeverity(resourceType, severity)\n\t*/\n\talerts, err := d.alertsManager.EnumerateWithFilters(d.getContext(), &api.SdkAlertsEnumerateWithFiltersRequest{\n\t\tQueries: []*api.SdkAlertsQuery{\n\t\t\t{\n\t\t\t\tQuery: &api.SdkAlertsQuery_ResourceTypeQuery{\n\t\t\t\t\tResourceTypeQuery: &api.SdkAlertsResourceTypeQuery{\n\t\t\t\t\t\tResourceType: resourceType,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tOpts: []*api.SdkAlertsOption{\n\t\t\t\t\t{Opt: &api.SdkAlertsOption_MinSeverityType{\n\t\t\t\t\t\tMinSeverityType: severity,\n\t\t\t\t\t}},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\talertsResp, err := alerts.Recv()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn alertsResp, nil\n}", "func (mg *MongoCollection) GetTerraformResourceType() string {\n\treturn \"azurerm_cosmosdb_mongo_collection\"\n}", "func (rs *ResourceSubscription) GetResourceType() ResourceType {\n\trs.e.mu.Lock()\n\tdefer rs.e.mu.Unlock()\n\treturn ResourceType(rs.state)\n}", "func ToResourceName(name string) string {\n\tif strings.HasPrefix(name, BuiltinGatewaySecretTypeURI) {\n\t\treturn \"default\"\n\t}\n\t// If they explicitly defined the type, keep it\n\tif strings.HasPrefix(name, KubernetesSecretTypeURI) || strings.HasPrefix(name, kubernetesGatewaySecretTypeURI) {\n\t\treturn name\n\t}\n\t// Otherwise, to kubernetes://\n\treturn KubernetesSecretTypeURI + name\n}", "func (o MachineInstanceSpecResourceRefOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MachineInstanceSpecResourceRef) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func (moduleBase) ResourceType() string {\n\treturn \"compose:module\"\n}", "func resourceTypeName(qualifiedName string) string {\n\ti := strings.LastIndex(qualifiedName, \".\")\n\tif i < 0 {\n\t\treturn qualifiedName\n\t}\n\n\treturn qualifiedName[i+1 : len(qualifiedName)]\n}", "func (resource *ResourceType) SpecType() Type {\n\treturn resource.spec\n}", "func (d DocLanguageHelper) GetDocLinkForResourceType(pkg *schema.Package, modName, typeName string) string {\n\t// The k8s module names contain the domain names. For now we are stripping them off manually so they link correctly.\n\tif modName != \"\" {\n\t\tmodName = strings.ReplaceAll(modName, \".k8s.io\", \"\")\n\t\tmodName = strings.ReplaceAll(modName, \".apiserver\", \"\")\n\t\tmodName = strings.ReplaceAll(modName, \".authorization\", \"\")\n\t}\n\n\tvar path string\n\tvar fqdnTypeName string\n\tswitch {\n\tcase pkg.Name != \"\" && modName != \"\":\n\t\tpath = fmt.Sprintf(\"pulumi_%s/%s\", pkg.Name, modName)\n\t\tfqdnTypeName = fmt.Sprintf(\"pulumi_%s.%s.%s\", pkg.Name, modName, typeName)\n\tcase pkg.Name == \"\" && modName != \"\":\n\t\tpath = modName\n\t\tfqdnTypeName = fmt.Sprintf(\"%s.%s\", modName, typeName)\n\tcase pkg.Name != \"\" && modName == \"\":\n\t\tpath = fmt.Sprintf(\"pulumi_%s\", pkg.Name)\n\t\tfqdnTypeName = fmt.Sprintf(\"pulumi_%s.%s\", pkg.Name, typeName)\n\t}\n\n\treturn fmt.Sprintf(\"/docs/reference/pkg/python/%s/#%s\", path, fqdnTypeName)\n}", "func (*Template) ResourceType() string {\n\treturn \"Template\"\n}", "func (i *Resource) Type() string {\n\treturn i.data.ResourceType\n}", "func (o KubernetesClusterSpecResourceRefOutput) Kind() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterSpecResourceRef) *string { return v.Kind }).(pulumi.StringPtrOutput)\n}", "func NewAzureResourceType(specType Type, statusType Type, typeName TypeName, scope ResourceScope) *ResourceType {\n\tif objectType, ok := specType.(*ObjectType); ok {\n\t\t// We have certain expectations about structure for resources\n\t\tvar nameProperty *PropertyDefinition\n\t\tvar typeProperty *PropertyDefinition\n\t\tvar apiVersionProperty *PropertyDefinition\n\n\t\tisNameOptional := false\n\t\tisTypeOptional := false\n\t\tobjectType.Properties().ForEach(func(property *PropertyDefinition) {\n\t\t\t// force this string because otherwise linter complains thinking it's an enum without an exhaustive switch...\n\t\t\t// It turns out there are other reasons to alias string than just to make an enum, seems like the linter doesn't\n\t\t\t// realize that\n\t\t\tswitch string(property.PropertyName()) {\n\t\t\tcase NameProperty:\n\t\t\t\tnameProperty = property\n\t\t\t\tif _, ok := AsOptionalType(property.PropertyType()); ok {\n\t\t\t\t\tisNameOptional = true\n\t\t\t\t}\n\t\t\tcase TypeProperty:\n\t\t\t\ttypeProperty = property\n\t\t\t\tif _, ok := AsOptionalType(property.PropertyType()); ok {\n\t\t\t\t\tisTypeOptional = true\n\t\t\t\t}\n\t\t\tcase APIVersionProperty:\n\t\t\t\tapiVersionProperty = property\n\t\t\t}\n\t\t})\n\n\t\tif typeProperty == nil {\n\t\t\tpanic(fmt.Sprintf(\"Resource %s is missing type property\", typeName))\n\t\t}\n\n\t\tif nameProperty == nil {\n\t\t\tnameProperty = NewPropertyDefinition(\"Name\", \"name\", StringType)\n\t\t\tnameProperty.WithDescription(\"The name of the resource\")\n\t\t\tisNameOptional = true\n\t\t}\n\n\t\tif apiVersionProperty == nil {\n\t\t\tpanic(fmt.Sprintf(\"Resource %s is missing apiVersion property\", typeName))\n\t\t}\n\n\t\tif isNameOptional {\n\t\t\t// Fix name to be required -- again this is an artifact of bad spec more than anything\n\t\t\tnameProperty = nameProperty.MakeTypeRequired()\n\t\t\tobjectType = objectType.WithProperty(nameProperty)\n\t\t}\n\n\t\t// Fix APIVersion to be required. Technically this isn't due to a bad specification, but in our\n\t\t// case forcing it to required makes our lives simpler (and the vast majority of resources specify\n\t\t// it as required anyway). The only time it's allowed to be optional is if you set apiProfile on\n\t\t// the ARM template instead, which we never do.\n\t\tapiVersionProperty = apiVersionProperty.MakeTypeRequired()\n\t\tobjectType = objectType.WithProperty(apiVersionProperty)\n\n\t\tif isTypeOptional {\n\t\t\ttypeProperty = typeProperty.MakeTypeRequired()\n\t\t\tobjectType = objectType.WithProperty(typeProperty)\n\t\t}\n\n\t\tspecType = objectType\n\t}\n\n\treturn NewResourceType(specType, statusType).WithScope(scope)\n}", "func ToCustomResourceKind(k *v1alpha1.CustomResourceKind) *CustomResourceKind {\n\tresult := &CustomResourceKind{\n\t\tSpec: &CustomResourceKindSpec{},\n\t}\n\tresult.MeshResource = NewCustomResourceKindResource(DefaultAPIVersion, k.Name)\n\tresult.Spec.JSONSchema = k.JsonSchema\n\treturn result\n}", "func getLabelToMatchForKind(kind string) []string {\n\tswitch kind {\n\tcase \"apiservice\": // API Services are not namespaced\n\t\treturn []string{\"apiservice\"}\n\tcase \"customresourcedefinition\": // CRD are not namespaced\n\t\treturn []string{\"customresourcedefinition\"}\n\tcase \"job\": // job metrics use specific label\n\t\treturn []string{\"job_name\", \"namespace\"}\n\tcase \"node\": // persistent nodes are not namespaced\n\t\treturn []string{\"node\"}\n\tcase \"persistentvolume\": // persistent volumes are not namespaced\n\t\treturn []string{\"persistentvolume\"}\n\tdefault:\n\t\treturn []string{kind, \"namespace\"}\n\t}\n}", "func SubresourceName(subresource interface{}, subresourceType SubresourceType) string {\n\tswitch res := subresource.(type) {\n\tcase *regv1.Notary:\n\t\tswitch subresourceType {\n\t\t// Notary DB\n\t\tcase SubTypeNotaryDBPod, SubTypeNotaryDBPVC, SubTypeNotaryDBService:\n\t\t\treturn regv1.K8sPrefix + regv1.K8sNotaryPrefix + NotaryDBPrefix + res.Name\n\n\t\t// Notary Server\n\t\tcase SubTypeNotaryServerIngress, SubTypeNotaryServerPod, SubTypeNotaryServerSecret, SubTypeNotaryServerService:\n\t\t\treturn regv1.K8sPrefix + regv1.K8sNotaryPrefix + NotaryServerPrefix + res.Name\n\n\t\t// Notary signer\n\t\tcase SubTypeNotarySignerPod, SubTypeNotarySignerSecret, SubTypeNotarySignerService:\n\t\t\treturn regv1.K8sPrefix + regv1.K8sNotaryPrefix + NotarySignerPrefix + res.Name\n\t\t}\n\n\tcase *regv1.Registry:\n\t\tswitch subresourceType {\n\t\tcase SubTypeRegistryNotary:\n\t\t\treturn res.Name\n\n\t\tcase SubTypeRegistryService, SubTypeRegistryPVC, SubTypeRegistryDeployment, SubTypeRegistryOpaqueSecret, SubTypeRegistryConfigmap, SubTypeRegistryIngress:\n\t\t\treturn regv1.K8sPrefix + res.Name\n\n\t\tcase SubTypeRegistryTLSSecret:\n\t\t\treturn regv1.K8sPrefix + regv1.TLSPrefix + res.Name\n\n\t\tcase SubTypeRegistryDCJSecret:\n\t\t\treturn regv1.K8sPrefix + regv1.K8sRegistryPrefix + res.Name\n\t\t}\n\n\tcase *regv1.ExternalRegistry:\n\t\tswitch subresourceType {\n\t\tcase SubTypeExternalRegistryLoginSecret:\n\t\t\treturn regv1.K8sPrefix + ExternalRegistryPrefix + LoginSecretPrefix + res.Name\n\t\tcase SubTypeExternalRegistryCronJob:\n\t\t\treturn regv1.K8sPrefix + ExternalRegistryPrefix + res.Name\n\t\tcase SubTypeExternalRegistryJob:\n\t\t\treturn regv1.K8sPrefix + ExternalRegistryPrefix + res.Name + \"-\" + utils.RandomString(10)\n\t\t}\n\n\tcase *regv1.ImageReplicate:\n\t\tswitch subresourceType {\n\t\tcase SubTypeImageReplicateJob:\n\t\t\treturn regv1.K8sPrefix + ImageReplicatePrefix + res.Name\n\t\tcase SubTypeImageReplicateSyncJob:\n\t\t\treturn regv1.K8sPrefix + ImageReplicatePrefix + SynchronizePrefix + res.Name\n\t\tcase SubTypeImageReplicateImageSignRequest:\n\t\t\tif res.Status.ImageSignRequestName != \"\" {\n\t\t\t\treturn res.Status.ImageSignRequestName\n\t\t\t}\n\t\t\treturn regv1.K8sPrefix + ImageReplicatePrefix + res.Name + \"-\" + utils.RandomString(10)\n\t\t}\n\t}\n\n\treturn \"\"\n}", "func detectResourceType(awsIdentityDoc *awsIdentityDocument) Interface {\n\tif awsIdentityDoc != nil {\n\t\treturn createEC2InstanceMonitoredResource(awsIdentityDoc)\n\t}\n\treturn nil\n}", "func kindType(rse string) string {\n\tname := strings.ToLower(rse)\n\tif strings.Contains(name, \"_tape\") || strings.Contains(name, \"_mss\") || strings.Contains(name, \"_export\") {\n\t\treturn \"TAPE\"\n\t}\n\treturn \"DISK\"\n}", "func populateResourceTypes() {\n\tresTypes := []unversioned.Resource{\n\t\tapiv1.NewBGPPeer(),\n\t\tapiv1.NewIPPool(),\n\t\tapiv1.NewHostEndpoint(),\n\t\tapiv1.NewNode(),\n\t\tapiv1.NewPolicy(),\n\t\tapiv1.NewProfile(),\n\t\tapiv1.NewWorkloadEndpoint(),\n\t\tNewK8sNetworkPolicy(),\n\t\tNewK8sNetworkPolicyList(),\n\t}\n\n\tfor _, rt := range resTypes {\n\t\tresourceToType[rt.GetTypeMetadata()] = reflect.ValueOf(rt).Elem().Type()\n\t}\n}", "func GetKind(node *yaml.RNode, path string) string {\n\treturn GetStringField(node, path, \"kind\")\n}", "func (mg *SQLStoredProcedure) GetTerraformResourceType() string {\n\treturn \"azurerm_cosmosdb_sql_stored_procedure\"\n}", "func (dgpr DirectoryGetPropertiesResponse) XMsResourceType() string {\n\treturn PathGetPropertiesResponse(dgpr).XMsResourceType()\n}", "func (d DocLanguageHelper) GetDocLinkForResourceType(pkg *schema.Package, modName, typeName string) string {\n\tvar path string\n\tswitch {\n\tcase pkg.Name != \"\" && modName != \"\":\n\t\tpath = fmt.Sprintf(\"%s/%s\", pkg.Name, modName)\n\tcase pkg.Name == \"\" && modName != \"\":\n\t\tpath = modName\n\tcase pkg.Name != \"\" && modName == \"\":\n\t\tpath = pkg.Name\n\t}\n\ttypeName = strings.ReplaceAll(typeName, \"?\", \"\")\n\treturn fmt.Sprintf(\"/docs/reference/pkg/nodejs/pulumi/%s/#%s\", path, typeName)\n}", "func (mg *MongoDatabase) GetTerraformResourceType() string {\n\treturn \"azurerm_cosmosdb_mongo_database\"\n}", "func (gppr GetPathPropertiesResponse) XMsResourceType() string {\n\treturn gppr.rawResponse.Header.Get(\"x-ms-resource-type\")\n}" ]
[ "0.71807873", "0.69749266", "0.68243074", "0.66789937", "0.63449645", "0.6294564", "0.6266824", "0.6245719", "0.6232153", "0.62168616", "0.6216163", "0.6209334", "0.6199783", "0.6177958", "0.61682874", "0.6163824", "0.6140141", "0.61386055", "0.61364526", "0.6111235", "0.61023086", "0.6080219", "0.6051447", "0.6044955", "0.6042995", "0.60418", "0.60303986", "0.60277814", "0.602754", "0.6013255", "0.60009027", "0.60009027", "0.59984106", "0.5998341", "0.59950835", "0.5992485", "0.597964", "0.5967094", "0.5951472", "0.59474957", "0.594718", "0.5938584", "0.5935711", "0.59180796", "0.5910708", "0.5904602", "0.58888453", "0.587727", "0.58748233", "0.5874236", "0.5864671", "0.58560944", "0.5854819", "0.5845968", "0.5835702", "0.58317107", "0.58194065", "0.58005774", "0.5784644", "0.5772762", "0.57493794", "0.5737566", "0.5734844", "0.5733606", "0.5729773", "0.57234925", "0.5673135", "0.5664855", "0.5654116", "0.5622373", "0.5563361", "0.55531526", "0.5549886", "0.5540148", "0.5539522", "0.55338764", "0.55166525", "0.5513951", "0.5503166", "0.5449591", "0.5430331", "0.54095864", "0.5397659", "0.5345288", "0.53346246", "0.5334246", "0.53233236", "0.53097254", "0.5306387", "0.5296433", "0.52869964", "0.5279043", "0.52777696", "0.52770174", "0.5270212", "0.5268103", "0.52613705", "0.5260166", "0.52592236", "0.52570504" ]
0.70482045
1
Encode will safely encode the id.
func (s SafeID) Encode() ([]byte, error) { id := platform.ID(s) b, _ := id.Encode() return b, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (id ID) Encode() string {\n\tbody := base32i.CheckEncode(id.Bytes())\n\treturn string(append([]byte{'b'}, body...))\n}", "func EncodeId(sentEmailId uint32) string {\n\tvar buf [4]byte\n\tbuf[0] = byte(sentEmailId >> 24)\n\tbuf[1] = byte(sentEmailId >> 16)\n\tbuf[2] = byte(sentEmailId >> 8)\n\tbuf[3] = byte(sentEmailId)\n\tstr := URLEncoding.EncodeToString(buf[:])\n\n\t// Strip out the leading A's so we have nice short ids in the url\n\tstr = strings.TrimLeft(str, \"A\")\n\treturn str\n}", "func (id ID) String() string {\n\treturn base64.URLEncoding.EncodeToString(id[:])\n}", "func (id ID) Encode(dst []byte) []byte {\n\tencode(dst, id[:])\n\treturn dst\n}", "func encode(dst, id []byte) {\n\tencoder.Encode(dst, id)\n}", "func Encode(id int) string {\n\tvar arr []string\n\tvalue := id\n\tfor value > 0 {\n\t\tsymbIndex := value % base\n\t\tif symbIndex < 10 {\n\t\t\tarr = append(arr, strconv.Itoa(symbIndex))\n\t\t} else if symbIndex < 36 {\n\t\t\tarr = append(arr, string(rune(97 + symbIndex - 10)))\n\t\t} else {\n\t\t\tarr = append(arr, string(rune(65 + symbIndex - 10 - 26)))\n\t\t}\n\t\tvalue /= base\n\t}\n\tvar b strings.Builder\n\tfor i := len(arr)-1; i >= 0; i-- {\n\t\tfmt.Fprintf(&b, arr[i])\n\t}\n\treturn b.String()\n}", "func encodeID(val string) string {\n\tif reValidAvatarID.MatchString(val) {\n\t\treturn strings.TrimSuffix(val, imgSfx) // already encoded, strip .image\n\t}\n\treturn store.EncodeID(val)\n}", "func (ed EncodeDecoder) Encode(id int64) string {\n\tid += ed.offset\n\tif id < 58 {\n\t\treturn string(ed.alphabet[id])\n\t}\n\n\tb := make([]byte, 0, 11)\n\tfor id >= 58 {\n\t\tb = append(b, ed.alphabet[id%58])\n\t\tid /= 58\n\t}\n\tb = append(b, ed.alphabet[id])\n\n\tfor x, y := 0, len(b)-1; x < y; x, y = x+1, y-1 {\n\t\tb[x], b[y] = b[y], b[x]\n\t}\n\n\treturn string(b)\n}", "func (id ID) String() string {\n\tvar buf bytes.Buffer\n\tbuf.Grow(21)\n\tenc := base64.NewEncoder(base64.StdEncoding, &buf)\n\tvar bytes [16]byte\n\tbinary.LittleEndian.PutUint64(bytes[0:8], id.High)\n\tbinary.LittleEndian.PutUint64(bytes[8:16], id.Low)\n\tenc.Write(bytes[:])\n\tenc.Close()\n\treturn buf.String()\n}", "func (_Contract *ContractCaller) EncodeTokenId(opts *bind.CallOpts, _x *big.Int, _y *big.Int) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _Contract.contract.Call(opts, out, \"encodeTokenId\", _x, _y)\n\treturn *ret0, err\n}", "func (m Mixer) EncodeID(password string, id uint64) string {\n\treturn m.EncodeIDPadding(password, id, DefaultPaddingLength)\n}", "func (id ID) String() string {\n\ttext := make([]byte, encodedLen)\n\tencode(text, id[:])\n\treturn *(*string)(unsafe.Pointer(&text))\n}", "func IDB58Encode(id ID) string {\n\treturn core.IDB58Encode(id)\n}", "func EncodeUnrestrictedSQLIdent(buf *bytes.Buffer, s string, flags EncodeFlags) {\n\tif flags.HasFlags(EncBareIdentifiers) || IsBareIdentifier(s) {\n\t\tbuf.WriteString(s)\n\t\treturn\n\t}\n\tEncodeEscapedSQLIdent(buf, s)\n}", "func Encode(userID, profileID string) string {\n\tif profileID == \"\" {\n\t\treturn userID\n\t}\n\treturn fmt.Sprintf(\"%s_%s\", userID, profileID)\n}", "func (p *pid) Encode() string {\n\tidStrs := make([]string, len(p.Ids))\n\tfor i, v := range p.Ids {\n\t\tidStrs[i] = fmt.Sprintf(\"%d.%d\", v.Pos, v.AgentId)\n\t}\n\treturn strings.Join(idStrs, \":\") + \"~\" + common.Itoa(p.Seq)\n}", "func (op *insert) Encode() string {\n\treturn fmt.Sprintf(\"i,%s,%s\", op.Pid.Encode(), op.Value)\n}", "func IDB58Encode(id ID) string {\n\treturn b58.Encode([]byte(id))\n}", "func (this *CallID) EncodeBody() string {\n\tif this.callIdentifier == nil {\n\t\treturn \"\"\n\t} else {\n\t\treturn this.callIdentifier.String()\n\t}\n}", "func (s SafeID) String() string {\n\treturn platform.ID(s).String()\n}", "func (u *UUID) Encode() string {\n\tif u == nil {\n\t\tu = &zeroUUID\n\t}\n\treturn fmt.Sprintf(\"%x-%x-%x-%x-%x\", u[0:4], u[4:6], u[6:8], u[8:10], u[10:])\n}", "func (_Contract *ContractCallerSession) EncodeTokenId(_x *big.Int, _y *big.Int) (*big.Int, error) {\n\treturn _Contract.Contract.EncodeTokenId(&_Contract.CallOpts, _x, _y)\n}", "func (id *Id) String() string {\n\treturn strconv.FormatUint(uint64(*id), 16)\n}", "func (_Contract *ContractSession) EncodeTokenId(_x *big.Int, _y *big.Int) (*big.Int, error) {\n\treturn _Contract.Contract.EncodeTokenId(&_Contract.CallOpts, _x, _y)\n}", "func EncodeRestrictedSQLIdent(buf *bytes.Buffer, s string, flags EncodeFlags) {\n\tif flags.HasFlags(EncBareIdentifiers) || (!isReservedKeyword(s) && IsBareIdentifier(s)) {\n\t\tbuf.WriteString(s)\n\t\treturn\n\t}\n\tEncodeEscapedSQLIdent(buf, s)\n}", "func (op *delete) Encode() string {\n\treturn fmt.Sprintf(\"d,%s\", op.Pid.Encode())\n}", "func EncodeDocID(docType string, id []byte) string {\n\treturn fmt.Sprintf(\"%s-%s\", docType, b64encoder.EncodeToString(id))\n}", "func IDHexEncode(id ID) string {\n\treturn core.IDHexEncode(id)\n}", "func identifiedEncode(dst []byte, id ID, t Type) (int, error) {\n\ttotal := 0\n\n\t// check packet id\n\tif !id.Valid() {\n\t\treturn total, makeError(t, \"packet id must be grater than zero\")\n\t}\n\n\t// encode header\n\tn, err := headerEncode(dst[total:], 0, 2, identifiedLen(), t)\n\ttotal += n\n\tif err != nil {\n\t\treturn total, err\n\t}\n\n\t// write packet id\n\tbinary.BigEndian.PutUint16(dst[total:], uint16(id))\n\ttotal += 2\n\n\treturn total, nil\n}", "func Encode(idFile *IdentityFile) ([]byte, error) {\n\tbuf := new(bytes.Buffer)\n\tif err := encodeIdentityFile(buf, idFile); err != nil {\n\t\treturn nil, trace.Wrap(err)\n\t}\n\n\treturn buf.Bytes(), nil\n}", "func (id ID) MarshalText() (text []byte, err error) {\n\ttext = make([]byte, encodedLen)\n\tbase64.URLEncoding.Encode(text, id[:])\n\treturn\n}", "func (id ShortID) String() string {\n\treturn hex.EncodeToString(id)\n}", "func EncodeKey(pk *[32]byte) string {\n\treturn base32.StdEncoding.EncodeToString(pk[:])\n}", "func (id MessageID) String() string {\n\treturn base58.Encode(id[:])\n}", "func (id ID) Pretty() string {\n\treturn IDB58Encode(id)\n}", "func (t InitProducerIdRequest) Encode(e *Encoder, version int16) {\n\te.PutString(t.TransactionalId) // TransactionalId\n\te.PutInt32(t.TransactionTimeoutMs) // TransactionTimeoutMs\n}", "func (r ManagedObjectReference) Encode() string {\n\treturn strings.Join([]string{r.Type, url.QueryEscape(r.Value)}, \"-\")\n}", "func (txid TransactionID) String() string {\n\treturn hex.EncodeToString(txid)\n}", "func (id ID) MarshalJSON() ([]byte, error) {\n\tif id.IsZero() {\n\t\treturn []byte(\"null\"), nil\n\t}\n\ttext := make([]byte, encodedLen+2)\n\tencode(text[1:encodedLen+1], id[:])\n\ttext[0], text[encodedLen+1] = '\"', '\"'\n\treturn text, nil\n}", "func (id ID) Marshal(enc *marshal.Encoder) {\n\tif id < symbols.preInterned {\n\t\tenc.PutByte(0)\n\t\tenc.PutVarint(int64(id))\n\t\treturn\n\t}\n\tenc.PutByte(1)\n\tenc.PutSymbol(id.Str())\n}", "func (id NodeID) Bytes() []byte { return id[:] }", "func (id ID) String() string {\n\tbuf := make([]byte, 10) // This will never be more than 9 bytes.\n\tl := binary.PutUvarint(buf, uint64(id))\n\treturn strings.ToUpper(hex.EncodeToString(buf[:l]))\n}", "func (v CoinIdentifier) EncodeJSON(b []byte) []byte {\n\tb = append(b, `{\"identifier\":`...)\n\tb = json.AppendString(b, v.Identifier)\n\treturn append(b, \"}\"...)\n}", "func encodeUUID(src [16]byte) string {\n\treturn fmt.Sprintf(\"%x-%x-%x-%x-%x\", src[0:4], src[4:6], src[6:8], src[8:10], src[10:16])\n}", "func (id ID) String() string {\n\treturn strconv.Itoa(int(id))\n}", "func (c ConnectionId) EncodeLength() byte {\n\tif len(c) == 0 {\n\t\treturn 0\n\t}\n\tassert(len(c) >= 4 && len(c) <= 18)\n\treturn byte(len(c) - 3)\n}", "func (op *clientInsert) Encode() string {\n\tvar prevPidStr, nextPidStr string\n\tif op.PrevPid != nil {\n\t\tprevPidStr = op.PrevPid.Encode()\n\t}\n\tif op.NextPid != nil {\n\t\tnextPidStr = op.NextPid.Encode()\n\t}\n\treturn fmt.Sprintf(\"ci,%s,%s,%s\", prevPidStr, nextPidStr, op.Value)\n}", "func encodeInt(s string) (json.RawMessage, error) {\n\tv, err := strconv.ParseInt(s, 0, 64)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn json.RawMessage(strconv.FormatInt(v, 10)), nil\n}", "func EncodeEscapedSQLIdent(buf *bytes.Buffer, s string) {\n\tbuf.WriteByte('\"')\n\tstart := 0\n\tfor i, n := 0, len(s); i < n; i++ {\n\t\tch := s[i]\n\t\t// The only character that requires escaping is a double quote.\n\t\tif ch == '\"' {\n\t\t\tif start != i {\n\t\t\t\tbuf.WriteString(s[start:i])\n\t\t\t}\n\t\t\tstart = i + 1\n\t\t\tbuf.WriteByte(ch)\n\t\t\tbuf.WriteByte(ch) // add extra copy of ch\n\t\t}\n\t}\n\tif start < len(s) {\n\t\tbuf.WriteString(s[start:])\n\t}\n\tbuf.WriteByte('\"')\n}", "func EncodeStringID(s string, start uint32) uint64 {\n\tbs := StringToNTBytes(s)\n\tps := unsafe.Pointer(&bs[0]) // #nosec\n\treturn uint64(C.encode_string_id((*C.char)(ps), C.size_t(len(s)), C.size_t(start)))\n}", "func makeAccountID(id int) []byte {\n\treturn []byte(fmt.Sprintf(\"%09d\", id))\n}", "func EncodeInt(i int64) string {\n\treturn strconv.FormatInt(i, 10)\n}", "func (kademliaID *KademliaID) String() string {\n\treturn hex.EncodeToString(kademliaID[0:ID_LEN])\n}", "func (id *Id) Base62String() string {\n\tb := make([]byte, 8)\n\tbinary.LittleEndian.PutUint64(b, uint64(*id))\n\treturn base62.EncodeToString(b)\n}", "func (kademliaID *KademliaID) String() string {\n\treturn hex.EncodeToString(kademliaID[0:IDLength])\n}", "func (kademliaID *KademliaID) String() string {\n\treturn hex.EncodeToString(kademliaID[0:IDLength])\n}", "func (kademliaID *KademliaID) String() string {\n\treturn hex.EncodeToString(kademliaID[0:IDLength])\n}", "func EncodeDocID(id *metapb.DocID) metapb.Key {\n\tkey := encoding.EncodeUvarintAscending(nil, uint64(id.SlotID))\n\tkey = encoding.EncodeUvarintAscending(key, id.SeqNo)\n\treturn key\n}", "func (id ID) String() string {\n\treturn string(id[:])\n}", "func (this *IdBuilder) NewId() (rv string) {\n\tarr := [24]byte{}\n\ts := arr[:]\n\n\t//\n\t// base32 encode all 64 bits of counter\n\t// - encode 5 bits at a time allows us to use base62 routine to get base32\n\t// - we stop at 10 because that's where date/time begins\n\t//\n\tu := atomic.AddUint64(&this.counter, 1)\n\tvar shift uint = 0\n\tfor i := len(arr) - 1; i > 10; i-- {\n\t\tb := byte(u >> shift)\n\t\tshift += 5\n\t\ts[i] = base62(b & 0x1f)\n\t}\n\n\t//\n\t// encode YMMDDHHMMSS\n\t//\n\t// we need the year as these can get databased\n\t//\n\t// we want the MMDDHHMMSS to be easily human readable\n\t//\n\tnow := time.Now().UTC()\n\tyear, month, day := now.Date()\n\thour, minute, second := now.Clock()\n\ts[0] = base62(byte(year - 2016))\n\ts[1], s[2] = base10(byte(month))\n\ts[3], s[4] = base10(byte(day))\n\ts[5], s[6] = base10(byte(hour))\n\ts[7], s[8] = base10(byte(minute))\n\ts[9], s[10] = base10(byte(second))\n\n\treturn string(s)\n}", "func (id ID) MarshalJSON() (_ []byte, err error) {\n\tname, ok := registry[id]\n\tif !ok {\n\t\treturn []byte(fmt.Sprintf(\"%d\", id)), nil\n\t}\n\treturn []byte(fmt.Sprintf(\"%q\", name)), nil\n}", "func (id GID) String() string {\n\treturn fmt.Sprintf(`Hex(\"%x\")`, string(id))\n}", "func (id KeyID) MarshalJSON() ([]byte, error) {\n\tif len(id) == 0 {\n\t\treturn []byte(`\"\"`), nil\n\t}\n\tdst := make([]byte, hex.EncodedLen(len(id))+2)\n\thex.Encode(dst[1:], id)\n\t// need to supply leading/trailing double quotes.\n\tdst[0], dst[len(dst)-1] = '\"', '\"'\n\treturn dst, nil\n}", "func (id *GUID) String() string {\n\treturn hex.EncodeToString(id[:])\n}", "func MarshalID(id external.ObjectID) graphql.Marshaler {\n\treturn graphql.WriterFunc(func(w io.Writer) {\n\t\tio.WriteString(w, strconv.Quote(fmt.Sprintf(\"=%d=\", id)))\n\t})\n}", "func (obj *ObjectBase) MarshalId(m map[string]*json.RawMessage) error {\n\t{\n\t\tvar value json.RawMessage\n\t\tvalue, err := json.Marshal(obj.fq_name)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tm[\"fq_name\"] = &value\n\t}\n\tif len(obj.uuid) > 0 {\n\t\tvar value json.RawMessage\n\t\tvalue, err := json.Marshal(&obj.uuid)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tm[\"uuid\"] = &value\n\t}\n\treturn nil\n}", "func (i ID) Raw() string {\n\treturn string(i)\n}", "func (m simple) Id() []byte {\n\tif m.marshaled == nil {\n\t\tbuf, err := Marshal(m)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tm.marshaled = &buf\n\t}\n\n\treturn m.marshaled.Bytes()\n}", "func (id *Identity) Bytes() []byte { return id[:] }", "func EncodeDocIDToString(id *metapb.DocID) string {\n\treturn strings.Join([]string{strconv.FormatUint(uint64(id.SlotID), 10), strconv.FormatUint(id.SeqNo, 10)}, docIDSeparator)\n}", "func (c ConnectionId) String() string {\n\treturn hex.EncodeToString(c)\n}", "func EncodeUUIDBase64(id *uuid.UUID) string {\n\tif id == nil {\n\t\treturn \"\"\n\t}\n\treturn base64.RawStdEncoding.EncodeToString(id.Bytes())\n}", "func (e *Encoder) encodeInteger(v int64) error {\n\treturn e.encodePrefixed(':', strconv.FormatInt(v, 10))\n}", "func EncodeOnChainExternalJobID(jobID uuid.UUID) [32]byte {\n\tvar ji [32]byte\n\tcopy(ji[:], strings.Replace(jobID.String(), \"-\", \"\", 4))\n\treturn ji\n}", "func (id ID) String() string {\n\treturn fmt.Sprintf(\"ID{EthAddr: %v, NetAddr: %v}\", id.EthAddress, id.NetworkAddress)\n}", "func Encode(value string) string {\n\t// Don't encode empty strings\n\tif value == \"\" {\n\t\treturn \"\"\n\t}\n\tvar buf bytes.Buffer\n\tcompressorWriter, err := flate.NewWriter(&buf, 1) // compression level 1 (fastest)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tcompressorWriter.Write([]byte(value))\n\tcompressorWriter.Close()\n\treturn hex.EncodeToString(buf.Bytes())\n\t//return base32.StdEncoding.EncodeToString(buf.Bytes())\n}", "func (v TransactionIdentifier) EncodeJSON(b []byte) []byte {\n\tb = append(b, `{\"hash\":`...)\n\tb = json.AppendString(b, v.Hash)\n\treturn append(b, \"}\"...)\n}", "func (m Mixer) EncodeIDPadding(password string, id uint64, paddingLen int) string {\n\treturn m.EncodeBase32Padding(password, strconv.FormatUint(id, 10), paddingLen)\n}", "func (t InitProducerIdResponse) Encode(e *Encoder, version int16) {\n\te.PutInt32(t.ThrottleTimeMs) // ThrottleTimeMs\n\te.PutInt16(t.ErrorCode) // ErrorCode\n\te.PutInt64(t.ProducerId) // ProducerId\n\te.PutInt16(t.ProducerEpoch) // ProducerEpoch\n}", "func (pp *Pubrel) Encode(dst []byte) (int, error) {\n\treturn identifiedEncode(dst, pp.ID, PUBREL)\n}", "func Encode(number int) (string, error) {\n\tif number == 0 {\n\t\treturn \"0\", nil\n\t}\n\n\tstr := \"\"\n\tfor number > 0 {\n\t\tdigit := number % 62\n\t\tchr, err := dehydrate(digit)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tstr = chr + str\n\t\tnumber = int(number / 62)\n\t}\n\treturn str, nil\n}", "func (k *Key) Encode() string {\n\te := make([]*pb.Path_Element, len(k.toks))\n\tfor i, t := range k.toks {\n\t\tt := t\n\t\te[i] = &pb.Path_Element{\n\t\t\tType: &t.Kind,\n\t\t}\n\t\tif t.StringID != \"\" {\n\t\t\te[i].Name = &t.StringID\n\t\t} else {\n\t\t\te[i].Id = &t.IntID\n\t\t}\n\t}\n\tvar namespace *string\n\tif ns := k.kc.Namespace; ns != \"\" {\n\t\tnamespace = &ns\n\t}\n\tr, err := proto.Marshal(&pb.Reference{\n\t\tApp: &k.kc.AppID,\n\t\tNameSpace: namespace,\n\t\tPath: &pb.Path{\n\t\t\tElement: e,\n\t\t},\n\t})\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// trim padding\n\treturn strings.TrimRight(base64.URLEncoding.EncodeToString(r), \"=\")\n}", "func (id NodeID) String() string {\n\tunversionedID := id.unversioned()\n\treturn base58.CheckEncode(unversionedID[:], byte(id.Version().Number))\n}", "func (id *ID) String() string {\n\treturn strings.Replace((*uuid.UUID)(id).String(), \"-\", \"\", -1)\n}", "func encodeCursor(i int) graphql.ID {\n\treturn graphql.ID(base64.StdEncoding.EncodeToString([]byte(fmt.Sprintf(\"cursor%d\", i))))\n}", "func EncodeFID(fid int64) string {\n\treturn base62.Encode(fid)\n}", "func (id ID) Hex() string {\n\treturn hex.EncodeToString(id[:])\n}", "func (x XID) String() string {\n\tdst := make([]byte, 20)\n\tx.encode(dst)\n\treturn b2s(dst)\n}", "func (e *Encode) Encode(s []int64) (string, error) {\n\tif e == nil || e.h == nil {\n\t\treturn \"\", ErrInvalidInstancePointer\n\t}\n\treturn e.h.EncodeInt64(s)\n}", "func (i ID) MarshalJSON() ([]byte, error) {\n\treturn []byte(fmt.Sprintf(\n\t\t\"%016q\", strconv.FormatUint(uint64(i), 16),\n\t)), nil\n}", "func (tid TransactionID) String() string {\n\treturn fmt.Sprintf(\"%x\", tid[:])\n}", "func (j *Job) GetIDShort() string { x, _ := hex.DecodeString(j.ID); return hex.EncodeToString(x[:8]) }", "func (kv *DisKV) encodeKey(key string) string {\n\treturn base32.StdEncoding.EncodeToString([]byte(key))\n}", "func (pkt *SubAddRequest) Encode(buffer *bytes.Buffer) (xID uint32) {\n\txID = XID()\n\tXdrPutInt32(buffer, int32(pkt.ID()))\n\tXdrPutUint32(buffer, xID)\n\tXdrPutString(buffer, pkt.Expression)\n\tXdrPutBool(buffer, pkt.AcceptInsecure)\n\tXdrPutKeys(buffer, pkt.Keys)\n\n\treturn\n}", "func (ep *epInfoCache) StringID() string {\n\treturn strconv.FormatUint(ep.id, 10)\n}", "func (e *Encoder) Encode(v interface{}) error {\n\treturn e.EncodeContext(context.Background(), v)\n}", "func (id ObjectId) Hex() string {\n\treturn hex.EncodeToString([]byte(id))\n}", "func toId(str string) string {\n\treturn IdRegex.ReplaceAllString(strings.ToLower(str), \"\")\n}", "func (id UUID) EncodeString(dst []byte) int {\n\tconst srcBase = 0x100000000\n\tconst dstBase = 62\n\n\tparts := [4]uint32{\n\t\tuint32(id[0])<<24 | uint32(id[1])<<16 | uint32(id[2])<<8 | uint32(id[3]),\n\t\tuint32(id[4])<<24 | uint32(id[5])<<16 | uint32(id[6])<<8 | uint32(id[7]),\n\t\tuint32(id[8])<<24 | uint32(id[9])<<16 | uint32(id[10])<<8 | uint32(id[11]),\n\t\tuint32(id[12])<<24 | uint32(id[13])<<16 | uint32(id[14])<<8 | uint32(id[15]),\n\t}\n\n\tn := len(dst)\n\tbp := parts[:]\n\tbq := [4]uint32{}\n\tdst[0] = '0'\n\n\tfor len(bp) != 0 {\n\t\tquotient := bq[:0]\n\t\tremainder := uint64(0)\n\n\t\tfor _, c := range bp {\n\t\t\tvalue := uint64(c) + uint64(remainder)*srcBase\n\t\t\tdigit := value / dstBase\n\t\t\tremainder = value % dstBase\n\n\t\t\tif len(quotient) != 0 || digit != 0 {\n\t\t\t\tquotient = append(quotient, uint32(digit))\n\t\t\t}\n\t\t}\n\n\t\t// Writes at the end of the destination buffer because we computed the\n\t\t// lowest bits first.\n\t\tn--\n\t\tdst[n] = base62Characters[remainder]\n\t\tbp = quotient\n\t}\n\n\treturn n\n}", "func (id ID) Bytes() []byte {\n\treturn id[:]\n}" ]
[ "0.77108073", "0.66864574", "0.66426456", "0.65644044", "0.6460869", "0.6443142", "0.6401249", "0.6338997", "0.62644243", "0.62184376", "0.6162067", "0.6094737", "0.6081955", "0.6073592", "0.6046869", "0.6023225", "0.599918", "0.59980166", "0.5976361", "0.5976155", "0.592403", "0.5907201", "0.5899911", "0.58560824", "0.5845897", "0.58250797", "0.58124983", "0.57852995", "0.5764488", "0.576138", "0.5737418", "0.5716009", "0.57130796", "0.5678257", "0.5663832", "0.56487155", "0.56374836", "0.5635045", "0.5624878", "0.56088614", "0.5576479", "0.5570211", "0.5549794", "0.5548357", "0.5516655", "0.54677355", "0.54581594", "0.5451906", "0.54489124", "0.5445619", "0.5445326", "0.5439322", "0.543005", "0.540628", "0.540064", "0.540064", "0.540064", "0.53946704", "0.5384873", "0.5380191", "0.53600013", "0.5321036", "0.5308839", "0.5301163", "0.5299299", "0.5292595", "0.5290939", "0.52907705", "0.528625", "0.52807325", "0.52766895", "0.5265031", "0.5260539", "0.5258529", "0.5258014", "0.5247816", "0.5246891", "0.5243073", "0.52423286", "0.5241124", "0.5239766", "0.5237229", "0.5231692", "0.5222784", "0.52087766", "0.52062535", "0.5200864", "0.5190246", "0.5183912", "0.5181981", "0.5180101", "0.5172702", "0.51680547", "0.5165741", "0.51573205", "0.51549023", "0.5152809", "0.5149226", "0.51418066", "0.5137293" ]
0.77954394
0
String prints a encoded string representation of the id.
func (s SafeID) String() string { return platform.ID(s).String() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (id ID) String() string {\n\treturn base64.URLEncoding.EncodeToString(id[:])\n}", "func (id *Id) String() string {\n\treturn strconv.FormatUint(uint64(*id), 16)\n}", "func (id ID) String() string {\n\ttext := make([]byte, encodedLen)\n\tencode(text, id[:])\n\treturn *(*string)(unsafe.Pointer(&text))\n}", "func (id ID) String() string {\n\tbuf := make([]byte, 10) // This will never be more than 9 bytes.\n\tl := binary.PutUvarint(buf, uint64(id))\n\treturn strings.ToUpper(hex.EncodeToString(buf[:l]))\n}", "func (id ID) String() string {\n\treturn strconv.Itoa(int(id))\n}", "func (id ID) String() string {\n\treturn fmt.Sprintf(\"ID{EthAddr: %v, NetAddr: %v}\", id.EthAddress, id.NetworkAddress)\n}", "func (id ID) String() string {\n\tvar buf bytes.Buffer\n\tbuf.Grow(21)\n\tenc := base64.NewEncoder(base64.StdEncoding, &buf)\n\tvar bytes [16]byte\n\tbinary.LittleEndian.PutUint64(bytes[0:8], id.High)\n\tbinary.LittleEndian.PutUint64(bytes[8:16], id.Low)\n\tenc.Write(bytes[:])\n\tenc.Close()\n\treturn buf.String()\n}", "func (id GID) String() string {\n\treturn fmt.Sprintf(`Hex(\"%x\")`, string(id))\n}", "func (id ID) String() string {\n\treturn string(id[:])\n}", "func (id ID) String() string {\n\tgen, seq := id>>idBits, id&idSeqMask\n\tif gen == 0 {\n\t\tif seq != 0 {\n\t\t\treturn fmt.Sprintf(\"INVALID_ZeroID(seq:%d)\", gen)\n\t\t}\n\t\treturn \"ZeroID\"\n\t}\n\treturn fmt.Sprintf(\"%d(gen:%d)\", seq, gen)\n}", "func (scoid SiacoinOutputID) String() string {\n\treturn fmt.Sprintf(\"%x\", scoid[:])\n}", "func (id ShortID) String() string {\n\treturn hex.EncodeToString(id)\n}", "func (id MessageID) String() string {\n\treturn base58.Encode(id[:])\n}", "func (kademliaID *KademliaID) String() string {\n\treturn hex.EncodeToString(kademliaID[0:ID_LEN])\n}", "func (kademliaID *KademliaID) String() string {\n\treturn hex.EncodeToString(kademliaID[0:IDLength])\n}", "func (kademliaID *KademliaID) String() string {\n\treturn hex.EncodeToString(kademliaID[0:IDLength])\n}", "func (kademliaID *KademliaID) String() string {\n\treturn hex.EncodeToString(kademliaID[0:IDLength])\n}", "func (id SessionId) String() string {\n\treturn fmt.Sprintf(\"%x-%x-%x-%x-%x\", id[0:4], id[4:6], id[6:8], id[8:10], id[10:])\n}", "func (i ID) String() string {\n\treturn string(i)\n}", "func (x XID) String() string {\n\tdst := make([]byte, 20)\n\tx.encode(dst)\n\treturn b2s(dst)\n}", "func (oid OutputID) String() string {\n\treturn fmt.Sprintf(\"%x\", oid[:])\n}", "func (i *ID) String() string {\n\treturn string(*i)\n}", "func (tid TransactionID) String() string {\n\treturn fmt.Sprintf(\"%x\", tid[:])\n}", "func (id *ID) String() string {\n\treturn strings.Replace((*uuid.UUID)(id).String(), \"-\", \"\", -1)\n}", "func (id ID) String() string {\n\tname, ok := registry[id]\n\tif !ok {\n\t\treturn fmt.Sprintf(\"%d\", id)\n\t}\n\treturn fmt.Sprintf(\"%s\", name)\n}", "func (n NodesID) String() string {\n\treturn fmt.Sprintf(\"%x\", n[:])\n}", "func (id *GUID) String() string {\n\treturn hex.EncodeToString(id[:])\n}", "func (v ID) String() string {\n\treturn v.v\n}", "func (id NodeID) String() string {\n\tunversionedID := id.unversioned()\n\treturn base58.CheckEncode(unversionedID[:], byte(id.Version().Number))\n}", "func (sid *Shortid) String() string {\n\treturn fmt.Sprintf(\"Shortid(worker=%v, epoch=%v, abc=%v)\", sid.worker, sid.epoch, sid.abc)\n}", "func (i ID) String() string {\n\treturn fmt.Sprintf(\"<round:%d,category:%s>\", i.Round(), i.Category().String())\n}", "func (c client) String() string {\n\tmsg := fmt.Sprintf(\"Id: %d\", c.id)\n\treturn msg\n}", "func (n NodeID) String() string {\n\treturn fmt.Sprintf(\"%x\", n[:])\n}", "func (sID SemanticID) String() string {\n\tif sID.IsNil() {\n\t\treturn \"\"\n\t}\n\n\treturn strings.Join([]string{sID.Namespace, sID.Collection, sID.ID}, Separator)\n}", "func (n NodeID) String() string {\n\treturn hexutil.Encode(n[:])\n}", "func (id ID) String() string {\n\tswitch id {\n\tcase Version:\n\t\treturn \"RT_VERSION\"\n\tdefault:\n\t\treturn \"\"\n\t}\n}", "func (nodeID *NodeID) String() string {\n\treturn hex.EncodeToString(nodeID[0:IDLength])\n}", "func (id ID) Pretty() string {\n\treturn IDB58Encode(id)\n}", "func (txid TransactionID) String() string {\n\treturn hex.EncodeToString(txid)\n}", "func (gID *GameID) String() string {\n\treturn fmt.Sprintf(\"%s:%s:%s\", gID.Prefix, gID.Source, gID.ID)\n}", "func (s SlotId) String() string {\n\treturn fmt.Sprintf(\"%s (%x)\", s.Name, s.Key)\n}", "func (t AuthenticatorID) String() string {\n\treturn string(t)\n}", "func (bid BlockID) String() string {\n\treturn fmt.Sprintf(\"%x\", bid[:])\n}", "func (p ObjectID) String() string {\n\treturn p.ID()\n}", "func (id UUID) String() string {\n\treturn string(id)\n}", "func (id UUID) String() string {\n\tvar buf [StringMaxLen]byte\n\tn := id.EncodeString(buf[:])\n\treturn string(buf[n:])\n}", "func (mwi *MessageWithID) String() string {\n\tvar builder strings.Builder\n\tbuilder.WriteString(\"MessageWithID(\")\n\tbuilder.WriteString(fmt.Sprintf(\"id=%v\", mwi.ID))\n\tbuilder.WriteByte(')')\n\treturn builder.String()\n}", "func (c ConnectionId) String() string {\n\treturn hex.EncodeToString(c)\n}", "func (a AccountID) String() string {\n\treturn string(a)\n}", "func (t TransactionID) String() string {\n\treturn string(t)\n}", "func (id *HashChainLinkIdentity) String() string {\n\tif id == nil {\n\t\treturn \"\"\n\t}\n\n\tswitch id.Type() {\n\tcase IdentityTypeLegacyID:\n\t\treturn fmt.Sprintf(\"'%s' (legacy)\", id.clientID)\n\tcase IdentityTypeMetadata:\n\t\treturn fmt.Sprintf(\"Client ID: '%s'; Machine ID: '%s'; Sequence number: %d; Request time: %d\",\n\t\t\tid.clientID, id.machineID, id.sequenceNr, id.requestTime)\n\tdefault:\n\t\treturn \"Unknown\"\n\t}\n}", "func (s StructureID) String() string {\n\treturn string(s[:])\n}", "func (id IscsiTargetId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Resource Group Name: %q\", id.ResourceGroupName),\n\t\tfmt.Sprintf(\"Disk Pool Name: %q\", id.DiskPoolName),\n\t\tfmt.Sprintf(\"Iscsi Target Name: %q\", id.IscsiTargetName),\n\t}\n\treturn fmt.Sprintf(\"Iscsi Target (%s)\", strings.Join(components, \"\\n\"))\n}", "func (id ID) String() string {\n\tif id.IsZero() {\n\t\treturn \"\"\n\t}\n\n\treturn id.URL().String()\n}", "func (id PacketCoreControlPlaneId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Resource Group Name: %q\", id.ResourceGroupName),\n\t\tfmt.Sprintf(\"Packet Core Control Plane Name: %q\", id.PacketCoreControlPlaneName),\n\t}\n\treturn fmt.Sprintf(\"Packet Core Control Plane (%s)\", strings.Join(components, \"\\n\"))\n}", "func (s *Siegfried) String() string {\n\tstr := fmt.Sprintf(\n\t\t\"%s (%v)\\nidentifiers: \\n\",\n\t\tconfig.Signature(),\n\t\ts.C.Format(time.RFC3339))\n\tfor _, id := range s.ids {\n\t\td := id.Describe()\n\t\tstr += fmt.Sprintf(\" - %v: %v\\n\", d[0], d[1])\n\t}\n\treturn str\n}", "func (id ScheduleID) String() string {\n\treturn fmt.Sprintf(\"%d.%d.%d\", id.Shard, id.Realm, id.Schedule)\n}", "func (id SubscriptionId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Group: %q\", id.GroupId),\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t}\n\treturn fmt.Sprintf(\"Subscription (%s)\", strings.Join(components, \"\\n\"))\n}", "func (id NodeConfigurationId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Resource Group Name: %q\", id.ResourceGroupName),\n\t\tfmt.Sprintf(\"Automation Account Name: %q\", id.AutomationAccountName),\n\t\tfmt.Sprintf(\"Node Configuration Name: %q\", id.NodeConfigurationName),\n\t}\n\treturn fmt.Sprintf(\"Node Configuration (%s)\", strings.Join(components, \"\\n\"))\n}", "func (c identifier) String() string {\n\treturn string(c)\n}", "func (ep *epInfoCache) StringID() string {\n\treturn strconv.FormatUint(ep.id, 10)\n}", "func (blockID BlockID) String() string {\n\treturn fmt.Sprintf(`%v:%v`, blockID.Hash, blockID.PartSetHeader)\n}", "func (blockID BlockID) String() string {\n\treturn fmt.Sprintf(`%v:%v`, blockID.Hash, blockID.PartSetHeader)\n}", "func (id WorkflowId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Resource Group Name: %q\", id.ResourceGroupName),\n\t\tfmt.Sprintf(\"Workflow Name: %q\", id.WorkflowName),\n\t}\n\treturn fmt.Sprintf(\"Workflow (%s)\", strings.Join(components, \"\\n\"))\n}", "func (id *ResourceID) String() string {\n\tif len(id.stringValue) > 0 {\n\t\treturn id.stringValue\n\t}\n\n\tif id.Parent == nil {\n\t\treturn \"\"\n\t}\n\n\tbuilder := strings.Builder{}\n\tbuilder.WriteString(id.Parent.String())\n\n\tif id.isChild {\n\t\tbuilder.WriteString(fmt.Sprintf(\"/%s\", id.ResourceType.lastType()))\n\t\tif len(id.Name) > 0 {\n\t\t\tbuilder.WriteString(fmt.Sprintf(\"/%s\", id.Name))\n\t\t}\n\t} else {\n\t\tbuilder.WriteString(fmt.Sprintf(\"/providers/%s/%s/%s\", id.ResourceType.Namespace, id.ResourceType.Type, id.Name))\n\t}\n\n\tid.stringValue = builder.String()\n\n\treturn id.stringValue\n}", "func (id TopicID) String() string {\n\treturn fmt.Sprintf(\"%d.%d.%d\", id.Shard, id.Realm, id.Topic)\n}", "func (m MarkerID) String() string {\n\treturn string(m)\n}", "func (id DeploymentID) String() string {\n\treturn fmt.Sprintf(\"%s/%d\", id.Owner, id.DSeq)\n}", "func IDToStr(id ID) (string, error) {\n\treturn string(id), nil\n}", "func (id StreamingLocatorId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Resource Group Name: %q\", id.ResourceGroupName),\n\t\tfmt.Sprintf(\"Media Service Name: %q\", id.MediaServiceName),\n\t\tfmt.Sprintf(\"Streaming Locator Name: %q\", id.StreamingLocatorName),\n\t}\n\treturn fmt.Sprintf(\"Streaming Locator (%s)\", strings.Join(components, \"\\n\"))\n}", "func (id *ObjectID) String() string {\n\treturn id.Value.String()\n}", "func (fcid FileContractID) String() string {\n\treturn fmt.Sprintf(\"%x\", fcid[:])\n}", "func (stateID StateID) String() string {\n\treturn fmt.Sprintf(`%v`, stateID.LastAppHash)\n}", "func (a *L3n4Addr) StringID() string {\n\t// This does not include the protocol right now as the datapath does\n\t// not include the protocol in the lookup of the service IP.\n\treturn a.String()\n}", "func (id ResourceGroupProviderId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Resource Group Name: %q\", id.ResourceGroupName),\n\t\tfmt.Sprintf(\"Provider Name: %q\", id.ProviderName),\n\t\tfmt.Sprintf(\"Resource Parent Type: %q\", id.ResourceParentType),\n\t\tfmt.Sprintf(\"Resource Parent Name: %q\", id.ResourceParentName),\n\t\tfmt.Sprintf(\"Resource Type: %q\", id.ResourceType),\n\t\tfmt.Sprintf(\"Resource Name: %q\", id.ResourceName),\n\t}\n\treturn fmt.Sprintf(\"Resource Group Provider (%s)\", strings.Join(components, \"\\n\"))\n}", "func (s CustomVocabularyEntryId) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (id GroupID) String() string {\n\treturn fmt.Sprintf(\"%s/%d\", id.DeploymentID(), id.GSeq)\n}", "func (p PortalID) String() string {\n\treturn string(p)\n}", "func (eid EID) String() string {\n\treturn \"EID: \" + strconv.FormatUint(uint64(eid), 10)\n}", "func (id JobID) String() string { return uuid.UUID(id).String() }", "func (t RequestID) String() string {\n\treturn string(t)\n}", "func (objectIDs *DbObjectIDs) String() string {\n\tid := objectIDs.ownerControllerName + \":\" + string(objectIDs.idsType.ownerObjectType)\n\tfor _, key := range objectIDs.idsType.GetExternalIDKeys() {\n\t\tid += \":\" + objectIDs.objectIDs[key]\n\t}\n\treturn id\n}", "func (tgid TelegramID) String() string {\n\treturn strconv.Itoa(int(tgid))\n}", "func (id DeletedManagedHSMId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Location Name: %q\", id.LocationName),\n\t\tfmt.Sprintf(\"Deleted Managed H S M Name: %q\", id.DeletedManagedHSMName),\n\t}\n\treturn fmt.Sprintf(\"Deleted Managed H S M (%s)\", strings.Join(components, \"\\n\"))\n}", "func (pp *Pubrec) String() string {\n\treturn fmt.Sprintf(\"<Pubrec ID=%d>\", pp.ID)\n}", "func (sid SpanID) HexString() string {\n\tif sid.IsEmpty() {\n\t\treturn \"\"\n\t}\n\treturn hex.EncodeToString(sid.id[:])\n}", "func (u UID) String() string {\n\treturn fmt.Sprintf(\"UID(%s)\", u.RDF())\n}", "func (v ServiceID) String() string {\n\tx := (int32)(v)\n\n\treturn fmt.Sprint(x)\n}", "func (id OfferPlanId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Publisher: %q\", id.PublisherId),\n\t\tfmt.Sprintf(\"Offer: %q\", id.OfferId),\n\t\tfmt.Sprintf(\"Plan: %q\", id.PlanId),\n\t}\n\treturn fmt.Sprintf(\"Offer Plan (%s)\", strings.Join(components, \"\\n\"))\n}", "func (ulid ULID) String() string {\n\tvar (\n\t\tbuf [26]byte\n\t\tx, y byte\n\t)\n\t// Backups [6] and [7] bytes to override them with their original values later.\n\tx, y, ulid[6], ulid[7] = ulid[6], ulid[7], x, y\n\ttimestamp := int64(binary.LittleEndian.Uint64(ulid[:8]))\n\t// This is useful to shave some nanoseconds from copy() operations.\n\tulid[6], ulid[7] = x, y\n\tfor x := encodedTimeLength - 1; x >= 0; x-- {\n\t\tmod := timestamp % alphabetSize\n\t\tbuf[x] = alphabet[mod]\n\t\ttimestamp = (timestamp - mod) / alphabetSize\n\t}\n\tfor x := encodedTimeLength; x < len(ulid); x++ {\n\t\tbuf[x] = alphabet[int64(ulid[x])%alphabetSize]\n\t}\n\treturn string(buf[:])\n}", "func (id SoldierID) String() string {\n\treturn fmt.Sprintf(\"%s (%s)\", id.Name, id.Faction)\n}", "func (s ServiceId) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (u UUID) String() string {\n\treturn fmt.Sprintf(\"#%d#%s.%s_%d\", u.Type, u.Origin, u.ID, u.Tick)\n}", "func (s ResourceIdentifier) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (p ProcessID) String() string {\n\treturn fmt.Sprintf(\"<%s:%s>\", p.Name, p.Node)\n}", "func (v ModuleID) String() string {\n\tx := (int32)(v)\n\n\treturn fmt.Sprint(x)\n}", "func (p PaymentID) String() string {\n\treturn strconv.FormatInt(p.ProjectID, 10) + \"-\" + strconv.FormatInt(p.PaymentID, 10)\n}", "func (g *GUID) String() string {\n\treturn fmt.Sprintf(\"%x-%x-%x-%x-%x\", g[0:4], g[4:6], g[6:8], g[8:10], g[10:])\n}", "func (i ID) String() string {\n\treturn names[i]\n}", "func (pp *Pubrel) String() string {\n\treturn fmt.Sprintf(\"<Pubrel ID=%d>\", pp.ID)\n}" ]
[ "0.8165663", "0.8161382", "0.8121128", "0.80728704", "0.80641985", "0.80392027", "0.7919137", "0.79033333", "0.7877943", "0.77691376", "0.7767631", "0.77567667", "0.7708006", "0.7594143", "0.75821745", "0.75821745", "0.75821745", "0.75687236", "0.7556296", "0.75527304", "0.75484645", "0.75422704", "0.7422379", "0.7410528", "0.7396791", "0.7395478", "0.7354025", "0.7334391", "0.73196536", "0.73115", "0.7308226", "0.72723573", "0.7265665", "0.72553056", "0.72423524", "0.7213874", "0.71939576", "0.7172145", "0.7165598", "0.7159519", "0.7142476", "0.7065682", "0.70556617", "0.7054565", "0.70284474", "0.70278525", "0.7002474", "0.6986691", "0.6939521", "0.6935491", "0.69156796", "0.6905865", "0.6904306", "0.6903368", "0.68992406", "0.68961334", "0.6886858", "0.68718654", "0.68613416", "0.68536645", "0.68476355", "0.68475825", "0.68475825", "0.6817735", "0.68068427", "0.68004686", "0.67826897", "0.67806077", "0.6775583", "0.67676747", "0.6759956", "0.67572874", "0.675571", "0.67495286", "0.67467344", "0.6745991", "0.6742222", "0.6737825", "0.6730688", "0.67277235", "0.6725874", "0.6723963", "0.6718095", "0.6687837", "0.6665714", "0.66568655", "0.665539", "0.66403574", "0.6639967", "0.6639647", "0.66278446", "0.6612773", "0.6602992", "0.65915114", "0.65846074", "0.6566819", "0.6550866", "0.6546891", "0.6534382", "0.65335673" ]
0.741215
23
IsNew indicates the resource is new to the platform.
func (d DiffIdentifier) IsNew() bool { return d.ID == 0 }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (me TxsdShow) IsNew() bool { return me == \"new\" }", "func (me TxsdShow) IsNew() bool { return me.String() == \"new\" }", "func (c Customer) IsNew() bool {\n\treturn c.ID == 0\n}", "func (c *Checkpoint) IsNew() bool {\n\treturn c.State == ActiveState && c.ID == 0 && c.EndTime == nil &&\n\t\tc.UUID == nil && len(c.Resources) == 0 && len(c.Metadata) == 0\n}", "func (rec *Recorder) IsNewCassette() bool {\n\treturn rec.cassette.IsNew\n}", "func (s *Step) IsNew() bool {\n\treturn s.State == ActiveState && s.EndTime == nil && len(s.Metrics) == 0\n}", "func (obj *Edge) GetIsNew() bool {\n\treturn obj.getIsNew()\n}", "func (v *Validation) IsNew() bool {\n\treturn v.State == ActiveState && v.ID == 0 && v.EndTime == nil && len(v.Metrics) == 0\n}", "func (p PodStatusInformation) IsNew(lastSeen PodStatusInformation, timeSince int) bool {\n\n\t// assume not a failure\n\tif p.Image == \"\" || p.ContainerName == \"\" || p.FinishedAt.IsZero() {\n\t\treturn false\n\t}\n\n\t// Check to see if its been over 'x' minutes, if so its new yet again.\n\tif ok := p.timeCheck(lastSeen, timeSince); ok {\n\t\treturn true\n\t}\n\n\t// Identical\n\tif reflect.DeepEqual(p, lastSeen) {\n\n\t\treturn false\n\t}\n\n\tif p.PodName == lastSeen.PodName && p.ContainerName == lastSeen.PodName {\n\n\t\treturn false\n\t}\n\n\t// Same pod, same start time\n\tif p.PodName == lastSeen.PodName && p.StartedAt == lastSeen.StartedAt {\n\n\t\treturn false\n\t}\n\n\t// same container, same exit code\n\tif p.ContainerName == lastSeen.ContainerName && p.ExitCode == lastSeen.ExitCode {\n\n\t\treturn false\n\t}\n\n\t// same container, same exit code\n\tif p.PodName == lastSeen.PodName && p.ExitCode == lastSeen.ExitCode {\n\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (rfp *RumoursFromPeer) IsNew(rm RumourMessage) bool {\n\tid := rm.ID\n\tif id <= rfp.Count {\n\t\treturn false\n\t}\n\treturn true\n}", "func (me TactionType) IsStatusNewInfo() bool { return me.String() == \"status-new-info\" }", "func (sp SnapshotPolicy) IsAllNew() bool { return sp == SnapshotPolicyAllNew || sp == \"\" }", "func (obj *Edge) SetIsNew(flag bool) {\n\tobj.setIsNew(flag)\n}", "func (p *Position) IsNew() bool {\n\treturn p.EntranceOrder() == nil && p.ExitOrder() == nil\n}", "func (c *CanaryDeployer) IsNewSpec(cd *flaggerv1.Canary) (bool, error) {\n\ttargetName := cd.Spec.TargetRef.Name\n\tcanary, err := c.kubeClient.AppsV1().Deployments(cd.Namespace).Get(targetName, metav1.GetOptions{})\n\tif err != nil {\n\t\tif errors.IsNotFound(err) {\n\t\t\treturn false, fmt.Errorf(\"deployment %s.%s not found\", targetName, cd.Namespace)\n\t\t}\n\t\treturn false, fmt.Errorf(\"deployment %s.%s query error %v\", targetName, cd.Namespace, err)\n\t}\n\n\tif cd.Status.LastAppliedSpec == \"\" {\n\t\treturn true, nil\n\t}\n\n\tnewSpec := &canary.Spec.Template.Spec\n\toldSpecJson, err := base64.StdEncoding.DecodeString(cd.Status.LastAppliedSpec)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"%s.%s decode error %v\", cd.Name, cd.Namespace, err)\n\t}\n\toldSpec := &corev1.PodSpec{}\n\terr = json.Unmarshal(oldSpecJson, oldSpec)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"%s.%s unmarshal error %v\", cd.Name, cd.Namespace, err)\n\t}\n\n\tif diff := cmp.Diff(*newSpec, *oldSpec, cmpopts.IgnoreUnexported(resource.Quantity{})); diff != \"\" {\n\t\t//fmt.Println(diff)\n\t\treturn true, nil\n\t}\n\n\treturn false, nil\n}", "func (rh *RumourHolder) IsNew(rm RumourMessage) bool {\n\trh.mux.Lock()\n\tdefer rh.mux.Unlock()\n\tsrc := rm.Origin\n\tif _, ok := rh.Rumours[src]; ok {\n\t\trfp := rh.Rumours[src]\n\t\treturn rfp.IsNew(rm)\n\t}\n\treturn true\n}", "func (s *Service) isNew(entry ytfeed.Entry, fi FeedInfo) (ok bool, err error) {\n\n\t// check if entry already exists in store\n\t// this method won't work after migration to locally altered published ts but have to stay for now\n\t// to avoid false-positives on old entries what never got set with SetProcessed\n\texists, exErr := s.Store.Exist(entry)\n\tif err != nil {\n\t\treturn false, errors.Wrapf(exErr, \"failed to check if entry %s exists\", entry.VideoID)\n\t}\n\tif exists {\n\t\treturn false, nil\n\t}\n\n\t// check if we already processed this entry.\n\t// this is needed to avoid infinite get/remove loop when the original feed is updated in place.\n\t// after migration to locally altered published ts, it is also the primary way to detect already processed entries\n\tfound, _, procErr := s.Store.CheckProcessed(entry)\n\tif procErr != nil {\n\t\tlog.Printf(\"[WARN] can't get processed status for %s, %+v\", entry.VideoID, fi)\n\t}\n\tif procErr == nil && found {\n\t\treturn false, nil\n\t}\n\n\treturn true, nil\n}", "func (o *UpdateServerCertificateRequest) HasNewName() bool {\n\tif o != nil && o.NewName != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *UpdateServerCertificateRequest) HasNewPath() bool {\n\tif o != nil && o.NewPath != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (_KeepRegistry *KeepRegistryCaller) IsNewOperatorContract(opts *bind.CallOpts, operatorContract common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _KeepRegistry.contract.Call(opts, out, \"isNewOperatorContract\", operatorContract)\n\treturn *ret0, err\n}", "func IsNewUpdateAvailable() bool {\n\tlist, err := fetchReleases()\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn list.Get(\"latest\").IsApplicable()\n}", "func (o *InlineObject633) HasNewPassword() bool {\n\tif o != nil && o.NewPassword != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (cs *CredentialSpecResource) KnownCreated() bool {\n\tcs.lock.RLock()\n\tdefer cs.lock.RUnlock()\n\n\treturn cs.knownStatusUnsafe == resourcestatus.ResourceStatus(CredentialSpecCreated)\n}", "func (_Registry *RegistryCaller) IsNewOperatorContract(opts *bind.CallOpts, operatorContract common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Registry.contract.Call(opts, out, \"isNewOperatorContract\", operatorContract)\n\treturn *ret0, err\n}", "func (_KeepRegistry *KeepRegistrySession) IsNewOperatorContract(operatorContract common.Address) (bool, error) {\n\treturn _KeepRegistry.Contract.IsNewOperatorContract(&_KeepRegistry.CallOpts, operatorContract)\n}", "func (_Registry *RegistrySession) IsNewOperatorContract(operatorContract common.Address) (bool, error) {\n\treturn _Registry.Contract.IsNewOperatorContract(&_Registry.CallOpts, operatorContract)\n}", "func (st *Store) IsCreated() bool {\n\treturn true\n}", "func (_KeepRegistry *KeepRegistryCallerSession) IsNewOperatorContract(operatorContract common.Address) (bool, error) {\n\treturn _KeepRegistry.Contract.IsNewOperatorContract(&_KeepRegistry.CallOpts, operatorContract)\n}", "func (_Registry *RegistryCallerSession) IsNewOperatorContract(operatorContract common.Address) (bool, error) {\n\treturn _Registry.Contract.IsNewOperatorContract(&_Registry.CallOpts, operatorContract)\n}", "func (s *deployerSuite) TestNew(c *gc.C) {\n\tdeployer := deployer.NewState(s.stateAPI)\n\tc.Assert(deployer, gc.NotNil)\n}", "func (r *MFRC522) PICC_IsNewCardPresent() bool {\n\n\t// Reset baud rates\n\tres, err := r.PICC_RequestA()\n\tlog.Printf(\"PICC_RequestA: len(%d)\\n\", len(res))\n\tif err != nil {\n\t\tlog.Printf(\"PICC_RequestA: %s\\n\", err.Error())\n\t}\n\treturn len(res) == 2\n}", "func AreNewIssuePinsAllowed(ctx *context.APIContext) {\n\t// swagger:operation GET /repos/{owner}/{repo}/new_pin_allowed repository repoNewPinAllowed\n\t// ---\n\t// summary: Returns if new Issue Pins are allowed\n\t// produces:\n\t// - application/json\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// responses:\n\t// \"200\":\n\t// \"$ref\": \"#/responses/RepoNewIssuePinsAllowed\"\n\tpinsAllowed := api.NewIssuePinsAllowed{}\n\tvar err error\n\n\tpinsAllowed.Issues, err = issues_model.IsNewPinAllowed(ctx, ctx.Repo.Repository.ID, false)\n\tif err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"IsNewIssuePinAllowed\", err)\n\t\treturn\n\t}\n\n\tpinsAllowed.PullRequests, err = issues_model.IsNewPinAllowed(ctx, ctx.Repo.Repository.ID, true)\n\tif err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"IsNewPullRequestPinAllowed\", err)\n\t\treturn\n\t}\n\n\tctx.JSON(http.StatusOK, pinsAllowed)\n}", "func (o *EnvironmentUsageDto) HasNewProblems() bool {\n\tif o != nil && o.NewProblems != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (g *Blockchain) IsNewfoundland(height uint64) bool {\n\treturn g.isPost(g.NewfoundlandBlockHeight, height)\n}", "func (e *FileEvent) IsCreate() bool { return (e.flags & EventFlagItemCreated) == EventFlagItemCreated}", "func (p *Page) IsNewer(other time.Time) bool {\n\treturn p.file.Modified.After(other) || p.meta.Modified.After(other)\n}", "func isNewWorker(worker *workerv1.Worker, currentDeployment *appsv1.Deployment) bool {\n\treturn currentDeployment == nil && worker.DeletionTimestamp == nil\n}", "func (o *Post) HasNewParticipants() bool {\n\tif o != nil && o.NewParticipants != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func IsErrNewIssueInsert(err error) bool {\n\t_, ok := err.(ErrNewIssueInsert)\n\treturn ok\n}", "func isNewServer(ctx context.Context, prevUUID string, db *sql.DB, flavor string) (bool, error) {\n\tif len(prevUUID) == 0 {\n\t\t// no sub dir exists before\n\t\treturn true, nil\n\t}\n\tuuid, err := utils.GetServerUUID(ctx, db, flavor)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif strings.HasPrefix(prevUUID, uuid) {\n\t\t// same server as before\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func (d *Desc) IsCreated() bool {\n\td.createLock.RLock()\n\tdefer d.createLock.RUnlock()\n\n\treturn d.isCreated\n}", "func Repository_IsResource(construct awscdk.IConstruct) *bool {\n\t_init_.Initialize()\n\n\tvar returns *bool\n\n\t_jsii_.StaticInvoke(\n\t\t\"monocdk.aws_ecr.Repository\",\n\t\t\"isResource\",\n\t\t[]interface{}{construct},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (st *Account) IsNewbieAccount() bool {\n\treturn st.nonce == 0\n}", "func (options *CreateWorkspaceDeletionJobOptions) SetNewDestroyResources(newDestroyResources bool) *CreateWorkspaceDeletionJobOptions {\n\toptions.NewDestroyResources = core.BoolPtr(newDestroyResources)\n\treturn options\n}", "func (d UserData) HasNewPassword() bool {\n\treturn d.ModelData.Has(models.NewFieldName(\"NewPassword\", \"new_password\"))\n}", "func (o *CreateRepository28Created) IsSuccess() bool {\n\treturn true\n}", "func (o *InlineObject633) GetNewPasswordOk() (string, bool) {\n\tif o == nil || o.NewPassword == nil {\n\t\tvar ret string\n\t\treturn ret, false\n\t}\n\treturn *o.NewPassword, true\n}", "func (o *CreateRepository8Created) IsSuccess() bool {\n\treturn true\n}", "func IsNewEvent(redisClient *redis.Client, source, eventKey string) (new bool) {\n\tset, err := redisClient.SetNX(eventKey+\":\"+source, true, time.Minute*5).Result()\n\tif err != nil {\n\t\tcache.GetLogger().Errorln(\"error doing deduplication:\", err.Error())\n\t\treturn false\n\t}\n\n\treturn set\n}", "func (o *InlineResponse20027Person) HasHasAccessToNewProjects() bool {\n\tif o != nil && o.HasAccessToNewProjects != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (e *Event) IsCreate() bool {\n\treturn e.Op == 1 || e.Op&CREATE == CREATE\n}", "func (o *UpdateServerCertificateRequest) GetNewNameOk() (*string, bool) {\n\tif o == nil || o.NewName == nil {\n\t\treturn nil, false\n\t}\n\treturn o.NewName, true\n}", "func (b *GroupsSetCallbackSettingsBuilder) PhotoNew(v bool) *GroupsSetCallbackSettingsBuilder {\n\tb.Params[\"photo_new\"] = v\n\treturn b\n}", "func New(name string) *Resource {\n\n\tlastId := (int64)(len(resourceMap))\n\n\treturn &Resource{\n\t\tId: lastId + 1,\n\t\tName: name,\n\t\tStatus: true,\n\t\tCreatedAt: time.Now(),\n\t\tUpdatedAt: time.Now(),\n\t}\n}", "func (m *Machine) StatefulSetNewGeneration(namespace string, name string, version int64) (bool, error) {\n\tclient := m.Clientset.AppsV1beta1().StatefulSets(namespace)\n\n\tss, err := client.Get(name, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn false, errors.Wrapf(err, \"failed to query for statefulSet by name: %v\", name)\n\t}\n\n\tif *ss.Status.ObservedGeneration > version {\n\t\treturn true, nil\n\t}\n\n\treturn false, nil\n}", "func newVersionAvailable(currentVersion string) (bool, *string, error) {\n\tgardenctlLatestURL := \"https://api.github.com/repos/gardener/gardenctl/releases/latest\"\n\tresp, err := http.Get(gardenctlLatestURL)\n\tif err != nil {\n\t\treturn false, nil, err\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn false, nil, err\n\t}\n\n\tdata := make(map[string]interface{})\n\tif err := json.Unmarshal([]byte(body), &data); err != nil {\n\t\treturn false, nil, err\n\t}\n\tvar latestVersion string\n\tif data[\"tag_name\"] != nil {\n\t\tlatestVersion = data[\"tag_name\"].(string)\n\t}\n\n\tc, err := semver.NewConstraint(\"> \" + currentVersion)\n\tif err != nil {\n\t\treturn false, nil, err\n\t}\n\n\tlatest, err := semver.NewVersion(latestVersion)\n\tif err != nil {\n\t\treturn false, nil, err\n\t}\n\n\treturn c.Check(latest), &latestVersion, nil\n}", "func (lh *LoginHandler) HandleStateNew(c *Condition) (string, interface{}) {\n\tswitch pld := c.Payload.(type) {\n\tcase *PreparePayload:\n\t\tlh.userId = pld.userId\n\t\tlh.password = pld.password\n\t\tlh.illegalLoginCounter = 0\n\t\tlh.locked = false\n\n\t\tlog.Printf(\"User '%v' prepared.\", lh.userId)\n\n\t\treturn \"Authenticating\", nil\n\tcase *LoginPayload:\n\t\tlog.Printf(\"Illegal login, handler not initialized!\")\n\n\t\treturn \"New\", false\n\tcase Timeout:\n\t\tlog.Printf(\"Timeout, terminate handler!\")\n\n\t\treturn \"Terminate\", nil\n\t}\n\n\tlog.Printf(\"Illegal payload '%v' during state 'new'!\", c.Payload)\n\n\treturn \"New\", nil\n}", "func newResource() *resource.Resource {\n\tr, _ := resource.Merge(\n\t\tresource.Default(),\n\t\tresource.NewWithAttributes(\n\t\t\tsemconv.SchemaURL,\n\t\t\tsemconv.ServiceNameKey.String(\"opentelemetry-server\"),\n\t\t\tsemconv.ServiceVersionKey.String(\"v0.1.0\"),\n\t\t\tattribute.String(\"environment\", \"demo\"),\n\t\t),\n\t)\n\treturn r\n}", "func IsCreated(name string) bool {\n\treturn exec.Command(execName, \"inspect\", \"vm\", name).Run() == nil\n}", "func (g *Goods) ListNew(c Context) {\n\t// TODO\n\tc.String(http.StatusOK, \"get new goods list\")\n}", "func (b *GroupsSetLongPollSettingsBuilder) PhotoNew(v bool) *GroupsSetLongPollSettingsBuilder {\n\tb.Params[\"photo_new\"] = v\n\treturn b\n}", "func (o *LocalDatabaseProvider) HasCreated() bool {\n\tif o != nil && o.Created != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (c *Container) GetNew(p *Provider) (interface{}, error) {\n\treturn (*p).New(c)\n}", "func (d *Driver) DiscoverNew(r *sdk.DiscoveryNotification) error {\n\treturn nil\n}", "func (m *Microservice) CheckForNewConfiguration() {\n\tzap.L().Info(\"checking pending operations\")\n\tdata, _, err := m.GetOperations(c8y.OperationStatusPending)\n\n\tif err != nil {\n\t\tlog.Printf(\"Error getting operations. %s\", err)\n\t\treturn\n\t}\n\n\tfor _, op := range data.Items {\n\n\t\t//\n\t\t// Update Configuration Operation\n\t\t//\n\t\tif c8yConfig := op.Get(\"c8y_Configuration.config\"); c8yConfig.Exists() {\n\t\t\tm.onUpdateConfigurationOperation(op.Get(\"id\").String(), c8yConfig.String())\n\t\t\tconfigurationChangeCount.Inc()\n\t\t}\n\t}\n}", "func CfnResourceVersion_IsConstruct(x interface{}) *bool {\n\t_init_.Initialize()\n\n\tvar returns *bool\n\n\t_jsii_.StaticInvoke(\n\t\t\"aws-cdk-lib.aws_cloudformation.CfnResourceVersion\",\n\t\t\"isConstruct\",\n\t\t[]interface{}{x},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *Controller) renderNew(ctx context.Context, w http.ResponseWriter, app *database.MobileApp) {\n\tm := templateMap(ctx)\n\tm.Title(\"New mobile app\")\n\tm[\"app\"] = app\n\tc.h.RenderHTML(w, \"mobileapps/new\", m)\n}", "func (o *CreateRepository32Created) IsSuccess() bool {\n\treturn true\n}", "func (o *Ga4ghChemotherapy) HasCreated() bool {\n\tif o != nil && o.Created != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (s *AlertChannel) IsCreated() bool {\n\treturn s.Status.IsCreated()\n}", "func New(i interface{}) *Resource {\n\tr := &Resource{}\n\tr.geth = mustMakeRpc(i, \"Get\")\n\tr.posth = mustMakeRpc(i, \"Post\")\n\tr.puth = mustMakeRpc(i, \"Put\")\n\tr.deleteh = mustMakeRpc(i, \"Delete\")\n\n\t// println(\"[debug]\", r.geth, r.posth, r.puth, r.deleteh)\n\treturn r\n}", "func CfnResourceVersion_IsCfnResource(construct constructs.IConstruct) *bool {\n\t_init_.Initialize()\n\n\tvar returns *bool\n\n\t_jsii_.StaticInvoke(\n\t\t\"aws-cdk-lib.aws_cloudformation.CfnResourceVersion\",\n\t\t\"isCfnResource\",\n\t\t[]interface{}{construct},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func RepositoryBase_IsResource(construct awscdk.IConstruct) *bool {\n\t_init_.Initialize()\n\n\tvar returns *bool\n\n\t_jsii_.StaticInvoke(\n\t\t\"monocdk.aws_ecr.RepositoryBase\",\n\t\t\"isResource\",\n\t\t[]interface{}{construct},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (t *Type) IsResource() bool {\n\treturn false\n}", "func canCreateOnUpdate(stub shim.ChaincodeStubInterface) (bool) {\n var createOnUpdate CreateOnUpdate\n createOnUpdateBytes, err := stub.GetState(\"CreateOnUpdate\")\n if err != nil {\n err = fmt.Errorf(\"GETSTATE for canCreateOnUpdate failed: %s\", err)\n log.Error(err)\n return true // true is the default\n }\n err = json.Unmarshal(createOnUpdateBytes, &createOnUpdate)\n if err != nil {\n err = fmt.Errorf(\"canCreateOnUpdate failed to marshal: %s\", err)\n log.Error(err)\n return true // true is the default\n }\n return createOnUpdate.CreateOnUpdate\n}", "func (o *UpdateServerCertificateRequest) GetNewPathOk() (*string, bool) {\n\tif o == nil || o.NewPath == nil {\n\t\treturn nil, false\n\t}\n\treturn o.NewPath, true\n}", "func New(w http.ResponseWriter, r *http.Request) {\r\n\ttmpl.ExecuteTemplate(w, \"New\", nil)\r\n}", "func New(w http.ResponseWriter, r *http.Request) {\r\n\ttmpl.ExecuteTemplate(w, \"New\", nil)\r\n}", "func NewNewData() *NewData {\n\tthis := NewData{}\n\treturn &this\n}", "func (c *Namespace) HasChanged(k K8sResource) bool {\n\treturn false\n}", "func (opts resourceOptions) newResource() *resource.Resource {\n\treturn &resource.Resource{\n\t\tGVK: resource.GVK{ // Remove whitespaces to prevent values like \" \" pass validation\n\t\t\tGroup: strings.TrimSpace(opts.Group),\n\t\t\tDomain: strings.TrimSpace(opts.Domain),\n\t\t\tVersion: strings.TrimSpace(opts.Version),\n\t\t\tKind: strings.TrimSpace(opts.Kind),\n\t\t},\n\t\tPlural: resource.RegularPlural(opts.Kind),\n\t\tAPI: &resource.API{},\n\t\tWebhooks: &resource.Webhooks{},\n\t}\n}", "func (o *CreateRepository38Created) IsSuccess() bool {\n\treturn true\n}", "func CfnSchemaVersionMetadata_IsCfnResource(construct constructs.IConstruct) *bool {\n\t_init_.Initialize()\n\n\tvar returns *bool\n\n\t_jsii_.StaticInvoke(\n\t\t\"aws-cdk-lib.aws_glue.CfnSchemaVersionMetadata\",\n\t\t\"isCfnResource\",\n\t\t[]interface{}{construct},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func TestNew(t *testing.T){\n\tclient := New(\"localhost\", 5555)\n\t\n\tif client.Server != \"localhost\" {\n\t\tt.Error(\"The server hasn't being initialized correctly\")\n\t}\n\t\n\tif client.Port != 5555 {\n\t\tt.Error(\"The port wasn't initialized propertly\")\n\t}\n\t\n\tif client.Opened {\n\t\tt.Error(\"The opened indicator wasn't initialized propertly\")\n\t}\n\t\n\tif client.conn != nil {\n\t\tt.Error(\"The TCP connection wasn't initialized propertly\")\n\t}\n}", "func (c *Controller) renderNew(ctx context.Context, w http.ResponseWriter, authApp *database.AuthorizedApp) {\n\tm := controller.TemplateMapFromContext(ctx)\n\tm.Title(\"New API key\")\n\tm[\"authApp\"] = authApp\n\tm[\"typeAdmin\"] = database.APIKeyTypeAdmin\n\tm[\"typeDevice\"] = database.APIKeyTypeDevice\n\tm[\"typeStats\"] = database.APIKeyTypeStats\n\tc.h.RenderHTML(w, \"apikeys/new\", m)\n}", "func (s *SymbolDictionary) NumberOfNewSymbols() int {\n\treturn s.numberOfNewSymbols\n}", "func CfnModuleVersion_IsCfnResource(construct constructs.IConstruct) *bool {\n\t_init_.Initialize()\n\n\tvar returns *bool\n\n\t_jsii_.StaticInvoke(\n\t\t\"aws-cdk-lib.aws_cloudformation.CfnModuleVersion\",\n\t\t\"isCfnResource\",\n\t\t[]interface{}{construct},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (bc *Catalog) IsChanged() bool {\n\tbc.lock.RLock()\n\tdefer bc.lock.RUnlock()\n\treturn bc.lastAcked < bc.epoch\n}", "func ServerApplication_IsResource(construct awscdk.IConstruct) *bool {\n\t_init_.Initialize()\n\n\tvar returns *bool\n\n\t_jsii_.StaticInvoke(\n\t\t\"monocdk.aws_codedeploy.ServerApplication\",\n\t\t\"isResource\",\n\t\t[]interface{}{construct},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func New(w http.ResponseWriter, r *http.Request) {\n\tgetTemplates().ExecuteTemplate(w, \"New\", nil)\n}", "func (v ProspectsResource) New(c buffalo.Context) error {\n\t// Make prospect available inside the html template\n\tc.Set(\"prospect\", &models.Prospect{})\n\n\treturn c.Render(200, r.HTML(\"prospects/new.html\"))\n}", "func (o *CreateRepository37Created) IsSuccess() bool {\n\treturn true\n}", "func (m *Resource) IsOK() bool {\n\tswitch {\n\tcase len(m.name) == 0:\n\t\treturn false\n\tcase len(m.description) == 0:\n\t\treturn false\n\tcase m.schema == nil:\n\t\treturn false\n\tcase m.model == nil:\n\t\treturn false\n\tcase m.store == nil:\n\t\treturn false\n\tcase len(m.methods) == 0:\n\t\treturn false\n\tdefault:\n\t\treturn true\n\t}\n}", "func (r *resource) Create(resource meta.ResourceObject) error {\n\tlog := log.With(\"kind\", resource.GetObjectKind().GroupVersionKind().Kind, \"name\", resource.GetName(), \"namespace\", resource.GetNamespace())\n\tlog.Debug(\"Creating\")\n\tif err := r.client.ControlCli.Create(context.TODO(), resource); err != nil {\n\t\tlog.Warn(\"Failed to create object. \", err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (cl *Client) Get_New_Status(do_read_lock bool) (s string, err error) {\n\tif do_read_lock {\n\t\tread_lock, xerr := cl.RLockNamed(\"Get_New_Status\")\n\t\tif xerr != nil {\n\t\t\terr = xerr\n\t\t\treturn\n\t\t}\n\t\tdefer cl.RUnlockNamed(read_lock)\n\t}\n\ts = cl.Status\n\treturn\n}", "func (o *UpdateServerCertificateRequest) GetNewName() string {\n\tif o == nil || o.NewName == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.NewName\n}", "func newResource(r interface{}) corev2.Resource {\n\treturn reflect.New(reflect.ValueOf(r).Elem().Type()).Interface().(corev2.Resource)\n}", "func CfnResourceDefaultVersion_IsConstruct(x interface{}) *bool {\n\t_init_.Initialize()\n\n\tvar returns *bool\n\n\t_jsii_.StaticInvoke(\n\t\t\"aws-cdk-lib.aws_cloudformation.CfnResourceDefaultVersion\",\n\t\t\"isConstruct\",\n\t\t[]interface{}{x},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (m Marker) isNewFile(f *os.File) (bool, error) {\n\tfi, inode, err := getInode(f)\n\tif err != nil {\n\t\treturn false, errors.WithStack(err)\n\t}\n\n\tif m.Inode != 0 {\n\t\tif inode != m.Inode {\n\t\t\treturn true, nil\n\t\t}\n\t}\n\n\tif fi.Size() < m.Offset {\n\t\treturn true, nil\n\t}\n\n\treturn false, nil\n}", "func (sp SnapshotPolicy) IsSwapNew4Old() bool { return sp == SnapshotPolicySwapNew4Old }" ]
[ "0.7124189", "0.71037585", "0.7013115", "0.6765984", "0.67634845", "0.6687251", "0.65053016", "0.64151883", "0.63380104", "0.63252676", "0.60556406", "0.60512", "0.60003394", "0.58730227", "0.5747134", "0.5649383", "0.5528927", "0.5502173", "0.5344137", "0.52407295", "0.5237779", "0.520098", "0.5171572", "0.51694596", "0.5135902", "0.51161957", "0.50910306", "0.5084502", "0.50663763", "0.5065055", "0.50551045", "0.50506675", "0.5013355", "0.49965173", "0.49699914", "0.49695772", "0.49413094", "0.49385798", "0.4913746", "0.49080268", "0.48746145", "0.48544374", "0.4851298", "0.48502287", "0.47973353", "0.47701117", "0.47638765", "0.4760137", "0.47467905", "0.4740033", "0.47024986", "0.46937406", "0.46785706", "0.4673832", "0.4651259", "0.46500647", "0.4648689", "0.46447033", "0.46382156", "0.46338114", "0.46272954", "0.46175978", "0.4607672", "0.4607238", "0.46066856", "0.46014574", "0.45927426", "0.45859268", "0.45810366", "0.45762065", "0.455736", "0.4548553", "0.45416537", "0.4541102", "0.45387423", "0.4532404", "0.4531994", "0.4531994", "0.45139626", "0.45135006", "0.45052987", "0.4503538", "0.4502778", "0.45004457", "0.44935736", "0.4489494", "0.44888622", "0.44886985", "0.44873214", "0.44870627", "0.44807693", "0.44781843", "0.44781843", "0.4474942", "0.44747737", "0.44611412", "0.44597492", "0.4452688", "0.4450789", "0.44430453" ]
0.6897515
3
HasConflicts provides a binary t/f if there are any changes within package after dry run is complete.
func (d Diff) HasConflicts() bool { for _, b := range d.Buckets { if b.hasConflict() { return true } } for _, l := range d.Labels { if l.hasConflict() { return true } } for _, v := range d.Variables { if v.hasConflict() { return true } } return false }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *Table) HasConflicts(ctx context.Context) (bool, error) {\n\tif t.Format() == types.Format_DOLT {\n\t\tart, err := t.GetArtifacts(ctx)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\treturn art.HasConflicts(ctx)\n\t}\n\treturn t.table.HasConflicts(ctx)\n}", "func (rnode *RuleNode) HasConflict() bool {\n\treturn false // TODO\n}", "func (env *Environment) checkIntegrity(ps ...*gdf.Package) (conflicts map[string]map[string][3]string) {\n\tconflicts = map[string]map[string][3]string{}\n\tconflicts[\"#dep-registry-orphan#\"] = map[string][3]string{}\n\tconflicts[\"#dep-registry-inconsistency#\"] = map[string][3]string{}\n\n\tdefer func() {\n\t\tif len(conflicts[\"#dep-registry-orphan#\"]) == 0 {\n\t\t\tdelete(conflicts, \"#dep-registry-orphan#\")\n\t\t}\n\t\tif len(conflicts[\"#dep-registry-inconsistency#\"]) == 0 {\n\t\t\tdelete(conflicts, \"#dep-registry-inconsistency#\")\n\t\t}\n\t}()\n\tpkgs := map[string]bool{}\n\n\tfor _, p := range ps {\n\t\tpkgs[p.Path] = true\n\t\td, er := env.Diff(p, false)\n\t\tif er != nil {\n\t\t\tconflicts[p.Path] = map[string][3]string{\n\t\t\t\t\"#dep-registry-inconsistency#\": [3]string{\"missing\", er.Error(), \"\"},\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tif d != nil && len(d.Exports) > 0 {\n\t\t\tconflicts[p.Path] = map[string][3]string{\n\t\t\t\t\"#dep-registry-inconsistency#\": [3]string{\"exports\", strings.Join(d.Exports, \"\\n\"), \"\"},\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tif d != nil && len(d.Imports) > 0 {\n\t\t\tconflicts[p.Path] = map[string][3]string{\n\t\t\t\t\"#dep-registry-inconsistency#\": [3]string{\"imports\", strings.Join(d.Imports, \"\\n\"), \"\"},\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\terrs := env.db.hasConflict(p, map[string]bool{})\n\t\tif len(errs) > 0 {\n\t\t\tconflicts[p.Path] = errs\n\t\t\tcontinue\n\t\t}\n\t}\n\n\tdbpkgs, err := env.db.GetAllPackages()\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\n\tfor _, dbp := range dbpkgs {\n\t\tif !pkgs[dbp.Package] {\n\t\t\tconflicts[\"#dep-registry-orphan#\"][dbp.Package] = [3]string{\"orphan\", dbp.Package, \"\"}\n\t\t\tcontinue\n\t\t}\n\t}\n\n\treturn\n}", "func (c *Client) HasDependency(ctx context.Context, change *gerritpb.ChangeInfo) (bool, error) {\n\trelatedChanges, err := c.getRelatedChanges(ctx, change)\n\tif err != nil {\n\t\treturn false, errors.Annotate(err, \"failed checking dependency\").Err()\n\t}\n\n\tfor _, relatedChange := range relatedChanges {\n\t\tif relatedChange.Status == gerritpb.ChangeStatus_MERGED {\n\t\t\t// relatedChange here is the newest merged. If relatedChange != change,\n\t\t\t// then there is a merged dependency\n\t\t\treturn relatedChange.Project != change.Project ||\n\t\t\t\trelatedChange.Number != change.Number, nil\n\t\t}\n\t}\n\n\t// none of the related changes are merged, so no merged dependencies\n\treturn false, nil\n}", "func (o *oracle) hasConflict(txn *Txn) bool {\n\tif len(txn.reads) == 0 {\n\t\treturn false\n\t}\n\tfor _, committedTxn := range o.committedTxns {\n\t\t// If the committedTxn.ts is less than txn.readTs that implies that the\n\t\t// committedTxn finished before the current transaction started.\n\t\t// We don't need to check for conflict in that case.\n\t\t// This change assumes linearizability. Lack of linearizability could\n\t\t// cause the read ts of a new txn to be lower than the commit ts of\n\t\t// a txn before it (@mrjn).\n\t\tif committedTxn.ts <= txn.readTs {\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, ro := range txn.reads {\n\t\t\tif _, has := committedTxn.conflictKeys[ro]; has {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false\n}", "func (o *TransactionResult) HasProblems() bool {\n\tif o != nil && o.Problems != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func hasDependency(bld *build.File, r *build.Rule, dep string) bool {\n\tpkg := filepath.Dir(bld.Path)\n\toldDeps := r.Attr(\"deps\")\n\tif edit.ListFind(oldDeps, dep, pkg) != nil {\n\t\treturn true\n\t}\n\truntimeDeps := r.Attr(\"runtime_deps\")\n\treturn edit.ListFind(runtimeDeps, dep, pkg) != nil\n}", "func (pd PackageDifference) Any() bool {\n\tif len(pd.Additions) > 0 || len(pd.Removals) > 0 || len(pd.Changes) > 0 {\n\t\treturn true\n\t}\n\treturn false\n}", "func (e *Eval) dirty(f *Flow) bool {\n\tif f.Op == Extern {\n\t\treturn true\n\t}\n\tfor _, dep := range f.Deps {\n\t\tif e.dirty(dep) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (self *AgentState) HasConflict(potentialJobName string, potentialConflicts []string) (bool, string) {\n\t// Iterate through each existing Job, asserting two things:\n\tfor existingJobName, existingConflicts := range self.conflicts {\n\n\t\t// 1. Each tracked Job does not conflict with the potential conflicts\n\t\tfor _, pc := range potentialConflicts {\n\t\t\tif globMatches(pc, existingJobName) {\n\t\t\t\treturn true, existingJobName\n\t\t\t}\n\t\t}\n\n\t\t// 2. The new Job does not conflict with any of the tracked confclits\n\t\tfor _, ec := range existingConflicts {\n\t\t\tif globMatches(ec, potentialJobName) {\n\t\t\t\treturn true, existingJobName\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false, \"\"\n}", "func ConflictingRequires(verbose bool) (bool, error) {\r\n\t// obtain the set of requires by all modules in our build (via 'go mod graph').\r\n\t// this takes into account replace directives.\r\n\trequires, err := modgraph.Requirements()\r\n\tif err != nil {\r\n\t\treturn false, err\r\n\t}\r\n\r\n\t// track our paths and versions in { path: {version, version, ...}, ... } map.\r\n\tpaths := make(map[string][]string)\r\n\tfor _, require := range requires {\r\n\t\tf := strings.Split(require, \"@\")\r\n\t\tif len(f) != 2 {\r\n\t\t\treturn false, fmt.Errorf(\"unexpected requirement: %s\", require)\r\n\t\t}\r\n\t\tpath, version := f[0], f[1]\r\n\t\tif !semver.IsValid(version) {\r\n\t\t\treturn false, fmt.Errorf(\"invalid semver version: %s\", require)\r\n\t\t}\r\n\r\n\t\t// Probably not needed, but might as well use the canonical semver version. That strips \"+incompatible\",\r\n\t\t// which we need to preserve. Thus, we check here for \"+incompatible\" and add it back if needed.\r\n\t\tif semver.Build(version) == \"+incompatible\" {\r\n\t\t\tpaths[path] = append(paths[path], semver.Canonical(version)+\"+incompatible\")\r\n\t\t} else {\r\n\t\t\tpaths[path] = append(paths[path], semver.Canonical(version))\r\n\t\t}\r\n\t}\r\n\r\n\t// for each path, loop over its versions (in semantic order) and build up a list\r\n\t// of potential conflicts.\r\n\tflagged := false\r\n\tfor path, versions := range paths {\r\n\t\tsort.Slice(versions, func(i, j int) bool { return -1 == semver.Compare(versions[i], versions[j]) })\r\n\r\n\t\tif verbose {\r\n\t\t\tfmt.Printf(\"gomodvet: conflictingrequires: module %q has require versions: %v\\n\", path, versions)\r\n\t\t}\r\n\r\n\t\tpriorVersion := \"\"\r\n\t\tvar potentialIncompats []string\r\n\t\tfor _, version := range versions {\r\n\t\t\tif version == priorVersion {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t\tif isBeforeV1(version) {\r\n\t\t\t\t// all pre-v1 versions are potentially incompatible\r\n\t\t\t\tpotentialIncompats = append(potentialIncompats, version)\r\n\t\t\t} else if isV1(version) && !isV1(priorVersion) {\r\n\t\t\t\t// the first v1 version seen is potentially incompatible with any v0, v2+incompatible, v3+incompatible, etc.\r\n\t\t\t\tpotentialIncompats = append(potentialIncompats, version)\r\n\t\t\t} else if isV2OrHigherIncompat(version) && semver.Major(version) != semver.Major(priorVersion) {\r\n\t\t\t\t// the first major version v2+incompatible, v3+incompatible, etc is potentially incompatible.\r\n\t\t\t\t// (If two v2+incompatible versions are seen, in theory they should be compatible with each other).\r\n\t\t\t\tpotentialIncompats = append(potentialIncompats, version)\r\n\t\t\t}\r\n\t\t\tpriorVersion = version\r\n\t\t}\r\n\t\tif len(potentialIncompats) > 1 {\r\n\t\t\t// mutiple potential incompatible versions, which means they can be incompatible with each other.\r\n\t\t\tfmt.Printf(\"gomodvet-004: module %q was required with potentially incompatible versions: %s\\n\",\r\n\t\t\t\tpath, strings.Join(potentialIncompats, \", \"))\r\n\t\t\tflagged = true\r\n\t\t}\r\n\t}\r\n\treturn flagged, nil\r\n}", "func (h *HealthStatus) HasFailingDeps() bool {\n\treturn len(h.FailingDeps) > 0\n}", "func HasBuilder(info os.FileInfo) bool {\n\tfor _, b := range builders {\n\t\tif b.canBuild(info) {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (i *interactor) IsDirty() (bool, error) {\n\ti.logger.Info(\"Checking is dirty.\")\n\tb, err := i.executor.Run(\"status\", \"--porcelain\")\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"error add -A: %v. output: %s\", err, string(b))\n\t}\n\treturn len(b) > 0, nil\n}", "func (o *UpdateLoadBalancerRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *UpdateVpnConnectionRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func IsConflict(err error) bool {\n\treturn errors.Cause(err) == errConflict\n}", "func (o *UpdateServerCertificateRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *EnvironmentUsageDto) HasNewProblems() bool {\n\tif o != nil && o.NewProblems != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (d *Deployment) NeedsUpdate() bool {\n\tfor name, config := range d.Configs {\n\t\tif config.Pending() {\n\t\t\tcontinue\n\t\t}\n\n\t\tif config.Checksum() != d.AppliedChecksums[name] {\n\t\t\treturn true\n\t\t}\n\t}\n\n\tfor name := range d.AppliedChecksums {\n\t\tif _, ok := d.Configs[name]; !ok {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (o *ApplianceImageBundleAllOf) HasAutoUpgrade() bool {\n\tif o != nil && o.AutoUpgrade != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *ConvergedinfraServerComplianceDetailsAllOf) HasFirmware() bool {\n\tif o != nil && o.Firmware != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (dm *DagModifier) HasChanges() bool {\n\treturn dm.wrBuf != nil\n}", "func (s *status) DependsReady() bool {\n\t// In case\n\ts.locker.Lock()\n\tdefer s.locker.Unlock()\n\n\tfor _, c := range s.data.Conditions {\n\t\tif c.Type == goharborv1.CacheReady ||\n\t\t\tc.Type == goharborv1.DatabaseReady ||\n\t\t\tc.Type == goharborv1.StorageReady {\n\t\t\tif c.Status != corev1.ConditionTrue {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\n\treturn true\n}", "func (t *triagedDeps) OK() bool {\n\treturn t.invalidDeps == nil\n}", "func (o *GlobalProblemStatus) HasOpenProblemCounts() bool {\n\tif o != nil && o.OpenProblemCounts != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (component *Component) IsDirty() (bool, error) {\n\tspecs, err := component.JSONSpecs()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tcurrentContentsHash := fmt.Sprintf(\"%x\", sha1.Sum([]byte(specs)))\n\treturn component.loadedContentsHash != currentContentsHash, nil\n}", "func areDepsComplete(s string, n map[string][]string) bool {\n\tcount := 0 // Count needs to be the length of the slice of dependencies.\n\tfor _, i := range n[s] {\n\t\tfor _, q := range donequeue {\n\t\t\tif s == q { // If this node is already in the donequeue, fail fast.\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tif q == i {\n\t\t\t\tcount++\n\t\t\t}\n\t\t}\n\t}\n\tif len(n[s]) == 0 { // if len is zero, these have already been addressed in the first iteration.\n\t\treturn false\n\t}\n\tif len(n[s]) == count { // All dependencies are in the donequeue.\n\t\treturn true\n\t}\n\n\treturn false\n}", "func alreadySatisfied(consumingRuleLabel bazel.Label, existingDeps map[bazel.Label]bool, satisfyingRules []*bazel.Rule) bool {\n\tfor _, r := range satisfyingRules {\n\t\tif r.Label() == consumingRuleLabel {\n\t\t\treturn true\n\t\t}\n\t\tif _, ok := existingDeps[r.Label()]; ok {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (o *CreateSnapshotExportTaskRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func subsetIsConflicted(subset [9]int) bool {\n\tcount := 0\n\tfor i := 1; i<=9; i++ {\n\t\tfor _, v := range subset {\n\t\t\tif v == 1 {\n\t\t\t\tcount++\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\treturn count != 9\n}", "func hasChanges(dir string) (bool, error) {\n\t// Write command output to both os.Stderr and local, so that we can check\n\t// whether there are modified files.\n\tinmem := bytes.NewBuffer([]byte{}) // TODO(deklerk): Try `var inmem bytes.Buffer`.\n\tw := io.MultiWriter(os.Stderr, inmem)\n\n\tc := exec.Command(\"bash\", \"-c\", \"git status --short\")\n\tc.Dir = dir\n\tc.Stdout = w\n\tc.Stderr = os.Stderr\n\tc.Stdin = os.Stdin // Prevents \"the input device is not a TTY\" error.\n\terr := c.Run()\n\n\treturn inmem.Len() > 0, err\n}", "func (i *multiNamespaceInformer) HasSynced() bool {\n\ti.lock.Lock()\n\tdefer i.lock.Unlock()\n\n\tfor _, informer := range i.informers {\n\t\tif synced := informer.HasSynced(); !synced {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func (o *ControllersUpdateGlobalConfigRequest) HasVersionUpdatability() bool {\n\tif o != nil && o.VersionUpdatability != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *FileversionFileversion) HasPendingFileRef() bool {\n\tif o != nil && o.PendingFileRef != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (obj *InstallPhase) IsSatisfied() bool {\n\treturn obj.Spec.TargetState == obj.Status.ActualState\n}", "func (o *ViewProjectBudget) HasNotifications() bool {\n\tif o != nil && o.Notifications != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (container *container) HasPendingUpdates() (bool, error) {\r\n\treturn false, nil\r\n}", "func (o *UpdateNetRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func IsErrMergeConflicts(err error) bool {\n\t_, ok := err.(ErrMergeConflicts)\n\treturn ok\n}", "func conflictCheckForDevfileFlag(args []string, registryName string) error {\n\t// More than one arguments should not be allowed when --devfile is used\n\tif len(args) > 1 {\n\t\treturn &DevfileExistsExtraArgsError{len(args)}\n\t}\n\t// Check if both --devfile and --registry flag are used, in which case raise an error\n\tif registryName != \"\" {\n\t\treturn &DevfileFlagWithRegistryFlagError{}\n\t}\n\treturn nil\n}", "func (o *DeregisterVmsInLoadBalancerRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (gh *GithubRequireAheadPrecondition) NeedsMerge(deployment common.Deployment, ctx *GithubRequireAheadContext) (bool, error) {\n\tif deployment.IsForce() {\n\t\treturn false, nil\n\t}\n\n\tbase := *ctx.RepoDetails.DefaultBranch\n\thead := deployment.Ref()\n\n\tcomparison, err := ctx.RepoClient.CompareCommits(base, head)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn (*comparison.BehindBy > 0), nil\n}", "func (o *GlobalProblemStatus) HasTotalOpenProblemsCount() bool {\n\tif o != nil && o.TotalOpenProblemsCount != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (s *Syncthing) HasBinary() bool {\n\tif _, err := os.Stat(s.binPath()); err != nil {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (_BREMICO *BREMICOCaller) HasClosed(opts *bind.CallOpts) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _BREMICO.contract.Call(opts, out, \"hasClosed\")\n\treturn *ret0, err\n}", "func HasPackage(name string) bool {\n\tif installers, ok := GetInstallers(name); ok {\n\t\tfor _, installer := range installers {\n\t\t\tif installer.Available() {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func (o *ViewProjectBudget) HasIsRepeating() bool {\n\tif o != nil && o.IsRepeating != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func CheckDependencies(ctx context.Context, clt client.Client, addon *types.Addon) bool {\n\tvar app v1beta1.Application\n\tfor _, dep := range addon.Dependencies {\n\t\terr := clt.Get(ctx, client.ObjectKey{\n\t\t\tNamespace: types.DefaultKubeVelaNS,\n\t\t\tName: Convert2AppName(dep.Name),\n\t\t}, &app)\n\t\tif err != nil {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (o *CreateClientGatewayRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (c *Changeset) IsOutdated() (bool, error) {\n\tcurrentTitle, err := c.Changeset.Title()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tif currentTitle != c.Title {\n\t\treturn true, nil\n\t}\n\n\tcurrentBody, err := c.Changeset.Body()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tif currentBody != c.Body {\n\t\treturn true, nil\n\t}\n\n\tcurrentBaseRef, err := c.Changeset.BaseRef()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tif git.EnsureRefPrefix(currentBaseRef) != git.EnsureRefPrefix(c.BaseRef) {\n\t\treturn true, nil\n\t}\n\n\treturn false, nil\n}", "func (o *BudgetProjectBudgetsResponseIncluded) HasNotifications() bool {\n\tif o != nil && o.Notifications != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *ApplianceImageBundleAllOf) HasDebugPackages() bool {\n\tif o != nil && o.DebugPackages != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (u *Updater) HasUpdate() (bool, error) {\n\treturn u.pkg.CanUpdate()\n}", "func (t *Table) GetConflicts(ctx context.Context) (conflict.ConflictSchema, durable.ConflictIndex, error) {\n\tif t.Format() == types.Format_DOLT {\n\t\tpanic(\"should use artifacts\")\n\t}\n\n\treturn t.table.GetConflicts(ctx)\n}", "func (checker *TimestampChecker) IsUpToDate(t *taskfile.Task) (bool, error) {\n\tif len(t.Sources) == 0 {\n\t\treturn false, nil\n\t}\n\n\tsources, err := Globs(t.Dir, t.Sources)\n\tif err != nil {\n\t\treturn false, nil\n\t}\n\tgenerates, err := Globs(t.Dir, t.Generates)\n\tif err != nil {\n\t\treturn false, nil\n\t}\n\n\ttimestampFile := checker.timestampFilePath(t)\n\n\t// If the file exists, add the file path to the generates.\n\t// If the generate file is old, the task will be executed.\n\t_, err = os.Stat(timestampFile)\n\tif err == nil {\n\t\tgenerates = append(generates, timestampFile)\n\t} else {\n\t\t// Create the timestamp file for the next execution when the file does not exist.\n\t\tif !checker.dry {\n\t\t\tif err := os.MkdirAll(filepath.Dir(timestampFile), 0o755); err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t\tf, err := os.Create(timestampFile)\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t\tf.Close()\n\t\t}\n\t}\n\n\ttaskTime := time.Now()\n\n\t// Compare the time of the generates and sources. If the generates are old, the task will be executed.\n\n\t// Get the max time of the generates.\n\tgenerateMaxTime, err := getMaxTime(generates...)\n\tif err != nil || generateMaxTime.IsZero() {\n\t\treturn false, nil\n\t}\n\n\t// Check if any of the source files is newer than the max time of the generates.\n\tshouldUpdate, err := anyFileNewerThan(sources, generateMaxTime)\n\tif err != nil {\n\t\treturn false, nil\n\t}\n\n\t// Modify the metadata of the file to the the current time.\n\tif !checker.dry {\n\t\tif err := os.Chtimes(timestampFile, taskTime, taskTime); err != nil {\n\t\t\treturn false, err\n\t\t}\n\t}\n\n\treturn !shouldUpdate, nil\n}", "func (ms *ManifestState) HasPendingChanges() (bool, time.Time) {\n\treturn ms.HasPendingChangesBefore(time.Now())\n}", "func isUpToDate(ctx context.Context, kube client.Client, in *v1alpha1.ReleaseParameters, observed *release.Release, s v1alpha1.ReleaseStatus) (bool, error) {\n\toc := observed.Chart\n\tif oc == nil {\n\t\treturn false, errors.New(errChartNilInObservedRelease)\n\t}\n\n\tocm := oc.Metadata\n\tif ocm == nil {\n\t\treturn false, errors.New(errChartMetaNilInObservedRelease)\n\t}\n\tif in.Chart.Name != ocm.Name {\n\t\treturn false, nil\n\t}\n\tif in.Chart.Version != ocm.Version {\n\t\treturn false, nil\n\t}\n\tdesiredConfig, err := composeValuesFromSpec(ctx, kube, in.ValuesSpec)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, errFailedToComposeValues)\n\t}\n\n\tif !reflect.DeepEqual(desiredConfig, observed.Config) {\n\t\treturn false, nil\n\t}\n\n\tchanged, err := newPatcher().hasUpdates(ctx, kube, in.PatchesFrom, s)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, errFailedToLoadPatches)\n\t}\n\n\tif changed {\n\t\treturn false, nil\n\t}\n\n\treturn true, nil\n}", "func isRevisionConflict(err error) bool {\n\tif err == nil {\n\t\treturn false\n\t}\n\t_, isConflictRevision := err.(kbfsmd.ServerErrorConflictRevision)\n\t_, isConflictPrevRoot := err.(kbfsmd.ServerErrorConflictPrevRoot)\n\t_, isConflictDiskUsage := err.(kbfsmd.ServerErrorConflictDiskUsage)\n\t_, isConditionFailed := err.(kbfsmd.ServerErrorConditionFailed)\n\t_, isConflictFolderMapping := err.(kbfsmd.ServerErrorConflictFolderMapping)\n\t_, isJournal := err.(MDJournalConflictError)\n\treturn isConflictRevision || isConflictPrevRoot ||\n\t\tisConflictDiskUsage || isConditionFailed ||\n\t\tisConflictFolderMapping || isJournal\n}", "func (o *CreateLoadBalancerRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func confirmShouldApplyExternalDependency(module *TerraformModule, dependency *TerraformModule, terragruntOptions *options.TerragruntOptions) (bool, error) {\n\tif terragruntOptions.IncludeExternalDependencies {\n\t\tterragruntOptions.Logger.Debugf(\"The --terragrunt-include-external-dependencies flag is set, so automatically including all external dependencies, and will run this command against module %s, which is a dependency of module %s.\", dependency.Path, module.Path)\n\t\treturn true, nil\n\t}\n\n\tif terragruntOptions.NonInteractive {\n\t\tterragruntOptions.Logger.Debugf(\"The --non-interactive flag is set. To avoid accidentally affecting external dependencies with a run-all command, will not run this command against module %s, which is a dependency of module %s.\", dependency.Path, module.Path)\n\t\treturn false, nil\n\t}\n\n\tstackCmd := terragruntOptions.TerraformCommand\n\tif stackCmd == \"destroy\" {\n\t\tterragruntOptions.Logger.Debugf(\"run-all command called with destroy. To avoid accidentally having destructive effects on external dependencies with run-all command, will not run this command against module %s, which is a dependency of module %s.\", dependency.Path, module.Path)\n\t\treturn false, nil\n\t}\n\n\tprompt := fmt.Sprintf(\"Module: \\t\\t %s\\nExternal dependency: \\t %s\\nShould Terragrunt apply the external dependency?\", module.Path, dependency.Path)\n\treturn shell.PromptUserForYesNo(prompt, terragruntOptions)\n}", "func (addon Addon) HasToBeApplied(addonConfiguration AddonConfiguration, skubaConfiguration *skuba.SkubaConfiguration) (bool, error) {\n\tif !addon.IsPresentForClusterVersion(addonConfiguration.ClusterVersion) {\n\t\t// TODO (ereslibre): this logic can be triggered if some registered\n\t\t// addons are not supported in all Kubernetes versions. Either:\n\t\t//\n\t\t// a) When rendering all addons on `skuba cluster init`, we skip those\n\t\t// that don't apply to the chosen Kubernetes version.\n\t\t//\n\t\t// b) When running `skuba addon upgrade apply`; in this case (hence the\n\t\t// TODO), should we trigger a deletion of the addons that are not present\n\t\t// in the new version but were present on the old Kubernetes version? For\n\t\t// now, just return that it doesn't have to be applied.\n\t\treturn false, nil\n\t}\n\t// Check whether this is a CNI addon and whether its base config has been rendered.\n\t// If it's a CNI plugin and the base config has not been rendered, we can assume\n\t// the user requested a different CNI plugin and this addon does not need to be\n\t// applied.\n\tif info, err := os.Stat(addon.addonDir()); addon.AddOnType == CniAddOn && (os.IsNotExist(err) || !info.IsDir()) {\n\t\treturn false, nil\n\t}\n\tif skubaConfiguration.AddonsVersion == nil {\n\t\treturn true, nil\n\t}\n\tcurrentAddonVersion, found := skubaConfiguration.AddonsVersion[addon.Addon]\n\tif !found {\n\t\treturn true, nil\n\t}\n\taddonVersion := kubernetes.AddonVersionForClusterVersion(addon.Addon, addonConfiguration.ClusterVersion)\n\treturn addonVersionLower(currentAddonVersion, addonVersion), nil\n}", "func (ch *compositeChangeDetector) IsSpinnakerUpToDate(ctx context.Context, svc interfaces.SpinnakerService) (bool, error) {\n\trLogger := ch.log.WithValues(\"Service\", svc.GetName())\n\tisUpToDate := true\n\tfor _, changeDetector := range ch.changeDetectors {\n\t\t// Don't run the change detector if we already know Spinnaker is not up to date\n\t\tif !isUpToDate && !changeDetector.AlwaysRun() {\n\t\t\tcontinue\n\t\t}\n\n\t\tupd, err := changeDetector.IsSpinnakerUpToDate(ctx, svc)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\tif !upd {\n\t\t\trLogger.Info(fmt.Sprintf(\"%T detected a change that needs to be reconciled\", changeDetector))\n\t\t\tch.evtRecorder.Eventf(svc, v1.EventTypeNormal, \"ConfigChanged\", \"%T detected a change that needs to be reconciled\", changeDetector)\n\t\t\tisUpToDate = false\n\t\t}\n\t}\n\treturn isUpToDate, nil\n}", "func (s RepoRevSpec) Resolved() bool {\n\treturn s.Rev != \"\" && len(s.CommitID) == 40\n}", "func checkOcBinaryCached() (bool, error) {\n\toc := oc.OcCached{}\n\tif !oc.IsCached() {\n\t\treturn false, errors.New(\"oc binary is not cached\")\n\t}\n\tlogging.Debug(\"oc binary already cached\")\n\treturn true, nil\n}", "func IsTiltPatchApplied() bool {\n\treturn true\n}", "func (s commitLogSource) shouldReturnUnfulfilled(\n\tencounteredCorruptData bool,\n\tns namespace.Metadata,\n\tshardsTimeRanges result.ShardTimeRanges,\n\topts bootstrap.RunOptions,\n) (bool, error) {\n\tif !s.opts.ReturnUnfulfilledForCorruptCommitLogFiles() {\n\t\ts.log.Info(\"returning not-unfulfilled: ReturnUnfulfilledForCorruptCommitLogFiles is false\")\n\t\treturn false, nil\n\t}\n\n\tif !encounteredCorruptData {\n\t\ts.log.Info(\"returning not-unfulfilled: no corrupt data encountered\")\n\t\treturn false, nil\n\t}\n\n\tareShardsReplicated := s.areShardsReplicated(\n\t\tns, shardsTimeRanges, opts)\n\tif !areShardsReplicated {\n\t\ts.log.Info(\"returning not-unfulfilled: replication is not enabled\")\n\t}\n\n\treturn areShardsReplicated, nil\n}", "func (o *FeatureFlag) HasModified() bool {\n\tif o != nil && o.Modified != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *OnpremUpgradePhase) HasFailed() bool {\n\tif o != nil && o.Failed != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *ProjectDeploymentRuleResponse) HasAutoDeploy() bool {\n\tif o != nil && o.AutoDeploy != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *ResourceVersion) HasChecksum() bool {\n\tif o != nil && o.Checksum != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (h *Helper) HasConflictingTypeName(typeName string) bool {\n\t// First grab the set of CRD struct names and the names of their Spec and\n\t// Status structs\n\tcleanTypeName := names.New(typeName).Camel\n\tcrdNames := h.GetCRDNames()\n\tcrdResourceNames := []string{}\n\tcrdSpecNames := []string{}\n\tcrdStatusNames := []string{}\n\n\tfor _, crdName := range crdNames {\n\t\tif h.IsIgnoredResource(crdName.Original) {\n\t\t\tcontinue\n\t\t}\n\t\tcleanResourceName := crdName.Camel\n\t\tcrdResourceNames = append(crdResourceNames, cleanResourceName)\n\t\tcrdSpecNames = append(crdSpecNames, cleanResourceName+\"Spec\")\n\t\tcrdStatusNames = append(crdStatusNames, cleanResourceName+\"Status\")\n\t}\n\treturn (inStrings(cleanTypeName, crdResourceNames) ||\n\t\tinStrings(cleanTypeName, crdSpecNames) ||\n\t\tinStrings(cleanTypeName, crdStatusNames))\n}", "func IsErrRebaseConflicts(err error) bool {\n\t_, ok := err.(ErrRebaseConflicts)\n\treturn ok\n}", "func (o *HyperflexHealthCheckPackageChecksum) HasChecksum() bool {\n\tif o != nil && o.Checksum != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (h *HealthStatus) HasChanges() bool {\n\treturn h.TaskStatus.HasChanges()\n}", "func (s *SpendJournalPruner) DependencyExists(blockHash *chainhash.Hash) bool {\n\ts.dependentsMtx.RLock()\n\tdefer s.dependentsMtx.RUnlock()\n\n\t_, ok := s.dependents[*blockHash]\n\treturn ok\n}", "func (o *CreateLoadBalancerListenersRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (d Drone) HasPackageAssigned() bool {\n\treturn len(d.Packages) > 0\n}", "func (o *Ga4ghTumourboard) HasPatientHasBeenReferredToAHereditaryCancerProgramBasedOnThisMolecularProfiling() bool {\n\tif o != nil && o.PatientHasBeenReferredToAHereditaryCancerProgramBasedOnThisMolecularProfiling != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (c *Client) IsPullRequestMerged(owner, repo string, index int64) (bool, *Response, error) {\n\tif err := escapeValidatePathSegments(&owner, &repo); err != nil {\n\t\treturn false, nil, err\n\t}\n\tstatus, resp, err := c.getStatusCode(\"GET\", fmt.Sprintf(\"/repos/%s/%s/pulls/%d/merge\", owner, repo, index), nil, nil)\n\n\tif err != nil {\n\t\treturn false, resp, err\n\t}\n\n\treturn status == 204, resp, nil\n}", "func (o *Invoice) HasCorrections() bool {\n\tif o != nil && o.Corrections != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *TransactionSplit) HasReconciled() bool {\n\tif o != nil && o.Reconciled != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func IsErrPullRequestHasMerged(err error) bool {\n\t_, ok := err.(ErrPullRequestHasMerged)\n\treturn ok\n}", "func (o *CreateUserRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (f *FlagSet) HasFlags() bool {\n\treturn len(f.formal) > 0\n}", "func (state *State) Has(name string) bool {\n\tif _, ok := state.Problem[name]; ok {\n\t\treturn true\n\t}\n\tif _, ok := state.Solution[name]; ok {\n\t\treturn true\n\t}\n\treturn false\n}", "func (f *Frugal) ContainsFrugalDefinitions() bool {\n\treturn len(f.Scopes)+len(f.Services) > 0\n}", "func IsUpToDate(actual, desired interface{}, opts ...LateInitOption) (bool, error) {\n\tvalDesired := reflect.ValueOf(desired)\n\tif valDesired.Kind() != reflect.Ptr {\n\t\treturn false, errors.Errorf(\"desired must be of pointer kind, got: %s\", valDesired.Kind().String())\n\t}\n\tnewDesired := func() interface{} {\n\t\treturn reflect.New(reflect.TypeOf(valDesired.Elem().Interface())).Interface()\n\t}\n\tactualConfig := newDesired()\n\tif _, err := LateInitializeFromResponse(\"\", actualConfig, actual, opts...); err != nil {\n\t\treturn false, err\n\t}\n\n\tjsonPatch, err := awsclients.CreateJSONPatch(actualConfig, desired)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tpatch := newDesired()\n\tif err := json.Unmarshal(jsonPatch, patch); err != nil {\n\t\treturn false, err\n\t}\n\treturn cmp.Equal(newDesired(), patch, cmpopts.EquateEmpty(),\n\t\tcmpopts.IgnoreTypes(&xpv1.Reference{}, &xpv1.Selector{}, []xpv1.Reference{})), nil\n}", "func (o *ApplianceImageBundleAllOf) HasStatusMessage() bool {\n\tif o != nil && o.StatusMessage != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *StoragePhysicalDiskAllOf) HasRunningFirmware() bool {\n\tif o != nil && o.RunningFirmware != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func Check(projectName string, parallel int, owner, repo string, prNumber int) error {\n\tif _, err := os.Stat(EffritFileName); err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn errors.New(EffritFileName)\n\t\t}\n\t}\n\tvar packages = struct {\n\t\tPackages []Package `json:\"packages\"`\n\t}{\n\t\tPackages: make([]Package, 0),\n\t}\n\tdata, err := os.ReadFile(EffritFileName)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := json.Unmarshal(data, &packages); err != nil {\n\t\treturn err\n\t}\n\n\t// Construct a map of package names and their imports.\n\tpackageMap := make(map[string][]string)\n\n\tfor _, p := range packages.Packages {\n\t\tsort.Strings(p.DependedOnByNames)\n\t\tpackageMap[p.FullName] = p.DependedOnByNames\n\t}\n\n\tpkgs, err := Scan(projectName, parallel)\n\tif err != nil {\n\t\treturn err\n\t}\n\tpkgs.Dump()\n\tpackages.Packages = make([]Package, 0)\n\tdata, _ = os.ReadFile(EffritFileName)\n\t_ = json.Unmarshal(data, &packages)\n\n\townersToContact := make(map[string]string)\n\t// Compare the new result with the old result's map data.\n\tfor _, p := range packages.Packages {\n\t\tdependents := packageMap[p.FullName]\n\t\tfor _, dep := range dependents {\n\t\t\tif !contains(p.DependedOnByNames, dep) {\n\t\t\t\towner, err := getOwnerForFile(p.Dir, p.GoFiles)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\townersToContact[owner] = dep\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\tif len(ownersToContact) > 0 {\n\t\tfmt.Print(\"Contacting owners about package dependency changes...\")\n\t\tif err := contactOwners(ownersToContact, owner, repo, prNumber); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfmt.Println(\"done.\")\n\t}\n\treturn nil\n}", "func (r *Reservation) IsSuccessfullyDeployed() bool {\n\tsucceeded := false\n\tif len(r.Results) >= len(r.Workloads(\"\")) {\n\t\tsucceeded = true\n\t\tfor _, result := range r.Results {\n\t\t\tif result.State != generated.ResultStateOK {\n\t\t\t\tsucceeded = false\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn succeeded\n}", "func IsConflict(err error) bool {\n\treturn ReasonForError(err) == http.StatusConflict\n}", "func IsConflict(err error) bool {\n\t// TODO(horwitz): This is supposed to be fixed soon. It's a bug in the OCI API that causes a 409 to\n\t// be returned instead of a 412.\n\treturn IsError(err, \"409\", \"Conflict\") || IsError(err, \"412\", \"NoEtagMatch\")\n}", "func (o *CreateRouteRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (t ResolvedPipelineRunTask) HasRemainingRetries() bool {\n\tvar retriesDone int\n\tif t.IsCustomTask() {\n\t\tif t.Run == nil {\n\t\t\treturn true\n\t\t}\n\t\tretriesDone = len(t.Run.Status.RetriesStatus)\n\t} else {\n\t\tif t.TaskRun == nil {\n\t\t\treturn true\n\t\t}\n\t\tretriesDone = len(t.TaskRun.Status.RetriesStatus)\n\t}\n\treturn retriesDone < t.PipelineTask.Retries\n}", "func (pkg *goPackage) isBuildable(c *config.Config) bool {\n\treturn pkg.firstGoFile() != \"\" || !pkg.proto.sources.isEmpty()\n}", "func (o *ReadConsumptionAccountRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *AccountDashboardStatistic) HasReferrals() bool {\n\tif o != nil && o.Referrals != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (bm *DockerBenchmarker) CheckRunUpdateOnly() {\n\tdfiles := []string{}\n\n\tfor file, df := range bm.dfiles {\n\t\t// apt\n\t\tupdateIdxs := df.LookupInstructionAndContent(dockerfile.Run, `apt\\s+(-y\\s)?update`)\n\t\tinstallIdx := df.LookupInstructionAndContent(dockerfile.Run, `apt\\s+(-y\\s)?install`)\n\n\t\tupdateOnly, _ := diffArray(updateIdxs, installIdx)\n\t\tif len(updateOnly) > 0 {\n\t\t\tdfiles = append(dfiles, file)\n\t\t}\n\n\t\t// apt-get\n\t\tupdateIdxs = df.LookupInstructionAndContent(dockerfile.Run, `apt-get\\s+(-y\\s)?update`)\n\t\tinstallIdx = df.LookupInstructionAndContent(dockerfile.Run, `apt-get\\s+(-y\\s)?install`)\n\n\t\tupdateOnly, _ = diffArray(updateIdxs, installIdx)\n\t\tif len(updateOnly) > 0 {\n\t\t\tdfiles = append(dfiles, file)\n\t\t}\n\n\t\t// apk\n\t\tupdateIdxs = df.LookupInstructionAndContent(dockerfile.Run, `apk\\s+(-y\\s)?update`)\n\t\tinstallIdx = df.LookupInstructionAndContent(dockerfile.Run, `apk\\s+(-y\\s)?add`)\n\n\t\tupdateOnly, _ = diffArray(updateIdxs, installIdx)\n\t\tif len(updateOnly) > 0 {\n\t\t\tdfiles = append(dfiles, file)\n\t\t}\n\n\t}\n\n\tbm.violationReport.AddViolation(benchmark.CIS_4_7, dfiles)\n}" ]
[ "0.6893376", "0.5742307", "0.5700375", "0.55312514", "0.5376664", "0.529813", "0.52914315", "0.5250208", "0.5233584", "0.52170324", "0.5205163", "0.5082322", "0.5065302", "0.50519806", "0.5041187", "0.5038167", "0.49985608", "0.4966648", "0.4961827", "0.49594378", "0.49502417", "0.49449858", "0.49447072", "0.49331287", "0.4921041", "0.49113634", "0.49006942", "0.49004707", "0.4898781", "0.48882565", "0.48875743", "0.48839173", "0.48714435", "0.48702884", "0.48681298", "0.48623174", "0.48576906", "0.4857065", "0.4843487", "0.4829589", "0.48227775", "0.48096287", "0.4806294", "0.47992086", "0.4797668", "0.47935528", "0.47864297", "0.478036", "0.4778528", "0.47781134", "0.47766116", "0.47709468", "0.4770734", "0.47680873", "0.4767343", "0.476508", "0.47640526", "0.47619042", "0.47603196", "0.47573024", "0.4748054", "0.47195926", "0.4719548", "0.4717024", "0.47079355", "0.47023633", "0.46831784", "0.46659875", "0.46640357", "0.46633366", "0.4659567", "0.4657563", "0.46553373", "0.4654615", "0.46536237", "0.46468812", "0.46327406", "0.4632736", "0.46322623", "0.462521", "0.46162271", "0.46132335", "0.46075204", "0.46046093", "0.45973572", "0.4596593", "0.4594224", "0.45822346", "0.45796967", "0.45739534", "0.45738387", "0.45729238", "0.45711342", "0.45677787", "0.45620325", "0.45600507", "0.4559082", "0.45564297", "0.4554789", "0.4545762" ]
0.67191136
1
MarshalJSON implementation here is forced by the embedded check value here.
func (d DiffCheckValues) MarshalJSON() ([]byte, error) { if d.Check == nil { return json.Marshal(nil) } return json.Marshal(d.Check) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func mustMarshalJSON(v interface{}) []byte {\n\tb, err := json.Marshal(v)\n\tif err != nil {\n\t\tpanic(\"marshal json: \" + err.Error())\n\t}\n\treturn b\n}", "func (sc SafetyCheck) MarshalJSON() ([]byte, error) {\n\tsc.Kind = KindSafetyCheck\n\tobjectMap := make(map[string]interface{})\n\tif sc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = sc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (v checkBadUARule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules5(w, v)\n}", "func (v AndroidCheckRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules12(w, v)\n}", "func (easc EnsureAvailabilitySafetyCheck) MarshalJSON() ([]byte, error) {\n\teasc.Kind = KindEnsureAvailability\n\tobjectMap := make(map[string]interface{})\n\tif easc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = easc.PartitionID\n\t}\n\tif easc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = easc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (cvpbo CheckValuePropertyBatchOperation) MarshalJSON() ([]byte, error) {\n\tcvpbo.Kind = KindCheckValue\n\tobjectMap := make(map[string]interface{})\n\tobjectMap[\"Value\"] = cvpbo.Value\n\tif cvpbo.PropertyName != nil {\n\t\tobjectMap[\"PropertyName\"] = cvpbo.PropertyName\n\t}\n\tif cvpbo.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = cvpbo.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (JSONIndent) IsMarshal7951Arg() {}", "func (*RFC7951JSONConfig) IsEncodeTypedValueOpt() {}", "func (j *JSON) CanMarshalWithCallback() bool {\n\treturn false\n}", "func (n NullJSON) MarshalJSON() ([]byte, error) { return nulljson(n.Valid, n.JSONVal) }", "func (v ItemCheckResponse) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson6a975c40EncodeJsonBenchmark2(w, v)\n}", "func (wfirsc WaitForInbuildReplicaSafetyCheck) MarshalJSON() ([]byte, error) {\n\twfirsc.Kind = KindWaitForInbuildReplica\n\tobjectMap := make(map[string]interface{})\n\tif wfirsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = wfirsc.PartitionID\n\t}\n\tif wfirsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = wfirsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func mustJSON(thing interface{}) string {\n\ts, err := json.Marshal(thing)\n\tif err != nil {\n\t\treturn \"invalid json: \" + err.Error()\n\t}\n\treturn string(s)\n}", "func (wfppsc WaitForPrimaryPlacementSafetyCheck) MarshalJSON() ([]byte, error) {\n\twfppsc.Kind = KindWaitForPrimaryPlacement\n\tobjectMap := make(map[string]interface{})\n\tif wfppsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = wfppsc.PartitionID\n\t}\n\tif wfppsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = wfppsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (v checkBadUARule) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson795c59c6EncodeGrapeGuardRules5(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (cepbo CheckExistsPropertyBatchOperation) MarshalJSON() ([]byte, error) {\n\tcepbo.Kind = KindCheckExists\n\tobjectMap := make(map[string]interface{})\n\tif cepbo.Exists != nil {\n\t\tobjectMap[\"Exists\"] = cepbo.Exists\n\t}\n\tif cepbo.PropertyName != nil {\n\t\tobjectMap[\"PropertyName\"] = cepbo.PropertyName\n\t}\n\tif cepbo.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = cepbo.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (psc PartitionSafetyCheck) MarshalJSON() ([]byte, error) {\n\tpsc.Kind = KindPartitionSafetyCheck\n\tobjectMap := make(map[string]interface{})\n\tif psc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = psc.PartitionID\n\t}\n\tif psc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = psc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (snsc SeedNodeSafetyCheck) MarshalJSON() ([]byte, error) {\n\tsnsc.Kind = KindEnsureSeedNodeQuorum\n\tobjectMap := make(map[string]interface{})\n\tif snsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = snsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (*RFC7951JSONConfig) IsMarshal7951Arg() {}", "func (v Allow) EncodeJSON(b []byte) []byte {\n\tb = append(b, '{', '\"', 'b', 'a', 'l', 'a', 'n', 'c', 'e', '_', 'e', 'x', 'e', 'm', 'p', 't', 'i', 'o', 'n', 's', '\"', ':', '[')\n\tfor i, elem := range v.BalanceExemptions {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = elem.EncodeJSON(b)\n\t}\n\tb = append(b, ']', ',', '\"', 'c', 'a', 'l', 'l', '_', 'm', 'e', 't', 'h', 'o', 'd', 's', '\"', ':', '[')\n\tfor i, elem := range v.CallMethods {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = json.AppendString(b, elem)\n\t}\n\tb = append(b, `],\"errors\":[`...)\n\tfor i, elem := range v.Errors {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = elem.EncodeJSON(b)\n\t}\n\tb = append(b, ']', ',', '\"', 'h', 'i', 's', 't', 'o', 'r', 'i', 'c', 'a', 'l', '_', 'b', 'a', 'l', 'a', 'n', 'c', 'e', '_', 'l', 'o', 'o', 'k', 'u', 'p', '\"', ':')\n\tb = json.AppendBool(b, v.HistoricalBalanceLookup)\n\tb = append(b, ',', '\"', 'm', 'e', 'm', 'p', 'o', 'o', 'l', '_', 'c', 'o', 'i', 'n', 's', '\"', ':')\n\tb = json.AppendBool(b, v.MempoolCoins)\n\tb = append(b, ',', '\"', 'o', 'p', 'e', 'r', 'a', 't', 'i', 'o', 'n', '_', 's', 't', 'a', 't', 'u', 's', 'e', 's', '\"', ':', '[')\n\tfor i, elem := range v.OperationStatuses {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = elem.EncodeJSON(b)\n\t}\n\tb = append(b, ']', ',', '\"', 'o', 'p', 'e', 'r', 'a', 't', 'i', 'o', 'n', '_', 't', 'y', 'p', 'e', 's', '\"', ':', '[')\n\tfor i, elem := range v.OperationTypes {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = json.AppendString(b, elem)\n\t}\n\tb = append(b, \"],\"...)\n\tif v.TimestampStartIndex.Set {\n\t\tb = append(b, '\"', 't', 'i', 'm', 'e', 's', 't', 'a', 'm', 'p', '_', 's', 't', 'a', 'r', 't', '_', 'i', 'n', 'd', 'e', 'x', '\"', ':')\n\t\tb = json.AppendInt(b, v.TimestampStartIndex.Value)\n\t\tb = append(b, \",\"...)\n\t}\n\tb[len(b)-1] = '}'\n\treturn b\n}", "func (v singleRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules1(w, v)\n}", "func (v commonRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules4(w, v)\n}", "func (v OrderCheckResponse) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson6a975c40EncodeJsonBenchmark1(w, v)\n}", "func (v NullableServiceCheckStatus) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(v.value)\n}", "func (o ServiceCheck) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\ttoSerialize[\"check\"] = o.Check\n\ttoSerialize[\"host_name\"] = o.HostName\n\tif o.Message != nil {\n\t\ttoSerialize[\"message\"] = o.Message\n\t}\n\ttoSerialize[\"status\"] = o.Status\n\ttoSerialize[\"tags\"] = o.Tags\n\tif o.Timestamp != nil {\n\t\ttoSerialize[\"timestamp\"] = o.Timestamp\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func jsonify(v interface{}) string { return string(mustMarshalJSON(v)) }", "func (v proxyRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules3(w, v)\n}", "func (epqsc EnsurePartitionQurumSafetyCheck) MarshalJSON() ([]byte, error) {\n\tepqsc.Kind = KindEnsurePartitionQuorum\n\tobjectMap := make(map[string]interface{})\n\tif epqsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = epqsc.PartitionID\n\t}\n\tif epqsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = epqsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (v ItemCheckResponse) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson6a975c40EncodeJsonBenchmark2(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v reqGuardRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules2(w, v)\n}", "func testJSONMarshal(t *testing.T, v interface{}, want string) {\n\tj, err := json.Marshal(v)\n\tif err != nil {\n\t\tt.Errorf(\"Unable to marshal JSON for %v\", v)\n\t}\n\n\tw := new(bytes.Buffer)\n\terr = json.Compact(w, []byte(want))\n\tif err != nil {\n\t\tt.Errorf(\"String is not valid json: %s\", want)\n\t}\n\n\tif w.String() != string(j) {\n\t\tt.Errorf(\"json.Marshal(%q) returned %s, want %s\", v, j, w)\n\t}\n\n\t// now go the other direction and make sure things unmarshal as expected\n\tu := reflect.ValueOf(v).Interface()\n\tif err := json.Unmarshal([]byte(want), u); err != nil {\n\t\tt.Errorf(\"Unable to unmarshal JSON for %v\", want)\n\t}\n\n\tif !reflect.DeepEqual(v, u) {\n\t\tt.Errorf(\"json.Unmarshal(%q) returned %s, want %s\", want, u, v)\n\t}\n}", "func testJSONMarshal(t *testing.T, v interface{}, want string) {\n\tj, err := json.Marshal(v)\n\tif err != nil {\n\t\tt.Errorf(\"Unable to marshal JSON for %v\", v)\n\t}\n\n\tw := new(bytes.Buffer)\n\terr = json.Compact(w, []byte(want))\n\tif err != nil {\n\t\tt.Errorf(\"String is not valid json: %s\", want)\n\t}\n\n\tif w.String() != string(j) {\n\t\tt.Errorf(\"json.Marshal(%q) returned %s, want %s\", v, j, w)\n\t}\n\n\t// now go the other direction and make sure things unmarshal as expected\n\tu := reflect.ValueOf(v).Interface()\n\tif err := json.Unmarshal([]byte(want), u); err != nil {\n\t\tt.Errorf(\"Unable to unmarshal JSON for %v: %v\", want, err)\n\t}\n\n\tif !reflect.DeepEqual(v, u) {\n\t\tt.Errorf(\"json.Unmarshal(%q) returned %s, want %s\", want, u, v)\n\t}\n}", "func MustMarshalJSON(v interface{}) string {\n\tbuf, err := json.Marshal(v)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn string(buf)\n}", "func marshallJSONWithoutError(value interface{}) []byte {\n\tj, err := json.Marshal(value)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn j\n}", "func (cspbo CheckSequencePropertyBatchOperation) MarshalJSON() ([]byte, error) {\n\tcspbo.Kind = KindCheckSequence\n\tobjectMap := make(map[string]interface{})\n\tif cspbo.SequenceNumber != nil {\n\t\tobjectMap[\"SequenceNumber\"] = cspbo.SequenceNumber\n\t}\n\tif cspbo.PropertyName != nil {\n\t\tobjectMap[\"PropertyName\"] = cspbo.PropertyName\n\t}\n\tif cspbo.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = cspbo.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (o HealthCheckResult) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.NullableMessage == nil {\n\t\tif o.isExplicitNullNullableMessage {\n\t\t\ttoSerialize[\"NullableMessage\"] = o.NullableMessage\n\t\t}\n\t} else {\n\t\ttoSerialize[\"NullableMessage\"] = o.NullableMessage\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (v UnloadCheckResponse) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson6a975c40EncodeJsonBenchmark(w, v)\n}", "func MustJsonEncode(v interface{}) string {\n\tbuf, err := ffjson.Marshal(v)\n\tif err != nil {\n\t\tfmt.Printf(\"json encode %v err: %v\", v, err)\n\t}\n\treturn string(buf)\n}", "func mustMarshal(val interface{}) []byte {\n\tif b, err := asn1.Marshal(val); err != nil {\n\t\tpanic(err)\n\t} else {\n\t\treturn b\n\t}\n}", "func (wfrsc WaitForReconfigurationSafetyCheck) MarshalJSON() ([]byte, error) {\n\twfrsc.Kind = KindWaitForReconfiguration\n\tobjectMap := make(map[string]interface{})\n\tif wfrsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = wfrsc.PartitionID\n\t}\n\tif wfrsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = wfrsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func MustMarshalJSON(data interface{}) []byte {\n\tb, err := json.Marshal(data)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn b\n}", "func (v ExtFilter) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules11(w, v)\n}", "func (v OrderCheckResponse) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson6a975c40EncodeJsonBenchmark1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v PackoutRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules8(w, v)\n}", "func (h ScanningOrFalse) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(h.Value)\n}", "func ensureAppliedConfigAnnotation(scheme *runtime.Scheme, obj GCRuntimeObject) ([]byte, error) {\n\t// Instead of encoding the json directly, we first convert it to \"Unstructured\", i.e. a map[string]interface{}\n\t// We do this to remove a \"status\" item if there is any. Not all status fields are marked as `json:\",omitempty`,\n\t// so their default value will be serialized too. The status field itself is marked as `json:\",omitempty`, which\n\t// is useless, as struct values are never empty in golang.\n\tobjUnstructured := unstructured.Unstructured{}\n\n\terr := scheme.Convert(obj, &objUnstructured, nil)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to convert to unstructured item: %w\", err)\n\t}\n\n\tdelete(objUnstructured.Object, \"status\")\n\n\tobjEncoded, err := runtime.Encode(unstructured.UnstructuredJSONScheme, &objUnstructured)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to prepare applied configuration metadata: %w\", err)\n\t}\n\n\tannotations := obj.GetAnnotations()\n\tif annotations == nil {\n\t\tannotations = map[string]string{}\n\t}\n\n\tannotations[lastAppliedAnnotation] = string(objEncoded)\n\tobj.SetAnnotations(annotations)\n\n\treturn objEncoded, nil\n}", "func (v AndroidCheckRule) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson795c59c6EncodeGrapeGuardRules12(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func MustMarshalJSON(o interface{}) []byte {\n\treturn cdc.MustMarshalJSON(o)\n}", "func (c CheckNameAvailabilityResponseBody) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"absCode\", c.AbsCode)\n\tpopulate(objectMap, \"message\", c.Message)\n\tpopulate(objectMap, \"valid\", c.Valid)\n\treturn json.Marshal(objectMap)\n}", "func MustMarshal(v interface{}) []byte {\n\tbytes, err := json.Marshal(v)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn bytes\n}", "func TestMarshal(t *testing.T) {\n\tfor _, c := range []struct {\n\t\tname string\n\t\tfieldPairs []interface{}\n\t\t// errNeedle is \"\" if we expect no error. Otherwise, it is a string that\n\t\t// we expect to see in the resulting err.Error().\n\t\terrNeedle string\n\t}{\n\t\t{\n\t\t\t\"no fields\",\n\t\t\t[]interface{}{},\n\t\t\t\"\",\n\t\t},\n\t\t{\n\t\t\t\"simple\",\n\t\t\t[]interface{}{\"k0\", \"v0\"},\n\t\t\t\"\",\n\t\t},\n\t\t{\n\t\t\t\"mixed value types\",\n\t\t\t// Numeric types turn into float64s in JSON.\n\t\t\t[]interface{}{\"k0\", \"v0\", \"k1\", float64(1), \"k2\", true},\n\t\t\t\"\",\n\t\t},\n\t\t{\n\t\t\t\"odd field pairs\",\n\t\t\t[]interface{}{\"k0\", \"v0\", \"k1\"},\n\t\t\t\"even\",\n\t\t},\n\t\t{\n\t\t\t\"non-string key\",\n\t\t\t[]interface{}{0, \"v0\"},\n\t\t\t\"string\",\n\t\t},\n\t\t{\n\t\t\t\"duplicate keys\",\n\t\t\t[]interface{}{\"k0\", \"v0\", \"k0\", \"v1\"},\n\t\t\t\"duplicate\",\n\t\t},\n\t} {\n\t\tt.Run(c.name, func(t *testing.T) {\n\t\t\tmarshalOK := c.errNeedle == \"\"\n\t\t\ts, err := Marshal(c.name, c.fieldPairs)\n\t\t\tif got, want := err == nil, marshalOK; got != want {\n\t\t\t\tt.Fatalf(\"got %v, want %v\", got, want)\n\t\t\t}\n\t\t\tif !marshalOK {\n\t\t\t\tif !strings.Contains(err.Error(), c.errNeedle) {\n\t\t\t\t\tt.Errorf(\"error %q does not contain expected substring %q\", err.Error(), c.errNeedle)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\t\t\tvar m map[string]interface{}\n\t\t\terr = json.Unmarshal([]byte(s), &m)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatalf(\"unmarshaling failed: %v\", err)\n\t\t\t}\n\t\t\t// The +1 is for the eventType.\n\t\t\tif got, want := len(m), (len(c.fieldPairs)/2)+1; got != want {\n\t\t\t\tt.Errorf(\"got %v, want %v\", got, want)\n\t\t\t}\n\t\t\ttyp, ok := m[eventTypeFieldKey]\n\t\t\tif ok {\n\t\t\t\tif got, want := typ, c.name; got != want {\n\t\t\t\t\tt.Errorf(\"got %v, want %v\", got, want)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tt.Errorf(\"eventType field not marshaled\")\n\t\t\t}\n\t\t\tfor i := 0; i < len(c.fieldPairs); i++ {\n\t\t\t\tkey := c.fieldPairs[i].(string)\n\t\t\t\ti++\n\t\t\t\tvalue := c.fieldPairs[i]\n\t\t\t\tmvalue, ok := m[key]\n\t\t\t\tif !ok {\n\t\t\t\t\tt.Errorf(\"field with key %q not marshaled\", key)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif got, want := mvalue, value; got != want {\n\t\t\t\t\tt.Errorf(\"got %v(%T), want %v(%T)\", got, got, want, want)\n\t\t\t\t}\n\t\t\t}\n\t\t})\n\t}\n}", "func (wfpssc WaitForPrimarySwapSafetyCheck) MarshalJSON() ([]byte, error) {\n\twfpssc.Kind = KindWaitForPrimarySwap\n\tobjectMap := make(map[string]interface{})\n\tif wfpssc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = wfpssc.PartitionID\n\t}\n\tif wfpssc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = wfpssc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (src UpdateVerificationFlowBody) MarshalJSON() ([]byte, error) {\n\tif src.UpdateVerificationFlowWithCodeMethod != nil {\n\t\treturn json.Marshal(&src.UpdateVerificationFlowWithCodeMethod)\n\t}\n\n\tif src.UpdateVerificationFlowWithLinkMethod != nil {\n\t\treturn json.Marshal(&src.UpdateVerificationFlowWithLinkMethod)\n\t}\n\n\treturn nil, nil // no data in oneOf schemas\n}", "func (n NullBool) MarshalJSON() ([]byte, error) { return nulljson(n.Valid, n.Bool) }", "func marshal() {\n\tfmt.Println(\"=== json.marshal ===\")\n\tryan := &Person{\"Ryan\", 25}\n\twire, err := json.Marshal(ryan)\n\tcheck(err)\n\tfmt.Println(string(wire))\n}", "func (v anyOfRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker47(w, v)\n}", "func (v GuardRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules10(w, v)\n}", "func (rmwaps ResourceModelWithAllowedPropertySet)MarshalJSON() ([]byte, error){\n objectMap := make(map[string]interface{})\n if(rmwaps.Location != nil) {\n objectMap[\"location\"] = rmwaps.Location\n }\n if(rmwaps.ManagedBy != nil) {\n objectMap[\"managedBy\"] = rmwaps.ManagedBy\n }\n if(rmwaps.Kind != nil) {\n objectMap[\"kind\"] = rmwaps.Kind\n }\n if(rmwaps.Tags != nil) {\n objectMap[\"tags\"] = rmwaps.Tags\n }\n if(rmwaps.Identity != nil) {\n objectMap[\"identity\"] = rmwaps.Identity\n }\n if(rmwaps.Sku != nil) {\n objectMap[\"sku\"] = rmwaps.Sku\n }\n if(rmwaps.Plan != nil) {\n objectMap[\"plan\"] = rmwaps.Plan\n }\n return json.Marshal(objectMap)\n }", "func (e EntityNameAvailabilityCheckOutput) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"message\", e.Message)\n\tpopulate(objectMap, \"nameAvailable\", e.NameAvailable)\n\tpopulate(objectMap, \"reason\", e.Reason)\n\treturn json.Marshal(objectMap)\n}", "func (v singleRule) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson795c59c6EncodeGrapeGuardRules1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (r *HealthCheck) marshal(c *Client) ([]byte, error) {\n\tm, err := expandHealthCheck(c, r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshalling HealthCheck: %w\", err)\n\t}\n\n\treturn json.Marshal(m)\n}", "func (t JSONTime)MarshalJSON(b []byte) (err error) {\n s := string(b)\n logging.Trace(currentCtx,\"JSONTime(String)=%s\",s)\n if s == \"null\" || s == \"\" {\n t.Time = time.Time{}\n return\n }\n tt, err := time.Parse(jsonTimeLayout, s)\n t = JSONTime{tt}\n return\n}", "func (c CheckNameAvailabilityRequestBody) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"name\", c.Name)\n\tpopulate(objectMap, \"type\", c.Type)\n\treturn json.Marshal(objectMap)\n}", "func mustToRawJson(v interface{}) (string, error) {\n\tbuf := new(bytes.Buffer)\n\tenc := json.NewEncoder(buf)\n\tenc.SetEscapeHTML(false)\n\terr := enc.Encode(&v)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn strings.TrimSuffix(buf.String(), \"\\n\"), nil\n}", "func (v prefixRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker13(w, v)\n}", "func (v matchRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker19(w, v)\n}", "func (e Effect) MarshalJSON() ([]byte, error) {\n\tif bool(e) {\n\t\treturn []byte(`\"Allow\"`), nil\n\t} else {\n\t\treturn []byte(`\"Deny\"`), nil\n\t}\n}", "func marshal(v interface{}) []byte {\n\tb, err := json.Marshal(v)\n\tassert(err == nil, \"marshal error: %s\", err)\n\treturn b\n}", "func (j JSON) MarshalJSON() ([]byte, error) {\n\tif j.Valid {\n\t\treturn json.Marshal(j.Map)\n\t}\n\n\treturn json.Marshal(nil)\n}", "func (v NickFail) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson42239ddeEncodeGithubComKhliengDispatchServer13(w, v)\n}", "func (rmwaps ResourceModelWithAllowedPropertySetIdentity)MarshalJSON() ([]byte, error){\n objectMap := make(map[string]interface{})\n if(rmwaps.Type != \"\") {\n objectMap[\"type\"] = rmwaps.Type\n }\n return json.Marshal(objectMap)\n }", "func (v empty) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdproto(w, v)\n}", "func (c *JsonCodec) CanMarshalWithCallback() bool {\n\treturn false\n}", "func (v JSONB) MarshalJSON() ([]byte, error) {\n\tif bytes.Equal(v, []byte{}) || bytes.Equal(v, []byte(\"null\")) {\n\t\treturn []byte(\"{}\"), nil\n\t}\n\n\treturn v, nil\n}", "func (spbi SuccessfulPropertyBatchInfo) MarshalJSON() ([]byte, error) {\n\tspbi.Kind = KindSuccessful\n\tobjectMap := make(map[string]interface{})\n\tif spbi.Properties != nil {\n\t\tobjectMap[\"Properties\"] = spbi.Properties\n\t}\n\tif spbi.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = spbi.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (o InlineObject901) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.LogicalTest == nil {\n\t\tif o.isExplicitNullLogicalTest {\n\t\t\ttoSerialize[\"logicalTest\"] = o.LogicalTest\n\t\t}\n\t} else {\n\t\ttoSerialize[\"logicalTest\"] = o.LogicalTest\n\t}\n\tif o.ValueIfTrue == nil {\n\t\tif o.isExplicitNullValueIfTrue {\n\t\t\ttoSerialize[\"valueIfTrue\"] = o.ValueIfTrue\n\t\t}\n\t} else {\n\t\ttoSerialize[\"valueIfTrue\"] = o.ValueIfTrue\n\t}\n\tif o.ValueIfFalse == nil {\n\t\tif o.isExplicitNullValueIfFalse {\n\t\t\ttoSerialize[\"valueIfFalse\"] = o.ValueIfFalse\n\t\t}\n\t} else {\n\t\ttoSerialize[\"valueIfFalse\"] = o.ValueIfFalse\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (v Bounds) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoBrowser28(w, v)\n}", "func (js JSONSerializable) MarshalJSON() ([]byte, error) {\n\tif !js.Valid {\n\t\treturn json.Marshal(nil)\n\t}\n\tjsWithHex := replaceBytesWithHex(js.Val)\n\treturn json.Marshal(jsWithHex)\n}", "func (v NullableSyntheticsAssertion) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(v.value)\n}", "func (j *HealthcheckRequest) MarshalJSON() ([]byte, error) {\n\tvar buf fflib.Buffer\n\tif j == nil {\n\t\tbuf.WriteString(\"null\")\n\t\treturn buf.Bytes(), nil\n\t}\n\terr := j.MarshalJSONBuf(&buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn buf.Bytes(), nil\n}", "func (out String) MarshalJSON() ([]byte, error) {\n\tif out.Valid {\n\t\treturn json.Marshal(out.String)\n\t}\n\treturn json.Marshal(nil)\n}", "func suppressEquivalentJSON(k, old, new string, d *schema.ResourceData) bool {\n\tvar oldObj, newObj interface{}\n\tif err := json.Unmarshal([]byte(old), &oldObj); err != nil {\n\t\treturn false\n\t}\n\tif err := json.Unmarshal([]byte(new), &newObj); err != nil {\n\t\treturn false\n\t}\n\treturn reflect.DeepEqual(oldObj, newObj)\n}", "func (f FailJSONEncoding) MarshalJSON() ([]byte, error) {\n\treturn nil, fmt.Errorf(\"error\")\n}", "func (c ContentKeyPolicyUnknownRestriction) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tobjectMap[\"@odata.type\"] = \"#Microsoft.Media.ContentKeyPolicyUnknownRestriction\"\n\treturn json.Marshal(objectMap)\n}", "func (v item) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonD2b7633eEncodeGithubComZhekabyGoGeneratorMongoRequestwrapperTests(w, v)\n}", "func (v booleanField) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker45(w, v)\n}", "func (out Bool) MarshalJSON() ([]byte, error) {\n\tif out.Valid {\n\t\treturn json.Marshal(out.Bool)\n\t}\n\treturn json.Marshal(nil)\n}", "func MarshalValid(obj MarshalValidater) ([]byte, error) {\n\terr := obj.Validate()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn obj.Marshal()\n}", "func (v InstallabilityErrorArgument) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage35(w, v)\n}", "func (v booleanField) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson390b7126EncodeGithubComChancedPicker45(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (n NullInt64) MarshalJSON() ([]byte, error) { return nulljson(n.Valid, n.Int64) }", "func (v allOfRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker48(w, v)\n}", "func (c CheckNameAvailabilityParameters) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"name\", c.Name)\n\tpopulate(objectMap, \"type\", c.Type)\n\treturn json.Marshal(objectMap)\n}", "func Marshal(v interface{}) ([]byte, error) {\n\tif ImplementsPreJSONMarshaler(v) {\n\t\terr := v.(PreJSONMarshaler).PreMarshalJSON()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn json.Marshal(v)\n}", "func (v PackoutRule) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson795c59c6EncodeGrapeGuardRules8(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (c CheckNameAvailabilityInput) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"name\", c.Name)\n\tpopulate(objectMap, \"type\", c.Type)\n\treturn json.Marshal(objectMap)\n}", "func (v Boo) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson42239ddeEncodeMsgpJson(w, v)\n}", "func (o InlineObject885) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.X == nil {\n\t\tif o.isExplicitNullX {\n\t\t\ttoSerialize[\"x\"] = o.X\n\t\t}\n\t} else {\n\t\ttoSerialize[\"x\"] = o.X\n\t}\n\tif o.Alpha == nil {\n\t\tif o.isExplicitNullAlpha {\n\t\t\ttoSerialize[\"alpha\"] = o.Alpha\n\t\t}\n\t} else {\n\t\ttoSerialize[\"alpha\"] = o.Alpha\n\t}\n\tif o.Beta == nil {\n\t\tif o.isExplicitNullBeta {\n\t\t\ttoSerialize[\"beta\"] = o.Beta\n\t\t}\n\t} else {\n\t\ttoSerialize[\"beta\"] = o.Beta\n\t}\n\tif o.Cumulative == nil {\n\t\tif o.isExplicitNullCumulative {\n\t\t\ttoSerialize[\"cumulative\"] = o.Cumulative\n\t\t}\n\t} else {\n\t\ttoSerialize[\"cumulative\"] = o.Cumulative\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func marshal(v interface{}) string {\n\tb, err := json.Marshal(v)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn string(b)\n}", "func (c CheckAvailabilityParameters) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"id\", c.ID)\n\tpopulate(objectMap, \"isAvailiable\", c.IsAvailiable)\n\tpopulate(objectMap, \"location\", c.Location)\n\tpopulate(objectMap, \"name\", c.Name)\n\tpopulate(objectMap, \"sku\", c.SKU)\n\tpopulate(objectMap, \"tags\", c.Tags)\n\tpopulate(objectMap, \"type\", c.Type)\n\treturn json.Marshal(objectMap)\n}" ]
[ "0.6703141", "0.6430447", "0.628579", "0.62582934", "0.6081891", "0.60665685", "0.6062087", "0.60332453", "0.60280246", "0.5984815", "0.5958768", "0.5944417", "0.5929845", "0.59293", "0.59254575", "0.5917242", "0.5909856", "0.58898854", "0.58615434", "0.5857886", "0.5848257", "0.5833942", "0.5829239", "0.5828556", "0.5822303", "0.58206034", "0.5807151", "0.5790106", "0.5789338", "0.5778644", "0.57691896", "0.57675546", "0.5761629", "0.5756817", "0.5754298", "0.57424414", "0.5735426", "0.57346183", "0.5730835", "0.572441", "0.5703111", "0.5671316", "0.5670845", "0.5664575", "0.56497437", "0.5644712", "0.5641398", "0.563322", "0.5625497", "0.56218076", "0.5620961", "0.55961734", "0.5594544", "0.5586744", "0.5577262", "0.55645096", "0.5563386", "0.5552156", "0.5550147", "0.55428064", "0.5542773", "0.5540824", "0.5536966", "0.5535833", "0.55270493", "0.5521415", "0.5513476", "0.54908967", "0.54894435", "0.5481205", "0.54682124", "0.54626006", "0.5460574", "0.54581594", "0.5458126", "0.54559916", "0.5455845", "0.54518884", "0.5440668", "0.54355174", "0.5431866", "0.54252845", "0.54144436", "0.5412935", "0.5403988", "0.53995645", "0.5395987", "0.5395067", "0.5391936", "0.53900343", "0.5383004", "0.5381485", "0.5381089", "0.5377517", "0.5376241", "0.53684646", "0.53679025", "0.536135", "0.5358345", "0.5351233" ]
0.6018383
9
UnmarshalJSON decodes the check values.
func (d *DiffCheckValues) UnmarshalJSON(b []byte) (err error) { d.Check, err = icheck.UnmarshalJSON(b) if errors2.EInternal == errors2.ErrorCode(err) { return nil } return err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (v *AndroidCheckRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules12(&r, v)\n\treturn r.Error()\n}", "func (v *ItemCheckResponse) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson6a975c40DecodeJsonBenchmark2(&r, v)\n\treturn r.Error()\n}", "func (o *ServiceCheck) UnmarshalJSON(bytes []byte) (err error) {\n\traw := map[string]interface{}{}\n\trequired := struct {\n\t\tCheck *string `json:\"check\"`\n\t\tHostName *string `json:\"host_name\"`\n\t\tStatus *ServiceCheckStatus `json:\"status\"`\n\t\tTags *[]string `json:\"tags\"`\n\t}{}\n\tall := struct {\n\t\tCheck string `json:\"check\"`\n\t\tHostName string `json:\"host_name\"`\n\t\tMessage *string `json:\"message,omitempty\"`\n\t\tStatus ServiceCheckStatus `json:\"status\"`\n\t\tTags []string `json:\"tags\"`\n\t\tTimestamp *int64 `json:\"timestamp,omitempty\"`\n\t}{}\n\terr = json.Unmarshal(bytes, &required)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif required.Check == nil {\n\t\treturn fmt.Errorf(\"Required field check missing\")\n\t}\n\tif required.HostName == nil {\n\t\treturn fmt.Errorf(\"Required field host_name missing\")\n\t}\n\tif required.Status == nil {\n\t\treturn fmt.Errorf(\"Required field status missing\")\n\t}\n\tif required.Tags == nil {\n\t\treturn fmt.Errorf(\"Required field tags missing\")\n\t}\n\terr = json.Unmarshal(bytes, &all)\n\tif err != nil {\n\t\terr = json.Unmarshal(bytes, &raw)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\to.UnparsedObject = raw\n\t\treturn nil\n\t}\n\tif v := all.Status; !v.IsValid() {\n\t\terr = json.Unmarshal(bytes, &raw)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\to.UnparsedObject = raw\n\t\treturn nil\n\t}\n\to.Check = all.Check\n\to.HostName = all.HostName\n\to.Message = all.Message\n\to.Status = all.Status\n\to.Tags = all.Tags\n\to.Timestamp = all.Timestamp\n\treturn nil\n}", "func (c *CheckNameAvailabilityResponseBody) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"absCode\":\n\t\t\terr = unpopulate(val, \"AbsCode\", &c.AbsCode)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &c.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"valid\":\n\t\t\terr = unpopulate(val, \"Valid\", &c.Valid)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (scw *SafetyCheckWrapper) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"SafetyCheck\":\n\t\t\tif v != nil {\n\t\t\t\tsafetyCheck, err := unmarshalBasicSafetyCheck(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tscw.SafetyCheck = safetyCheck\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (v *NullableServiceCheckStatus) UnmarshalJSON(src []byte) error {\n\tv.isSet = true\n\treturn json.Unmarshal(src, &v.value)\n}", "func (c *CheckAvailabilityResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &c.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isAvailiable\":\n\t\t\terr = unpopulate(val, \"IsAvailiable\", &c.IsAvailiable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &c.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &c.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &c.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *UnloadCheckResponse) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson6a975c40DecodeJsonBenchmark(&r, v)\n\treturn r.Error()\n}", "func (v *checkBadUARule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules5(&r, v)\n\treturn r.Error()\n}", "func UnmarshalAndCheckValue(data []byte, v interface{}) error {\n\treturn json.Unmarshal(data, v)\n}", "func (v *ServiceCheckStatus) UnmarshalJSON(src []byte) error {\n\tvar value int32\n\terr := json.Unmarshal(src, &value)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*v = ServiceCheckStatus(value)\n\treturn nil\n}", "func (v *AndroidCheckRule) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson795c59c6DecodeGrapeGuardRules12(l, v)\n}", "func (e *EntityNameAvailabilityCheckOutput) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &e.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nameAvailable\":\n\t\t\terr = unpopulate(val, \"NameAvailable\", &e.NameAvailable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reason\":\n\t\t\terr = unpopulate(val, \"Reason\", &e.Reason)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *HealthcheckResponse) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (v *OrderCheckResponse) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson6a975c40DecodeJsonBenchmark1(&r, v)\n\treturn r.Error()\n}", "func (c *CheckNameAvailabilityRequestBody) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckNameAvailabilityInput) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CreateJobValidations) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"individualRequestDetails\":\n\t\t\tc.IndividualRequestDetails, err = unmarshalValidationInputRequestClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"validationCategory\":\n\t\t\terr = unpopulate(val, \"ValidationCategory\", &c.ValidationCategory)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *ItemCheckResponse) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson6a975c40DecodeJsonBenchmark2(l, v)\n}", "func (v *VerificationIPFlowResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"access\":\n\t\t\terr = unpopulate(val, \"Access\", &v.Access)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ruleName\":\n\t\t\terr = unpopulate(val, \"RuleName\", &v.RuleName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FirewallPolicyIntrusionDetection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"configuration\":\n\t\t\terr = unpopulate(val, \"Configuration\", &f.Configuration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"mode\":\n\t\t\terr = unpopulate(val, \"Mode\", &f.Mode)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *HealthcheckRequest) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (c *CheckAvailabilityParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &c.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isAvailiable\":\n\t\t\terr = unpopulate(val, \"IsAvailiable\", &c.IsAvailiable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &c.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &c.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &c.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckNameAvailabilityResponse) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &c.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nameAvailable\":\n\t\t\terr = unpopulate(val, \"NameAvailable\", &c.NameAvailable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reason\":\n\t\t\terr = unpopulate(val, \"Reason\", &c.Reason)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CertificateVerificationDescription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"certificate\":\n\t\t\terr = unpopulate(val, \"Certificate\", &c.Certificate)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckSKUAvailabilityParameter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"kind\":\n\t\t\terr = unpopulate(val, \"Kind\", &c.Kind)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"skus\":\n\t\t\terr = unpopulate(val, \"SKUs\", &c.SKUs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (o *CheckCanDeleteMonitorResponse) UnmarshalJSON(bytes []byte) (err error) {\n\traw := map[string]interface{}{}\n\trequired := struct {\n\t\tData *CheckCanDeleteMonitorResponseData `json:\"data\"`\n\t}{}\n\tall := struct {\n\t\tData CheckCanDeleteMonitorResponseData `json:\"data\"`\n\t\tErrors map[string][]string `json:\"errors,omitempty\"`\n\t}{}\n\terr = json.Unmarshal(bytes, &required)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif required.Data == nil {\n\t\treturn fmt.Errorf(\"Required field data missing\")\n\t}\n\terr = json.Unmarshal(bytes, &all)\n\tif err != nil {\n\t\terr = json.Unmarshal(bytes, &raw)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\to.UnparsedObject = raw\n\t\treturn nil\n\t}\n\tif all.Data.UnparsedObject != nil && o.UnparsedObject == nil {\n\t\terr = json.Unmarshal(bytes, &raw)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\to.UnparsedObject = raw\n\t}\n\to.Data = all.Data\n\to.Errors = all.Errors\n\treturn nil\n}", "func (c *CheckNameAvailabilityParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Validator) UnmarshalJSON(data []byte) error {\n\tbv := &bechValidator{}\n\tif err := codec.Cdc.UnmarshalJSON(data, bv); err != nil {\n\t\treturn err\n\t}\n\tconsPubKey, err := sdk.GetConsPubKeyBech32(bv.ConsPubKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*v = Validator{\n\t\tAddress: bv.Address,\n\t\tConsPubKey: consPubKey,\n\t\tJailed: bv.Jailed,\n\t\tStakedTokens: bv.StakedTokens,\n\t\tStatus: bv.Status,\n\t\tUnstakingCompletionTime: bv.UnstakingCompletionTime,\n\t}\n\treturn nil\n}", "func (c *CheckResourceNameResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, \"Status\", &c.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *booleanField) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson390b7126DecodeGithubComChancedPicker45(&r, v)\n\treturn r.Error()\n}", "func (s *SKUAvailabilityValidationResponseProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"error\":\n\t\t\terr = unpopulate(val, \"Error\", &s.Error)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, \"Status\", &s.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"validationType\":\n\t\t\terr = unpopulate(val, \"ValidationType\", &s.ValidationType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *commonRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules4(&r, v)\n\treturn r.Error()\n}", "func (c *ClientCheckFeatureSupportResponse) UnmarshalJSON(data []byte) error {\n\tres, err := unmarshalFeatureValidationResponseBaseClassification(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\tc.FeatureValidationResponseBaseClassification = res\n\treturn nil\n}", "func (a *ApplicationGatewayAvailableWafRuleSetsResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &a.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (d *DataTransferDetailsValidationResponseProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"error\":\n\t\t\terr = unpopulate(val, \"Error\", &d.Error)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, \"Status\", &d.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"validationType\":\n\t\t\terr = unpopulate(val, \"ValidationType\", &d.ValidationType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (w *WebApplicationFirewallScrubbingRules) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"matchVariable\":\n\t\t\terr = unpopulate(val, \"MatchVariable\", &w.MatchVariable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"selector\":\n\t\t\terr = unpopulate(val, \"Selector\", &w.Selector)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"selectorMatchOperator\":\n\t\t\terr = unpopulate(val, \"SelectorMatchOperator\", &w.SelectorMatchOperator)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"state\":\n\t\t\terr = unpopulate(val, \"State\", &w.State)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckDomainAvailabilityParameter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"kind\":\n\t\t\terr = unpopulate(val, \"Kind\", &c.Kind)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subdomainName\":\n\t\t\terr = unpopulate(val, \"SubdomainName\", &c.SubdomainName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *GuardXml) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules9(&r, v)\n\treturn r.Error()\n}", "func (v *captchaRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules6(&r, v)\n\treturn r.Error()\n}", "func (b *BackupSecretResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"value\":\n\t\t\terr = runtime.DecodeByteArray(string(val), &b.Value, runtime.Base64URLFormat)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (w *Entry) UnmarshalJSON(bb []byte) error {\n\t<<!!YOUR_CODE!!>>\n}", "func (v *ExtFilter) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules11(&r, v)\n\treturn r.Error()\n}", "func (v *singleRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules1(&r, v)\n\treturn r.Error()\n}", "func (d *DataTransferDetailsValidationRequest) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"dataExportDetails\":\n\t\t\terr = unpopulate(val, \"DataExportDetails\", &d.DataExportDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataImportDetails\":\n\t\t\terr = unpopulate(val, \"DataImportDetails\", &d.DataImportDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deviceType\":\n\t\t\terr = unpopulate(val, \"DeviceType\", &d.DeviceType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"transferType\":\n\t\t\terr = unpopulate(val, \"TransferType\", &d.TransferType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"validationType\":\n\t\t\terr = unpopulate(val, \"ValidationType\", &d.ValidationType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (d *DeviceServiceCheckNameAvailabilityParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &d.Name)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SKUAvailabilityListResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &s.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *condition) UnmarshalJSON(data []byte) error {\n\tvar d []json.RawMessage\n\tif err := json.Unmarshal(data, &d); err != nil {\n\t\treturn err\n\t}\n\tif len(d) != 2 {\n\t\treturn errors.New(\"condition must be a list with two items.\")\n\t}\n\tif err := json.Unmarshal(d[0], &p.Condition); err != nil {\n\t\treturn err\n\t}\n\tm := map[string]variables{}\n\tif err := json.Unmarshal(d[1], &m); err != nil {\n\t\treturn err\n\t}\n\tvar ok bool\n\tif p.Variables, ok = m[\"variables\"]; !ok {\n\t\treturn errors.New(\"variables item is required in condition.\")\n\t}\n\treturn nil\n}", "func (s *SubscriptionIsAllowedToCreateJobValidationRequest) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"validationType\":\n\t\t\terr = unpopulate(val, \"ValidationType\", &s.ValidationType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Status) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecodeGithubComMailcoursesTechnoparkDbmsForumGeneratedModels5(&r, v)\n\treturn r.Error()\n}", "func (v *Validator) UnmarshalJSON(data []byte) error {\n\tbv := &bechValidator{}\n\tif err := codec.Cdc.UnmarshalJSON(data, bv); err != nil {\n\t\treturn err\n\t}\n\tconsPubKey, err := GetConsPubKeyBech32(bv.ConsPubKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*v = Validator{\n\t\tOperatorAddress: bv.OperatorAddress,\n\t\tConsPubKey: consPubKey,\n\t\tJailed: bv.Jailed,\n\t\tTokens: bv.Tokens,\n\t\tStatus: bv.Status,\n\t\tDelegatorShares: bv.DelegatorShares,\n\t\tDescription: bv.Description,\n\t\tUnbondingHeight: bv.UnbondingHeight,\n\t\tUnbondingCompletionTime: bv.UnbondingCompletionTime,\n\t\tCommission: bv.Commission,\n\t\tMinSelfDelegation: bv.MinSelfDelegation,\n\t}\n\treturn nil\n}", "func (c *CheckServiceProviderAvailabilityInput) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"peeringServiceLocation\":\n\t\t\terr = unpopulate(val, \"PeeringServiceLocation\", &c.PeeringServiceLocation)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"peeringServiceProvider\":\n\t\t\terr = unpopulate(val, \"PeeringServiceProvider\", &c.PeeringServiceProvider)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (cvpbo *CheckValuePropertyBatchOperation) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"Value\":\n\t\t\tif v != nil {\n\t\t\t\tvalue, err := unmarshalBasicPropertyValue(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tcvpbo.Value = value\n\t\t\t}\n\t\tcase \"PropertyName\":\n\t\t\tif v != nil {\n\t\t\t\tvar propertyName string\n\t\t\t\terr = json.Unmarshal(*v, &propertyName)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tcvpbo.PropertyName = &propertyName\n\t\t\t}\n\t\tcase \"Kind\":\n\t\t\tif v != nil {\n\t\t\t\tvar kind KindBasicPropertyBatchOperation\n\t\t\t\terr = json.Unmarshal(*v, &kind)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tcvpbo.Kind = kind\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (b *Bool) UnmarshalJSON(data []byte) error {\n\treturn b.Scan(string(data))\n}", "func (c *CheckNameAvailabilityRequest) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SubscriptionIsAllowedToCreateJobValidationResponseProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"error\":\n\t\t\terr = unpopulate(val, \"Error\", &s.Error)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, \"Status\", &s.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"validationType\":\n\t\t\terr = unpopulate(val, \"ValidationType\", &s.ValidationType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *SwapFillsInfo) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson25363b2dDecodeGithubComDarkfoxs96OpenApiV3SdkOkexGoSdkApi32(&r, v)\n\treturn r.Error()\n}", "func (j *LuaBool) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (t *TestLineUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &t.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (ts *TransferStatus) UnmarshalJSON(b []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(b, &s); err != nil {\n\t\treturn err\n\t}\n\treturn ts.Parse(s)\n}", "func (cat *CatAPTrue) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"friendly\":\n\t\t\tif v != nil {\n\t\t\t\tvar friendly bool\n\t\t\t\terr = json.Unmarshal(*v, &friendly)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tcat.Friendly = &friendly\n\t\t\t}\n\t\tdefault:\n\t\t\tif v != nil {\n\t\t\t\tvar additionalProperties interface{}\n\t\t\t\terr = json.Unmarshal(*v, &additionalProperties)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tif cat.AdditionalProperties == nil {\n\t\t\t\t\tcat.AdditionalProperties = make(map[string]interface{})\n\t\t\t\t}\n\t\t\t\tcat.AdditionalProperties[k] = additionalProperties\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID int32\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tcat.ID = &ID\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tcat.Name = &name\n\t\t\t}\n\t\tcase \"status\":\n\t\t\tif v != nil {\n\t\t\t\tvar status bool\n\t\t\t\terr = json.Unmarshal(*v, &status)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tcat.Status = &status\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *AuthorizationListResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"nextLink\":\n\t\t\terr = unpopulate(val, \"NextLink\", &a.NextLink)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &a.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FilterItems) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"field\":\n\t\t\terr = unpopulate(val, \"Field\", &f.Field)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"values\":\n\t\t\terr = unpopulate(val, \"Values\", &f.Values)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *MultipleActivationKeyUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &m.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *allOfRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson390b7126DecodeGithubComChancedPicker48(&r, v)\n\treturn r.Error()\n}", "func (h *ScanningOrFalse) UnmarshalJSON(data []byte) error {\n\tvar unmarshalled interface{}\n\tif err := json.Unmarshal(data, &unmarshalled); err != nil {\n\t\treturn err\n\t}\n\n\tswitch v := unmarshalled.(type) {\n\tcase bool:\n\t\th.Value = v\n\tcase map[string]interface{}:\n\t\th.Value = ScanProgress{\n\t\t\tDuration: int(v[\"duration\"].(float64)),\n\t\t\tProgress: v[\"progress\"].(float64),\n\t\t}\n\tdefault:\n\t\treturn fmt.Errorf(\"invalid scanning value: %v\", unmarshalled)\n\t}\n\n\treturn nil\n}", "func (v *checkBadUARule) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson795c59c6DecodeGrapeGuardRules5(l, v)\n}", "func (m *ManagerDeploymentStatusListResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"skipToken\":\n\t\t\terr = unpopulate(val, \"SkipToken\", &m.SkipToken)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &m.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *BaseFillInfo) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson25363b2dDecodeGithubComDarkfoxs96OpenApiV3SdkOkexGoSdkApi131(&r, v)\n\treturn r.Error()\n}", "func (v *GuardRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules10(&r, v)\n\treturn r.Error()\n}", "func (v *VirtualNetworkDdosProtectionStatusResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"nextLink\":\n\t\t\terr = unpopulate(val, \"NextLink\", &v.NextLink)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &v.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (this *OutlierDetection) UnmarshalJSON(b []byte) error {\n\treturn DestinationRuleUnmarshaler.Unmarshal(bytes.NewReader(b), this)\n}", "func (v *UnloadCheckResponse) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson6a975c40DecodeJsonBenchmark(l, v)\n}", "func (v *ValidationResponseProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"individualResponseDetails\":\n\t\t\tv.IndividualResponseDetails, err = unmarshalValidationInputResponseClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, \"Status\", &v.Status)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *userAgentRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules(&r, v)\n\treturn r.Error()\n}", "func (b *BgpPeerStatusListResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", b, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &b.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", b, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Bool) UnmarshalJSON(data []byte) error {\n\tif bytes.Equal(data, []byte{'n', 'u', 'l', 'l'}) {\n\t\tv.present = false\n\t\tv.initialized = true\n\t\treturn nil\n\t}\n\terr := json.Unmarshal(data, &v.v)\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\tv.present = true\n\tv.initialized = true\n\treturn nil\n}", "func (v *Status) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecodeGithubComDbProjectPkgModels6(&r, v)\n\treturn r.Error()\n}", "func (s *SecretUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &s.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *User) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson2bc03518DecodeLangTaskOnBench(&r, v)\n\treturn r.Error()\n}", "func (s *SecretRestoreParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"value\":\n\t\t\terr = runtime.DecodeByteArray(string(val), &s.SecretBundleBackup, runtime.Base64URLFormat)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FirewallPolicyIntrusionDetectionBypassTrafficSpecifications) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"description\":\n\t\t\terr = unpopulate(val, \"Description\", &f.Description)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"destinationAddresses\":\n\t\t\terr = unpopulate(val, \"DestinationAddresses\", &f.DestinationAddresses)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"destinationIpGroups\":\n\t\t\terr = unpopulate(val, \"DestinationIPGroups\", &f.DestinationIPGroups)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"destinationPorts\":\n\t\t\terr = unpopulate(val, \"DestinationPorts\", &f.DestinationPorts)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &f.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"protocol\":\n\t\t\terr = unpopulate(val, \"Protocol\", &f.Protocol)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sourceAddresses\":\n\t\t\terr = unpopulate(val, \"SourceAddresses\", &f.SourceAddresses)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sourceIpGroups\":\n\t\t\terr = unpopulate(val, \"SourceIPGroups\", &f.SourceIPGroups)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *XJSONYesNo) UnmarshalJSON(b []byte) error {\n\t*t = XJSONYesNo(Stripchars(string(b), \"\\\"\"))\n\treturn nil\n}", "func (t *XJSONYesNo) UnmarshalJSON(b []byte) error {\n\t*t = XJSONYesNo(Stripchars(string(b), \"\\\"\"))\n\treturn nil\n}", "func (t *TestLineListResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"nextLink\":\n\t\t\terr = unpopulate(val, \"NextLink\", &t.NextLink)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &t.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckPrivateLinkServiceVisibilityRequest) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"privateLinkServiceAlias\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceAlias\", &c.PrivateLinkServiceAlias)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *AppAvailabilityInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &a.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nameAvailable\":\n\t\t\terr = unpopulate(val, \"NameAvailable\", &a.NameAvailable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reason\":\n\t\t\terr = unpopulate(val, \"Reason\", &a.Reason)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *ResourceSKURestrictionInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"locations\":\n\t\t\terr = unpopulate(val, \"Locations\", &r.Locations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"zones\":\n\t\t\terr = unpopulate(val, \"Zones\", &r.Zones)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FilterTrackPropertyCondition) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"operation\":\n\t\t\terr = unpopulate(val, \"Operation\", &f.Operation)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"property\":\n\t\t\terr = unpopulate(val, \"Property\", &f.Property)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &f.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FirewallPolicyIntrusionDetectionConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"bypassTrafficSettings\":\n\t\t\terr = unpopulate(val, \"BypassTrafficSettings\", &f.BypassTrafficSettings)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateRanges\":\n\t\t\terr = unpopulate(val, \"PrivateRanges\", &f.PrivateRanges)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"signatureOverrides\":\n\t\t\terr = unpopulate(val, \"SignatureOverrides\", &f.SignatureOverrides)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *IPRule) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &i.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *ExclusionManagedRule) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"ruleId\":\n\t\t\terr = unpopulate(val, \"RuleID\", &e.RuleID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *proxyRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules3(&r, v)\n\treturn r.Error()\n}", "func (s *SKUAvailabilityValidationRequest) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"country\":\n\t\t\terr = unpopulate(val, \"Country\", &s.Country)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deviceType\":\n\t\t\terr = unpopulate(val, \"DeviceType\", &s.DeviceType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &s.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"transferType\":\n\t\t\terr = unpopulate(val, \"TransferType\", &s.TransferType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"validationType\":\n\t\t\terr = unpopulate(val, \"ValidationType\", &s.ValidationType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SyncIdentityProviderUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &s.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SignatureOverridesFilterValuesResponse) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"filterValues\":\n\t\t\terr = unpopulate(val, \"FilterValues\", &s.FilterValues)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *UnbindParams) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeGithubComChromedpCdprotoTethering(&r, v)\n\treturn r.Error()\n}", "func (s *SyncSetUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &s.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *anyOfRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson390b7126DecodeGithubComChancedPicker47(&r, v)\n\treturn r.Error()\n}", "func (t *TryjobStatus) UnmarshalJSON(data []byte) error {\n\tstrStatus := strings.Trim(string(data), \"\\\"\")\n\t*t = tryjobStatusFromString(strStatus)\n\treturn nil\n}" ]
[ "0.6849288", "0.6813789", "0.67794085", "0.65396965", "0.6531918", "0.64446926", "0.63754225", "0.6328835", "0.6325424", "0.6293037", "0.62720406", "0.6265991", "0.6262729", "0.62530154", "0.62403893", "0.6238377", "0.6227791", "0.61955345", "0.6188854", "0.6162939", "0.6148716", "0.61237776", "0.6109798", "0.6105271", "0.60968167", "0.60932446", "0.6068361", "0.60587543", "0.60531336", "0.60411483", "0.60262114", "0.6006905", "0.60036665", "0.59654474", "0.59607327", "0.59547275", "0.5944235", "0.59048796", "0.58999467", "0.5899035", "0.58919245", "0.5874731", "0.587338", "0.5862666", "0.58615243", "0.58609045", "0.58590716", "0.5854654", "0.5847212", "0.58470666", "0.5844605", "0.5840455", "0.5839069", "0.58349013", "0.58114487", "0.5806775", "0.5805979", "0.57889086", "0.57873833", "0.57838017", "0.57791936", "0.57774806", "0.5768114", "0.57659847", "0.575845", "0.5757933", "0.5756892", "0.5753198", "0.5752486", "0.57377493", "0.57364947", "0.57360077", "0.57328326", "0.5732356", "0.5729729", "0.5724923", "0.57212883", "0.57176775", "0.571355", "0.57078", "0.5705937", "0.57035136", "0.57005984", "0.57005984", "0.5698723", "0.5694025", "0.5692641", "0.569165", "0.56838375", "0.5681509", "0.56806654", "0.56797194", "0.56790763", "0.56744516", "0.5665621", "0.566328", "0.56627595", "0.56480116", "0.5647453", "0.56404" ]
0.75885034
0
MarshalJSON implementation here is forced by the embedded check value here.
func (d DiffNotificationEndpointValues) MarshalJSON() ([]byte, error) { if d.NotificationEndpoint == nil { return json.Marshal(nil) } return json.Marshal(d.NotificationEndpoint) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func mustMarshalJSON(v interface{}) []byte {\n\tb, err := json.Marshal(v)\n\tif err != nil {\n\t\tpanic(\"marshal json: \" + err.Error())\n\t}\n\treturn b\n}", "func (sc SafetyCheck) MarshalJSON() ([]byte, error) {\n\tsc.Kind = KindSafetyCheck\n\tobjectMap := make(map[string]interface{})\n\tif sc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = sc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (v checkBadUARule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules5(w, v)\n}", "func (v AndroidCheckRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules12(w, v)\n}", "func (easc EnsureAvailabilitySafetyCheck) MarshalJSON() ([]byte, error) {\n\teasc.Kind = KindEnsureAvailability\n\tobjectMap := make(map[string]interface{})\n\tif easc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = easc.PartitionID\n\t}\n\tif easc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = easc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (cvpbo CheckValuePropertyBatchOperation) MarshalJSON() ([]byte, error) {\n\tcvpbo.Kind = KindCheckValue\n\tobjectMap := make(map[string]interface{})\n\tobjectMap[\"Value\"] = cvpbo.Value\n\tif cvpbo.PropertyName != nil {\n\t\tobjectMap[\"PropertyName\"] = cvpbo.PropertyName\n\t}\n\tif cvpbo.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = cvpbo.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (JSONIndent) IsMarshal7951Arg() {}", "func (*RFC7951JSONConfig) IsEncodeTypedValueOpt() {}", "func (j *JSON) CanMarshalWithCallback() bool {\n\treturn false\n}", "func (d DiffCheckValues) MarshalJSON() ([]byte, error) {\n\tif d.Check == nil {\n\t\treturn json.Marshal(nil)\n\t}\n\treturn json.Marshal(d.Check)\n}", "func (n NullJSON) MarshalJSON() ([]byte, error) { return nulljson(n.Valid, n.JSONVal) }", "func (v ItemCheckResponse) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson6a975c40EncodeJsonBenchmark2(w, v)\n}", "func (wfirsc WaitForInbuildReplicaSafetyCheck) MarshalJSON() ([]byte, error) {\n\twfirsc.Kind = KindWaitForInbuildReplica\n\tobjectMap := make(map[string]interface{})\n\tif wfirsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = wfirsc.PartitionID\n\t}\n\tif wfirsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = wfirsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (wfppsc WaitForPrimaryPlacementSafetyCheck) MarshalJSON() ([]byte, error) {\n\twfppsc.Kind = KindWaitForPrimaryPlacement\n\tobjectMap := make(map[string]interface{})\n\tif wfppsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = wfppsc.PartitionID\n\t}\n\tif wfppsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = wfppsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func mustJSON(thing interface{}) string {\n\ts, err := json.Marshal(thing)\n\tif err != nil {\n\t\treturn \"invalid json: \" + err.Error()\n\t}\n\treturn string(s)\n}", "func (v checkBadUARule) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson795c59c6EncodeGrapeGuardRules5(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (cepbo CheckExistsPropertyBatchOperation) MarshalJSON() ([]byte, error) {\n\tcepbo.Kind = KindCheckExists\n\tobjectMap := make(map[string]interface{})\n\tif cepbo.Exists != nil {\n\t\tobjectMap[\"Exists\"] = cepbo.Exists\n\t}\n\tif cepbo.PropertyName != nil {\n\t\tobjectMap[\"PropertyName\"] = cepbo.PropertyName\n\t}\n\tif cepbo.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = cepbo.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (psc PartitionSafetyCheck) MarshalJSON() ([]byte, error) {\n\tpsc.Kind = KindPartitionSafetyCheck\n\tobjectMap := make(map[string]interface{})\n\tif psc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = psc.PartitionID\n\t}\n\tif psc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = psc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (snsc SeedNodeSafetyCheck) MarshalJSON() ([]byte, error) {\n\tsnsc.Kind = KindEnsureSeedNodeQuorum\n\tobjectMap := make(map[string]interface{})\n\tif snsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = snsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (*RFC7951JSONConfig) IsMarshal7951Arg() {}", "func (v Allow) EncodeJSON(b []byte) []byte {\n\tb = append(b, '{', '\"', 'b', 'a', 'l', 'a', 'n', 'c', 'e', '_', 'e', 'x', 'e', 'm', 'p', 't', 'i', 'o', 'n', 's', '\"', ':', '[')\n\tfor i, elem := range v.BalanceExemptions {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = elem.EncodeJSON(b)\n\t}\n\tb = append(b, ']', ',', '\"', 'c', 'a', 'l', 'l', '_', 'm', 'e', 't', 'h', 'o', 'd', 's', '\"', ':', '[')\n\tfor i, elem := range v.CallMethods {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = json.AppendString(b, elem)\n\t}\n\tb = append(b, `],\"errors\":[`...)\n\tfor i, elem := range v.Errors {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = elem.EncodeJSON(b)\n\t}\n\tb = append(b, ']', ',', '\"', 'h', 'i', 's', 't', 'o', 'r', 'i', 'c', 'a', 'l', '_', 'b', 'a', 'l', 'a', 'n', 'c', 'e', '_', 'l', 'o', 'o', 'k', 'u', 'p', '\"', ':')\n\tb = json.AppendBool(b, v.HistoricalBalanceLookup)\n\tb = append(b, ',', '\"', 'm', 'e', 'm', 'p', 'o', 'o', 'l', '_', 'c', 'o', 'i', 'n', 's', '\"', ':')\n\tb = json.AppendBool(b, v.MempoolCoins)\n\tb = append(b, ',', '\"', 'o', 'p', 'e', 'r', 'a', 't', 'i', 'o', 'n', '_', 's', 't', 'a', 't', 'u', 's', 'e', 's', '\"', ':', '[')\n\tfor i, elem := range v.OperationStatuses {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = elem.EncodeJSON(b)\n\t}\n\tb = append(b, ']', ',', '\"', 'o', 'p', 'e', 'r', 'a', 't', 'i', 'o', 'n', '_', 't', 'y', 'p', 'e', 's', '\"', ':', '[')\n\tfor i, elem := range v.OperationTypes {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = json.AppendString(b, elem)\n\t}\n\tb = append(b, \"],\"...)\n\tif v.TimestampStartIndex.Set {\n\t\tb = append(b, '\"', 't', 'i', 'm', 'e', 's', 't', 'a', 'm', 'p', '_', 's', 't', 'a', 'r', 't', '_', 'i', 'n', 'd', 'e', 'x', '\"', ':')\n\t\tb = json.AppendInt(b, v.TimestampStartIndex.Value)\n\t\tb = append(b, \",\"...)\n\t}\n\tb[len(b)-1] = '}'\n\treturn b\n}", "func (v singleRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules1(w, v)\n}", "func (v commonRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules4(w, v)\n}", "func (v OrderCheckResponse) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson6a975c40EncodeJsonBenchmark1(w, v)\n}", "func (v NullableServiceCheckStatus) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(v.value)\n}", "func (o ServiceCheck) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\ttoSerialize[\"check\"] = o.Check\n\ttoSerialize[\"host_name\"] = o.HostName\n\tif o.Message != nil {\n\t\ttoSerialize[\"message\"] = o.Message\n\t}\n\ttoSerialize[\"status\"] = o.Status\n\ttoSerialize[\"tags\"] = o.Tags\n\tif o.Timestamp != nil {\n\t\ttoSerialize[\"timestamp\"] = o.Timestamp\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func jsonify(v interface{}) string { return string(mustMarshalJSON(v)) }", "func (v proxyRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules3(w, v)\n}", "func (epqsc EnsurePartitionQurumSafetyCheck) MarshalJSON() ([]byte, error) {\n\tepqsc.Kind = KindEnsurePartitionQuorum\n\tobjectMap := make(map[string]interface{})\n\tif epqsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = epqsc.PartitionID\n\t}\n\tif epqsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = epqsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (v ItemCheckResponse) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson6a975c40EncodeJsonBenchmark2(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v reqGuardRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules2(w, v)\n}", "func testJSONMarshal(t *testing.T, v interface{}, want string) {\n\tj, err := json.Marshal(v)\n\tif err != nil {\n\t\tt.Errorf(\"Unable to marshal JSON for %v\", v)\n\t}\n\n\tw := new(bytes.Buffer)\n\terr = json.Compact(w, []byte(want))\n\tif err != nil {\n\t\tt.Errorf(\"String is not valid json: %s\", want)\n\t}\n\n\tif w.String() != string(j) {\n\t\tt.Errorf(\"json.Marshal(%q) returned %s, want %s\", v, j, w)\n\t}\n\n\t// now go the other direction and make sure things unmarshal as expected\n\tu := reflect.ValueOf(v).Interface()\n\tif err := json.Unmarshal([]byte(want), u); err != nil {\n\t\tt.Errorf(\"Unable to unmarshal JSON for %v\", want)\n\t}\n\n\tif !reflect.DeepEqual(v, u) {\n\t\tt.Errorf(\"json.Unmarshal(%q) returned %s, want %s\", want, u, v)\n\t}\n}", "func testJSONMarshal(t *testing.T, v interface{}, want string) {\n\tj, err := json.Marshal(v)\n\tif err != nil {\n\t\tt.Errorf(\"Unable to marshal JSON for %v\", v)\n\t}\n\n\tw := new(bytes.Buffer)\n\terr = json.Compact(w, []byte(want))\n\tif err != nil {\n\t\tt.Errorf(\"String is not valid json: %s\", want)\n\t}\n\n\tif w.String() != string(j) {\n\t\tt.Errorf(\"json.Marshal(%q) returned %s, want %s\", v, j, w)\n\t}\n\n\t// now go the other direction and make sure things unmarshal as expected\n\tu := reflect.ValueOf(v).Interface()\n\tif err := json.Unmarshal([]byte(want), u); err != nil {\n\t\tt.Errorf(\"Unable to unmarshal JSON for %v: %v\", want, err)\n\t}\n\n\tif !reflect.DeepEqual(v, u) {\n\t\tt.Errorf(\"json.Unmarshal(%q) returned %s, want %s\", want, u, v)\n\t}\n}", "func MustMarshalJSON(v interface{}) string {\n\tbuf, err := json.Marshal(v)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn string(buf)\n}", "func marshallJSONWithoutError(value interface{}) []byte {\n\tj, err := json.Marshal(value)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn j\n}", "func (cspbo CheckSequencePropertyBatchOperation) MarshalJSON() ([]byte, error) {\n\tcspbo.Kind = KindCheckSequence\n\tobjectMap := make(map[string]interface{})\n\tif cspbo.SequenceNumber != nil {\n\t\tobjectMap[\"SequenceNumber\"] = cspbo.SequenceNumber\n\t}\n\tif cspbo.PropertyName != nil {\n\t\tobjectMap[\"PropertyName\"] = cspbo.PropertyName\n\t}\n\tif cspbo.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = cspbo.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (o HealthCheckResult) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.NullableMessage == nil {\n\t\tif o.isExplicitNullNullableMessage {\n\t\t\ttoSerialize[\"NullableMessage\"] = o.NullableMessage\n\t\t}\n\t} else {\n\t\ttoSerialize[\"NullableMessage\"] = o.NullableMessage\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (v UnloadCheckResponse) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson6a975c40EncodeJsonBenchmark(w, v)\n}", "func MustJsonEncode(v interface{}) string {\n\tbuf, err := ffjson.Marshal(v)\n\tif err != nil {\n\t\tfmt.Printf(\"json encode %v err: %v\", v, err)\n\t}\n\treturn string(buf)\n}", "func mustMarshal(val interface{}) []byte {\n\tif b, err := asn1.Marshal(val); err != nil {\n\t\tpanic(err)\n\t} else {\n\t\treturn b\n\t}\n}", "func (wfrsc WaitForReconfigurationSafetyCheck) MarshalJSON() ([]byte, error) {\n\twfrsc.Kind = KindWaitForReconfiguration\n\tobjectMap := make(map[string]interface{})\n\tif wfrsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = wfrsc.PartitionID\n\t}\n\tif wfrsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = wfrsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func MustMarshalJSON(data interface{}) []byte {\n\tb, err := json.Marshal(data)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn b\n}", "func (v ExtFilter) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules11(w, v)\n}", "func (v OrderCheckResponse) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson6a975c40EncodeJsonBenchmark1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v PackoutRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules8(w, v)\n}", "func (h ScanningOrFalse) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(h.Value)\n}", "func ensureAppliedConfigAnnotation(scheme *runtime.Scheme, obj GCRuntimeObject) ([]byte, error) {\n\t// Instead of encoding the json directly, we first convert it to \"Unstructured\", i.e. a map[string]interface{}\n\t// We do this to remove a \"status\" item if there is any. Not all status fields are marked as `json:\",omitempty`,\n\t// so their default value will be serialized too. The status field itself is marked as `json:\",omitempty`, which\n\t// is useless, as struct values are never empty in golang.\n\tobjUnstructured := unstructured.Unstructured{}\n\n\terr := scheme.Convert(obj, &objUnstructured, nil)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to convert to unstructured item: %w\", err)\n\t}\n\n\tdelete(objUnstructured.Object, \"status\")\n\n\tobjEncoded, err := runtime.Encode(unstructured.UnstructuredJSONScheme, &objUnstructured)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to prepare applied configuration metadata: %w\", err)\n\t}\n\n\tannotations := obj.GetAnnotations()\n\tif annotations == nil {\n\t\tannotations = map[string]string{}\n\t}\n\n\tannotations[lastAppliedAnnotation] = string(objEncoded)\n\tobj.SetAnnotations(annotations)\n\n\treturn objEncoded, nil\n}", "func (v AndroidCheckRule) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson795c59c6EncodeGrapeGuardRules12(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func MustMarshalJSON(o interface{}) []byte {\n\treturn cdc.MustMarshalJSON(o)\n}", "func (c CheckNameAvailabilityResponseBody) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"absCode\", c.AbsCode)\n\tpopulate(objectMap, \"message\", c.Message)\n\tpopulate(objectMap, \"valid\", c.Valid)\n\treturn json.Marshal(objectMap)\n}", "func MustMarshal(v interface{}) []byte {\n\tbytes, err := json.Marshal(v)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn bytes\n}", "func TestMarshal(t *testing.T) {\n\tfor _, c := range []struct {\n\t\tname string\n\t\tfieldPairs []interface{}\n\t\t// errNeedle is \"\" if we expect no error. Otherwise, it is a string that\n\t\t// we expect to see in the resulting err.Error().\n\t\terrNeedle string\n\t}{\n\t\t{\n\t\t\t\"no fields\",\n\t\t\t[]interface{}{},\n\t\t\t\"\",\n\t\t},\n\t\t{\n\t\t\t\"simple\",\n\t\t\t[]interface{}{\"k0\", \"v0\"},\n\t\t\t\"\",\n\t\t},\n\t\t{\n\t\t\t\"mixed value types\",\n\t\t\t// Numeric types turn into float64s in JSON.\n\t\t\t[]interface{}{\"k0\", \"v0\", \"k1\", float64(1), \"k2\", true},\n\t\t\t\"\",\n\t\t},\n\t\t{\n\t\t\t\"odd field pairs\",\n\t\t\t[]interface{}{\"k0\", \"v0\", \"k1\"},\n\t\t\t\"even\",\n\t\t},\n\t\t{\n\t\t\t\"non-string key\",\n\t\t\t[]interface{}{0, \"v0\"},\n\t\t\t\"string\",\n\t\t},\n\t\t{\n\t\t\t\"duplicate keys\",\n\t\t\t[]interface{}{\"k0\", \"v0\", \"k0\", \"v1\"},\n\t\t\t\"duplicate\",\n\t\t},\n\t} {\n\t\tt.Run(c.name, func(t *testing.T) {\n\t\t\tmarshalOK := c.errNeedle == \"\"\n\t\t\ts, err := Marshal(c.name, c.fieldPairs)\n\t\t\tif got, want := err == nil, marshalOK; got != want {\n\t\t\t\tt.Fatalf(\"got %v, want %v\", got, want)\n\t\t\t}\n\t\t\tif !marshalOK {\n\t\t\t\tif !strings.Contains(err.Error(), c.errNeedle) {\n\t\t\t\t\tt.Errorf(\"error %q does not contain expected substring %q\", err.Error(), c.errNeedle)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\t\t\tvar m map[string]interface{}\n\t\t\terr = json.Unmarshal([]byte(s), &m)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatalf(\"unmarshaling failed: %v\", err)\n\t\t\t}\n\t\t\t// The +1 is for the eventType.\n\t\t\tif got, want := len(m), (len(c.fieldPairs)/2)+1; got != want {\n\t\t\t\tt.Errorf(\"got %v, want %v\", got, want)\n\t\t\t}\n\t\t\ttyp, ok := m[eventTypeFieldKey]\n\t\t\tif ok {\n\t\t\t\tif got, want := typ, c.name; got != want {\n\t\t\t\t\tt.Errorf(\"got %v, want %v\", got, want)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tt.Errorf(\"eventType field not marshaled\")\n\t\t\t}\n\t\t\tfor i := 0; i < len(c.fieldPairs); i++ {\n\t\t\t\tkey := c.fieldPairs[i].(string)\n\t\t\t\ti++\n\t\t\t\tvalue := c.fieldPairs[i]\n\t\t\t\tmvalue, ok := m[key]\n\t\t\t\tif !ok {\n\t\t\t\t\tt.Errorf(\"field with key %q not marshaled\", key)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif got, want := mvalue, value; got != want {\n\t\t\t\t\tt.Errorf(\"got %v(%T), want %v(%T)\", got, got, want, want)\n\t\t\t\t}\n\t\t\t}\n\t\t})\n\t}\n}", "func (wfpssc WaitForPrimarySwapSafetyCheck) MarshalJSON() ([]byte, error) {\n\twfpssc.Kind = KindWaitForPrimarySwap\n\tobjectMap := make(map[string]interface{})\n\tif wfpssc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = wfpssc.PartitionID\n\t}\n\tif wfpssc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = wfpssc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (src UpdateVerificationFlowBody) MarshalJSON() ([]byte, error) {\n\tif src.UpdateVerificationFlowWithCodeMethod != nil {\n\t\treturn json.Marshal(&src.UpdateVerificationFlowWithCodeMethod)\n\t}\n\n\tif src.UpdateVerificationFlowWithLinkMethod != nil {\n\t\treturn json.Marshal(&src.UpdateVerificationFlowWithLinkMethod)\n\t}\n\n\treturn nil, nil // no data in oneOf schemas\n}", "func (n NullBool) MarshalJSON() ([]byte, error) { return nulljson(n.Valid, n.Bool) }", "func marshal() {\n\tfmt.Println(\"=== json.marshal ===\")\n\tryan := &Person{\"Ryan\", 25}\n\twire, err := json.Marshal(ryan)\n\tcheck(err)\n\tfmt.Println(string(wire))\n}", "func (v anyOfRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker47(w, v)\n}", "func (v GuardRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson795c59c6EncodeGrapeGuardRules10(w, v)\n}", "func (rmwaps ResourceModelWithAllowedPropertySet)MarshalJSON() ([]byte, error){\n objectMap := make(map[string]interface{})\n if(rmwaps.Location != nil) {\n objectMap[\"location\"] = rmwaps.Location\n }\n if(rmwaps.ManagedBy != nil) {\n objectMap[\"managedBy\"] = rmwaps.ManagedBy\n }\n if(rmwaps.Kind != nil) {\n objectMap[\"kind\"] = rmwaps.Kind\n }\n if(rmwaps.Tags != nil) {\n objectMap[\"tags\"] = rmwaps.Tags\n }\n if(rmwaps.Identity != nil) {\n objectMap[\"identity\"] = rmwaps.Identity\n }\n if(rmwaps.Sku != nil) {\n objectMap[\"sku\"] = rmwaps.Sku\n }\n if(rmwaps.Plan != nil) {\n objectMap[\"plan\"] = rmwaps.Plan\n }\n return json.Marshal(objectMap)\n }", "func (e EntityNameAvailabilityCheckOutput) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"message\", e.Message)\n\tpopulate(objectMap, \"nameAvailable\", e.NameAvailable)\n\tpopulate(objectMap, \"reason\", e.Reason)\n\treturn json.Marshal(objectMap)\n}", "func (r *HealthCheck) marshal(c *Client) ([]byte, error) {\n\tm, err := expandHealthCheck(c, r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshalling HealthCheck: %w\", err)\n\t}\n\n\treturn json.Marshal(m)\n}", "func (v singleRule) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson795c59c6EncodeGrapeGuardRules1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (t JSONTime)MarshalJSON(b []byte) (err error) {\n s := string(b)\n logging.Trace(currentCtx,\"JSONTime(String)=%s\",s)\n if s == \"null\" || s == \"\" {\n t.Time = time.Time{}\n return\n }\n tt, err := time.Parse(jsonTimeLayout, s)\n t = JSONTime{tt}\n return\n}", "func (c CheckNameAvailabilityRequestBody) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"name\", c.Name)\n\tpopulate(objectMap, \"type\", c.Type)\n\treturn json.Marshal(objectMap)\n}", "func mustToRawJson(v interface{}) (string, error) {\n\tbuf := new(bytes.Buffer)\n\tenc := json.NewEncoder(buf)\n\tenc.SetEscapeHTML(false)\n\terr := enc.Encode(&v)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn strings.TrimSuffix(buf.String(), \"\\n\"), nil\n}", "func (v prefixRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker13(w, v)\n}", "func (v matchRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker19(w, v)\n}", "func (e Effect) MarshalJSON() ([]byte, error) {\n\tif bool(e) {\n\t\treturn []byte(`\"Allow\"`), nil\n\t} else {\n\t\treturn []byte(`\"Deny\"`), nil\n\t}\n}", "func marshal(v interface{}) []byte {\n\tb, err := json.Marshal(v)\n\tassert(err == nil, \"marshal error: %s\", err)\n\treturn b\n}", "func (j JSON) MarshalJSON() ([]byte, error) {\n\tif j.Valid {\n\t\treturn json.Marshal(j.Map)\n\t}\n\n\treturn json.Marshal(nil)\n}", "func (v NickFail) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson42239ddeEncodeGithubComKhliengDispatchServer13(w, v)\n}", "func (rmwaps ResourceModelWithAllowedPropertySetIdentity)MarshalJSON() ([]byte, error){\n objectMap := make(map[string]interface{})\n if(rmwaps.Type != \"\") {\n objectMap[\"type\"] = rmwaps.Type\n }\n return json.Marshal(objectMap)\n }", "func (v empty) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdproto(w, v)\n}", "func (c *JsonCodec) CanMarshalWithCallback() bool {\n\treturn false\n}", "func (v JSONB) MarshalJSON() ([]byte, error) {\n\tif bytes.Equal(v, []byte{}) || bytes.Equal(v, []byte(\"null\")) {\n\t\treturn []byte(\"{}\"), nil\n\t}\n\n\treturn v, nil\n}", "func (spbi SuccessfulPropertyBatchInfo) MarshalJSON() ([]byte, error) {\n\tspbi.Kind = KindSuccessful\n\tobjectMap := make(map[string]interface{})\n\tif spbi.Properties != nil {\n\t\tobjectMap[\"Properties\"] = spbi.Properties\n\t}\n\tif spbi.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = spbi.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (v Bounds) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoBrowser28(w, v)\n}", "func (o InlineObject901) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.LogicalTest == nil {\n\t\tif o.isExplicitNullLogicalTest {\n\t\t\ttoSerialize[\"logicalTest\"] = o.LogicalTest\n\t\t}\n\t} else {\n\t\ttoSerialize[\"logicalTest\"] = o.LogicalTest\n\t}\n\tif o.ValueIfTrue == nil {\n\t\tif o.isExplicitNullValueIfTrue {\n\t\t\ttoSerialize[\"valueIfTrue\"] = o.ValueIfTrue\n\t\t}\n\t} else {\n\t\ttoSerialize[\"valueIfTrue\"] = o.ValueIfTrue\n\t}\n\tif o.ValueIfFalse == nil {\n\t\tif o.isExplicitNullValueIfFalse {\n\t\t\ttoSerialize[\"valueIfFalse\"] = o.ValueIfFalse\n\t\t}\n\t} else {\n\t\ttoSerialize[\"valueIfFalse\"] = o.ValueIfFalse\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (js JSONSerializable) MarshalJSON() ([]byte, error) {\n\tif !js.Valid {\n\t\treturn json.Marshal(nil)\n\t}\n\tjsWithHex := replaceBytesWithHex(js.Val)\n\treturn json.Marshal(jsWithHex)\n}", "func (v NullableSyntheticsAssertion) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(v.value)\n}", "func (j *HealthcheckRequest) MarshalJSON() ([]byte, error) {\n\tvar buf fflib.Buffer\n\tif j == nil {\n\t\tbuf.WriteString(\"null\")\n\t\treturn buf.Bytes(), nil\n\t}\n\terr := j.MarshalJSONBuf(&buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn buf.Bytes(), nil\n}", "func (out String) MarshalJSON() ([]byte, error) {\n\tif out.Valid {\n\t\treturn json.Marshal(out.String)\n\t}\n\treturn json.Marshal(nil)\n}", "func suppressEquivalentJSON(k, old, new string, d *schema.ResourceData) bool {\n\tvar oldObj, newObj interface{}\n\tif err := json.Unmarshal([]byte(old), &oldObj); err != nil {\n\t\treturn false\n\t}\n\tif err := json.Unmarshal([]byte(new), &newObj); err != nil {\n\t\treturn false\n\t}\n\treturn reflect.DeepEqual(oldObj, newObj)\n}", "func (f FailJSONEncoding) MarshalJSON() ([]byte, error) {\n\treturn nil, fmt.Errorf(\"error\")\n}", "func (c ContentKeyPolicyUnknownRestriction) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tobjectMap[\"@odata.type\"] = \"#Microsoft.Media.ContentKeyPolicyUnknownRestriction\"\n\treturn json.Marshal(objectMap)\n}", "func (v item) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonD2b7633eEncodeGithubComZhekabyGoGeneratorMongoRequestwrapperTests(w, v)\n}", "func (v booleanField) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker45(w, v)\n}", "func MarshalValid(obj MarshalValidater) ([]byte, error) {\n\terr := obj.Validate()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn obj.Marshal()\n}", "func (out Bool) MarshalJSON() ([]byte, error) {\n\tif out.Valid {\n\t\treturn json.Marshal(out.Bool)\n\t}\n\treturn json.Marshal(nil)\n}", "func (v InstallabilityErrorArgument) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage35(w, v)\n}", "func (v booleanField) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson390b7126EncodeGithubComChancedPicker45(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (n NullInt64) MarshalJSON() ([]byte, error) { return nulljson(n.Valid, n.Int64) }", "func (v allOfRule) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson390b7126EncodeGithubComChancedPicker48(w, v)\n}", "func (c CheckNameAvailabilityParameters) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"name\", c.Name)\n\tpopulate(objectMap, \"type\", c.Type)\n\treturn json.Marshal(objectMap)\n}", "func Marshal(v interface{}) ([]byte, error) {\n\tif ImplementsPreJSONMarshaler(v) {\n\t\terr := v.(PreJSONMarshaler).PreMarshalJSON()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn json.Marshal(v)\n}", "func (v PackoutRule) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson795c59c6EncodeGrapeGuardRules8(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v Boo) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson42239ddeEncodeMsgpJson(w, v)\n}", "func (c CheckNameAvailabilityInput) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"name\", c.Name)\n\tpopulate(objectMap, \"type\", c.Type)\n\treturn json.Marshal(objectMap)\n}", "func (o InlineObject885) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.X == nil {\n\t\tif o.isExplicitNullX {\n\t\t\ttoSerialize[\"x\"] = o.X\n\t\t}\n\t} else {\n\t\ttoSerialize[\"x\"] = o.X\n\t}\n\tif o.Alpha == nil {\n\t\tif o.isExplicitNullAlpha {\n\t\t\ttoSerialize[\"alpha\"] = o.Alpha\n\t\t}\n\t} else {\n\t\ttoSerialize[\"alpha\"] = o.Alpha\n\t}\n\tif o.Beta == nil {\n\t\tif o.isExplicitNullBeta {\n\t\t\ttoSerialize[\"beta\"] = o.Beta\n\t\t}\n\t} else {\n\t\ttoSerialize[\"beta\"] = o.Beta\n\t}\n\tif o.Cumulative == nil {\n\t\tif o.isExplicitNullCumulative {\n\t\t\ttoSerialize[\"cumulative\"] = o.Cumulative\n\t\t}\n\t} else {\n\t\ttoSerialize[\"cumulative\"] = o.Cumulative\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func marshal(v interface{}) string {\n\tb, err := json.Marshal(v)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn string(b)\n}", "func (c CheckAvailabilityParameters) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"id\", c.ID)\n\tpopulate(objectMap, \"isAvailiable\", c.IsAvailiable)\n\tpopulate(objectMap, \"location\", c.Location)\n\tpopulate(objectMap, \"name\", c.Name)\n\tpopulate(objectMap, \"sku\", c.SKU)\n\tpopulate(objectMap, \"tags\", c.Tags)\n\tpopulate(objectMap, \"type\", c.Type)\n\treturn json.Marshal(objectMap)\n}" ]
[ "0.6702867", "0.64294523", "0.6286051", "0.6258406", "0.6081727", "0.6065949", "0.6061521", "0.6033792", "0.60271937", "0.601817", "0.5984657", "0.5958404", "0.5944886", "0.59299093", "0.59280074", "0.5925472", "0.5916274", "0.5909515", "0.5889001", "0.58614844", "0.5856765", "0.58479005", "0.5834139", "0.5828814", "0.5828304", "0.58204126", "0.58201355", "0.5807023", "0.57906497", "0.57882106", "0.57785", "0.5769039", "0.57673883", "0.57613593", "0.57562864", "0.5753424", "0.574225", "0.5736402", "0.57329875", "0.57328606", "0.5724386", "0.57030654", "0.56702137", "0.56696993", "0.56650347", "0.56489974", "0.5645565", "0.5640649", "0.563292", "0.5625287", "0.56227136", "0.56210566", "0.55960935", "0.5595587", "0.55863243", "0.5576685", "0.55645806", "0.5563234", "0.55521834", "0.55499506", "0.5542916", "0.55418324", "0.5539396", "0.5536451", "0.5535558", "0.55268747", "0.55218357", "0.551219", "0.54910636", "0.54893494", "0.5481431", "0.54680026", "0.5463481", "0.54597306", "0.54577935", "0.5457747", "0.5457213", "0.5456877", "0.54519343", "0.5440569", "0.543449", "0.5432166", "0.5427301", "0.54138196", "0.5413408", "0.5403867", "0.53988135", "0.5397886", "0.53959614", "0.53929746", "0.5388998", "0.5383442", "0.53818214", "0.53809047", "0.53778243", "0.5376309", "0.53684974", "0.5368223", "0.53620225", "0.53579843", "0.5351151" ]
0.0
-1
UnmarshalJSON decodes the notification endpoint. This is necessary unfortunately.
func (d *DiffNotificationEndpointValues) UnmarshalJSON(b []byte) (err error) { d.NotificationEndpoint, err = endpoint.UnmarshalJSON(b) if errors2.EInvalid == errors2.ErrorCode(err) { return nil } return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (n *Notification) UnmarshalJSON(b []byte) error {\n\tnotification := map[string]interface{}{}\n\n\terr := json.Unmarshal(b, &notification)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif notification[\"_id\"] != nil && bson.IsObjectIdHex(notification[\"_id\"].(string)) {\n\t\tn.ID = bson.ObjectIdHex(notification[\"_id\"].(string))\n\t}\n\tif notification[\"id\"] != nil && bson.IsObjectIdHex(notification[\"id\"].(string)) {\n\t\tn.ID = bson.ObjectIdHex(notification[\"id\"].(string))\n\t}\n\tif notification[\"recipient\"] == nil {\n\t\t// return errors.New(\"Order Hash is not set\")\n\t} else {\n\t\tn.Recipient = common.HexToAddress(notification[\"recipient\"].(string))\n\t}\n\n\tif notification[\"message\"] != nil {\n\t\tn.Message = notification[\"message\"].(Message)\n\t}\n\n\tif notification[\"type\"] != nil {\n\t\tn.Type = notification[\"type\"].(string)\n\t}\n\n\tif notification[\"status\"] != nil {\n\t\tn.Status = notification[\"status\"].(string)\n\t}\n\n\tif notification[\"createdAt\"] != nil {\n\t\tnm, _ := time.Parse(time.RFC3339Nano, notification[\"createdAt\"].(string))\n\t\tn.CreatedAt = nm\n\t}\n\n\tif notification[\"updatedAt\"] != nil {\n\t\tnm, _ := time.Parse(time.RFC3339Nano, notification[\"updatedAt\"].(string))\n\t\tn.UpdatedAt = nm\n\t}\n\n\treturn nil\n}", "func (v *Notification) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson9806e1DecodeGithubComKonstantinProninEmailSendingServicePkgModel1(&r, v)\n\treturn r.Error()\n}", "func (s *SummaryNotificationEndpoint) UnmarshalJSON(b []byte) error {\n\tvar a struct {\n\t\tSummaryIdentifier\n\t\tNotificationEndpoint json.RawMessage `json:\"notificationEndpoint\"`\n\t\tLabelAssociations []SummaryLabel `json:\"labelAssociations\"`\n\t}\n\tif err := json.Unmarshal(b, &a); err != nil {\n\t\treturn err\n\t}\n\ts.SummaryIdentifier = a.SummaryIdentifier\n\ts.LabelAssociations = a.LabelAssociations\n\n\te, err := endpoint.UnmarshalJSON(a.NotificationEndpoint)\n\ts.NotificationEndpoint = e\n\treturn err\n}", "func (l *LiveEventEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"protocol\":\n\t\t\terr = unpopulate(val, \"Protocol\", &l.Protocol)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"url\":\n\t\t\terr = unpopulate(val, \"URL\", &l.URL)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NotificationHubResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &n.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &n.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &n.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &n.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &n.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &n.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &n.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *MessagingEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"lockDurationAsIso8601\":\n\t\t\terr = unpopulate(val, \"LockDurationAsIso8601\", &m.LockDurationAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"maxDeliveryCount\":\n\t\t\terr = unpopulate(val, \"MaxDeliveryCount\", &m.MaxDeliveryCount)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ttlAsIso8601\":\n\t\t\terr = unpopulate(val, \"TTLAsIso8601\", &m.TTLAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NotificationChannelProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"createdDate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &n.CreatedDate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"description\":\n\t\t\terr = unpopulate(val, &n.Description)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"emailRecipient\":\n\t\t\terr = unpopulate(val, &n.EmailRecipient)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"events\":\n\t\t\terr = unpopulate(val, &n.Events)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"notificationLocale\":\n\t\t\terr = unpopulate(val, &n.NotificationLocale)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, &n.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"uniqueIdentifier\":\n\t\t\terr = unpopulate(val, &n.UniqueIdentifier)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"webHookUrl\":\n\t\t\terr = unpopulate(val, &n.WebHookURL)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalNotificationResponse(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationResponse)\n\terr = core.UnmarshalPrimitive(m, \"notification_id\", &obj.NotificationID)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (p *PrivateEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"extendedLocation\":\n\t\t\terr = unpopulate(val, \"ExtendedLocation\", &p.ExtendedLocation)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &p.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &p.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *NotificationList) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson9806e1DecodeGithubComKonstantinProninEmailSendingServicePkgModel(&r, v)\n\treturn r.Error()\n}", "func (w *WebHookEventSubscriptionDestination) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &w.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &w.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &p.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &p.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EndpointAccessResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"relay\":\n\t\t\terr = unpopulate(val, \"Relay\", &e.Relay)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &e.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceId\":\n\t\t\terr = unpopulate(val, \"ResourceID\", &e.ResourceID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &e.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"privateEndpoint\":\n\t\t\terr = unpopulate(val, \"PrivateEndpoint\", &p.PrivateEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &p.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NotificationHubPatchParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &n.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &n.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &n.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &n.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &n.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &n.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &n.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *HTTPNotificationEndpoint_Headers) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]string)\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal string\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, fmt.Sprintf(\"error unmarshaling field %s\", fieldName))\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NotificationHubProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"admCredential\":\n\t\t\terr = unpopulate(val, \"AdmCredential\", &n.AdmCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"apnsCredential\":\n\t\t\terr = unpopulate(val, \"ApnsCredential\", &n.ApnsCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"authorizationRules\":\n\t\t\terr = unpopulate(val, \"AuthorizationRules\", &n.AuthorizationRules)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"baiduCredential\":\n\t\t\terr = unpopulate(val, \"BaiduCredential\", &n.BaiduCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"gcmCredential\":\n\t\t\terr = unpopulate(val, \"GCMCredential\", &n.GCMCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"mpnsCredential\":\n\t\t\terr = unpopulate(val, \"MpnsCredential\", &n.MpnsCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &n.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"registrationTtl\":\n\t\t\terr = unpopulate(val, \"RegistrationTTL\", &n.RegistrationTTL)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"wnsCredential\":\n\t\t\terr = unpopulate(val, \"WnsCredential\", &n.WnsCredential)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *JobDeliveryInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"scheduledDateTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"ScheduledDateTime\", &j.ScheduledDateTime)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *ApplicationGatewayPrivateEndpointConnection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &a.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &a.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &a.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &a.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"groupIds\":\n\t\t\terr = unpopulate(val, \"GroupIDs\", &p.GroupIDs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateEndpoint\":\n\t\t\terr = unpopulate(val, \"PrivateEndpoint\", &p.PrivateEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &p.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"groupIds\":\n\t\t\terr = unpopulate(val, \"GroupIDs\", &p.GroupIDs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateEndpoint\":\n\t\t\terr = unpopulate(val, \"PrivateEndpoint\", &p.PrivateEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &p.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (h *HybridConnectionEventSubscriptionDestination) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &h.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &h.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CreateEmailSignInURLResponse) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &c.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &c.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &c.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (pm *PingMessage) Unmarshal(bytes []byte) error {\n\tvar message PingMessage\n\tif err := json.Unmarshal(bytes, &message); err != nil {\n\t\treturn err\n\t}\n\tpm.Sender = message.Sender\n\tpm.Payload = message.Payload\n\n\treturn nil\n}", "func (a *ApplicationGatewayPrivateEndpointConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"linkIdentifier\":\n\t\t\terr = unpopulate(val, \"LinkIdentifier\", &a.LinkIdentifier)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateEndpoint\":\n\t\t\terr = unpopulate(val, \"PrivateEndpoint\", &a.PrivateEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &a.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &a.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"linkIdentifier\":\n\t\t\terr = unpopulate(val, \"LinkIdentifier\", &p.LinkIdentifier)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateEndpoint\":\n\t\t\terr = unpopulate(val, \"PrivateEndpoint\", &p.PrivateEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateEndpointLocation\":\n\t\t\terr = unpopulate(val, \"PrivateEndpointLocation\", &p.PrivateEndpointLocation)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &p.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingServiceBusTopicEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"connectionString\":\n\t\t\terr = unpopulate(val, \"ConnectionString\", &r.ConnectionString)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceGroup\":\n\t\t\terr = unpopulate(val, \"ResourceGroup\", &r.ResourceGroup)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, \"SubscriptionID\", &r.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CreateEmailSignInURLResponseProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"url\":\n\t\t\terr = unpopulate(val, \"URL\", &c.URL)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Notification) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson9806e1DecodeGithubComKonstantinProninEmailSendingServicePkgModel1(l, v)\n}", "func (p *P2SVPNGateway) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &p.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &p.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *PublishMessagesResponse) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (e *EnqueueServerProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"health\":\n\t\t\terr = unpopulate(val, \"Health\", &e.Health)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"hostname\":\n\t\t\terr = unpopulate(val, \"Hostname\", &e.Hostname)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipAddress\":\n\t\t\terr = unpopulate(val, \"IPAddress\", &e.IPAddress)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"port\":\n\t\t\terr = unpopulate(val, \"Port\", &e.Port)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *ServiceEndpointPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"locations\":\n\t\t\terr = unpopulate(val, \"Locations\", &s.Locations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &s.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"service\":\n\t\t\terr = unpopulate(val, \"Service\", &s.Service)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (w *WebhookUpdatePartnerDestinationInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &w.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &w.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EndpointResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &e.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &e.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &e.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &e.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &e.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateLinkServiceIPConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalNotificationsIntegrationGetResponse(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationsIntegrationGetResponse)\n\terr = core.UnmarshalPrimitive(m, \"description\", &obj.Description)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"name\", &obj.Name)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"enabled\", &obj.Enabled)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"type\", &obj.Type)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"updated_at\", &obj.UpdatedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"topic_count\", &obj.TopicCount)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"topic_names\", &obj.TopicNames)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (n *NotificationHubListResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"nextLink\":\n\t\t\terr = unpopulate(val, \"NextLink\", &n.NextLink)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &n.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (h *HybridConnectionEventSubscriptionDestinationProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"deliveryAttributeMappings\":\n\t\t\th.DeliveryAttributeMappings, err = unmarshalDeliveryAttributeMappingClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceId\":\n\t\t\terr = unpopulate(val, \"ResourceID\", &h.ResourceID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointIPConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *ConnectionMonitorEndpointFilter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"items\":\n\t\t\terr = unpopulate(val, \"Items\", &c.Items)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateLinkResourceBase) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *AutoApprovedPrivateLinkService) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"privateLinkService\":\n\t\t\terr = unpopulate(val, \"PrivateLinkService\", &a.PrivateLinkService)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateLinkService) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"extendedLocation\":\n\t\t\terr = unpopulate(val, \"ExtendedLocation\", &p.ExtendedLocation)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &p.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &p.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateLinkServiceConnection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (x *NotifyType) UnmarshalJSON(b []byte) error {\n\tnum, err := protoimpl.X.UnmarshalJSONEnum(x.Descriptor(), b)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*x = NotifyType(num)\n\treturn nil\n}", "func UnmarshalNotificationsIntegrationPostResponse(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationsIntegrationPostResponse)\n\terr = core.UnmarshalPrimitive(m, \"description\", &obj.Description)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"name\", &obj.Name)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"enabled\", &obj.Enabled)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"type\", &obj.Type)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"created_at\", &obj.CreatedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (s *ServiceEndpointPolicy) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &s.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"kind\":\n\t\t\terr = unpopulate(val, \"Kind\", &s.Kind)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &s.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &s.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *AvailablePrivateEndpointType) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, \"DisplayName\", &a.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &a.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &a.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceName\":\n\t\t\terr = unpopulate(val, \"ResourceName\", &a.ResourceName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &a.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EventSubscriptionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"deadLetterDestination\":\n\t\t\te.DeadLetterDestination, err = unmarshalDeadLetterDestinationClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deadLetterWithResourceIdentity\":\n\t\t\terr = unpopulate(val, \"DeadLetterWithResourceIdentity\", &e.DeadLetterWithResourceIdentity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deliveryWithResourceIdentity\":\n\t\t\terr = unpopulate(val, \"DeliveryWithResourceIdentity\", &e.DeliveryWithResourceIdentity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"destination\":\n\t\t\te.Destination, err = unmarshalEventSubscriptionDestinationClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"eventDeliverySchema\":\n\t\t\terr = unpopulate(val, \"EventDeliverySchema\", &e.EventDeliverySchema)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"expirationTimeUtc\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"ExpirationTimeUTC\", &e.ExpirationTimeUTC)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"filter\":\n\t\t\terr = unpopulate(val, \"Filter\", &e.Filter)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"labels\":\n\t\t\terr = unpopulate(val, \"Labels\", &e.Labels)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &e.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"retryPolicy\":\n\t\t\terr = unpopulate(val, \"RetryPolicy\", &e.RetryPolicy)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"topic\":\n\t\t\terr = unpopulate(val, \"Topic\", &e.Topic)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"applicationSecurityGroups\":\n\t\t\terr = unpopulate(val, \"ApplicationSecurityGroups\", &p.ApplicationSecurityGroups)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"customDnsConfigs\":\n\t\t\terr = unpopulate(val, \"CustomDNSConfigs\", &p.CustomDNSConfigs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"customNetworkInterfaceName\":\n\t\t\terr = unpopulate(val, \"CustomNetworkInterfaceName\", &p.CustomNetworkInterfaceName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipConfigurations\":\n\t\t\terr = unpopulate(val, \"IPConfigurations\", &p.IPConfigurations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"manualPrivateLinkServiceConnections\":\n\t\t\terr = unpopulate(val, \"ManualPrivateLinkServiceConnections\", &p.ManualPrivateLinkServiceConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"networkInterfaces\":\n\t\t\terr = unpopulate(val, \"NetworkInterfaces\", &p.NetworkInterfaces)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnections\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnections\", &p.PrivateLinkServiceConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subnet\":\n\t\t\terr = unpopulate(val, \"Subnet\", &p.Subnet)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (w *WebHookEventSubscriptionDestinationProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"azureActiveDirectoryApplicationIdOrUri\":\n\t\t\terr = unpopulate(val, \"AzureActiveDirectoryApplicationIDOrURI\", &w.AzureActiveDirectoryApplicationIDOrURI)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"azureActiveDirectoryTenantId\":\n\t\t\terr = unpopulate(val, \"AzureActiveDirectoryTenantID\", &w.AzureActiveDirectoryTenantID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deliveryAttributeMappings\":\n\t\t\tw.DeliveryAttributeMappings, err = unmarshalDeliveryAttributeMappingClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpointBaseUrl\":\n\t\t\terr = unpopulate(val, \"EndpointBaseURL\", &w.EndpointBaseURL)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpointUrl\":\n\t\t\terr = unpopulate(val, \"EndpointURL\", &w.EndpointURL)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"maxEventsPerBatch\":\n\t\t\terr = unpopulate(val, \"MaxEventsPerBatch\", &w.MaxEventsPerBatch)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"preferredBatchSizeInKilobytes\":\n\t\t\terr = unpopulate(val, \"PreferredBatchSizeInKilobytes\", &w.PreferredBatchSizeInKilobytes)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *StorageEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"connectionString\":\n\t\t\terr = unpopulate(val, \"ConnectionString\", &s.ConnectionString)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"containerName\":\n\t\t\terr = unpopulate(val, \"ContainerName\", &s.ContainerName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sasTtlAsIso8601\":\n\t\t\terr = unpopulate(val, \"SasTTLAsIso8601\", &s.SasTTLAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CommunicationsGatewayUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &c.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EndpointResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &e.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &e.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\te.Properties, err = unmarshalEndpointResourcePropertiesClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &e.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &e.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingMessage) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"appProperties\":\n\t\t\terr = unpopulate(val, \"AppProperties\", &r.AppProperties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"body\":\n\t\t\terr = unpopulate(val, \"Body\", &r.Body)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemProperties\":\n\t\t\terr = unpopulate(val, \"SystemProperties\", &r.SystemProperties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *InterfaceIPConfigurationPrivateLinkConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"fqdns\":\n\t\t\terr = unpopulate(val, \"Fqdns\", &i.Fqdns)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"groupId\":\n\t\t\terr = unpopulate(val, \"GroupID\", &i.GroupID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"requiredMemberName\":\n\t\t\terr = unpopulate(val, \"RequiredMemberName\", &i.RequiredMemberName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *IncomingMessage) UnmarshalJSON(data []byte) error {\n\tincomingMessageTypeOnly := struct {\n\t\tType string `json:\"type\"`\n\t}{}\n\tif err := json.Unmarshal(data, &incomingMessageTypeOnly); err != nil {\n\t\treturn err\n\t}\n\n\tswitch incomingMessageTypeOnly.Type {\n\tcase \"webhook_event\":\n\t\tvar evt WebhookEvent\n\t\tif err := json.Unmarshal(data, &evt); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tm.WebhookEvent = &evt\n\tcase \"request_log_event\":\n\t\tvar evt RequestLogEvent\n\t\tif err := json.Unmarshal(data, &evt); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tm.RequestLogEvent = &evt\n\tdefault:\n\t\treturn fmt.Errorf(\"Unexpected message type: %s\", incomingMessageTypeOnly.Type)\n\t}\n\n\treturn nil\n}", "func (a *AppPatch) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"identity\":\n\t\t\terr = unpopulate(val, \"Identity\", &a.Identity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &a.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &a.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *StorageQueueEventSubscriptionDestination) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &s.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *ConnectionMonitorEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"address\":\n\t\t\terr = unpopulate(val, \"Address\", &c.Address)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"coverageLevel\":\n\t\t\terr = unpopulate(val, \"CoverageLevel\", &c.CoverageLevel)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"filter\":\n\t\t\terr = unpopulate(val, \"Filter\", &c.Filter)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceId\":\n\t\t\terr = unpopulate(val, \"ResourceID\", &c.ResourceID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"scope\":\n\t\t\terr = unpopulate(val, \"Scope\", &c.Scope)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *MediaServiceUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"identity\":\n\t\t\terr = unpopulate(val, \"Identity\", &m.Identity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &m.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &m.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EventHubEventSubscriptionDestination) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &e.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &e.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FirewallPolicyIntrusionDetection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"configuration\":\n\t\t\terr = unpopulate(val, \"Configuration\", &f.Configuration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"mode\":\n\t\t\terr = unpopulate(val, \"Mode\", &f.Mode)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (future *PrivateEndpointConnectionsUpdateFuture) UnmarshalJSON(body []byte) error {\n\tvar azFuture azure.Future\n\tif err := json.Unmarshal(body, &azFuture); err != nil {\n\t\treturn err\n\t}\n\tfuture.FutureAPI = &azFuture\n\tfuture.Result = future.result\n\treturn nil\n}", "func (m *HookMessage) Unmarshal(b []byte) error {\n\terr := json.Unmarshal(b, m)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (e *EndpointServiceResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &e.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &e.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &e.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (g *GatewayServerProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", g, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"health\":\n\t\t\terr = unpopulate(val, \"Health\", &g.Health)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"port\":\n\t\t\terr = unpopulate(val, \"Port\", &g.Port)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", g, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (q *QnAMakerEndpointKeysResponse) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", q, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"installedVersion\":\n\t\t\terr = unpopulate(val, \"InstalledVersion\", &q.InstalledVersion)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastStableVersion\":\n\t\t\terr = unpopulate(val, \"LastStableVersion\", &q.LastStableVersion)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"primaryEndpointKey\":\n\t\t\terr = unpopulate(val, \"PrimaryEndpointKey\", &q.PrimaryEndpointKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"secondaryEndpointKey\":\n\t\t\terr = unpopulate(val, \"SecondaryEndpointKey\", &q.SecondaryEndpointKey)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", q, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SyncIdentityProviderUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &s.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *UnInstallRespPacket) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (s *ServiceBusTopicEventSubscriptionDestination) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &s.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *StorageInformation) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (w *WatcherPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &w.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *MessageServerProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"httpPort\":\n\t\t\terr = unpopulate(val, \"HTTPPort\", &m.HTTPPort)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"httpsPort\":\n\t\t\terr = unpopulate(val, \"HTTPSPort\", &m.HTTPSPort)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"health\":\n\t\t\terr = unpopulate(val, \"Health\", &m.Health)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"hostname\":\n\t\t\terr = unpopulate(val, \"Hostname\", &m.Hostname)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipAddress\":\n\t\t\terr = unpopulate(val, \"IPAddress\", &m.IPAddress)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"internalMsPort\":\n\t\t\terr = unpopulate(val, \"InternalMsPort\", &m.InternalMsPort)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"msPort\":\n\t\t\terr = unpopulate(val, \"MSPort\", &m.MSPort)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PartnerEventSubscriptionDestination) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &p.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingServiceBusQueueEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"connectionString\":\n\t\t\terr = unpopulate(val, \"ConnectionString\", &r.ConnectionString)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceGroup\":\n\t\t\terr = unpopulate(val, \"ResourceGroup\", &r.ResourceGroup)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, \"SubscriptionID\", &r.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (ep *Endpoint) UnmarshalJSON(raw []byte) error {\n\t// We may have to populate structures in the Endpoint manually to do the\n\t// translation from serializableEndpoint --> Endpoint.\n\trestoredEp := &serializableEndpoint{\n\t\tOpLabels: labels.NewOpLabels(),\n\t\tDNSHistory: fqdn.NewDNSCacheWithLimit(option.Config.ToFQDNsMinTTL, option.Config.ToFQDNsMaxIPsPerHost),\n\t\tDNSZombies: fqdn.NewDNSZombieMappings(option.Config.ToFQDNsMaxDeferredConnectionDeletes, option.Config.ToFQDNsMaxIPsPerHost),\n\t}\n\tif err := json.Unmarshal(raw, restoredEp); err != nil {\n\t\treturn fmt.Errorf(\"error unmarshaling serializableEndpoint from base64 representation: %s\", err)\n\t}\n\n\tep.fromSerializedEndpoint(restoredEp)\n\treturn nil\n}", "func (v *DeviceShadowUpdateMsg) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonB7ed31d3DecodeMevericcoreMccommon6(&r, v)\n\treturn r.Error()\n}", "func (s *ServiceTagInformationPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"addressPrefixes\":\n\t\t\terr = unpopulate(val, \"AddressPrefixes\", &s.AddressPrefixes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"changeNumber\":\n\t\t\terr = unpopulate(val, \"ChangeNumber\", &s.ChangeNumber)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"region\":\n\t\t\terr = unpopulate(val, \"Region\", &s.Region)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"state\":\n\t\t\terr = unpopulate(val, \"State\", &s.State)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemService\":\n\t\t\terr = unpopulate(val, \"SystemService\", &s.SystemService)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateLinkServiceConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"groupIds\":\n\t\t\terr = unpopulate(val, \"GroupIDs\", &p.GroupIDs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &p.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceId\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceID\", &p.PrivateLinkServiceID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"requestMessage\":\n\t\t\terr = unpopulate(val, \"RequestMessage\", &p.RequestMessage)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *FactoryPluginRespPacket) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (s *ServiceEndpointPolicyDefinition) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &s.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalNotificationEvent(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationEvent)\n\terr = core.UnmarshalPrimitive(m, \"event\", &obj.Event)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"target\", &obj.Target)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"source\", &obj.Source)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"triggered_by\", &obj.TriggeredBy)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"action_url\", &obj.ActionURL)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"data\", &obj.Data)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (a *AzureFunctionEventSubscriptionDestination) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &a.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *PurgeQueueResponse) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (e *EndpointResourceProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"authenticationType\":\n\t\t\terr = unpopulate(val, \"AuthenticationType\", &e.AuthenticationType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"createdTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"CreatedTime\", &e.CreatedTime)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deadLetterSecret\":\n\t\t\terr = unpopulate(val, \"DeadLetterSecret\", &e.DeadLetterSecret)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deadLetterUri\":\n\t\t\terr = unpopulate(val, \"DeadLetterURI\", &e.DeadLetterURI)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &e.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &e.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (d *DeliveryWithResourceIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"destination\":\n\t\t\td.Destination, err = unmarshalEventSubscriptionDestinationClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"identity\":\n\t\t\terr = unpopulate(val, \"Identity\", &d.Identity)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Message) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer18(&r, v)\n\treturn r.Error()\n}", "func (v *closeMessage) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson2802b09fDecodeGithubComPhilippseithSignalr6(&r, v)\n\treturn r.Error()\n}", "func (p *PrivateLinkServicePropertiesVisibility) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"subscriptions\":\n\t\t\terr = unpopulate(val, \"Subscriptions\", &p.Subscriptions)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *BootInitiationRespPacket) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (s *ServiceUpdate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}" ]
[ "0.7236028", "0.7215726", "0.6961847", "0.67759407", "0.67417806", "0.67417806", "0.67417806", "0.6726022", "0.66012084", "0.65831494", "0.6509571", "0.6504241", "0.6405176", "0.64035046", "0.63879925", "0.63658404", "0.63658404", "0.63457674", "0.6283621", "0.62764734", "0.62665147", "0.6260423", "0.62573195", "0.6238474", "0.6233371", "0.6227856", "0.6224738", "0.6224738", "0.62183636", "0.6207705", "0.6200474", "0.618839", "0.6184851", "0.61830777", "0.61790603", "0.6159602", "0.61371833", "0.6135999", "0.61356735", "0.6134781", "0.61209315", "0.61163586", "0.6113407", "0.61059886", "0.6105743", "0.61027527", "0.6096996", "0.60888237", "0.6082394", "0.60798335", "0.607423", "0.6070171", "0.60659724", "0.60632145", "0.60619736", "0.6054212", "0.60518956", "0.6049075", "0.60487103", "0.604525", "0.6039735", "0.60304785", "0.6029786", "0.60275644", "0.6023814", "0.6019317", "0.6017994", "0.60100406", "0.60096145", "0.60001904", "0.59983057", "0.59983015", "0.59945065", "0.59857357", "0.5980753", "0.5969409", "0.59667814", "0.59662825", "0.596599", "0.59590095", "0.5956654", "0.5943625", "0.59429324", "0.5940191", "0.5932365", "0.5927289", "0.59257257", "0.5922632", "0.59223884", "0.5921914", "0.5921729", "0.5921603", "0.59182084", "0.59139395", "0.5912921", "0.59122884", "0.5912258", "0.5911674", "0.5909736", "0.59059733" ]
0.7383068
0
MarshalJSON marshals a summary chart.
func (s *SummaryChart) MarshalJSON() ([]byte, error) { b, err := influxdb.MarshalViewPropertiesJSON(s.Properties) if err != nil { return nil, err } type alias SummaryChart out := struct { Props json.RawMessage `json:"properties"` alias }{ Props: b, alias: alias(*s), } return json.Marshal(out) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (ss StatsSummary) MarshalJSON() ([]byte, error) {\n\ttype Alias StatsSummary\n\tresp := struct {\n\t\tStatDate *string `json:\"statDate\"`\n\t\tSummaryTime string `json:\"summaryTime\"`\n\t\tAlias\n\t}{\n\t\tSummaryTime: ss.SummaryTime.Format(TimeLayout),\n\t\tAlias: (Alias)(ss),\n\t}\n\tif ss.StatDate != nil {\n\t\tresp.StatDate = util.StrPtr(ss.StatDate.Format(dateFormat))\n\t}\n\treturn json.Marshal(&resp)\n}", "func (a *Action) AggregateSummaryJSON() ([]byte, error) {\n\ta.Lock()\n\tdefer a.Unlock()\n\n\tif a.agg == nil {\n\t\ta.agg = newActionAggregators(a)\n\t}\n\n\treturn a.agg.action.agg.resultJSON(), nil\n}", "func (s SimulationResults) MarshalJSON() ([]byte, error) {\n\tsummary := make(map[string]int)\n\n\tfor _, ops := range s {\n\t\tvar keyParts []string\n\t\tfor _, op := range ops {\n\t\t\tkeyParts = append(keyParts, fmt.Sprintf(\"%d\", op.Length))\n\t\t}\n\t\tsummary[strings.Join(keyParts, \" \")]++\n\t}\n\n\treturn json.Marshal(summary)\n}", "func (s *Series) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(&jsonMarshallableSeries{\n\t\tTarget: s.Alias(),\n\t\tDataPoints: s.Points(),\n\t})\n}", "func (s *SummaryChart) UnmarshalJSON(b []byte) error {\n\ttype alias SummaryChart\n\ta := (*alias)(s)\n\tif err := json.Unmarshal(b, a); err != nil {\n\t\treturn err\n\t}\n\ts.XPosition = a.XPosition\n\ts.XPosition = a.YPosition\n\ts.Height = a.Height\n\ts.Width = a.Width\n\n\tvp, err := influxdb.UnmarshalViewPropertiesJSON(b)\n\tif err != nil {\n\t\treturn err\n\t}\n\ts.Properties = vp\n\treturn nil\n}", "func (m *Monitoring) Summary(w http.ResponseWriter, r *http.Request) {\n\tb, err := stats.Summary()\n\tif err != nil {\n\t\tError(w, http.StatusNotFound, err, \"failed to get metrics\")\n\t\treturn\n\t}\n\tJSON(w, http.StatusOK, b)\n}", "func (es ExecutionStatistics) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tif es.ExecutionCount != nil {\n\t\tobjectMap[\"executionCount\"] = es.ExecutionCount\n\t}\n\tif es.CPUTimeMs != nil {\n\t\tobjectMap[\"cpuTimeMs\"] = es.CPUTimeMs\n\t}\n\tif es.ElapsedTimeMs != nil {\n\t\tobjectMap[\"elapsedTimeMs\"] = es.ElapsedTimeMs\n\t}\n\tif es.WaitStats != nil {\n\t\tobjectMap[\"waitStats\"] = es.WaitStats\n\t}\n\tif es.HasErrors != nil {\n\t\tobjectMap[\"hasErrors\"] = es.HasErrors\n\t}\n\tif es.SQLErrors != nil {\n\t\tobjectMap[\"sqlErrors\"] = es.SQLErrors\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (j *Jsonnet) Summarize() (Summary, error) {\n\treturn Summary{\n\t\tComponentName: j.Name(false),\n\t\tType: \"jsonnet\",\n\t}, nil\n}", "func (rs RegistryStatistics) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\treturn json.Marshal(objectMap)\n}", "func (ss StatsSummaryLastUpdated) MarshalJSON() ([]byte, error) {\n\tresp := struct {\n\t\tSummaryTime *string `json:\"summaryTime\"`\n\t}{}\n\tif ss.SummaryTime != nil {\n\t\tresp.SummaryTime = util.StrPtr(ss.SummaryTime.Format(TimeLayout))\n\t}\n\treturn json.Marshal(&resp)\n}", "func (v SFMetric) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson51bca34dEncodeGithubComSkydiveProjectSkydiveSflow2(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (h *Histogram) MarshalJSON() ([]byte, error) {\n\tb := bytes.NewBuffer(make([]byte, 0, 4096))\n\tfmt.Fprintf(b, \"{\")\n\ttotalCount := int64(0)\n\tfor i, label := range h.labels {\n\t\tcount := h.buckets[i].Load()\n\t\ttotalCount += count\n\t\tfmt.Fprintf(b, \"\\\"%v\\\": %v, \", label, count)\n\t}\n\tfmt.Fprintf(b, \"\\\"%s\\\": %v, \", h.countLabel, totalCount)\n\tfmt.Fprintf(b, \"\\\"%s\\\": %v\", h.totalLabel, h.total.Load())\n\tfmt.Fprintf(b, \"}\")\n\treturn b.Bytes(), nil\n}", "func (v Metric) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPerformance(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v EventMetrics) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPerformance3(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v OvsMetric) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson51bca34dEncodeGithubComSkydiveProjectSkydiveSflow3(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (dsphe DeployedServicePackagesHealthEvaluation) MarshalJSON() ([]byte, error) {\n\tdsphe.Kind = KindDeployedServicePackages\n\tobjectMap := make(map[string]interface{})\n\tif dsphe.TotalCount != nil {\n\t\tobjectMap[\"TotalCount\"] = dsphe.TotalCount\n\t}\n\tif dsphe.UnhealthyEvaluations != nil {\n\t\tobjectMap[\"UnhealthyEvaluations\"] = dsphe.UnhealthyEvaluations\n\t}\n\tif dsphe.AggregatedHealthState != \"\" {\n\t\tobjectMap[\"AggregatedHealthState\"] = dsphe.AggregatedHealthState\n\t}\n\tif dsphe.Description != nil {\n\t\tobjectMap[\"Description\"] = dsphe.Description\n\t}\n\tif dsphe.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = dsphe.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (dsphe DeployedServicePackageHealthEvaluation) MarshalJSON() ([]byte, error) {\n\tdsphe.Kind = KindDeployedServicePackage\n\tobjectMap := make(map[string]interface{})\n\tif dsphe.NodeName != nil {\n\t\tobjectMap[\"NodeName\"] = dsphe.NodeName\n\t}\n\tif dsphe.ApplicationName != nil {\n\t\tobjectMap[\"ApplicationName\"] = dsphe.ApplicationName\n\t}\n\tif dsphe.ServiceManifestName != nil {\n\t\tobjectMap[\"ServiceManifestName\"] = dsphe.ServiceManifestName\n\t}\n\tif dsphe.UnhealthyEvaluations != nil {\n\t\tobjectMap[\"UnhealthyEvaluations\"] = dsphe.UnhealthyEvaluations\n\t}\n\tif dsphe.AggregatedHealthState != \"\" {\n\t\tobjectMap[\"AggregatedHealthState\"] = dsphe.AggregatedHealthState\n\t}\n\tif dsphe.Description != nil {\n\t\tobjectMap[\"Description\"] = dsphe.Description\n\t}\n\tif dsphe.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = dsphe.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (s MonitoringJsonDatasetFormat) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (v Histogram) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoBrowser6(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func marshalStationviewsStationDataSummaryViewToStationDataSummaryResponseBody(v *stationviews.StationDataSummaryView) *StationDataSummaryResponseBody {\n\tif v == nil {\n\t\treturn nil\n\t}\n\tres := &StationDataSummaryResponseBody{\n\t\tStart: *v.Start,\n\t\tEnd: *v.End,\n\t\tNumberOfSamples: *v.NumberOfSamples,\n\t}\n\n\treturn res\n}", "func (r RegistryStatistics) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"disabledDeviceCount\", r.DisabledDeviceCount)\n\tpopulate(objectMap, \"enabledDeviceCount\", r.EnabledDeviceCount)\n\tpopulate(objectMap, \"totalDeviceCount\", r.TotalDeviceCount)\n\treturn json.Marshal(objectMap)\n}", "func Summary(jsonIn []byte) ([]byte, error) {\n\tvar s summary\n\tkeycache.Refresh()\n\n\tif err := json.Unmarshal(jsonIn, &s); err != nil {\n\t\treturn jsonStatusError(err)\n\t}\n\n\tif passvault.NumRecords() == 0 {\n\t\treturn jsonStatusError(errors.New(\"Vault is not created yet\"))\n\t}\n\n\tif err := validateAdmin(s.Name, s.Password); err != nil {\n\t\tlog.Printf(\"Error validating admin status of %s: %s\", s.Name, err)\n\t\treturn jsonStatusError(err)\n\t}\n\n\treturn jsonSummary()\n}", "func (v Metric) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPerformance(w, v)\n}", "func (o *Knowledgegroupstatistics) String() string {\n \n \n \n\n j, _ := json.Marshal(o)\n str, _ := strconv.Unquote(strings.Replace(strconv.Quote(string(j)), `\\\\u`, `\\u`, -1))\n\n return str\n}", "func (v GetMetricsReturns) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPerformance1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (ua UsageAggregation) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tif ua.ID != nil {\n\t\tobjectMap[\"id\"] = ua.ID\n\t}\n\tif ua.Name != nil {\n\t\tobjectMap[\"name\"] = ua.Name\n\t}\n\tif ua.Type != nil {\n\t\tobjectMap[\"type\"] = ua.Type\n\t}\n\tif ua.UsageSample != nil {\n\t\tobjectMap[\"properties\"] = ua.UsageSample\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (m *Metrics) String() string {\n\tb, _ := json.Marshal(m)\n\treturn string(b)\n}", "func (o MicrosoftGraphUsageDetails) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.LastAccessedDateTime == nil {\n\t\tif o.isExplicitNullLastAccessedDateTime {\n\t\t\ttoSerialize[\"lastAccessedDateTime\"] = o.LastAccessedDateTime\n\t\t}\n\t} else {\n\t\ttoSerialize[\"lastAccessedDateTime\"] = o.LastAccessedDateTime\n\t}\n\tif o.LastModifiedDateTime == nil {\n\t\tif o.isExplicitNullLastModifiedDateTime {\n\t\t\ttoSerialize[\"lastModifiedDateTime\"] = o.LastModifiedDateTime\n\t\t}\n\t} else {\n\t\ttoSerialize[\"lastModifiedDateTime\"] = o.LastModifiedDateTime\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func JSONEncode(metrics []FakeMetric) (string, error) {\n\tencoded, err := json.Marshal(metrics)\n\treturn string(encoded), err\n}", "func (s Sampling) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(s.String())\n}", "func (o MicrosoftGraphVisualInfo) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.Attribution == nil {\n\t\tif o.isExplicitNullAttribution {\n\t\t\ttoSerialize[\"attribution\"] = o.Attribution\n\t\t}\n\t} else {\n\t\ttoSerialize[\"attribution\"] = o.Attribution\n\t}\n\tif o.BackgroundColor == nil {\n\t\tif o.isExplicitNullBackgroundColor {\n\t\t\ttoSerialize[\"backgroundColor\"] = o.BackgroundColor\n\t\t}\n\t} else {\n\t\ttoSerialize[\"backgroundColor\"] = o.BackgroundColor\n\t}\n\tif o.Description == nil {\n\t\tif o.isExplicitNullDescription {\n\t\t\ttoSerialize[\"description\"] = o.Description\n\t\t}\n\t} else {\n\t\ttoSerialize[\"description\"] = o.Description\n\t}\n\tif o.DisplayText != nil {\n\t\ttoSerialize[\"displayText\"] = o.DisplayText\n\t}\n\tif o.Content == nil {\n\t\tif o.isExplicitNullContent {\n\t\t\ttoSerialize[\"content\"] = o.Content\n\t\t}\n\t} else {\n\t\ttoSerialize[\"content\"] = o.Content\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (ts *TimeStats) JSON() string {\n\tout, _ := json.Marshal(ts)\n\tts.Total = 0\n\tts.Earliest = time.Time{}\n\tts.Latest = time.Time{}\n\treturn string(out)\n\n}", "func (v GetHistogramsReturns) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoBrowser14(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (s Series) String() string {\n\tjs, _ := json.Marshal(s)\n\treturn string(js)\n}", "func (o *Metrics) String() string {\n \n \n \n \n \n \n \n \n \n \n \n\n j, _ := json.Marshal(o)\n str, _ := strconv.Unquote(strings.Replace(strconv.Quote(string(j)), `\\\\u`, `\\u`, -1))\n\n return str\n}", "func (v Histogram) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoBrowser6(w, v)\n}", "func (s MetricsSummary) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (metrics *Metrics) MarshalJSON() ([]byte, error) {\n\tif metrics == nil {\n\t\treturn nil, fmt.Errorf(\"cannot marshall nil\")\n\t}\n\n\tif metrics.promisedTransactions == nil || metrics.promisedTransfers == nil ||\n\t\tmetrics.committedTransactions == nil || metrics.committedTransfers == nil ||\n\t\tmetrics.rollbackedTransactions == nil || metrics.rollbackedTransfers == nil ||\n\t\tmetrics.transactionFinalizerCronLatency == nil {\n\t\treturn nil, fmt.Errorf(\"cannot marshall nil references\")\n\t}\n\n\tvar buffer bytes.Buffer\n\n\tbuffer.WriteString(\"{\\\"transactionFinalizerCronLatency\\\":\")\n\tbuffer.WriteString(strconv.FormatFloat(metrics.transactionFinalizerCronLatency.Percentile(0.95), 'f', -1, 64))\n\tbuffer.WriteString(\",\\\"promisedTransactions\\\":\")\n\tbuffer.WriteString(strconv.FormatInt(metrics.promisedTransactions.Count(), 10))\n\tbuffer.WriteString(\",\\\"promisedTransfers\\\":\")\n\tbuffer.WriteString(strconv.FormatInt(metrics.promisedTransfers.Count(), 10))\n\tbuffer.WriteString(\",\\\"committedTransactions\\\":\")\n\tbuffer.WriteString(strconv.FormatInt(metrics.committedTransactions.Count(), 10))\n\tbuffer.WriteString(\",\\\"committedTransfers\\\":\")\n\tbuffer.WriteString(strconv.FormatInt(metrics.committedTransfers.Count(), 10))\n\tbuffer.WriteString(\",\\\"rollbackedTransactions\\\":\")\n\tbuffer.WriteString(strconv.FormatInt(metrics.rollbackedTransactions.Count(), 10))\n\tbuffer.WriteString(\",\\\"rollbackedTransfers\\\":\")\n\tbuffer.WriteString(strconv.FormatInt(metrics.rollbackedTransfers.Count(), 10))\n\tbuffer.WriteString(\"}\")\n\n\treturn buffer.Bytes(), nil\n}", "func (v GetLayoutMetricsReturns) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage48(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v GetHistogramReturns) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoBrowser16(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (o SyntheticsAPITestResultShortResult) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.Passed != nil {\n\t\ttoSerialize[\"passed\"] = o.Passed\n\t}\n\tif o.Timings != nil {\n\t\ttoSerialize[\"timings\"] = o.Timings\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func ReportJSON(results []Result) ([]byte, error) {\n\treturn json.Marshal(NewMetrics(results))\n}", "func (m *GetAPICallSummaryReq) ToJSON() (string, error) {\n\treturn codec.ToJSON(m)\n}", "func (she ServiceHealthEvaluation) MarshalJSON() ([]byte, error) {\n\tshe.Kind = KindService\n\tobjectMap := make(map[string]interface{})\n\tif she.ServiceName != nil {\n\t\tobjectMap[\"ServiceName\"] = she.ServiceName\n\t}\n\tif she.UnhealthyEvaluations != nil {\n\t\tobjectMap[\"UnhealthyEvaluations\"] = she.UnhealthyEvaluations\n\t}\n\tif she.AggregatedHealthState != \"\" {\n\t\tobjectMap[\"AggregatedHealthState\"] = she.AggregatedHealthState\n\t}\n\tif she.Description != nil {\n\t\tobjectMap[\"Description\"] = she.Description\n\t}\n\tif she.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = she.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (vr *VulnerabilityReport) MarshalJSON() ([]byte, error) {\n\tvrs := vr.Spec\n\n\tfor _, vuln := range vrs.Vulnerabilities {\n\t\tswitch vuln.Severity {\n\t\tcase \"UNKNOWN\":\n\t\t\tvrs.Summary.UnknownCount++\n\t\tcase \"LOW\":\n\t\t\tvrs.Summary.LowCount++\n\t\tcase \"MEDIUM\":\n\t\t\tvrs.Summary.MediumCount++\n\t\tcase \"HIGH\":\n\t\t\tvrs.Summary.HighCount++\n\t\tcase \"CRITICAL\":\n\t\t\tvrs.Summary.CriticalCount++\n\t\t}\n\t}\n\n\tif vrs.Summary.CriticalCount > 0 {\n\t\tvrs.Summary.HighestSeverity = SeverityCritical\n\t} else if vrs.Summary.HighCount > 0 {\n\t\tvrs.Summary.HighestSeverity = SeverityHigh\n\t} else if vrs.Summary.MediumCount > 0 {\n\t\tvrs.Summary.HighestSeverity = SeverityMedium\n\t} else if vrs.Summary.LowCount > 0 {\n\t\tvrs.Summary.HighestSeverity = SeverityLow\n\t} else if vrs.Summary.NoneCount > 0 {\n\t\tvrs.Summary.HighestSeverity = SeverityNone\n\t} else if vrs.Summary.UnknownCount > 0 {\n\t\tvrs.Summary.HighestSeverity = SeverityUnknown\n\t}\n\n\tmetadata := vr.ObjectMeta\n\n\tif metadata.Labels == nil {\n\t\tmetadata.Labels = make(map[string]string)\n\t}\n\n\tmetadata.Labels[\"severity\"] = vrs.Summary.HighestSeverity\n\n\tfor pod := range vr.Status.Pods {\n\t\tmetadata.Labels[pod] = \"true\"\n\t}\n\n\ttype Alias VulnerabilityReport\n\treturn json.Marshal(&struct {\n\t\tObjectMeta metav1.ObjectMeta `json:\"metadata,omitempty\"`\n\t\tSpec VulnerabilityReportSpec `json:\"spec\"`\n\t\t*Alias\n\t}{\n\t\tObjectMeta: metadata,\n\t\tSpec: vrs,\n\t\tAlias: (*Alias)(vr),\n\t})\n}", "func (s *CountMinSketch) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(CountMinSketchJSON{s.d, s.w, s.count})\n}", "func (c *AverageDuration) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(struct {\n\t\tCount int64 `json:\"total\"`\n\t\tAverage int64 `json:\"average\"`\n\t}{\n\t\tc.Count(),\n\t\tc.Average(),\n\t})\n}", "func (h *Histogram) String() string {\n\tb, _ := h.MarshalJSON()\n\treturn string(b)\n}", "func (v GetHistogramReturns) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoBrowser16(w, v)\n}", "func (p point) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal([]interface{}{\n\t\tp.metric,\n\t\tp.value,\n\t})\n}", "func (sl SketchSeriesList) MarshalJSON() ([]byte, error) {\n\tdata := map[string][]*SketchSeries{\n\t\t\"sketch_series\": sl,\n\t}\n\treqBody := &bytes.Buffer{}\n\terr := json.NewEncoder(reqBody).Encode(data)\n\treturn reqBody.Bytes(), err\n}", "func (hst FioJsonHistogram) MarshalJSON() ([]byte, error) {\n\tstarted := false\n\tbuf := make([]byte, 8192) // lazy\n\tsep := \"\"\n\tbuf[0] = '{'\n\tbufidx := 1\n\tfor key, val := range hst {\n\t\tif started {\n\t\t\tsep = \",\"\n\t\t} else {\n\t\t\tstarted = true\n\t\t}\n\t\tout := []byte(fmt.Sprintf(\"\\\"%g\\\":%g%s\", key, val, sep))\n\t\tbufidx += copy(buf[bufidx:bufidx+len(out)], out)\n\t}\n\tbuf[bufidx] = '}'\n\n\treturn buf[0 : bufidx+1], nil\n}", "func (o UsageTopAvgMetricsHour) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.AvgMetricHour != nil {\n\t\ttoSerialize[\"avg_metric_hour\"] = o.AvgMetricHour\n\t}\n\tif o.MaxMetricHour != nil {\n\t\ttoSerialize[\"max_metric_hour\"] = o.MaxMetricHour\n\t}\n\tif o.MetricCategory != nil {\n\t\ttoSerialize[\"metric_category\"] = o.MetricCategory\n\t}\n\tif o.MetricName != nil {\n\t\ttoSerialize[\"metric_name\"] = o.MetricName\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (s SapLandscapeMonitorMetricThresholds) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"green\", s.Green)\n\tpopulate(objectMap, \"name\", s.Name)\n\tpopulate(objectMap, \"red\", s.Red)\n\tpopulate(objectMap, \"yellow\", s.Yellow)\n\treturn json.Marshal(objectMap)\n}", "func (v GetMetricsReturns) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPerformance1(w, v)\n}", "func (she ServicesHealthEvaluation) MarshalJSON() ([]byte, error) {\n\tshe.Kind = KindServices\n\tobjectMap := make(map[string]interface{})\n\tif she.ServiceTypeName != nil {\n\t\tobjectMap[\"ServiceTypeName\"] = she.ServiceTypeName\n\t}\n\tif she.MaxPercentUnhealthyServices != nil {\n\t\tobjectMap[\"MaxPercentUnhealthyServices\"] = she.MaxPercentUnhealthyServices\n\t}\n\tif she.TotalCount != nil {\n\t\tobjectMap[\"TotalCount\"] = she.TotalCount\n\t}\n\tif she.UnhealthyEvaluations != nil {\n\t\tobjectMap[\"UnhealthyEvaluations\"] = she.UnhealthyEvaluations\n\t}\n\tif she.AggregatedHealthState != \"\" {\n\t\tobjectMap[\"AggregatedHealthState\"] = she.AggregatedHealthState\n\t}\n\tif she.Description != nil {\n\t\tobjectMap[\"Description\"] = she.Description\n\t}\n\tif she.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = she.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (o SyntheticsBrowserTestResultShortResult) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.Device != nil {\n\t\ttoSerialize[\"device\"] = o.Device\n\t}\n\tif o.Duration != nil {\n\t\ttoSerialize[\"duration\"] = o.Duration\n\t}\n\tif o.ErrorCount != nil {\n\t\ttoSerialize[\"errorCount\"] = o.ErrorCount\n\t}\n\tif o.StepCountCompleted != nil {\n\t\ttoSerialize[\"stepCountCompleted\"] = o.StepCountCompleted\n\t}\n\tif o.StepCountTotal != nil {\n\t\ttoSerialize[\"stepCountTotal\"] = o.StepCountTotal\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (m MetricDimension) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"displayName\", m.DisplayName)\n\tpopulate(objectMap, \"name\", m.Name)\n\tpopulate(objectMap, \"toBeExportedForShoebox\", m.ToBeExportedForShoebox)\n\treturn json.Marshal(objectMap)\n}", "func (v SFMetric) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson51bca34dEncodeGithubComSkydiveProjectSkydiveSflow2(w, v)\n}", "func (s AuthorizedAppStats) MarshalJSON() ([]byte, error) {\n\t// Do nothing if there's no records\n\tif len(s) == 0 {\n\t\treturn json.Marshal(struct{}{})\n\t}\n\n\tstats := make([]*jsonAuthorizedAppStatstats, 0, len(s))\n\tfor _, stat := range s {\n\t\tstats = append(stats, &jsonAuthorizedAppStatstats{\n\t\t\tDate: stat.Date,\n\t\t\tData: &jsonAuthorizedAppStatstatsData{\n\t\t\t\tCodesIssued: stat.CodesIssued,\n\t\t\t\tCodesClaimed: stat.CodesClaimed,\n\t\t\t\tCodesInvalid: stat.CodesInvalid,\n\t\t\t\tTokensClaimed: stat.TokensClaimed,\n\t\t\t\tTokensInvalid: stat.TokensInvalid,\n\t\t\t},\n\t\t})\n\t}\n\n\t// Sort in descending order.\n\tsort.Slice(stats, func(i, j int) bool {\n\t\treturn stats[i].Date.After(stats[j].Date)\n\t})\n\n\tvar result jsonAuthorizedAppStat\n\tresult.AuthorizedAppID = s[0].AuthorizedAppID\n\tresult.AuthorizedAppName = s[0].AuthorizedAppName\n\tresult.AuthorizedAppType = s[0].AuthorizedAppType\n\tresult.Stats = stats\n\n\tb, err := json.Marshal(result)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to marshal json: %w\", err)\n\t}\n\treturn b, nil\n}", "func (o HostMetrics) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.Cpu != nil {\n\t\ttoSerialize[\"cpu\"] = o.Cpu\n\t}\n\tif o.Iowait != nil {\n\t\ttoSerialize[\"iowait\"] = o.Iowait\n\t}\n\tif o.Load != nil {\n\t\ttoSerialize[\"load\"] = o.Load\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (m MetricDimension) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"displayName\", m.DisplayName)\n\tpopulate(objectMap, \"name\", m.Name)\n\treturn json.Marshal(objectMap)\n}", "func (csr CampaignStatisticsResult) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tif csr.Queued != nil {\n\t\tobjectMap[\"queued\"] = csr.Queued\n\t}\n\tif csr.Pushed != nil {\n\t\tobjectMap[\"pushed\"] = csr.Pushed\n\t}\n\tif csr.PushedNative != nil {\n\t\tobjectMap[\"pushed-native\"] = csr.PushedNative\n\t}\n\tif csr.PushedNativeGoogle != nil {\n\t\tobjectMap[\"pushed-native-google\"] = csr.PushedNativeGoogle\n\t}\n\tif csr.PushedNativeAdm != nil {\n\t\tobjectMap[\"pushed-native-adm\"] = csr.PushedNativeAdm\n\t}\n\tif csr.Delivered != nil {\n\t\tobjectMap[\"delivered\"] = csr.Delivered\n\t}\n\tif csr.Dropped != nil {\n\t\tobjectMap[\"dropped\"] = csr.Dropped\n\t}\n\tif csr.SystemNotificationDisplayed != nil {\n\t\tobjectMap[\"system-notification-displayed\"] = csr.SystemNotificationDisplayed\n\t}\n\tif csr.InAppNotificationDisplayed != nil {\n\t\tobjectMap[\"in-app-notification-displayed\"] = csr.InAppNotificationDisplayed\n\t}\n\tif csr.ContentDisplayed != nil {\n\t\tobjectMap[\"content-displayed\"] = csr.ContentDisplayed\n\t}\n\tif csr.SystemNotificationActioned != nil {\n\t\tobjectMap[\"system-notification-actioned\"] = csr.SystemNotificationActioned\n\t}\n\tif csr.SystemNotificationExited != nil {\n\t\tobjectMap[\"system-notification-exited\"] = csr.SystemNotificationExited\n\t}\n\tif csr.InAppNotificationActioned != nil {\n\t\tobjectMap[\"in-app-notification-actioned\"] = csr.InAppNotificationActioned\n\t}\n\tif csr.InAppNotificationExited != nil {\n\t\tobjectMap[\"in-app-notification-exited\"] = csr.InAppNotificationExited\n\t}\n\tif csr.ContentActioned != nil {\n\t\tobjectMap[\"content-actioned\"] = csr.ContentActioned\n\t}\n\tif csr.ContentExited != nil {\n\t\tobjectMap[\"content-exited\"] = csr.ContentExited\n\t}\n\tif csr.Answers != nil {\n\t\tobjectMap[\"answers\"] = csr.Answers\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (a *MetricAggregation) MarshalJSON() ([]byte, error) {\n\troot := map[string]interface{}{\n\t\t\"field\": a.Field,\n\t}\n\n\tfor k, v := range a.Settings {\n\t\tif k != \"\" && v != nil {\n\t\t\troot[k] = v\n\t\t}\n\t}\n\n\treturn json.Marshal(root)\n}", "func (aslst AverageServiceLoadScalingTrigger) MarshalJSON() ([]byte, error) {\n\taslst.Kind = KindAverageServiceLoad\n\tobjectMap := make(map[string]interface{})\n\tif aslst.MetricName != nil {\n\t\tobjectMap[\"MetricName\"] = aslst.MetricName\n\t}\n\tif aslst.LowerLoadThreshold != nil {\n\t\tobjectMap[\"LowerLoadThreshold\"] = aslst.LowerLoadThreshold\n\t}\n\tif aslst.UpperLoadThreshold != nil {\n\t\tobjectMap[\"UpperLoadThreshold\"] = aslst.UpperLoadThreshold\n\t}\n\tif aslst.ScaleIntervalInSeconds != nil {\n\t\tobjectMap[\"ScaleIntervalInSeconds\"] = aslst.ScaleIntervalInSeconds\n\t}\n\tif aslst.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = aslst.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (bs *blplStats) statsJSON() string {\n\tbuf := bytes.NewBuffer(make([]byte, 0, 128))\n\tfmt.Fprintf(buf, \"{\")\n\tfmt.Fprintf(buf, \"\\n \\\"TxnCount\\\": %v,\", bs.txnCount)\n\tfmt.Fprintf(buf, \"\\n \\\"QueryCount\\\": %v,\", bs.queryCount)\n\tfmt.Fprintf(buf, \"\\n \\\"QueriesPerSec\\\": %v,\", bs.queriesPerSec)\n\tfmt.Fprintf(buf, \"\\n \\\"TxnPerSec\\\": %v\", bs.txnsPerSec)\n\tfmt.Fprintf(buf, \"\\n \\\"TxnTime\\\": %v,\", bs.txnTime)\n\tfmt.Fprintf(buf, \"\\n \\\"QueryTime\\\": %v,\", bs.queryTime)\n\tfmt.Fprintf(buf, \"\\n}\")\n\treturn buf.String()\n}", "func (v count) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson390b7126EncodeGithubComChancedPicker42(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (s *PrometheusSerializer) Flush() error {\n\twr := &prompb.WriteRequest{\n\t\tTimeseries: s.series[:s.cur],\n\t}\n\tdata, err := wr.Marshal()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsb := bytebufferpool.Get()\n\tsb.B = snappy.Encode(sb.B, data)\n\n\tvar sizeBuf []byte\n\tsizeBuf = marshalUint64(sizeBuf[:0], uint64(sb.Len()))\n\tif _, err := s.w.Write(sizeBuf); err != nil {\n\t\tbytebufferpool.Put(sb)\n\t\treturn err\n\t}\n\n\t_, err = s.w.Write(sb.Bytes())\n\ts.cur = 0\n\tbytebufferpool.Put(sb)\n\treturn err\n}", "func (s ResourceCountsSummary) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (v GetHistogramsReturns) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoBrowser14(w, v)\n}", "func (s *SeriesStruct) UnmarshalJSON(data []byte) error {\n\t// find the outer-most keys\n\tvar partial map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &partial); err != nil {\n\t\treturn err\n\t}\n\tsummary := partial[\"summary\"]\n\n\t// This is not strictly necessary but it is faster\n\tdelete(partial, \"summary\")\n\tdata, _ = json.Marshal(partial)\n\n\tvar ss seriesStruct\n\tif err := json.Unmarshal(data, &ss); err != nil {\n\t\treturn err\n\t}\n\n\t// \"null\" is 4 bytes\n\tif len(summary) > 4 {\n\t\tswitch ss.Game.Id {\n\t\t// Dota\n\t\tcase 1:\n\t\t\tvar tmp DotaSeriesSummary\n\t\t\tif err := json.Unmarshal(summary, &tmp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tss.Summary = tmp\n\t\t// Lol\n\t\tcase 2:\n\t\t\tvar tmp LolSeriesSummary\n\t\t\tif err := json.Unmarshal(summary, &tmp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tss.Summary = tmp\n\t\t//Cs\n\t\tcase 5:\n\t\t\tvar tmp CsSeriesSummary\n\t\t\tif err := json.Unmarshal(summary, &tmp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tss.Summary = tmp\n\t\tdefault:\n\t\t}\n\t}\n\n\t*s = SeriesStruct(ss)\n\treturn nil\n\n}", "func (e ExpressRouteCrossConnectionRoutesTableSummary) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"asn\", e.Asn)\n\tpopulate(objectMap, \"neighbor\", e.Neighbor)\n\tpopulate(objectMap, \"stateOrPrefixesReceived\", e.StateOrPrefixesReceived)\n\tpopulate(objectMap, \"upDown\", e.UpDown)\n\treturn json.Marshal(objectMap)\n}", "func (s *Summary) String() string {\n\treturn fmt.Sprintf(\n\t\t\"\\n{Connections: %d, RequestRate: %d, RequestTotal: %d, SuccessTotal: %d, ErrorTotal: %d, TimeElapsed: %s, Throughput: %.2f/s}\",\n\t\ts.Connections, s.RequestRate, (s.SuccessTotal + s.ErrorTotal), s.SuccessTotal, s.ErrorTotal, s.TimeElapsed, s.Throughput)\n}", "func (v GetLayoutMetricsReturns) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage48(w, v)\n}", "func (s CountsSummary) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (e ExpressRouteCircuitStats) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"primarybytesIn\", e.PrimarybytesIn)\n\tpopulate(objectMap, \"primarybytesOut\", e.PrimarybytesOut)\n\tpopulate(objectMap, \"secondarybytesIn\", e.SecondarybytesIn)\n\tpopulate(objectMap, \"secondarybytesOut\", e.SecondarybytesOut)\n\treturn json.Marshal(objectMap)\n}", "func (m *SensitiveDataCount) ToJSON() (string, error) {\n\treturn codec.ToJSON(m)\n}", "func SerializeIntSummaryValue(min, max, sum, count int64) string {\n\treturn fmt.Sprintf(\"gauge,min=%d,max=%d,sum=%d,count=%d\", min, max, sum, count)\n}", "func (f ForecastDataset) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"aggregation\", f.Aggregation)\n\tpopulate(objectMap, \"configuration\", f.Configuration)\n\tpopulate(objectMap, \"filter\", f.Filter)\n\tpopulate(objectMap, \"granularity\", f.Granularity)\n\treturn json.Marshal(objectMap)\n}", "func (o WorkbookChart) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.Height != nil {\n\t\ttoSerialize[\"height\"] = o.Height\n\t}\n\tif o.Left != nil {\n\t\ttoSerialize[\"left\"] = o.Left\n\t}\n\tif o.Name == nil {\n\t\tif o.isExplicitNullName {\n\t\t\ttoSerialize[\"name\"] = o.Name\n\t\t}\n\t} else {\n\t\ttoSerialize[\"name\"] = o.Name\n\t}\n\tif o.Top != nil {\n\t\ttoSerialize[\"top\"] = o.Top\n\t}\n\tif o.Width != nil {\n\t\ttoSerialize[\"width\"] = o.Width\n\t}\n\tif o.Axes == nil {\n\t\tif o.isExplicitNullAxes {\n\t\t\ttoSerialize[\"axes\"] = o.Axes\n\t\t}\n\t} else {\n\t\ttoSerialize[\"axes\"] = o.Axes\n\t}\n\tif o.DataLabels == nil {\n\t\tif o.isExplicitNullDataLabels {\n\t\t\ttoSerialize[\"dataLabels\"] = o.DataLabels\n\t\t}\n\t} else {\n\t\ttoSerialize[\"dataLabels\"] = o.DataLabels\n\t}\n\tif o.Format == nil {\n\t\tif o.isExplicitNullFormat {\n\t\t\ttoSerialize[\"format\"] = o.Format\n\t\t}\n\t} else {\n\t\ttoSerialize[\"format\"] = o.Format\n\t}\n\tif o.Legend == nil {\n\t\tif o.isExplicitNullLegend {\n\t\t\ttoSerialize[\"legend\"] = o.Legend\n\t\t}\n\t} else {\n\t\ttoSerialize[\"legend\"] = o.Legend\n\t}\n\tif o.Series != nil {\n\t\ttoSerialize[\"series\"] = o.Series\n\t}\n\tif o.Title == nil {\n\t\tif o.isExplicitNullTitle {\n\t\t\ttoSerialize[\"title\"] = o.Title\n\t\t}\n\t} else {\n\t\ttoSerialize[\"title\"] = o.Title\n\t}\n\tif o.Worksheet == nil {\n\t\tif o.isExplicitNullWorksheet {\n\t\t\ttoSerialize[\"worksheet\"] = o.Worksheet\n\t\t}\n\t} else {\n\t\ttoSerialize[\"worksheet\"] = o.Worksheet\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (ms *MetricSet) MarshalJSON() ([]byte, error) {\n\ttype toJSON MetricSet\n\treturn json.Marshal(&struct {\n\t\tType string `json:\"type\"`\n\t\t*toJSON\n\t}{\n\t\tType: strings.ToLower(ms.Type.String()),\n\t\ttoJSON: (*toJSON)(ms),\n\t})\n}", "func marshalChatterviewsChatSummaryViewToChatSummaryResponse(v *chatterviews.ChatSummaryView) *ChatSummaryResponse {\n\tres := &ChatSummaryResponse{\n\t\tMessage: *v.Message,\n\t\tLength: v.Length,\n\t\tSentAt: *v.SentAt,\n\t}\n\n\treturn res\n}", "func (v OvsMetric) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson51bca34dEncodeGithubComSkydiveProjectSkydiveSflow3(w, v)\n}", "func StatsSummary() *client.StatsSummaryResponse {\n\treturn &client.StatsSummaryResponse{\n\t\tResponse: []client.StatsSummary{\n\t\t\tclient.StatsSummary{\n\t\t\t\tSummaryTime: \"2015-05-14 14:39:47\",\n\t\t\t\tDeliveryService: \"test-ds1\",\n\t\t\t\tStatName: \"test-stat\",\n\t\t\t\tStatValue: \"3.1415\",\n\t\t\t\tCDNName: \"test-cdn\",\n\t\t\t},\n\t\t},\n\t}\n}", "func (v ScreencastFrameMetadata) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage20(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (a *Analysis) JSON() string {\n\tj, err := json.Marshal(*a)\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\treturn string(j)\n}", "func (v PlantainerShadowMetadataSt) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson5bd79fa1EncodeMevericcoreMcplantainer9(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (o ApplicationStatusOperationStateOperationSyncSourceOutput) Chart() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSyncSource) *string { return v.Chart }).(pulumi.StringPtrOutput)\n}", "func (v EventMetrics) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPerformance3(w, v)\n}", "func (c Count) JSON() cast.JSON {\n\tjson := cast.JSON{}\n\tfor k, v := range c {\n\t\tjson[cast.StringI(int(k))] = v\n\t}\n\treturn json\n}", "func (m Calls) MarshalJSON() ([]byte, error) {\n\tcalls := make([]*CallStats, 0, len(m))\n\tfor _, s := range m {\n\t\tcalls = append(calls, s)\n\t}\n\tsort.Slice(calls, func(i, j int) bool {\n\t\treturn calls[i].Count > calls[j].Count\n\t})\n\n\treturn json.Marshal(calls)\n}", "func (m MetricSpecification) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"aggregationType\", m.AggregationType)\n\tpopulate(objectMap, \"dimensions\", m.Dimensions)\n\tpopulate(objectMap, \"displayDescription\", m.DisplayDescription)\n\tpopulate(objectMap, \"displayName\", m.DisplayName)\n\tpopulate(objectMap, \"name\", m.Name)\n\tpopulate(objectMap, \"supportedTimeGrainTypes\", m.SupportedTimeGrainTypes)\n\tpopulate(objectMap, \"unit\", m.Unit)\n\treturn json.Marshal(objectMap)\n}", "func DeploySummary(ctx context.Context, objs Objects) (string, error) {\n\t// Sort values\n\tvar sorted []*Object\n\tfor _, obj := range objs {\n\t\tsorted = append(sorted, obj)\n\t}\n\tsorted = sortObjectsByKindAndName(sorted)\n\n\t// Create table\n\tpadding := 4\n\tbuf := new(bytes.Buffer)\n\tw := tabwriter.NewWriter(buf, 0, 0, padding, ' ', 0)\n\n\tif _, err := fmt.Fprintln(w, \"NAMESPACE\\tKIND\\tNAME\\tREADY\\t\"); err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to write to writer: %v\", err)\n\t}\n\n\tfor _, obj := range sorted {\n\t\tkind := ObjectKind(obj)\n\t\tname, err := ObjectName(obj)\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"failed to get resource name: %v\", err)\n\t\t}\n\t\tnamespace, err := ObjectNamespace(obj)\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"failed to get namespace of object: %v\", err)\n\t\t}\n\t\tif namespace == \"\" {\n\t\t\tnamespace = \"default\"\n\t\t}\n\n\t\textraInfo, err := deploySummaryExtraInfo(obj)\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"failed to get resource summary extra info: %v\", err)\n\t\t}\n\n\t\tvar ready string\n\t\tok, err := IsReady(ctx, obj)\n\t\tif err != nil {\n\t\t\tready = \"Unknown\"\n\t\t} else if ok {\n\t\t\tready = \"Yes\"\n\t\t} else {\n\t\t\tready = \"No\"\n\t\t}\n\n\t\tif _, err := fmt.Fprintf(w, \"%s\\t%s\\t%s\\t%s\\t%s\\n\", namespace, kind, name, ready, extraInfo); err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"failed to write to writer: %v\", err)\n\t\t}\n\t}\n\tif err := w.Flush(); err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to flush writer: %v\", err)\n\t}\n\n\treturn buf.String(), nil\n}", "func SerializeDoubleSummaryValue(min, max, sum float64, count int64) string {\n\treturn fmt.Sprintf(\"gauge,min=%s,max=%s,sum=%s,count=%d\", serializeFloat64(min), serializeFloat64(max), serializeFloat64(sum), count)\n}", "func printJson(ag *alertGroup, m *sync.Mutex) {\n\tm.Lock()\n\tfor _, alert := range ag.Alerts {\n\t\tout := map[string]string{\"status\": alert.Status}\n\n\t\tfor k, v := range alert.Labels {\n\t\t\tout[k] = v\n\t\t}\n\t\tfor k, v := range alert.Annotations {\n\t\t\tout[k] = v\n\t\t}\n\t\tout[\"startsAt\"] = alert.StartsAt.Truncate(time.Millisecond).String()\n\t\tout[\"endsAt\"] = alert.EndsAt.Truncate(time.Millisecond).String()\n\n\t\tjout, err := json.Marshal(out)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t\tfmt.Printf(\"%s\\n\", jout)\n\t}\n\tm.Unlock()\n}", "func (m *ApplicationSignInDetailedSummary) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.Entity.Serialize(writer)\n if err != nil {\n return err\n }\n {\n err = writer.WriteTimeValue(\"aggregatedEventDateTime\", m.GetAggregatedEventDateTime())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"appDisplayName\", m.GetAppDisplayName())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"appId\", m.GetAppId())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteInt64Value(\"signInCount\", m.GetSignInCount())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteObjectValue(\"status\", m.GetStatus())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func GetTestSummary(w http.ResponseWriter, r *http.Request) {\n\tlog.Printf(\"GetTestSummary\\n\")\n\n\t// Get rid of warning\n\t_ = r\n\n\t// Marshal array of struct\n\tcurrentTestSummariesJson, err := json.MarshalIndent(currentTestSummary, \"\", \" \")\n\tif err != nil {\n\t\thttp.Error(w, fmt.Sprintf(\"failed to write marshal currentTestEvaluation: %s\", err),\n\t\t\thttp.StatusInternalServerError)\n\t}\n\n\t_, err = fmt.Fprintf(w, string(currentTestSummariesJson))\n\tif err != nil {\n\t\thttp.Error(w, fmt.Sprintf(\"failed to write response: %s\", err),\n\t\t\thttp.StatusInternalServerError)\n\t}\n}", "func (si StorageInsight) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tif si.StorageInsightProperties != nil {\n\t\tobjectMap[\"properties\"] = si.StorageInsightProperties\n\t}\n\tif si.ETag != nil {\n\t\tobjectMap[\"eTag\"] = si.ETag\n\t}\n\tif si.Tags != nil {\n\t\tobjectMap[\"tags\"] = si.Tags\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (o MicrosoftGraphItemActionStat) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.ActionCount == nil {\n\t\tif o.isExplicitNullActionCount {\n\t\t\ttoSerialize[\"actionCount\"] = o.ActionCount\n\t\t}\n\t} else {\n\t\ttoSerialize[\"actionCount\"] = o.ActionCount\n\t}\n\tif o.ActorCount == nil {\n\t\tif o.isExplicitNullActorCount {\n\t\t\ttoSerialize[\"actorCount\"] = o.ActorCount\n\t\t}\n\t} else {\n\t\ttoSerialize[\"actorCount\"] = o.ActorCount\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (v VlanMetric) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson51bca34dEncodeGithubComSkydiveProjectSkydiveSflow(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}" ]
[ "0.63141054", "0.6245847", "0.5888697", "0.5545922", "0.5509937", "0.54198235", "0.5396004", "0.5379865", "0.53409183", "0.5336062", "0.5316955", "0.53086776", "0.52871865", "0.5250192", "0.5250104", "0.52132285", "0.5193233", "0.51777637", "0.51521647", "0.515026", "0.510514", "0.5084872", "0.5081824", "0.5078855", "0.5058627", "0.5051174", "0.5048088", "0.5032795", "0.5029943", "0.50247216", "0.5024125", "0.50202197", "0.50033665", "0.5001131", "0.499699", "0.4986647", "0.4982144", "0.49819827", "0.49797326", "0.4970961", "0.49696228", "0.49675837", "0.49619493", "0.49618718", "0.4959515", "0.4955355", "0.4954352", "0.49474588", "0.49451378", "0.49356827", "0.49326804", "0.49294183", "0.49145004", "0.49049467", "0.48909262", "0.48882", "0.48806608", "0.48777273", "0.48723945", "0.48713502", "0.4857923", "0.4857014", "0.48494697", "0.4847801", "0.48478", "0.48442844", "0.4839303", "0.483144", "0.48305675", "0.48264354", "0.48226756", "0.48218766", "0.4821076", "0.481946", "0.48151848", "0.4789827", "0.47895017", "0.4787201", "0.4785496", "0.47851267", "0.47666997", "0.47630468", "0.47510582", "0.47396144", "0.47357282", "0.4732841", "0.47196177", "0.471726", "0.47092918", "0.47083855", "0.4705959", "0.47030967", "0.47008625", "0.4685235", "0.46845227", "0.46759585", "0.46752596", "0.46717733", "0.46716455", "0.46643785" ]
0.69818693
0
UnmarshalJSON unmarshals a view properties and other data.
func (s *SummaryChart) UnmarshalJSON(b []byte) error { type alias SummaryChart a := (*alias)(s) if err := json.Unmarshal(b, a); err != nil { return err } s.XPosition = a.XPosition s.XPosition = a.YPosition s.Height = a.Height s.Width = a.Width vp, err := influxdb.UnmarshalViewPropertiesJSON(b) if err != nil { return err } s.Properties = vp return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (v *ViewProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"accumulated\":\n\t\t\terr = unpopulate(val, &v.Accumulated)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"chart\":\n\t\t\terr = unpopulate(val, &v.Chart)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"createdOn\":\n\t\t\terr = unpopulateTimeRFC3339(val, &v.CreatedOn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"currency\":\n\t\t\terr = unpopulate(val, &v.Currency)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dateRange\":\n\t\t\terr = unpopulate(val, &v.DateRange)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, &v.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"kpis\":\n\t\t\terr = unpopulate(val, &v.Kpis)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"metric\":\n\t\t\terr = unpopulate(val, &v.Metric)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"modifiedOn\":\n\t\t\terr = unpopulateTimeRFC3339(val, &v.ModifiedOn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"pivots\":\n\t\t\terr = unpopulate(val, &v.Pivots)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"query\":\n\t\t\terr = unpopulate(val, &v.Query)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"scope\":\n\t\t\terr = unpopulate(val, &v.Scope)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (v *PostViews) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson783c1624DecodeGithubComGobwasVk1(&r, v)\n\treturn r.Error()\n}", "func (v *PostViews) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson783c1624DecodeGithubComGobwasVk1(l, v)\n}", "func (v *Visit) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonE564fc13DecodeGithubComLa0rgHighloadcupModel1(&r, v)\n\treturn r.Error()\n}", "func (v *VM) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &v.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &v.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &v.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &v.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &v.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TableResponseProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"odata.editLink\":\n\t\t\terr = unpopulate(val, \"ODataEditLink\", &t.ODataEditLink)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"odata.id\":\n\t\t\terr = unpopulate(val, \"ODataID\", &t.ODataID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &t.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"TableName\":\n\t\t\terr = unpopulate(val, \"TableName\", &t.TableName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SecurityGroupViewParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"targetResourceId\":\n\t\t\terr = unpopulate(val, \"TargetResourceID\", &s.TargetResourceID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TestLineProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"phoneNumber\":\n\t\t\terr = unpopulate(val, \"PhoneNumber\", &t.PhoneNumber)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &t.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"purpose\":\n\t\t\terr = unpopulate(val, \"Purpose\", &t.Purpose)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (d *DashboardPartMetadata) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &d.Type)\n\t\t\tdelete(rawMsg, key)\n\t\tdefault:\n\t\t\tif d.AdditionalProperties == nil {\n\t\t\t\td.AdditionalProperties = map[string]interface{}{}\n\t\t\t}\n\t\t\tif val != nil {\n\t\t\t\tvar aux interface{}\n\t\t\t\terr = json.Unmarshal(val, &aux)\n\t\t\t\td.AdditionalProperties[key] = aux\n\t\t\t}\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingTwin) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &r.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &r.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Viewport) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeGithubComChromedpCdprotoPage2(&r, v)\n\treturn r.Error()\n}", "func (v *LayoutViewport) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeGithubComChromedpCdprotoPage34(&r, v)\n\treturn r.Error()\n}", "func (t *TransformProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"created\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"Created\", &t.Created)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"description\":\n\t\t\terr = unpopulate(val, \"Description\", &t.Description)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastModified\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"LastModified\", &t.LastModified)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"outputs\":\n\t\t\terr = unpopulate(val, \"Outputs\", &t.Outputs)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VisualViewport) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeGithubComChromedpCdprotoPage1(&r, v)\n\treturn r.Error()\n}", "func (o *PatientProfileViewOKBody) UnmarshalJSON(raw []byte) error {\n\t// PatientProfileViewOKBodyAO0\n\tvar patientProfileViewOKBodyAO0 models.SuccessData\n\tif err := swag.ReadJSON(raw, &patientProfileViewOKBodyAO0); err != nil {\n\t\treturn err\n\t}\n\to.SuccessData = patientProfileViewOKBodyAO0\n\n\t// PatientProfileViewOKBodyAO1\n\tvar dataPatientProfileViewOKBodyAO1 struct {\n\t\tData []*ProfileDataItem `json:\"data\"`\n\t}\n\tif err := swag.ReadJSON(raw, &dataPatientProfileViewOKBodyAO1); err != nil {\n\t\treturn err\n\t}\n\n\to.Data = dataPatientProfileViewOKBodyAO1.Data\n\n\treturn nil\n}", "func (v *TaskDetail) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeCommonSystoolbox4(&r, v)\n\treturn r.Error()\n}", "func (s *SecurityGroupViewResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"networkInterfaces\":\n\t\t\terr = unpopulate(val, \"NetworkInterfaces\", &s.NetworkInterfaces)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Metadata) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonBa0ee0e3DecodeGithubComSkydiveProjectSkydiveTopologyProbesLxd(&r, v)\n\treturn r.Error()\n}", "func (v *VirtualHubRouteTableV2Properties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"attachedConnections\":\n\t\t\terr = unpopulate(val, \"AttachedConnections\", &v.AttachedConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"routes\":\n\t\t\terr = unpopulate(val, \"Routes\", &v.Routes)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func unmarshalFounderResponseBodyToWarehouseviewsFounderView(v *FounderResponseBody) *warehouseviews.FounderView {\n\tres := &warehouseviews.FounderView{\n\t\tID: v.ID,\n\t\tName: v.Name,\n\t}\n\n\treturn res\n}", "func (a *Meta_Up) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]interface{})\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal interface{}\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, fmt.Sprintf(\"error unmarshaling field %s\", fieldName))\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingTwinProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"desired\":\n\t\t\terr = unpopulate(val, \"Desired\", &r.Desired)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reported\":\n\t\t\terr = unpopulate(val, \"Reported\", &r.Reported)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Part) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer12(&r, v)\n\treturn r.Error()\n}", "func (v *VirtualNetworkTapPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"destinationLoadBalancerFrontEndIPConfiguration\":\n\t\t\terr = unpopulate(val, \"DestinationLoadBalancerFrontEndIPConfiguration\", &v.DestinationLoadBalancerFrontEndIPConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"destinationNetworkInterfaceIPConfiguration\":\n\t\t\terr = unpopulate(val, \"DestinationNetworkInterfaceIPConfiguration\", &v.DestinationNetworkInterfaceIPConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"destinationPort\":\n\t\t\terr = unpopulate(val, \"DestinationPort\", &v.DestinationPort)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"networkInterfaceTapConfigurations\":\n\t\t\terr = unpopulate(val, \"NetworkInterfaceTapConfigurations\", &v.NetworkInterfaceTapConfigurations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceGuid\":\n\t\t\terr = unpopulate(val, \"ResourceGUID\", &v.ResourceGUID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *Transform) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &t.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &t.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &t.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &t.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &t.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VPNSiteLinkProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"bgpProperties\":\n\t\t\terr = unpopulate(val, \"BgpProperties\", &v.BgpProperties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"fqdn\":\n\t\t\terr = unpopulate(val, \"Fqdn\", &v.Fqdn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipAddress\":\n\t\t\terr = unpopulate(val, \"IPAddress\", &v.IPAddress)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"linkProperties\":\n\t\t\terr = unpopulate(val, \"LinkProperties\", &v.LinkProperties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TransformCollection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"@odata.nextLink\":\n\t\t\terr = unpopulate(val, \"ODataNextLink\", &t.ODataNextLink)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &t.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *ExportProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"definition\":\n\t\t\terr = unpopulate(val, &e.Definition)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deliveryInfo\":\n\t\t\terr = unpopulate(val, &e.DeliveryInfo)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"format\":\n\t\t\terr = unpopulate(val, &e.Format)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nextRunTimeEstimate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &e.NextRunTimeEstimate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"partitionData\":\n\t\t\terr = unpopulate(val, &e.PartitionData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"runHistory\":\n\t\t\terr = unpopulate(val, &e.RunHistory)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"schedule\":\n\t\t\terr = unpopulate(val, &e.Schedule)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Post) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson5a72dc82DecodeGithubComTimRazumovTechnoparkDBAppModels6(&r, v)\n\treturn r.Error()\n}", "func (v *VirtualApplianceSKUProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"bundledScaleUnit\":\n\t\t\terr = unpopulate(val, \"BundledScaleUnit\", &v.BundledScaleUnit)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"marketPlaceVersion\":\n\t\t\terr = unpopulate(val, \"MarketPlaceVersion\", &v.MarketPlaceVersion)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"vendor\":\n\t\t\terr = unpopulate(val, \"Vendor\", &v.Vendor)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (dst *ViewFieldOptions) UnmarshalJSON(data []byte) error {\n\tvar err error\n\t// try to unmarshal JSON data into CalendarViewFieldOptions\n\terr = json.Unmarshal(data, &dst.CalendarViewFieldOptions);\n\tif err == nil {\n\t\tjsonCalendarViewFieldOptions, _ := json.Marshal(dst.CalendarViewFieldOptions)\n\t\tif string(jsonCalendarViewFieldOptions) == \"{}\" { // empty struct\n\t\t\tdst.CalendarViewFieldOptions = nil\n\t\t} else {\n\t\t\treturn nil // data stored in dst.CalendarViewFieldOptions, return on the first match\n\t\t}\n\t} else {\n\t\tdst.CalendarViewFieldOptions = nil\n\t}\n\n\t// try to unmarshal JSON data into FormViewFieldOptions\n\terr = json.Unmarshal(data, &dst.FormViewFieldOptions);\n\tif err == nil {\n\t\tjsonFormViewFieldOptions, _ := json.Marshal(dst.FormViewFieldOptions)\n\t\tif string(jsonFormViewFieldOptions) == \"{}\" { // empty struct\n\t\t\tdst.FormViewFieldOptions = nil\n\t\t} else {\n\t\t\treturn nil // data stored in dst.FormViewFieldOptions, return on the first match\n\t\t}\n\t} else {\n\t\tdst.FormViewFieldOptions = nil\n\t}\n\n\t// try to unmarshal JSON data into GalleryViewFieldOptions\n\terr = json.Unmarshal(data, &dst.GalleryViewFieldOptions);\n\tif err == nil {\n\t\tjsonGalleryViewFieldOptions, _ := json.Marshal(dst.GalleryViewFieldOptions)\n\t\tif string(jsonGalleryViewFieldOptions) == \"{}\" { // empty struct\n\t\t\tdst.GalleryViewFieldOptions = nil\n\t\t} else {\n\t\t\treturn nil // data stored in dst.GalleryViewFieldOptions, return on the first match\n\t\t}\n\t} else {\n\t\tdst.GalleryViewFieldOptions = nil\n\t}\n\n\t// try to unmarshal JSON data into GridViewFieldOptions\n\terr = json.Unmarshal(data, &dst.GridViewFieldOptions);\n\tif err == nil {\n\t\tjsonGridViewFieldOptions, _ := json.Marshal(dst.GridViewFieldOptions)\n\t\tif string(jsonGridViewFieldOptions) == \"{}\" { // empty struct\n\t\t\tdst.GridViewFieldOptions = nil\n\t\t} else {\n\t\t\treturn nil // data stored in dst.GridViewFieldOptions, return on the first match\n\t\t}\n\t} else {\n\t\tdst.GridViewFieldOptions = nil\n\t}\n\n\t// try to unmarshal JSON data into KanbanViewFieldOptions\n\terr = json.Unmarshal(data, &dst.KanbanViewFieldOptions);\n\tif err == nil {\n\t\tjsonKanbanViewFieldOptions, _ := json.Marshal(dst.KanbanViewFieldOptions)\n\t\tif string(jsonKanbanViewFieldOptions) == \"{}\" { // empty struct\n\t\t\tdst.KanbanViewFieldOptions = nil\n\t\t} else {\n\t\t\treturn nil // data stored in dst.KanbanViewFieldOptions, return on the first match\n\t\t}\n\t} else {\n\t\tdst.KanbanViewFieldOptions = nil\n\t}\n\n\treturn fmt.Errorf(\"data failed to match schemas in anyOf(ViewFieldOptions)\")\n}", "func (v *ResponceUserVisit) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson15d499e9DecodeGithubComElBroomHighloadCupAppSchema1(&r, v)\n\treturn r.Error()\n}", "func (v *VisitArray) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonE564fc13DecodeGithubComLa0rgHighloadcupModel(&r, v)\n\treturn r.Error()\n}", "func (o *PatientProfileViewNotFoundBody) UnmarshalJSON(raw []byte) error {\n\t// PatientProfileViewNotFoundBodyAO0\n\tvar patientProfileViewNotFoundBodyAO0 models.Error404Data\n\tif err := swag.ReadJSON(raw, &patientProfileViewNotFoundBodyAO0); err != nil {\n\t\treturn err\n\t}\n\to.Error404Data = patientProfileViewNotFoundBodyAO0\n\n\treturn nil\n}", "func (v *VirtualApplianceSiteProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"addressPrefix\":\n\t\t\terr = unpopulate(val, \"AddressPrefix\", &v.AddressPrefix)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"o365Policy\":\n\t\t\terr = unpopulate(val, \"O365Policy\", &v.O365Policy)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TrafficAnalyticsProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"networkWatcherFlowAnalyticsConfiguration\":\n\t\t\terr = unpopulate(val, \"NetworkWatcherFlowAnalyticsConfiguration\", &t.NetworkWatcherFlowAnalyticsConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *PostGet) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson5a72dc82DecodeGithubComTimRazumovTechnoparkDBAppModels1(&r, v)\n\treturn r.Error()\n}", "func (v *VirtualRouter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &v.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &v.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &v.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &v.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &v.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &v.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &v.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VPNSite) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &v.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &v.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &v.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &v.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &v.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &v.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &v.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (o *ExportDataPartial) UnmarshalJSON(data []byte) error {\n\tkv := make(map[string]interface{})\n\tif err := json.Unmarshal(data, &kv); err != nil {\n\t\treturn err\n\t}\n\to.FromMap(kv)\n\treturn nil\n}", "func (v *VirtualHubRouteTable) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"routes\":\n\t\t\terr = unpopulate(val, \"Routes\", &v.Routes)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (l *LiveEventPreview) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"accessControl\":\n\t\t\terr = unpopulate(val, \"AccessControl\", &l.AccessControl)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"alternativeMediaId\":\n\t\t\terr = unpopulate(val, \"AlternativeMediaID\", &l.AlternativeMediaID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpoints\":\n\t\t\terr = unpopulate(val, \"Endpoints\", &l.Endpoints)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"previewLocator\":\n\t\t\terr = unpopulate(val, \"PreviewLocator\", &l.PreviewLocator)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"streamingPolicyName\":\n\t\t\terr = unpopulate(val, \"StreamingPolicyName\", &l.StreamingPolicyName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TestRouteResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"details\":\n\t\t\terr = unpopulate(val, \"Details\", &t.Details)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"result\":\n\t\t\terr = unpopulate(val, \"Result\", &t.Result)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *AppManifestParsedProperties) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeGithubComChromedpCdprotoPage103(&r, v)\n\treturn r.Error()\n}", "func (h *HeatMapProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, &h.EndTime)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpoints\":\n\t\t\terr = unpopulate(val, &h.Endpoints)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"startTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, &h.StartTime)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"trafficFlows\":\n\t\t\terr = unpopulate(val, &h.TrafficFlows)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (v *TaskDetailItem) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeCommonSystoolbox3(&r, v)\n\treturn r.Error()\n}", "func (c *CopyVideo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"label\":\n\t\t\terr = unpopulate(val, \"Label\", &c.Label)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &c.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (l *LocationPropertiesExchange) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"peeringFacilities\":\n\t\t\terr = unpopulate(val, \"PeeringFacilities\", &l.PeeringFacilities)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (o *OpenShiftVersionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", o, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"version\":\n\t\t\terr = unpopulate(val, \"Version\", &o.Version)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", o, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateLinkServicePropertiesVisibility) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"subscriptions\":\n\t\t\terr = unpopulate(val, \"Subscriptions\", &p.Subscriptions)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (o *ClientPermissionViewOKBody) UnmarshalJSON(raw []byte) error {\n\t// ClientPermissionViewOKBodyAO0\n\tvar clientPermissionViewOKBodyAO0 models.SuccessData\n\tif err := swag.ReadJSON(raw, &clientPermissionViewOKBodyAO0); err != nil {\n\t\treturn err\n\t}\n\to.SuccessData = clientPermissionViewOKBodyAO0\n\n\t// ClientPermissionViewOKBodyAO1\n\tvar dataClientPermissionViewOKBodyAO1 struct {\n\t\tData []*DataItems0 `json:\"data\"`\n\t}\n\tif err := swag.ReadJSON(raw, &dataClientPermissionViewOKBodyAO1); err != nil {\n\t\treturn err\n\t}\n\n\to.Data = dataClientPermissionViewOKBodyAO1.Data\n\n\treturn nil\n}", "func (v *VideoTrack) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &v.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TestRouteResultDetails) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"compilationErrors\":\n\t\t\terr = unpopulate(val, \"CompilationErrors\", &t.CompilationErrors)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Info) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC80ae7adDecodeGithubComDeiklovTechDbRomanovAndrGolangModels13(&r, v)\n\treturn r.Error()\n}", "func (a *AttestationProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"comments\":\n\t\t\terr = unpopulate(val, \"Comments\", &a.Comments)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"complianceState\":\n\t\t\terr = unpopulate(val, \"ComplianceState\", &a.ComplianceState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"evidence\":\n\t\t\terr = unpopulate(val, \"Evidence\", &a.Evidence)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"expiresOn\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"ExpiresOn\", &a.ExpiresOn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastComplianceStateChangeAt\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"LastComplianceStateChangeAt\", &a.LastComplianceStateChangeAt)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"owner\":\n\t\t\terr = unpopulate(val, \"Owner\", &a.Owner)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"policyAssignmentId\":\n\t\t\terr = unpopulate(val, \"PolicyAssignmentID\", &a.PolicyAssignmentID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"policyDefinitionReferenceId\":\n\t\t\terr = unpopulate(val, \"PolicyDefinitionReferenceID\", &a.PolicyDefinitionReferenceID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &a.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TableProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"TableName\":\n\t\t\terr = unpopulate(val, \"TableName\", &t.TableName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VirtualHubRouteTableV2) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &v.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &v.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &v.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &v.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VirtualNetworkTap) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &v.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &v.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &v.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &v.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &v.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &v.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &v.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *MatchedRoute) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &m.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FeatureProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"state\":\n\t\t\terr = unpopulate(val, \"State\", &f.State)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *PostGets) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson5a72dc82DecodeGithubComTimRazumovTechnoparkDBAppModels(&r, v)\n\treturn r.Error()\n}", "func (h *HubRouteTableProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"associatedConnections\":\n\t\t\terr = unpopulate(val, \"AssociatedConnections\", &h.AssociatedConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"labels\":\n\t\t\terr = unpopulate(val, \"Labels\", &h.Labels)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"propagatingConnections\":\n\t\t\terr = unpopulate(val, \"PropagatingConnections\", &h.PropagatingConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &h.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"routes\":\n\t\t\terr = unpopulate(val, \"Routes\", &h.Routes)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Topic) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer5(&r, v)\n\treturn r.Error()\n}", "func (v *EventPlayerPropertiesChanged) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeGithubComChromedpCdprotoMedia3(&r, v)\n\treturn r.Error()\n}", "func (a *TemplateApply_EnvRefs) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]interface{})\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal interface{}\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, fmt.Sprintf(\"error unmarshaling field %s\", fieldName))\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RelayNamespaceAccessProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"accessKey\":\n\t\t\terr = unpopulate(val, \"AccessKey\", &r.AccessKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"expiresOn\":\n\t\t\terr = unpopulate(val, \"ExpiresOn\", &r.ExpiresOn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"hybridConnectionName\":\n\t\t\terr = unpopulate(val, \"HybridConnectionName\", &r.HybridConnectionName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"namespaceName\":\n\t\t\terr = unpopulate(val, \"NamespaceName\", &r.NamespaceName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"namespaceNameSuffix\":\n\t\t\terr = unpopulate(val, \"NamespaceNameSuffix\", &r.NamespaceNameSuffix)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VirtualHubRoute) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"addressPrefixes\":\n\t\t\terr = unpopulate(val, \"AddressPrefixes\", &v.AddressPrefixes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nextHopIpAddress\":\n\t\t\terr = unpopulate(val, \"NextHopIPAddress\", &v.NextHopIPAddress)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *Route) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &r.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &r.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &r.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RouteTable) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &r.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &r.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &r.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &r.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &r.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CommonExportProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"definition\":\n\t\t\terr = unpopulate(val, &c.Definition)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deliveryInfo\":\n\t\t\terr = unpopulate(val, &c.DeliveryInfo)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"format\":\n\t\t\terr = unpopulate(val, &c.Format)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nextRunTimeEstimate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &c.NextRunTimeEstimate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"partitionData\":\n\t\t\terr = unpopulate(val, &c.PartitionData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"runHistory\":\n\t\t\terr = unpopulate(val, &c.RunHistory)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RouteMap) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &r.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &r.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &r.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VnetRoute) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"bgpConnections\":\n\t\t\terr = unpopulate(val, \"BgpConnections\", &v.BgpConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"staticRoutes\":\n\t\t\terr = unpopulate(val, \"StaticRoutes\", &v.StaticRoutes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"staticRoutesConfig\":\n\t\t\terr = unpopulate(val, \"StaticRoutesConfig\", &v.StaticRoutesConfig)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func ChartViewFromJSON(jsonView *view.ChartView) (*render.ChartView, error) {\n\tif jsonView == nil {\n\t\treturn nil, ErrViewIsEmpty\n\t}\n\n\tvar (\n\t\tbarLabelVisible bool\n\t\tpointVisible bool\n\t\tpointLabelVisible bool\n\t)\n\n\tif jsonView.BarLabelVisible == nil {\n\t\tbarLabelVisible = barLabelVisibleDefault\n\t} else {\n\t\tbarLabelVisible = *jsonView.BarLabelVisible\n\t}\n\n\tif jsonView.PointVisible == nil {\n\t\tpointVisible = pointVisibleDefault\n\t} else {\n\t\tpointVisible = *jsonView.PointVisible\n\t}\n\n\tif jsonView.PointLabelVisible == nil {\n\t\tpointLabelVisible = pointLabelVisibleDefault\n\t} else {\n\t\tpointLabelVisible = *jsonView.PointLabelVisible\n\t}\n\n\tcolors, err := viewColorsFromJSON(jsonView.Colors)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresult := &render.ChartView{\n\t\tKind: viewKindFromJSON(jsonView.Kind),\n\t\tValues: nil,\n\t\tColors: colors,\n\t\tBarLabelVisible: &wrapperspb.BoolValue{Value: barLabelVisible},\n\t\tBarLabelPosition: barLabelPositionFromJSON(jsonView.BarLabelPosition),\n\t\tPointVisible: &wrapperspb.BoolValue{Value: pointVisible},\n\t\tPointType: pointTypeFromJSON(jsonView.PointType),\n\t\tPointLabelVisible: &wrapperspb.BoolValue{Value: pointLabelVisible},\n\t\tPointLabelPosition: pointLabelPositionFromJSON(jsonView.PointLabelPosition),\n\t}\n\n\tvaluesAreSet := false\n\n\tif jsonView.BarsValues != nil {\n\t\tbarsValues, err := barsValuesFromJSON(jsonView)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tresult.Values = &render.ChartView_BarsValues{\n\t\t\tBarsValues: barsValues,\n\t\t}\n\n\t\tvaluesAreSet = true\n\t}\n\n\tif jsonView.PointsValues != nil {\n\t\tif valuesAreSet {\n\t\t\treturn nil, ErrOnlyOneOfValuesKindShouldBeSpecified\n\t\t}\n\n\t\tpointsValues, err := pointValuesFromJSON(jsonView)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tresult.Values = &render.ChartView_PointsValues{\n\t\t\tPointsValues: pointsValues,\n\t\t}\n\n\t\tvaluesAreSet = true\n\t}\n\n\tif jsonView.ScalarValues != nil {\n\t\tif valuesAreSet {\n\t\t\treturn nil, ErrOnlyOneOfValuesKindShouldBeSpecified\n\t\t}\n\n\t\tresult.Values = &render.ChartView_ScalarValues{\n\t\t\tScalarValues: scalarValuesFromJSON(jsonView),\n\t\t}\n\n\t\tvaluesAreSet = true\n\t}\n\n\tif !valuesAreSet {\n\t\treturn nil, ErrValuesShouldBeSpecified\n\t}\n\n\treturn result, nil\n}", "func (r *RouteMapProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"associatedInboundConnections\":\n\t\t\terr = unpopulate(val, \"AssociatedInboundConnections\", &r.AssociatedInboundConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"associatedOutboundConnections\":\n\t\t\terr = unpopulate(val, \"AssociatedOutboundConnections\", &r.AssociatedOutboundConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &r.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"rules\":\n\t\t\terr = unpopulate(val, \"Rules\", &r.Rules)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VirtualRouterPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"hostedGateway\":\n\t\t\terr = unpopulate(val, \"HostedGateway\", &v.HostedGateway)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"hostedSubnet\":\n\t\t\terr = unpopulate(val, \"HostedSubnet\", &v.HostedSubnet)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"peerings\":\n\t\t\terr = unpopulate(val, \"Peerings\", &v.Peerings)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualRouterAsn\":\n\t\t\terr = unpopulate(val, \"VirtualRouterAsn\", &v.VirtualRouterAsn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualRouterIps\":\n\t\t\terr = unpopulate(val, \"VirtualRouterIPs\", &v.VirtualRouterIPs)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *TemplateApply_Secrets) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]string)\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal string\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, fmt.Sprintf(\"error unmarshaling field %s\", fieldName))\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (v *ShadowModelSt) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonB7ed31d3DecodeMevericcoreMccommon5(&r, v)\n\treturn r.Error()\n}", "func (v *Item) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecodeDrhyuComIndexerModels2(&r, v)\n\treturn r.Error()\n}", "func (u *UserProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"createdDate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &u.CreatedDate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"identity\":\n\t\t\terr = unpopulate(val, &u.Identity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, &u.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"secretStore\":\n\t\t\terr = unpopulate(val, &u.SecretStore)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"uniqueIdentifier\":\n\t\t\terr = unpopulate(val, &u.UniqueIdentifier)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VPNSiteLink) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &v.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &v.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &v.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &v.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &v.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (l *LocationProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"azureRegion\":\n\t\t\terr = unpopulate(val, \"AzureRegion\", &l.AzureRegion)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"country\":\n\t\t\terr = unpopulate(val, \"Country\", &l.Country)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"direct\":\n\t\t\terr = unpopulate(val, \"Direct\", &l.Direct)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"exchange\":\n\t\t\terr = unpopulate(val, \"Exchange\", &l.Exchange)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"peeringLocation\":\n\t\t\terr = unpopulate(val, \"PeeringLocation\", &l.PeeringLocation)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (l *LiveEventProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"created\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"Created\", &l.Created)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"crossSiteAccessPolicies\":\n\t\t\terr = unpopulate(val, \"CrossSiteAccessPolicies\", &l.CrossSiteAccessPolicies)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"description\":\n\t\t\terr = unpopulate(val, \"Description\", &l.Description)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"encoding\":\n\t\t\terr = unpopulate(val, \"Encoding\", &l.Encoding)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"hostnamePrefix\":\n\t\t\terr = unpopulate(val, \"HostnamePrefix\", &l.HostnamePrefix)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"input\":\n\t\t\terr = unpopulate(val, \"Input\", &l.Input)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastModified\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"LastModified\", &l.LastModified)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"preview\":\n\t\t\terr = unpopulate(val, \"Preview\", &l.Preview)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &l.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceState\":\n\t\t\terr = unpopulate(val, \"ResourceState\", &l.ResourceState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"streamOptions\":\n\t\t\terr = unpopulate(val, \"StreamOptions\", &l.StreamOptions)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"transcriptions\":\n\t\t\terr = unpopulate(val, \"Transcriptions\", &l.Transcriptions)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"useStaticHostname\":\n\t\t\terr = unpopulate(val, \"UseStaticHostname\", &l.UseStaticHostname)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *TimeEntryActivity) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson44273644DecodeGithubComSomniSomGoRedmine1(&r, v)\n\treturn r.Error()\n}", "func (o *PatientProfileViewInternalServerErrorBody) UnmarshalJSON(raw []byte) error {\n\t// PatientProfileViewInternalServerErrorBodyAO0\n\tvar patientProfileViewInternalServerErrorBodyAO0 models.Error500Data\n\tif err := swag.ReadJSON(raw, &patientProfileViewInternalServerErrorBodyAO0); err != nil {\n\t\treturn err\n\t}\n\to.Error500Data = patientProfileViewInternalServerErrorBodyAO0\n\n\treturn nil\n}", "func (a *LabelUpdate_Properties) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]string)\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal string\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, fmt.Sprintf(\"error unmarshaling field %s\", fieldName))\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (mVar *Model) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"sku\":\n\t\t\tif v != nil {\n\t\t\t\tvar sku Sku\n\t\t\t\terr = json.Unmarshal(*v, &sku)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmVar.Sku = &sku\n\t\t\t}\n\t\tcase \"kind\":\n\t\t\tif v != nil {\n\t\t\t\tvar kind Kind\n\t\t\t\terr = json.Unmarshal(*v, &kind)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmVar.Kind = kind\n\t\t\t}\n\t\tcase \"properties\":\n\t\t\tif v != nil {\n\t\t\t\tvar properties Properties\n\t\t\t\terr = json.Unmarshal(*v, &properties)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmVar.Properties = &properties\n\t\t\t}\n\t\tcase \"location\":\n\t\t\tif v != nil {\n\t\t\t\tvar location string\n\t\t\t\terr = json.Unmarshal(*v, &location)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmVar.Location = &location\n\t\t\t}\n\t\tcase \"tags\":\n\t\t\tif v != nil {\n\t\t\t\tvar tags map[string]*string\n\t\t\t\terr = json.Unmarshal(*v, &tags)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmVar.Tags = tags\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmVar.Name = &name\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmVar.ID = &ID\n\t\t\t}\n\t\tcase \"type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar string\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmVar.Type = &typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (r *RouteProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"condition\":\n\t\t\terr = unpopulate(val, \"Condition\", &r.Condition)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpointNames\":\n\t\t\terr = unpopulate(val, \"EndpointNames\", &r.EndpointNames)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isEnabled\":\n\t\t\terr = unpopulate(val, \"IsEnabled\", &r.IsEnabled)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"source\":\n\t\t\terr = unpopulate(val, \"Source\", &r.Source)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func unmarshalWineryResponseBodyToStorageviewsWineryView(v *WineryResponseBody) *storageviews.WineryView {\n\tres := &storageviews.WineryView{\n\t\tName: v.Name,\n\t\tRegion: v.Region,\n\t\tCountry: v.Country,\n\t\tURL: v.URL,\n\t}\n\n\treturn res\n}", "func (a *Meta_UsState) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]interface{})\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal interface{}\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, fmt.Sprintf(\"error unmarshaling field %s\", fieldName))\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VPNSiteProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"addressSpace\":\n\t\t\terr = unpopulate(val, \"AddressSpace\", &v.AddressSpace)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"bgpProperties\":\n\t\t\terr = unpopulate(val, \"BgpProperties\", &v.BgpProperties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deviceProperties\":\n\t\t\terr = unpopulate(val, \"DeviceProperties\", &v.DeviceProperties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipAddress\":\n\t\t\terr = unpopulate(val, \"IPAddress\", &v.IPAddress)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isSecuritySite\":\n\t\t\terr = unpopulate(val, \"IsSecuritySite\", &v.IsSecuritySite)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"o365Policy\":\n\t\t\terr = unpopulate(val, \"O365Policy\", &v.O365Policy)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"siteKey\":\n\t\t\terr = unpopulate(val, \"SiteKey\", &v.SiteKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"vpnSiteLinks\":\n\t\t\terr = unpopulate(val, \"VPNSiteLinks\", &v.VPNSiteLinks)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualWan\":\n\t\t\terr = unpopulate(val, \"VirtualWan\", &v.VirtualWan)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Model) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson82a45abeDecodeGithubComOvhCdsSdk(&r, v)\n\treturn r.Error()\n}", "func (v *Msg) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecodeBackendInternalModels6(&r, v)\n\treturn r.Error()\n}", "func (o *ProjectWebhookPartial) UnmarshalJSON(data []byte) error {\n\tkv := make(map[string]interface{})\n\tif err := json.Unmarshal(data, &kv); err != nil {\n\t\treturn err\n\t}\n\to.FromMap(kv)\n\treturn nil\n}", "func (v *EventInterstitialShown) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeGithubComChromedpCdprotoPage74(&r, v)\n\treturn r.Error()\n}", "func (v *VotesInfo) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC80ae7adDecodeGithubComDeiklovTechDbRomanovAndrGolangModels(&r, v)\n\treturn r.Error()\n}", "func (v *VirtualAppliance) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &v.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &v.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"identity\":\n\t\t\terr = unpopulate(val, \"Identity\", &v.Identity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &v.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &v.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &v.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &v.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &v.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RouteTablePropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"disableBgpRoutePropagation\":\n\t\t\terr = unpopulate(val, \"DisableBgpRoutePropagation\", &r.DisableBgpRoutePropagation)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &r.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceGuid\":\n\t\t\terr = unpopulate(val, \"ResourceGUID\", &r.ResourceGUID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"routes\":\n\t\t\terr = unpopulate(val, \"Routes\", &r.Routes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subnets\":\n\t\t\terr = unpopulate(val, \"Subnets\", &r.Subnets)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *ArmTemplateProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"contents\":\n\t\t\terr = unpopulate(val, &a.Contents)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"createdDate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &a.CreatedDate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"description\":\n\t\t\terr = unpopulate(val, &a.Description)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, &a.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enabled\":\n\t\t\terr = unpopulate(val, &a.Enabled)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"icon\":\n\t\t\terr = unpopulate(val, &a.Icon)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"parametersValueFilesInfo\":\n\t\t\terr = unpopulate(val, &a.ParametersValueFilesInfo)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"publisher\":\n\t\t\terr = unpopulate(val, &a.Publisher)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalJSON(body io.Reader, v interface{}) error {\n\tdecoder := json.NewDecoder(body)\n\treturn decoder.Decode(v)\n}", "func (v *Vote) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecodeGithubComMailcoursesTechnoparkDbmsForumGeneratedModels(&r, v)\n\treturn r.Error()\n}" ]
[ "0.7521359", "0.69134307", "0.6182066", "0.60080624", "0.5860861", "0.56737185", "0.56453896", "0.5628084", "0.5626881", "0.5614671", "0.5613706", "0.56116426", "0.5608741", "0.5591743", "0.5576611", "0.5567037", "0.55564475", "0.5552273", "0.5540684", "0.5507903", "0.54966867", "0.5482834", "0.5468084", "0.54641813", "0.54412085", "0.54346025", "0.5431878", "0.5428448", "0.5410218", "0.54091835", "0.539753", "0.5389643", "0.53873307", "0.5387152", "0.53778267", "0.5376558", "0.5373828", "0.5369401", "0.5364752", "0.53644943", "0.53608614", "0.5351265", "0.5339851", "0.5329527", "0.53237605", "0.53203", "0.5320041", "0.5317824", "0.53143674", "0.53115165", "0.5307707", "0.5305985", "0.5298789", "0.5297029", "0.5295996", "0.5292895", "0.5290934", "0.5289561", "0.528785", "0.52868646", "0.5285782", "0.5281198", "0.52800053", "0.52764326", "0.52759385", "0.5271771", "0.52544", "0.5253631", "0.52490705", "0.52444166", "0.5237697", "0.5237549", "0.5233898", "0.52333975", "0.52300274", "0.522976", "0.52277225", "0.52267706", "0.5221663", "0.5220781", "0.52197874", "0.5219235", "0.5213395", "0.5208769", "0.52078015", "0.52069837", "0.52056164", "0.5202107", "0.5197858", "0.5197313", "0.51908815", "0.5190056", "0.5186193", "0.5184764", "0.5182142", "0.51745903", "0.5174393", "0.5171372", "0.5170627", "0.51693815" ]
0.57720464
5
UnmarshalJSON unmarshals the notificatio endpoint. This is necessary b/c of the notification endpoint does not have a means ot unmarshal itself.
func (s *SummaryNotificationEndpoint) UnmarshalJSON(b []byte) error { var a struct { SummaryIdentifier NotificationEndpoint json.RawMessage `json:"notificationEndpoint"` LabelAssociations []SummaryLabel `json:"labelAssociations"` } if err := json.Unmarshal(b, &a); err != nil { return err } s.SummaryIdentifier = a.SummaryIdentifier s.LabelAssociations = a.LabelAssociations e, err := endpoint.UnmarshalJSON(a.NotificationEndpoint) s.NotificationEndpoint = e return err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (n *Notification) UnmarshalJSON(b []byte) error {\n\tnotification := map[string]interface{}{}\n\n\terr := json.Unmarshal(b, &notification)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif notification[\"_id\"] != nil && bson.IsObjectIdHex(notification[\"_id\"].(string)) {\n\t\tn.ID = bson.ObjectIdHex(notification[\"_id\"].(string))\n\t}\n\tif notification[\"id\"] != nil && bson.IsObjectIdHex(notification[\"id\"].(string)) {\n\t\tn.ID = bson.ObjectIdHex(notification[\"id\"].(string))\n\t}\n\tif notification[\"recipient\"] == nil {\n\t\t// return errors.New(\"Order Hash is not set\")\n\t} else {\n\t\tn.Recipient = common.HexToAddress(notification[\"recipient\"].(string))\n\t}\n\n\tif notification[\"message\"] != nil {\n\t\tn.Message = notification[\"message\"].(Message)\n\t}\n\n\tif notification[\"type\"] != nil {\n\t\tn.Type = notification[\"type\"].(string)\n\t}\n\n\tif notification[\"status\"] != nil {\n\t\tn.Status = notification[\"status\"].(string)\n\t}\n\n\tif notification[\"createdAt\"] != nil {\n\t\tnm, _ := time.Parse(time.RFC3339Nano, notification[\"createdAt\"].(string))\n\t\tn.CreatedAt = nm\n\t}\n\n\tif notification[\"updatedAt\"] != nil {\n\t\tnm, _ := time.Parse(time.RFC3339Nano, notification[\"updatedAt\"].(string))\n\t\tn.UpdatedAt = nm\n\t}\n\n\treturn nil\n}", "func (v *Notification) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson9806e1DecodeGithubComKonstantinProninEmailSendingServicePkgModel1(&r, v)\n\treturn r.Error()\n}", "func (d *DiffNotificationEndpointValues) UnmarshalJSON(b []byte) (err error) {\n\td.NotificationEndpoint, err = endpoint.UnmarshalJSON(b)\n\tif errors2.EInvalid == errors2.ErrorCode(err) {\n\t\treturn nil\n\t}\n\treturn\n}", "func (n *NotificationChannelProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"createdDate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &n.CreatedDate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"description\":\n\t\t\terr = unpopulate(val, &n.Description)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"emailRecipient\":\n\t\t\terr = unpopulate(val, &n.EmailRecipient)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"events\":\n\t\t\terr = unpopulate(val, &n.Events)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"notificationLocale\":\n\t\t\terr = unpopulate(val, &n.NotificationLocale)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, &n.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"uniqueIdentifier\":\n\t\t\terr = unpopulate(val, &n.UniqueIdentifier)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"webHookUrl\":\n\t\t\terr = unpopulate(val, &n.WebHookURL)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NotificationHubResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &n.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &n.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &n.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &n.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &n.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &n.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &n.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NotificationHubProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"admCredential\":\n\t\t\terr = unpopulate(val, \"AdmCredential\", &n.AdmCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"apnsCredential\":\n\t\t\terr = unpopulate(val, \"ApnsCredential\", &n.ApnsCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"authorizationRules\":\n\t\t\terr = unpopulate(val, \"AuthorizationRules\", &n.AuthorizationRules)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"baiduCredential\":\n\t\t\terr = unpopulate(val, \"BaiduCredential\", &n.BaiduCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"gcmCredential\":\n\t\t\terr = unpopulate(val, \"GCMCredential\", &n.GCMCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"mpnsCredential\":\n\t\t\terr = unpopulate(val, \"MpnsCredential\", &n.MpnsCredential)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &n.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"registrationTtl\":\n\t\t\terr = unpopulate(val, \"RegistrationTTL\", &n.RegistrationTTL)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"wnsCredential\":\n\t\t\terr = unpopulate(val, \"WnsCredential\", &n.WnsCredential)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PropagatedRouteTableNfv) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"ids\":\n\t\t\terr = unpopulate(val, \"IDs\", &p.IDs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"labels\":\n\t\t\terr = unpopulate(val, \"Labels\", &p.Labels)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NotificationHubPatchParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &n.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &n.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &n.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &n.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &n.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &n.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &n.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *MessagingEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"lockDurationAsIso8601\":\n\t\t\terr = unpopulate(val, \"LockDurationAsIso8601\", &m.LockDurationAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"maxDeliveryCount\":\n\t\t\terr = unpopulate(val, \"MaxDeliveryCount\", &m.MaxDeliveryCount)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ttlAsIso8601\":\n\t\t\terr = unpopulate(val, \"TTLAsIso8601\", &m.TTLAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (l *LiveEventEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"protocol\":\n\t\t\terr = unpopulate(val, \"Protocol\", &l.Protocol)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"url\":\n\t\t\terr = unpopulate(val, \"URL\", &l.URL)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingTwinProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"desired\":\n\t\t\terr = unpopulate(val, \"Desired\", &r.Desired)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reported\":\n\t\t\terr = unpopulate(val, \"Reported\", &r.Reported)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FirewallPolicyIntrusionDetection) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"configuration\":\n\t\t\terr = unpopulate(val, \"Configuration\", &f.Configuration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"mode\":\n\t\t\terr = unpopulate(val, \"Mode\", &f.Mode)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FallbackRouteProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"condition\":\n\t\t\terr = unpopulate(val, \"Condition\", &f.Condition)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpointNames\":\n\t\t\terr = unpopulate(val, \"EndpointNames\", &f.EndpointNames)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isEnabled\":\n\t\t\terr = unpopulate(val, \"IsEnabled\", &f.IsEnabled)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &f.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"source\":\n\t\t\terr = unpopulate(val, \"Source\", &f.Source)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NamespaceProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"createdAt\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"CreatedAt\", &n.CreatedAt)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"critical\":\n\t\t\terr = unpopulate(val, \"Critical\", &n.Critical)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataCenter\":\n\t\t\terr = unpopulate(val, \"DataCenter\", &n.DataCenter)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enabled\":\n\t\t\terr = unpopulate(val, \"Enabled\", &n.Enabled)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"metricId\":\n\t\t\terr = unpopulate(val, \"MetricID\", &n.MetricID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &n.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"namespaceType\":\n\t\t\terr = unpopulate(val, \"NamespaceType\", &n.NamespaceType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &n.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"region\":\n\t\t\terr = unpopulate(val, \"Region\", &n.Region)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"scaleUnit\":\n\t\t\terr = unpopulate(val, \"ScaleUnit\", &n.ScaleUnit)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"serviceBusEndpoint\":\n\t\t\terr = unpopulate(val, \"ServiceBusEndpoint\", &n.ServiceBusEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, \"Status\", &n.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, \"SubscriptionID\", &n.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"updatedAt\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"UpdatedAt\", &n.UpdatedAt)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NatGateway) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &n.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &n.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &n.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &n.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &n.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &n.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &n.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &n.Type)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"zones\":\n\t\t\terr = unpopulate(val, \"Zones\", &n.Zones)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpoint) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"extendedLocation\":\n\t\t\terr = unpopulate(val, \"ExtendedLocation\", &p.ExtendedLocation)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &p.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &p.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *NotificationList) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson9806e1DecodeGithubComKonstantinProninEmailSendingServicePkgModel(&r, v)\n\treturn r.Error()\n}", "func (a *Meta_Note) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]interface{})\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal interface{}\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, fmt.Sprintf(\"error unmarshaling field %s\", fieldName))\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NatGatewayPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"idleTimeoutInMinutes\":\n\t\t\terr = unpopulate(val, \"IdleTimeoutInMinutes\", &n.IdleTimeoutInMinutes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &n.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"publicIpAddresses\":\n\t\t\terr = unpopulate(val, \"PublicIPAddresses\", &n.PublicIPAddresses)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"publicIpPrefixes\":\n\t\t\terr = unpopulate(val, \"PublicIPPrefixes\", &n.PublicIPPrefixes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceGuid\":\n\t\t\terr = unpopulate(val, \"ResourceGUID\", &n.ResourceGUID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subnets\":\n\t\t\terr = unpopulate(val, \"Subnets\", &n.Subnets)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FirewallPolicySNAT) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"autoLearnPrivateRanges\":\n\t\t\terr = unpopulate(val, \"AutoLearnPrivateRanges\", &f.AutoLearnPrivateRanges)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateRanges\":\n\t\t\terr = unpopulate(val, \"PrivateRanges\", &f.PrivateRanges)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FeedbackProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"lockDurationAsIso8601\":\n\t\t\terr = unpopulate(val, \"LockDurationAsIso8601\", &f.LockDurationAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"maxDeliveryCount\":\n\t\t\terr = unpopulate(val, \"MaxDeliveryCount\", &f.MaxDeliveryCount)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ttlAsIso8601\":\n\t\t\terr = unpopulate(val, \"TTLAsIso8601\", &f.TTLAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *HTTPNotificationEndpoint_Headers) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]string)\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal string\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, fmt.Sprintf(\"error unmarshaling field %s\", fieldName))\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (h *HeatMapProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, &h.EndTime)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpoints\":\n\t\t\terr = unpopulate(val, &h.Endpoints)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"startTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, &h.StartTime)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"trafficFlows\":\n\t\t\terr = unpopulate(val, &h.TrafficFlows)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NetworkNamespace) UnmarshalJSON(b []byte) error {\n\tvar s struct {\n\t\tNetNsPath string\n\t\tNetNsCreated bool\n\t\tEndpoints json.RawMessage\n\t}\n\n\tif err := json.Unmarshal(b, &s); err != nil {\n\t\treturn err\n\t}\n\n\t(*n).NetNsPath = s.NetNsPath\n\t(*n).NetNsCreated = s.NetNsCreated\n\n\tvar typedEndpoints []TypedJSONEndpoint\n\tif err := json.Unmarshal([]byte(string(s.Endpoints)), &typedEndpoints); err != nil {\n\t\treturn err\n\t}\n\n\tvar endpoints []Endpoint\n\n\tfor _, e := range typedEndpoints {\n\t\tswitch e.Type {\n\t\tcase PhysicalEndpointType:\n\t\t\tvar endpoint PhysicalEndpoint\n\t\t\terr := json.Unmarshal(e.Data, &endpoint)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tendpoints = append(endpoints, &endpoint)\n\t\t\tnetworkLogger().WithFields(logrus.Fields{\n\t\t\t\t\"endpoint\": endpoint,\n\t\t\t\t\"endpoint-type\": \"physical\",\n\t\t\t}).Info(\"endpoint unmarshalled\")\n\n\t\tcase VirtualEndpointType:\n\t\t\tvar endpoint VirtualEndpoint\n\t\t\terr := json.Unmarshal(e.Data, &endpoint)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tendpoints = append(endpoints, &endpoint)\n\t\t\tnetworkLogger().WithFields(logrus.Fields{\n\t\t\t\t\"endpoint\": endpoint,\n\t\t\t\t\"endpoint-type\": \"virtual\",\n\t\t\t}).Info(\"endpoint unmarshalled\")\n\n\t\tcase VhostUserEndpointType:\n\t\t\tvar endpoint VhostUserEndpoint\n\t\t\terr := json.Unmarshal(e.Data, &endpoint)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tendpoints = append(endpoints, &endpoint)\n\t\t\tnetworkLogger().WithFields(logrus.Fields{\n\t\t\t\t\"endpoint\": endpoint,\n\t\t\t\t\"endpoint-type\": \"vhostuser\",\n\t\t\t}).Info(\"endpoint unmarshalled\")\n\n\t\tcase BridgedMacvlanEndpointType:\n\t\t\tvar endpoint BridgedMacvlanEndpoint\n\t\t\terr := json.Unmarshal(e.Data, &endpoint)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tnetworkLogger().WithFields(logrus.Fields{\n\t\t\t\t\"endpoint\": endpoint,\n\t\t\t\t\"endpoint-type\": \"macvlan\",\n\t\t\t}).Info(\"endpoint unmarshalled\")\n\n\t\tcase MacvtapEndpointType:\n\t\t\tvar endpoint MacvtapEndpoint\n\t\t\terr := json.Unmarshal(e.Data, &endpoint)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tnetworkLogger().WithFields(logrus.Fields{\n\t\t\t\t\"endpoint\": endpoint,\n\t\t\t\t\"endpoint-type\": \"macvtap\",\n\t\t\t}).Info(\"endpoint unmarshalled\")\n\n\t\tdefault:\n\t\t\tnetworkLogger().WithField(\"endpoint-type\", e.Type).Error(\"Ignoring unknown endpoint type\")\n\t\t}\n\t}\n\n\t(*n).Endpoints = endpoints\n\treturn nil\n}", "func UnmarshalNotificationsIntegrationGetResponse(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationsIntegrationGetResponse)\n\terr = core.UnmarshalPrimitive(m, \"description\", &obj.Description)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"name\", &obj.Name)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"enabled\", &obj.Enabled)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"type\", &obj.Type)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"updated_at\", &obj.UpdatedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"topic_count\", &obj.TopicCount)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"topic_names\", &obj.TopicNames)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (r *RoutingConfigurationNfv) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"associatedRouteTable\":\n\t\t\terr = unpopulate(val, \"AssociatedRouteTable\", &r.AssociatedRouteTable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"inboundRouteMap\":\n\t\t\terr = unpopulate(val, \"InboundRouteMap\", &r.InboundRouteMap)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"outboundRouteMap\":\n\t\t\terr = unpopulate(val, \"OutboundRouteMap\", &r.OutboundRouteMap)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"propagatedRouteTables\":\n\t\t\terr = unpopulate(val, \"PropagatedRouteTables\", &r.PropagatedRouteTables)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NotificationHubListResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"nextLink\":\n\t\t\terr = unpopulate(val, \"NextLink\", &n.NextLink)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &n.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NonComplianceMessage) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &n.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"policyDefinitionReferenceId\":\n\t\t\terr = unpopulate(val, \"PolicyDefinitionReferenceID\", &n.PolicyDefinitionReferenceID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingTwin) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &r.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &r.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *AppPatch) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"identity\":\n\t\t\terr = unpopulate(val, \"Identity\", &a.Identity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &a.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &a.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointIPConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &e.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceId\":\n\t\t\terr = unpopulate(val, \"ResourceID\", &e.ResourceID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &e.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NamespacePatchParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &n.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &n.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (w *WebhookUpdatePartnerDestinationInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &w.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &w.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (w *WatcherPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &w.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *ProbePropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"intervalInSeconds\":\n\t\t\terr = unpopulate(val, \"IntervalInSeconds\", &p.IntervalInSeconds)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"loadBalancingRules\":\n\t\t\terr = unpopulate(val, \"LoadBalancingRules\", &p.LoadBalancingRules)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"numberOfProbes\":\n\t\t\terr = unpopulate(val, \"NumberOfProbes\", &p.NumberOfProbes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"port\":\n\t\t\terr = unpopulate(val, \"Port\", &p.Port)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"probeThreshold\":\n\t\t\terr = unpopulate(val, \"ProbeThreshold\", &p.ProbeThreshold)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"protocol\":\n\t\t\terr = unpopulate(val, \"Protocol\", &p.Protocol)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"requestPath\":\n\t\t\terr = unpopulate(val, \"RequestPath\", &p.RequestPath)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalNotificationsIntegrationPostResponse(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationsIntegrationPostResponse)\n\terr = core.UnmarshalPrimitive(m, \"description\", &obj.Description)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"name\", &obj.Name)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"enabled\", &obj.Enabled)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"type\", &obj.Type)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"created_at\", &obj.CreatedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (p *PrivateEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"applicationSecurityGroups\":\n\t\t\terr = unpopulate(val, \"ApplicationSecurityGroups\", &p.ApplicationSecurityGroups)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"customDnsConfigs\":\n\t\t\terr = unpopulate(val, \"CustomDNSConfigs\", &p.CustomDNSConfigs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"customNetworkInterfaceName\":\n\t\t\terr = unpopulate(val, \"CustomNetworkInterfaceName\", &p.CustomNetworkInterfaceName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipConfigurations\":\n\t\t\terr = unpopulate(val, \"IPConfigurations\", &p.IPConfigurations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"manualPrivateLinkServiceConnections\":\n\t\t\terr = unpopulate(val, \"ManualPrivateLinkServiceConnections\", &p.ManualPrivateLinkServiceConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"networkInterfaces\":\n\t\t\terr = unpopulate(val, \"NetworkInterfaces\", &p.NetworkInterfaces)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnections\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnections\", &p.PrivateLinkServiceConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subnet\":\n\t\t\terr = unpopulate(val, \"Subnet\", &p.Subnet)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FlowLogInformation) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"flowAnalyticsConfiguration\":\n\t\t\terr = unpopulate(val, \"FlowAnalyticsConfiguration\", &f.FlowAnalyticsConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &f.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"targetResourceId\":\n\t\t\terr = unpopulate(val, \"TargetResourceID\", &f.TargetResourceID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *ContentKeyPolicyWidevineConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &c.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"widevineTemplate\":\n\t\t\terr = unpopulate(val, \"WidevineTemplate\", &c.WidevineTemplate)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *JobDeliveryInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"scheduledDateTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"ScheduledDateTime\", &j.ScheduledDateTime)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalNotificationsIntegrationTestPostResponse(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationsIntegrationTestPostResponse)\n\terr = core.UnmarshalPrimitive(m, \"datacontenttype\", &obj.Datacontenttype)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"ibmendefaultlong\", &obj.Ibmendefaultlong)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"ibmendefaultshort\", &obj.Ibmendefaultshort)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"ibmensourceid\", &obj.Ibmensourceid)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"source\", &obj.Source)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"specversion\", &obj.Specversion)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"type\", &obj.Type)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (t *TopicProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"dataResidencyBoundary\":\n\t\t\terr = unpopulate(val, \"DataResidencyBoundary\", &t.DataResidencyBoundary)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"disableLocalAuth\":\n\t\t\terr = unpopulate(val, \"DisableLocalAuth\", &t.DisableLocalAuth)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpoint\":\n\t\t\terr = unpopulate(val, \"Endpoint\", &t.Endpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"inboundIpRules\":\n\t\t\terr = unpopulate(val, \"InboundIPRules\", &t.InboundIPRules)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"inputSchema\":\n\t\t\terr = unpopulate(val, \"InputSchema\", &t.InputSchema)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"inputSchemaMapping\":\n\t\t\tt.InputSchemaMapping, err = unmarshalInputSchemaMappingClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"metricResourceId\":\n\t\t\terr = unpopulate(val, \"MetricResourceID\", &t.MetricResourceID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateEndpointConnections\":\n\t\t\terr = unpopulate(val, \"PrivateEndpointConnections\", &t.PrivateEndpointConnections)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &t.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"publicNetworkAccess\":\n\t\t\terr = unpopulate(val, \"PublicNetworkAccess\", &t.PublicNetworkAccess)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (h *HubIPConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &h.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &h.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &h.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &h.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &h.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingServiceBusTopicEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"connectionString\":\n\t\t\terr = unpopulate(val, \"ConnectionString\", &r.ConnectionString)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceGroup\":\n\t\t\terr = unpopulate(val, \"ResourceGroup\", &r.ResourceGroup)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, \"SubscriptionID\", &r.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EventHubProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpoint\":\n\t\t\terr = unpopulate(val, \"Endpoint\", &e.Endpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"partitionCount\":\n\t\t\terr = unpopulate(val, \"PartitionCount\", &e.PartitionCount)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"partitionIds\":\n\t\t\terr = unpopulate(val, \"PartitionIDs\", &e.PartitionIDs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"path\":\n\t\t\terr = unpopulate(val, \"Path\", &e.Path)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"retentionTimeInDays\":\n\t\t\terr = unpopulate(val, \"RetentionTimeInDays\", &e.RetentionTimeInDays)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *ReminderV1) UnmarshalJSON(data []byte) error {\n\tif string(data) == \"null\" { return nil }\n\tobj := map[string]interface{}{}\n\terr := json.Unmarshal(data, &obj)\n\tif err != nil {\n\t\tfmt.Println(obj)\n\t\treturn fmt.Errorf(\"inital unmarshal failed: %w\", err)\n\t}\n\tfor key, i := range obj {\n\t\tswitch key {\n\t\tcase \"version\":\n\t\t\tvalue, ok := i.(string)\n\t\t\tif ! ok {\n\t\t\t\treturn fmt.Errorf(\"failed to parse value of key \\\"version\\\" into string\")\n\t\t\t}\n\t\t\tif value != \"v1\" {\n\t\t\t\treturn fmt.Errorf(\"expected value \\\"v1\\\" but got value \\\"%s\\\"\", value)\n\t\t\t}\n\t\t\tr.Version = value\n\n\t\tcase \"message\":\n\t\t\tvalue, ok := i.(string)\n\t\t\tif ! ok {\n\t\t\t\tmsg := \"failed to parse value of key \\\"message\\\" into string\"\n\t\t\t\treturn fmt.Errorf(msg)\n\t\t\t}\n\t\t\tr.Message = value\n\n\t\tcase \"triggers\":\n\t\t\tinterface_list, ok := i.([]interface{})\n\t\t\tif ! ok {\n\t\t\t\tmsg := \"failed to parse value of key \\\"triggers\\\" into []interface{}\"\n\t\t\t\treturn fmt.Errorf(msg)\n\t\t\t}\n\t\t\tr.Triggers = make([]Trigger, 0)\n\t\t\tfor _, i := range interface_list {\n\t\t\t\tobj_map, ok := i.(map[string]interface{})\n\t\t\t\tif ! ok {\n\t\t\t\t\tmsg := \"failed to parse a trigger into a map[string]interface{}\"\n\t\t\t\t\treturn fmt.Errorf(msg)\n\t\t\t\t}\n\t\t\t\ttrigger, err := parseTriggerFromInterface(obj_map)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"parseTriggerFromInterface: %w\", err)\n\t\t\t\t}\n\t\t\t\tr.Triggers = append(r.Triggers, trigger)\n\t\t\t}\n\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"ReminderV1.UnmarshalJSON: key %s is invalid\", key)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PNGFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"filenamePattern\":\n\t\t\terr = unpopulate(val, \"FilenamePattern\", &p.FilenamePattern)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &p.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *Format) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"filenamePattern\":\n\t\t\terr = unpopulate(val, \"FilenamePattern\", &f.FilenamePattern)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &f.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateLinkServiceIPConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VPNSiteLinkProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"bgpProperties\":\n\t\t\terr = unpopulate(val, \"BgpProperties\", &v.BgpProperties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"fqdn\":\n\t\t\terr = unpopulate(val, \"Fqdn\", &v.Fqdn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipAddress\":\n\t\t\terr = unpopulate(val, \"IPAddress\", &v.IPAddress)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"linkProperties\":\n\t\t\terr = unpopulate(val, \"LinkProperties\", &v.LinkProperties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalNotificationCreate(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationCreate)\n\terr = core.UnmarshalPrimitive(m, \"specversion\", &obj.Specversion)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"specversion\")\n\terr = core.UnmarshalPrimitive(m, \"time\", &obj.Time)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"time\")\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"id\")\n\terr = core.UnmarshalPrimitive(m, \"source\", &obj.Source)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"source\")\n\terr = core.UnmarshalPrimitive(m, \"type\", &obj.Type)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"type\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenseverity\", &obj.Ibmenseverity)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenseverity\")\n\terr = core.UnmarshalPrimitive(m, \"ibmensourceid\", &obj.Ibmensourceid)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmensourceid\")\n\terr = core.UnmarshalPrimitive(m, \"ibmendefaultshort\", &obj.Ibmendefaultshort)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmendefaultshort\")\n\terr = core.UnmarshalPrimitive(m, \"ibmendefaultlong\", &obj.Ibmendefaultlong)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmendefaultlong\")\n\terr = core.UnmarshalPrimitive(m, \"subject\", &obj.Subject)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"subject\")\n\terr = core.UnmarshalPrimitive(m, \"data\", &obj.Data)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"data\")\n\terr = core.UnmarshalPrimitive(m, \"datacontenttype\", &obj.Datacontenttype)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"datacontenttype\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenpushto\", &obj.Ibmenpushto)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenpushto\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenfcmbody\", &obj.Ibmenfcmbody)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenfcmbody\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenapnsbody\", &obj.Ibmenapnsbody)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenapnsbody\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenapnsheaders\", &obj.Ibmenapnsheaders)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenapnsheaders\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenchromebody\", &obj.Ibmenchromebody)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenchromebody\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenchromeheaders\", &obj.Ibmenchromeheaders)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenchromeheaders\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenfirefoxbody\", &obj.Ibmenfirefoxbody)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenfirefoxbody\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenfirefoxheaders\", &obj.Ibmenfirefoxheaders)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenfirefoxheaders\")\n\terr = core.UnmarshalPrimitive(m, \"ibmenhuaweibody\", &obj.Ibmenhuaweibody)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmenhuaweibody\")\n\terr = core.UnmarshalPrimitive(m, \"ibmensafaribody\", &obj.Ibmensafaribody)\n\tif err != nil {\n\t\treturn\n\t}\n\tdelete(m, \"ibmensafaribody\")\n\tfor k := range m {\n\t\tvar v interface{}\n\t\te := core.UnmarshalPrimitive(m, k, &v)\n\t\tif e != nil {\n\t\t\terr = e\n\t\t\treturn\n\t\t}\n\t\tobj.SetProperty(k, v)\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (f *FirewallPolicyIntrusionDetectionConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"bypassTrafficSettings\":\n\t\t\terr = unpopulate(val, \"BypassTrafficSettings\", &f.BypassTrafficSettings)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateRanges\":\n\t\t\terr = unpopulate(val, \"PrivateRanges\", &f.PrivateRanges)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"signatureOverrides\":\n\t\t\terr = unpopulate(val, \"SignatureOverrides\", &f.SignatureOverrides)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalNotificationResponse(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(NotificationResponse)\n\terr = core.UnmarshalPrimitive(m, \"notification_id\", &obj.NotificationID)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (r *RoutingConfigurationNfvSubResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"resourceUri\":\n\t\t\terr = unpopulate(val, \"ResourceURI\", &r.ResourceURI)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TroubleshootingProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"storageId\":\n\t\t\terr = unpopulate(val, \"StorageID\", &t.StorageID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"storagePath\":\n\t\t\terr = unpopulate(val, \"StoragePath\", &t.StoragePath)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *ImageTemplateProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"buildTimeoutInMinutes\":\n\t\t\terr = unpopulate(val, &i.BuildTimeoutInMinutes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"customize\":\n\t\t\ti.Customize, err = unmarshalImageTemplateCustomizerClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"distribute\":\n\t\t\ti.Distribute, err = unmarshalImageTemplateDistributorClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastRunStatus\":\n\t\t\terr = unpopulate(val, &i.LastRunStatus)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningError\":\n\t\t\terr = unpopulate(val, &i.ProvisioningError)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, &i.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"source\":\n\t\t\ti.Source, err = unmarshalImageTemplateSourceClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"vmProfile\":\n\t\t\terr = unpopulate(val, &i.VMProfile)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (m *ModelDeprecationInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"fineTune\":\n\t\t\terr = unpopulate(val, \"FineTune\", &m.FineTune)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"inference\":\n\t\t\terr = unpopulate(val, \"Inference\", &m.Inference)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *ServiceTagInformationPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"addressPrefixes\":\n\t\t\terr = unpopulate(val, \"AddressPrefixes\", &s.AddressPrefixes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"changeNumber\":\n\t\t\terr = unpopulate(val, \"ChangeNumber\", &s.ChangeNumber)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"region\":\n\t\t\terr = unpopulate(val, \"Region\", &s.Region)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"state\":\n\t\t\terr = unpopulate(val, \"State\", &s.State)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemService\":\n\t\t\terr = unpopulate(val, \"SystemService\", &s.SystemService)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *P2SVPNGateway) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &p.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &p.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &p.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &p.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &p.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &p.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *ContentKeyPolicyUnknownConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &c.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Notification) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson9806e1DecodeGithubComKonstantinProninEmailSendingServicePkgModel1(l, v)\n}", "func (c *CreateEmailSignInURLResponseProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"url\":\n\t\t\terr = unpopulate(val, \"URL\", &c.URL)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func UnmarshalNotification(reader io.Reader, consumer runtime.Consumer) (Notification, error) {\n\t// we need to read this twice, so first into a buffer\n\tdata, err := io.ReadAll(reader)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn unmarshalNotification(data, consumer)\n}", "func (i *InboundNatPoolPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"backendPort\":\n\t\t\terr = unpopulate(val, \"BackendPort\", &i.BackendPort)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enableFloatingIP\":\n\t\t\terr = unpopulate(val, \"EnableFloatingIP\", &i.EnableFloatingIP)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enableTcpReset\":\n\t\t\terr = unpopulate(val, \"EnableTCPReset\", &i.EnableTCPReset)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"frontendIPConfiguration\":\n\t\t\terr = unpopulate(val, \"FrontendIPConfiguration\", &i.FrontendIPConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"frontendPortRangeEnd\":\n\t\t\terr = unpopulate(val, \"FrontendPortRangeEnd\", &i.FrontendPortRangeEnd)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"frontendPortRangeStart\":\n\t\t\terr = unpopulate(val, \"FrontendPortRangeStart\", &i.FrontendPortRangeStart)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"idleTimeoutInMinutes\":\n\t\t\terr = unpopulate(val, \"IdleTimeoutInMinutes\", &i.IdleTimeoutInMinutes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"protocol\":\n\t\t\terr = unpopulate(val, \"Protocol\", &i.Protocol)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &i.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (k *KikChannelProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", k, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"apiKey\":\n\t\t\terr = unpopulate(val, \"APIKey\", &k.APIKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isEnabled\":\n\t\t\terr = unpopulate(val, \"IsEnabled\", &k.IsEnabled)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isValidated\":\n\t\t\terr = unpopulate(val, \"IsValidated\", &k.IsValidated)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"userName\":\n\t\t\terr = unpopulate(val, \"UserName\", &k.UserName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", k, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *ServiceEndpointPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"locations\":\n\t\t\terr = unpopulate(val, \"Locations\", &s.Locations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &s.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"service\":\n\t\t\terr = unpopulate(val, \"Service\", &s.Service)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *StorageEndpointProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"connectionString\":\n\t\t\terr = unpopulate(val, \"ConnectionString\", &s.ConnectionString)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"containerName\":\n\t\t\terr = unpopulate(val, \"ContainerName\", &s.ContainerName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sasTtlAsIso8601\":\n\t\t\terr = unpopulate(val, \"SasTTLAsIso8601\", &s.SasTTLAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateLinkServicePropertiesVisibility) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"subscriptions\":\n\t\t\terr = unpopulate(val, \"Subscriptions\", &p.Subscriptions)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CloudToDeviceProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"defaultTtlAsIso8601\":\n\t\t\terr = unpopulate(val, \"DefaultTTLAsIso8601\", &c.DefaultTTLAsIso8601)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"feedback\":\n\t\t\terr = unpopulate(val, \"Feedback\", &c.Feedback)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"maxDeliveryCount\":\n\t\t\terr = unpopulate(val, \"MaxDeliveryCount\", &c.MaxDeliveryCount)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TelephonyChannelProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"apiConfigurations\":\n\t\t\terr = unpopulate(val, \"APIConfigurations\", &t.APIConfigurations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"cognitiveServiceRegion\":\n\t\t\terr = unpopulate(val, \"CognitiveServiceRegion\", &t.CognitiveServiceRegion)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"cognitiveServiceSubscriptionKey\":\n\t\t\terr = unpopulate(val, \"CognitiveServiceSubscriptionKey\", &t.CognitiveServiceSubscriptionKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"defaultLocale\":\n\t\t\terr = unpopulate(val, \"DefaultLocale\", &t.DefaultLocale)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isEnabled\":\n\t\t\terr = unpopulate(val, \"IsEnabled\", &t.IsEnabled)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"phoneNumbers\":\n\t\t\terr = unpopulate(val, \"PhoneNumbers\", &t.PhoneNumbers)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"premiumSKU\":\n\t\t\terr = unpopulate(val, \"PremiumSKU\", &t.PremiumSKU)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RemediationProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"correlationId\":\n\t\t\terr = unpopulate(val, \"CorrelationID\", &r.CorrelationID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"createdOn\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"CreatedOn\", &r.CreatedOn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deploymentStatus\":\n\t\t\terr = unpopulate(val, \"DeploymentStatus\", &r.DeploymentStatus)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"failureThreshold\":\n\t\t\terr = unpopulate(val, \"FailureThreshold\", &r.FailureThreshold)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"filters\":\n\t\t\terr = unpopulate(val, \"Filters\", &r.Filters)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastUpdatedOn\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"LastUpdatedOn\", &r.LastUpdatedOn)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"parallelDeployments\":\n\t\t\terr = unpopulate(val, \"ParallelDeployments\", &r.ParallelDeployments)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"policyAssignmentId\":\n\t\t\terr = unpopulate(val, \"PolicyAssignmentID\", &r.PolicyAssignmentID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"policyDefinitionReferenceId\":\n\t\t\terr = unpopulate(val, \"PolicyDefinitionReferenceID\", &r.PolicyDefinitionReferenceID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &r.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceCount\":\n\t\t\terr = unpopulate(val, \"ResourceCount\", &r.ResourceCount)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceDiscoveryMode\":\n\t\t\terr = unpopulate(val, \"ResourceDiscoveryMode\", &r.ResourceDiscoveryMode)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"statusMessage\":\n\t\t\terr = unpopulate(val, \"StatusMessage\", &r.StatusMessage)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *InboundNatRulePropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"backendAddressPool\":\n\t\t\terr = unpopulate(val, \"BackendAddressPool\", &i.BackendAddressPool)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"backendIPConfiguration\":\n\t\t\terr = unpopulate(val, \"BackendIPConfiguration\", &i.BackendIPConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"backendPort\":\n\t\t\terr = unpopulate(val, \"BackendPort\", &i.BackendPort)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enableFloatingIP\":\n\t\t\terr = unpopulate(val, \"EnableFloatingIP\", &i.EnableFloatingIP)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enableTcpReset\":\n\t\t\terr = unpopulate(val, \"EnableTCPReset\", &i.EnableTCPReset)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"frontendIPConfiguration\":\n\t\t\terr = unpopulate(val, \"FrontendIPConfiguration\", &i.FrontendIPConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"frontendPort\":\n\t\t\terr = unpopulate(val, \"FrontendPort\", &i.FrontendPort)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"frontendPortRangeEnd\":\n\t\t\terr = unpopulate(val, \"FrontendPortRangeEnd\", &i.FrontendPortRangeEnd)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"frontendPortRangeStart\":\n\t\t\terr = unpopulate(val, \"FrontendPortRangeStart\", &i.FrontendPortRangeStart)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"idleTimeoutInMinutes\":\n\t\t\terr = unpopulate(val, \"IdleTimeoutInMinutes\", &i.IdleTimeoutInMinutes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"protocol\":\n\t\t\terr = unpopulate(val, \"Protocol\", &i.Protocol)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &i.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *AzureFirewallFqdnTagPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"fqdnTagName\":\n\t\t\terr = unpopulate(val, \"FqdnTagName\", &a.FqdnTagName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &a.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (w *WebhookPartnerDestinationProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"clientAuthentication\":\n\t\t\tw.ClientAuthentication, err = unmarshalPartnerClientAuthenticationClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpointBaseUrl\":\n\t\t\terr = unpopulate(val, \"EndpointBaseURL\", &w.EndpointBaseURL)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"endpointUrl\":\n\t\t\terr = unpopulate(val, \"EndpointURL\", &w.EndpointURL)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (o *ProjectWebhookPartial) UnmarshalJSON(data []byte) error {\n\tkv := make(map[string]interface{})\n\tif err := json.Unmarshal(data, &kv); err != nil {\n\t\treturn err\n\t}\n\to.FromMap(kv)\n\treturn nil\n}", "func (h *Hub) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"resourceId\":\n\t\t\terr = unpopulate(val, \"ResourceID\", &h.ResourceID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceType\":\n\t\t\terr = unpopulate(val, \"ResourceType\", &h.ResourceType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TrackBase) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &t.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (g *GatewayServerProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", g, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"health\":\n\t\t\terr = unpopulate(val, \"Health\", &g.Health)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"port\":\n\t\t\terr = unpopulate(val, \"Port\", &g.Port)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", g, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EnqueueReplicationServerProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"ersVersion\":\n\t\t\terr = unpopulate(val, \"ErsVersion\", &e.ErsVersion)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"health\":\n\t\t\terr = unpopulate(val, \"Health\", &e.Health)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"hostname\":\n\t\t\terr = unpopulate(val, \"Hostname\", &e.Hostname)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipAddress\":\n\t\t\terr = unpopulate(val, \"IPAddress\", &e.IPAddress)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"instanceNo\":\n\t\t\terr = unpopulate(val, \"InstanceNo\", &e.InstanceNo)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"kernelPatch\":\n\t\t\terr = unpopulate(val, \"KernelPatch\", &e.KernelPatch)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"kernelVersion\":\n\t\t\terr = unpopulate(val, \"KernelVersion\", &e.KernelVersion)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *AvailablePrivateEndpointType) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, \"DisplayName\", &a.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &a.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &a.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceName\":\n\t\t\terr = unpopulate(val, \"ResourceName\", &a.ResourceName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &a.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"privateEndpoint\":\n\t\t\terr = unpopulate(val, \"PrivateEndpoint\", &p.PrivateEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &p.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *IPConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &i.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &i.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &i.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &i.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *InterfaceIPConfigurationPrivateLinkConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"fqdns\":\n\t\t\terr = unpopulate(val, \"Fqdns\", &i.Fqdns)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"groupId\":\n\t\t\terr = unpopulate(val, \"GroupID\", &i.GroupID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"requiredMemberName\":\n\t\t\terr = unpopulate(val, \"RequiredMemberName\", &i.RequiredMemberName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PartnerTopicProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"activationState\":\n\t\t\terr = unpopulate(val, \"ActivationState\", &p.ActivationState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"eventTypeInfo\":\n\t\t\terr = unpopulate(val, \"EventTypeInfo\", &p.EventTypeInfo)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"expirationTimeIfNotActivatedUtc\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"ExpirationTimeIfNotActivatedUTC\", &p.ExpirationTimeIfNotActivatedUTC)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"messageForActivation\":\n\t\t\terr = unpopulate(val, \"MessageForActivation\", &p.MessageForActivation)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"partnerRegistrationImmutableId\":\n\t\t\terr = unpopulate(val, \"PartnerRegistrationImmutableID\", &p.PartnerRegistrationImmutableID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"partnerTopicFriendlyDescription\":\n\t\t\terr = unpopulate(val, \"PartnerTopicFriendlyDescription\", &p.PartnerTopicFriendlyDescription)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"source\":\n\t\t\terr = unpopulate(val, \"Source\", &p.Source)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *Properties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"authorizationPolicies\":\n\t\t\terr = unpopulate(val, \"AuthorizationPolicies\", &p.AuthorizationPolicies)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"cloudToDevice\":\n\t\t\terr = unpopulate(val, \"CloudToDevice\", &p.CloudToDevice)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"comments\":\n\t\t\terr = unpopulate(val, \"Comments\", &p.Comments)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deviceStreams\":\n\t\t\terr = unpopulate(val, \"DeviceStreams\", &p.DeviceStreams)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enableFileUploadNotifications\":\n\t\t\terr = unpopulate(val, \"EnableFileUploadNotifications\", &p.EnableFileUploadNotifications)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"eventHubEndpoints\":\n\t\t\terr = unpopulate(val, \"EventHubEndpoints\", &p.EventHubEndpoints)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"features\":\n\t\t\terr = unpopulate(val, \"Features\", &p.Features)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"hostName\":\n\t\t\terr = unpopulate(val, \"HostName\", &p.HostName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipFilterRules\":\n\t\t\terr = unpopulate(val, \"IPFilterRules\", &p.IPFilterRules)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"locations\":\n\t\t\terr = unpopulate(val, \"Locations\", &p.Locations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"messagingEndpoints\":\n\t\t\terr = unpopulate(val, \"MessagingEndpoints\", &p.MessagingEndpoints)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"routing\":\n\t\t\terr = unpopulate(val, \"Routing\", &p.Routing)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"state\":\n\t\t\terr = unpopulate(val, \"State\", &p.State)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"storageEndpoints\":\n\t\t\terr = unpopulate(val, \"StorageEndpoints\", &p.StorageEndpoints)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TelegramChannelProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"accessToken\":\n\t\t\terr = unpopulate(val, \"AccessToken\", &t.AccessToken)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isEnabled\":\n\t\t\terr = unpopulate(val, \"IsEnabled\", &t.IsEnabled)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"isValidated\":\n\t\t\terr = unpopulate(val, \"IsValidated\", &t.IsValidated)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (w *WebHookEventSubscriptionDestination) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"endpointType\":\n\t\t\terr = unpopulate(val, \"EndpointType\", &w.EndpointType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &w.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", w, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *Future) UnmarshalJSON(data []byte) error {\n\t// unmarshal into JSON object to determine the tracker type\n\tobj := map[string]interface{}{}\n\terr := json.Unmarshal(data, &obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif obj[\"method\"] == nil {\n\t\treturn autorest.NewError(\"Future\", \"UnmarshalJSON\", \"missing 'method' property\")\n\t}\n\tmethod := obj[\"method\"].(string)\n\tswitch strings.ToUpper(method) {\n\tcase http.MethodDelete:\n\t\tf.pt = &pollingTrackerDelete{}\n\tcase http.MethodPatch:\n\t\tf.pt = &pollingTrackerPatch{}\n\tcase http.MethodPost:\n\t\tf.pt = &pollingTrackerPost{}\n\tcase http.MethodPut:\n\t\tf.pt = &pollingTrackerPut{}\n\tdefault:\n\t\treturn autorest.NewError(\"Future\", \"UnmarshalJSON\", \"unsupoorted method '%s'\", method)\n\t}\n\t// now unmarshal into the tracker\n\treturn json.Unmarshal(data, &f.pt)\n}", "func (p *PrivateEndpointConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"groupIds\":\n\t\t\terr = unpopulate(val, \"GroupIDs\", &p.GroupIDs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateEndpoint\":\n\t\t\terr = unpopulate(val, \"PrivateEndpoint\", &p.PrivateEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &p.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (p *PrivateEndpointConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"groupIds\":\n\t\t\terr = unpopulate(val, \"GroupIDs\", &p.GroupIDs)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateEndpoint\":\n\t\t\terr = unpopulate(val, \"PrivateEndpoint\", &p.PrivateEndpoint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"privateLinkServiceConnectionState\":\n\t\t\terr = unpopulate(val, \"PrivateLinkServiceConnectionState\", &p.PrivateLinkServiceConnectionState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &p.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TransportStreamFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"filenamePattern\":\n\t\t\terr = unpopulate(val, \"FilenamePattern\", &t.FilenamePattern)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &t.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"outputFiles\":\n\t\t\terr = unpopulate(val, \"OutputFiles\", &t.OutputFiles)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VirtualNetworkTapPropertiesFormat) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"destinationLoadBalancerFrontEndIPConfiguration\":\n\t\t\terr = unpopulate(val, \"DestinationLoadBalancerFrontEndIPConfiguration\", &v.DestinationLoadBalancerFrontEndIPConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"destinationNetworkInterfaceIPConfiguration\":\n\t\t\terr = unpopulate(val, \"DestinationNetworkInterfaceIPConfiguration\", &v.DestinationNetworkInterfaceIPConfiguration)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"destinationPort\":\n\t\t\terr = unpopulate(val, \"DestinationPort\", &v.DestinationPort)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"networkInterfaceTapConfigurations\":\n\t\t\terr = unpopulate(val, \"NetworkInterfaceTapConfigurations\", &v.NetworkInterfaceTapConfigurations)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"resourceGuid\":\n\t\t\terr = unpopulate(val, \"ResourceGUID\", &v.ResourceGUID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *StreamingPolicyWidevineConfiguration) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"customLicenseAcquisitionUrlTemplate\":\n\t\t\terr = unpopulate(val, \"CustomLicenseAcquisitionURLTemplate\", &s.CustomLicenseAcquisitionURLTemplate)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RoutingMessage) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"appProperties\":\n\t\t\terr = unpopulate(val, \"AppProperties\", &r.AppProperties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"body\":\n\t\t\terr = unpopulate(val, \"Body\", &r.Body)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemProperties\":\n\t\t\terr = unpopulate(val, \"SystemProperties\", &r.SystemProperties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *ImageTemplateWindowsUpdateCustomizer) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"filters\":\n\t\t\terr = unpopulate(val, &i.Filters)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, &i.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"searchCriteria\":\n\t\t\terr = unpopulate(val, &i.SearchCriteria)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, &i.Type)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"updateLimit\":\n\t\t\terr = unpopulate(val, &i.UpdateLimit)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (v *VPNGateway) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &v.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &v.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &v.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &v.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &v.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &v.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &v.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}" ]
[ "0.69373727", "0.65384173", "0.6510438", "0.62654644", "0.6248569", "0.6122419", "0.604768", "0.6027493", "0.60112524", "0.59933054", "0.5926565", "0.5925177", "0.5888255", "0.5877787", "0.58761215", "0.58761215", "0.58761215", "0.5857041", "0.5844848", "0.5835386", "0.58060557", "0.58037406", "0.57821596", "0.576952", "0.5768042", "0.5764205", "0.57605934", "0.57593584", "0.57411486", "0.573664", "0.57309127", "0.5727368", "0.57271343", "0.5727061", "0.5721508", "0.570718", "0.56995225", "0.569872", "0.56860214", "0.56833476", "0.5680605", "0.5673292", "0.56477785", "0.5646842", "0.56385726", "0.5620001", "0.561706", "0.56135374", "0.5613187", "0.5612997", "0.5604462", "0.56031626", "0.55973905", "0.5595734", "0.5592253", "0.5591124", "0.5591052", "0.55857044", "0.5584081", "0.55769503", "0.55737996", "0.55653125", "0.55612165", "0.556109", "0.5547108", "0.55459", "0.5545517", "0.5542135", "0.5540312", "0.5534946", "0.5531783", "0.5520351", "0.5518121", "0.5508719", "0.5507462", "0.5506888", "0.5502665", "0.5502572", "0.5494765", "0.5494007", "0.54925305", "0.54913974", "0.5488776", "0.5484372", "0.5483733", "0.5482749", "0.5479516", "0.5477801", "0.54752624", "0.5472332", "0.5464759", "0.5463527", "0.5463393", "0.5463393", "0.54599327", "0.54582727", "0.5448854", "0.54451007", "0.54359126", "0.5434979" ]
0.62947893
3
Checks that the given request has a valid session
func CheckAuth(c *gin.Context) { }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func IsSession(r *http.Request) bool {\n\tval := r.Context().Value(authSessionActiveKey)\n\tswitch val.(type) {\n\tcase bool:\n\t\treturn val.(bool)\n\tdefault:\n\t\treturn false\n\t}\n}", "func SessionCheck(writer http.ResponseWriter, request *http.Request) (sess Session, err error) {\n\tcookie, err := request.Cookie(\"_cookie\")\n\tif err == nil {\n\t\tsess = Session{Uuid: cookie.Value}\n\t\tif ok, _ := sess.Valid(); ok {\n\t\t\terr = errors.New(\"invalid session\")\n\t\t}\n\t}\n\treturn\n}", "func GoodSession(r *http.Request) bool {\n\tstore, err := pgstore.NewPGStore(os.Getenv(\"PGURL\"), key)\n\tcheck(err)\n\tdefer store.Close()\n\n\tsession, err := store.Get(r, \"scheduler-session\")\n\tcheck(err)\n\n\t// Check if user is authenticated\n\tif auth, ok := session.Values[\"authenticated\"].(bool); !ok || !auth {\n\t\tlog.Info(\"Stale session rejected: \", session)\n\t\treturn false\n\t}\n\tlog.Info(\"Session OK: \", session)\n\treturn true\n}", "func ValidateSession(c appengine.Context, r *http.Request) error {\n\tsession, _ := store.Get(r, \"session-inventory\")\n\ttoken, _ := session.Values[\"token\"].(string)\n\tip := strings.Split(r.RemoteAddr, \":\")[0]\n\tresult, err := memcache.Get(c, \"sessions:\"+token)\n\tif err == memcache.ErrCacheMiss {\n\t\treturn errors.New(\"validate-session: token not recognized\")\n\t}\n\n\tif string(result.Value) != ip {\n\t\treturn errors.New(\"validate-session: token is from different IP\")\n\t}\n\n\treturn nil\n}", "func (session ClientSession) Valid() error {\n\tif time.Now().Unix() > session.ExpiresAt {\n\t\treturn errors.New(\"invalid session, session has expired\")\n\t}\n\n\treturn nil\n}", "func ValidateSession(w http.ResponseWriter, r *http.Request) (session Session, err error) {\n\tcookie, err := r.Cookie(\"goblog_session_cookie\")\n\tif err == nil {\n\t\tsession = Session{UUID: cookie.Value}\n\t\tif valid, _ := session.Validate(); !valid {\n\t\t\tcookie := http.Cookie{\n\t\t\t\tName: \"goblog_session_cookie\",\n\t\t\t\tValue: session.UUID,\n\t\t\t\tHttpOnly: true,\n\t\t\t\tExpires: time.Now(),\n\t\t\t\tMaxAge: -1,\n\t\t\t}\n\t\t\thttp.SetCookie(w, &cookie)\n\t\t\tlog.Println(\"Invalid session\", err)\n\t\t\thttp.Redirect(w, r, \"/login\", 302)\n\t\t}\n\t} else {\n\t\tlog.Println(\"Session doesn't exists\", err)\n\t\thttp.Redirect(w, r, \"/login\", 302)\n\t}\n\treturn\n}", "func isAuthSession(r *http.Request, w http.ResponseWriter) bool {\n\tloggedIn, loggedInMat := loggedIn(r)\n\tloggedInUser, err := user.FromMatrikel(loggedInMat)\n\n\tif !loggedIn || loggedInUser.Usertype == user.STUDENT || err != nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func CheckSession(c *revel.Controller) {\n\tresult := false\n\tif value, ok := c.Session[SESSION_KEY]; ok {\n\t\tresult = VerifySession(value)\n\t}\n\tif !result {\n\t\tc.Flash.Error(\"Session invalid. Please login.\")\n\t\tc.Response.Status = 302\n\t\tc.Response.Out.Header().Add(\"Location\", \"/session/create\")\n\t}\n}", "func isAuthenticated(req *http.Request) bool {\n\tif _, err := sessionStore.Get(req, sessionName); err == nil {\n\t\treturn true\n\t}\n\treturn false\n}", "func getSessionFromRequest(r *http.Request) *Session {\n\tkey := getSession(r)\n\ts := getSessionByKey(key)\n\n\tif s != nil && s.ID != 0 {\n\t\treturn s\n\t}\n\treturn nil\n}", "func session(request *http.Request, cookName string) (sess data.Session, err error) {\n\tcookie, err := request.Cookie(cookName)\n\tif err == nil {\n\t\tsess = data.Session{Uuid: cookie.Value}\n\t\tif ok, _ := sess.Check(); !ok {\n\t\t\terr = errors.New(\"Invalid session\")\n\t\t}\n\t}\n\treturn\n}", "func (oc *Client) validateSession(loginDetails *creds.LoginDetails) error {\n\tlogger.Debug(\"validate session func called\")\n\n\tif loginDetails == nil {\n\t\tlogger.Debug(\"unable to validate the okta session, nil input\")\n\t\treturn fmt.Errorf(\"unable to validate the okta session, nil input\")\n\t}\n\n\tsessionCookie := loginDetails.OktaSessionCookie\n\n\toktaURL, err := url.Parse(loginDetails.URL)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"error building oktaURL\")\n\t}\n\n\toktaOrgHost := oktaURL.Host\n\n\tsessionReqURL := fmt.Sprintf(\"https://%s/api/v1/sessions/me\", oktaOrgHost) // This api endpoint returns user details\n\tsessionReqBody := new(bytes.Buffer)\n\n\treq, err := http.NewRequest(\"GET\", sessionReqURL, sessionReqBody)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"error building new session request\")\n\t}\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Add(\"Cookie\", fmt.Sprintf(\"sid=%s\", sessionCookie))\n\n\tres, err := oc.client.Do(req)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"error retrieving session response\")\n\t}\n\n\tbody, err := io.ReadAll(res.Body)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"error retrieving body from response\")\n\t}\n\n\tresp := string(body)\n\n\tif res.StatusCode != 200 {\n\t\tlogger.Debug(\"invalid okta session\")\n\t\treturn fmt.Errorf(\"invalid okta session\")\n\t} else {\n\t\tsessionResponseStatus := gjson.Get(resp, \"status\").String()\n\t\tswitch sessionResponseStatus {\n\t\tcase \"ACTIVE\":\n\t\t\tlogger.Debug(\"okta session established\")\n\t\tcase \"MFA_REQUIRED\":\n\t\t\t_, err := verifyMfa(oc, oktaOrgHost, loginDetails, resp)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, \"error verifying MFA\")\n\t\t\t}\n\t\tcase \"MFA_ENROLL\":\n\t\t\t// Not yet fully implemented, so just return the status as the error string...\n\t\t\treturn fmt.Errorf(\"MFA_ENROLL\")\n\t\t}\n\t}\n\n\tlogger.Debug(\"valid okta session\")\n\treturn nil\n}", "func RequestIsAuth(r * http.Request) bool {\n\tif r.FormValue(\"username\") != \"\" && r.FormValue(\"key\") != \"\" {\n\t\tuser := UserForName(r.FormValue(\"username\"))\n\t\tif IsUser(user) {\n\t\t\tfor i := 0 ; i < len(Keys); i++ {\n\t\t\t\tif Keys[i].User == user.ID && Keys[i].Key == r.FormValue(\"key\") {\n\t\t\t\t\ttimeNow := time.Now()\n\t\t\t\t\tif timeNow.After(Keys[i].StartTime) && timeNow.Before(Keys[i].EndTime) {\n\t\t\t\t\t\treturn true\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func (session *Session) Valid() (valid bool, err error) {\n\terr = Db.QueryRow(\"SELECT id, uuid, email, user_id, created_at FROM sessions WHERE uuid=?\", session.Uuid).\n\t\tScan(&session.Id, &session.Uuid, &session.Email, &session.UserId, &session.CreatedAt)\n\tif err != nil {\n\t\tvalid = false\n\t\treturn\n\t}\n\tif session.Id != 0 {\n\t\tvalid = true\n\t}\n\treturn\n}", "func verifyLogin(req *restful.Request, resp *restful.Response ) bool {\n\tcookie, err := req.Request.Cookie(\"session-id\")\n\tif cookie.Value != \"\" {\n\t\t_, exists := sessions[cookie.Value]\n\t\tif !exists {\n\t\t\thttp.Redirect(resp.ResponseWriter, req.Request, \"/\", 302)\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\t} else if err != nil {\n\t\tfmt.Println(err.Error())\n\t\thttp.Redirect(resp.ResponseWriter, req.Request, \"/\", 302)\n\t\treturn false\n\t} else {\n\t\thttp.Redirect(resp.ResponseWriter, req.Request, \"/\", 302)\n\t\treturn false\n\t}\n}", "func (s *SessionManager) ValidateSession(w http.ResponseWriter, r *http.Request) (*SessionInfo, error) {\n\tsession, err := s.Store.Get(r, \"session\")\n\tif err != nil {\n\t\terr = errors.Errorf(errors.CodeUnauthorized, \"Previous session no longer valid: %s\", err)\n\t\tlog.Println(err)\n\t\ts.DeleteSession(w, r)\n\t\tutil.ErrorHandler(err, w)\n\t\treturn nil, err\n\t}\n\tif session.IsNew {\n\t\terr = errors.Errorf(errors.CodeUnauthorized, \"No valid session\")\n\t\tlog.Println(err)\n\t\tutil.ErrorHandler(err, w)\n\t\treturn nil, err\n\t}\n\tvar si SessionInfo\n\tvar exists bool\n\tusername, exists := session.Values[\"username\"]\n\tif !exists || username.(string) == \"\" {\n\t\terr = errors.Errorf(errors.CodeUnauthorized, \"Existing session invalid: (username: %s)\", si.Username)\n\t\tlog.Println(err)\n\t\ts.DeleteSession(w, r)\n\t\tutil.ErrorHandler(err, w)\n\t\treturn nil, err\n\t}\n\tsi.Username = username.(string)\n\tuserID, exists := session.Values[\"user_id\"]\n\tif !exists || userID == nil || userID.(string) == \"\" {\n\t\terr = errors.Errorf(errors.CodeUnauthorized, \"Existing session invalid: (id: %s)\", si.UserID)\n\t\tlog.Println(err)\n\t\ts.DeleteSession(w, r)\n\t\tutil.ErrorHandler(err, w)\n\t\treturn nil, err\n\t}\n\tsi.UserID = userID.(string)\n\treturn &si, nil\n}", "func (ctx *RequestContext) HasSession(key string) bool {\n\tvars := ctx.token.Claims[\"vars\"].(map[string]interface{})\n\t_, ok := vars[key]\n\treturn ok\n}", "func (o *LoginRequest) HasSessionId() bool {\n\tif o != nil && o.SessionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (session *Session) IsValid() error {\n\tif err := mongo.Execute(\"monotonic\", SessionCollectionName(session.TenantID),\n\t\tfunc(collection *mgo.Collection) error {\n\t\t\tselector := bson.M{\n\t\t\t\t\"token\": session.Token,\n\t\t\t}\n\t\t\treturn collection.Find(selector).One(session)\n\t\t}); err != nil {\n\t\treturn fmt.Errorf(\"Error[%s] while finding session with token[%s]\", err, session.Token)\n\t}\n\t// if session.Expired() {\n\t// \treturn fmt.Errorf(\"Session expired\")\n\t// }\n\t// extend expired Time\n\tselector := bson.M{\n\t\t\"_id\": session.ID,\n\t}\n\tupdator := bson.M{\n\t\t\"$set\": bson.M{\n\t\t\t\"e_at\": time.Now().Add(time.Duration(1) * time.Hour),\n\t\t\t\"u_at\": time.Now(),\n\t\t},\n\t}\n\treturn session.Update(session.TenantID, selector, updator)\n}", "func (handler *IdentityProviderHandler) CheckSession(sessionID string, userAgent string, remoteAddr string) (r bool, err error) {\n\thandler.log.Printf(\"checkSession(%v, %v, %v)\", sessionID, userAgent, remoteAddr)\n\n\tsession, err := handler.SessionInteractor.Find(sessionID)\n\tif err != nil {\n\t\te := err.(*errs.Error)\n\t\treturn false, errorToServiceError(e)\n\t}\n\n\tif !session.Domain.Enabled || !session.User.Enabled {\n\t\te := services.NewUnauthorizedError()\n\t\te.Msg = \"Domain and/or user disabled\"\n\t\treturn false, e\n\t}\n\n\tif session.UserAgent != userAgent || session.RemoteAddr != remoteAddr {\n\t\te := services.NewNotFoundError()\n\t\te.Msg = \"Session not found\"\n\t\treturn false, e\n\t}\n\n\tif session.IsExpired() {\n\t\te := services.NewUnauthorizedError()\n\t\te.Msg = \"Session expired\"\n\t\treturn false, e\n\t}\n\n\terr = handler.SessionInteractor.Retain(*session)\n\tif err != nil {\n\t\te := err.(*errs.Error)\n\t\treturn false, errorToServiceError(e)\n\t}\n\n\treturn true, nil\n}", "func session(w http.ResponseWriter, r *http.Request) (sess data.Session, err error) {\n\t\n\t// リクエストからクッキーを取得\n\tcookie, err := r.Cookie(\"_cookie\")\n\t// ユーザーがログインしているならクッキーがあるはず\n\tif err == nil {\n\t\t// データベースを検索\n\t\t// ユニークIDが存在してるか?\n\t\tsess = data.Session{ Uuid: cookie.Value }\n\t\t\n\t\tif ok, _ := sess.Check(); !ok {\n\t\t\terr = errors.New(\"Invalid session\")\n\t\t}\n\t}\n\treturn\n}", "func ValidateSession(auth *Authenticator, otp string) (bool, error) {\n\tif (auth == nil) || (auth.Type != TypeSession) {\n\t\treturn false, ErrInvalidAuthenticator\n\t}\n\n\totpBytes, err := hex.DecodeString(otp)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tif len(otpBytes) != 2*sessionLength {\n\t\treturn false, ErrValidationFail\n\t}\n\n\tlastBytes, err := hex.DecodeString(auth.Last)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\th := hmac.New(sha256.New, auth.Secret)\n\th.Write(lastBytes)\n\texpected := h.Sum(nil)\n\n\tif !bytes.Equal(otpBytes[:sessionLength], lastBytes) {\n\t\treturn false, ErrValidationFail\n\t}\n\n\tif !hmac.Equal(otpBytes[sessionLength:], expected) {\n\t\treturn false, ErrValidationFail\n\t}\n\n\tnext := util.RandBytes(sessionLength)\n\tif next == nil {\n\t\treturn false, errors.New(\"auth: PRNG failure\")\n\t}\n\n\tauth.Last = hex.EncodeToString(next)\n\treturn true, nil\n}", "func (authSvc *AuthService) sessionIdIsValid(sessionId string) bool {\n\t\n\treturn authSvc.validateSessionId(sessionId)\n}", "func verifyLogin(r *http.Request) bool {\n\tsession, err := store.Get(r, sessionName)\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to get session: %s\", err)\n\t\treturn false\n\t}\n\tif session.Values[\"LoggedIn\"] != \"yes\" {\n\t\treturn false\n\t}\n\treturn true\n}", "func SessionFromRequest(r *http.Request) *Session {\n\tsession, _ := r.Context().Value(\"session\").(*Session)\n\treturn session\n}", "func IsValid(req *http.Request) bool {\n\t// check if token exists in cookie\n\tcookie, err := req.Cookie(\"_token\")\n\tif err != nil {\n\t\treturn false\n\t}\n\t// validate it and allow or redirect request\n\ttoken := cookie.Value\n\treturn jwt.Passes(token)\n}", "func ValidateSession(session Session, secret string) bool {\n\treturn validSignature(session.ID+session.IssuedAt, session.Signature, secret)\n}", "func IsAuthenticated(r *http.Request) *Session {\n\tkey := getSession(r)\n\n\tif strings.HasPrefix(key, \"nouser:\") {\n\t\treturn nil\n\t}\n\n\ts := getSessionByKey(key)\n\tif isValidSession(r, s) {\n\t\treturn s\n\t}\n\treturn nil\n}", "func CheckLogin(w http.ResponseWriter, r *http.Request) (*sessions.Session, bool, error) {\n\tsession, err := store.Get(r, \"piplayer-session\")\n\tif err != nil {\n\t\treturn nil, false, err\n\t}\n\n\treturn session, session.Values[\"x-forwarded-for\"] != nil, nil\n}", "func returnSessionErr(w http.ResponseWriter, r *http.Request) bool {\n\ts, _ := session.GetCurrentSession(r)\n\treturn s.Active()\n}", "func (claims *IdentityProviderJwt) SessionRequest() SessionRequest { return claims.Request.Request }", "func AlreadyLoggedIn(r *http.Request) bool {\n\tc, err := r.Cookie(COOKIE_NAME)\n\tif err != nil {\n\t\treturn false\n\t}\n\tsess, _ := sessionStore.data[c.Value]\n\tif sess.Username != \"\" {\n\t\treturn true\n\t}\n\treturn false\n}", "func isLoggedIn(req *http.Request) bool {\n\tloginCookie, err := req.Cookie(\"loginCookie\")\n\tif err != nil {\n\t\treturn false\n\t}\n\tusername := mapSessions[loginCookie.Value]\n\t_, ok := mapUsers[username]\n\treturn ok\n}", "func (env *Env) CheckSessionReplacement() (bool) {\n if env.replacingSession != nil {\n session := env.session\n log.Debugf(\"The new session (str=%+v) is replacing the current one (str=%+v)\", env.replacingSession.String(), env.session.String())\n env.RenewEnv(env.context, env.replacingSession)\n session.SessionRelease()\n\t\tlog.Debugf(\"Restarted connection successfully with new session: %+v.\", env.session.String())\n return true\n }\n return false\n}", "func checkAuth(w http.ResponseWriter, r *http.Request, s *MemorySessionStore) bool {\n\tauth := r.Header.Get(\"Authorization\")\n\tif auth == \"\" {\n\t\treturnHTTP(w, http.StatusUnauthorized, nil)\n\t\treturn false\n\t}\n\n\tmatch := authRegexp.FindStringSubmatch(auth)\n\tif len(match) != 2 {\n\t\treturnHTTP(w, http.StatusBadRequest, nil)\n\t\treturn false\n\t}\n\n\tid := match[1]\n\tif !s.Check(id) {\n\t\treturnHTTP(w, http.StatusUnauthorized, nil)\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (s *Session) Valid() bool {\n\treturn !time.Now().After(s.Expires)\n}", "func Session(w http.ResponseWriter, r *http.Request) (models.Session, error) {\n\tvar s models.Session\n\tcookie, err := r.Cookie(\"_cookie\")\n\tif err != nil {\n\t\tfmt.Printf(\"%s\", err.Error())\n\t\treturn s, err\n\t}\n\ts = models.Session{UUID: cookie.Value}\n\tif ok, _ := s.Check(); !ok {\n\t\treturn s, errors.New(\"Invalid session. \")\n\t}\n\treturn s, nil\n}", "func (claims *ServiceProviderJwt) SessionRequest() SessionRequest { return claims.Request.Request }", "func (claims *SignatureRequestorJwt) SessionRequest() SessionRequest { return claims.Request.Request }", "func CheckRunningSession(userID int64) bool {\n\tif err := GetSessionByUserID(&Session{}, userID); err != nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func (p *OktaProvider) ValidateSessionState(s *sessions.SessionState) bool {\n\tif s.AccessToken == \"\" {\n\t\treturn false\n\t}\n\n\tvar response struct {\n\t\tActive bool `json:\"active\"`\n\t}\n\n\tform := url.Values{}\n\tform.Add(\"token\", s.AccessToken)\n\tform.Add(\"token_type_hint\", \"access_token\")\n\tform.Add(\"client_id\", p.ClientID)\n\tform.Add(\"client_secret\", p.ClientSecret)\n\n\terr := p.oktaRequest(\"POST\", p.ValidateURL.String(), form, []string{\"action:validate\"}, nil, &response)\n\tif err != nil {\n\t\treturn false\n\t}\n\tif !response.Active {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func validateSessionToken(token string) (username string, found bool) {\n\tusername, found = sessions[token]\n\treturn\n}", "func (o *LoginRequest) HasRequestedScope() bool {\n\tif o != nil && o.RequestedScope != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (s *Sessions) IsExist(r *http.Request) bool {\n\t// Getting cookie\n\tcookie, err := r.Cookie(cfg.SessionIDKey)\n\t// No sessionID in request's cookies.\n\tif err != nil {\n\t\treturn false\n\t}\n\tsessionID := cookie.Value\n\t// Atomic Check if sssion with given id contains in sessions map\n\ts.RLock()\n\t_, ok := s.sessions[sessionID]\n\ts.RUnlock()\n\n\treturn ok\n}", "func (p *AzureProvider) ValidateSession(ctx context.Context, s *sessions.SessionState) bool {\n\treturn validateToken(ctx, p, s.AccessToken, makeAzureHeader(s.AccessToken))\n}", "func verifyLoggedIn(resp *http.Response) bool {\n\tif resp.Request != nil && resp.Request.URL != nil {\n\t\treturn strings.HasPrefix(resp.Request.URL.String(), loggedinURLPrefix)\n\t}\n\treturn false\n}", "func IsLoggedIn(r *http.Request) bool {\n\tsession, err := loggedUserSession.Get(r, \"authenticated-user-session\")\n\tif err != nil || session.Values[\"username\"] != \"admin\" {\n\t\treturn false\n\t}\n\treturn true\n}", "func checkRequest(req *http.Request) error {\n\treturn nil\n}", "func VerifySession(sid string) bool {\n\tvar session Session\n\tif err := cache.Get(SessionId+SESSION_KEY, &session); err != nil {\n\t\treturn false\n\t}\n\treturn sid == session.Id\n}", "func IsSessionTokenValid(token string) bool {\n\t_, err := SessionTokenAuthenticate(token)\n\treturn err == nil\n}", "func (env *Env) session(r *http.Request) (*models.Session, error) {\n\tcValue := \"\"\n\tfor _, c := range r.Cookies() {\n\t\tif c.Name == \"cryptotax\" && c.Value != \"\" {\n\t\t\tcValue = c.Value\n\t\t}\n\t}\n\tif cValue == \"\" {\n\t\treturn nil, errors.New(\"unable to read cryptotax cookie\")\n\t}\n\n\treturn env.db.Session(cValue)\n}", "func AlreadyLoggedIn(w http.ResponseWriter, r *http.Request) bool {\n\tcookie, err := r.Cookie(\"session\")\n\tif err != nil {\n\t\treturn false\n\t}\n\ts := gs.sessions[cookie.Value]\n\t_, ok := gs.users[s.userName]\n\treturn ok\n}", "func (f *aclFilter) allowSession(node string) bool {\n\tif !f.enforceVersion8 {\n\t\treturn true\n\t}\n\treturn f.authorizer.SessionRead(node)\n}", "func CheckLogin(w http.ResponseWriter, r *http.Request) bool {\n\tCookieSession, err := r.Cookie(\"sessionid\")\n\tif err != nil {\n\t\tfmt.Println(\"No Such Cookies\")\n\t\tSession.Create()\n\t\tfmt.Println(Session.ID)\n\t\tSession.Expire = time.Now().Local()\n\t\tSession.Expire.Add(time.Hour)\n\t\treturn false\n\t}\n\tfmt.Println(\"Cookki Found\")\n\ttempSession := session.UserSession{UID: 0}\n\tLoggedIn := database.QueryRow(\"select user_id from sessions where session_id = ?\",\n\t\tCookieSession).Scan(&tempSession)\n\tif LoggedIn == nil {\n\t\treturn false\n\t}\n\treturn true\n\n}", "func ValidateSession(result *Session) (err error) {\n\tswitch result.View {\n\tcase \"default\", \"\":\n\t\terr = ValidateSessionView(result.Projected)\n\tdefault:\n\t\terr = goa.InvalidEnumValueError(\"view\", result.View, []interface{}{\"default\"})\n\t}\n\treturn\n}", "func SessionId(r *http.Request) string {\n\tif r == nil {\n\t\treturn \"\"\n\t}\n\n\tvar sessId string\n\n\t// Try to obtain from the query string\n\tif sessId = r.URL.Query().Get(\"session_id\"); sessId != \"\" {\n\t\treturn sessId\n\t}\n\tif sessId = r.URL.Query().Get(\"api_token\"); sessId != \"\" {\n\t\treturn sessId\n\t}\n\n\t// Try request body if we don't have a session from the query string\n\tif sessId = r.Form.Get(\"session_id\"); sessId != \"\" {\n\t\treturn sessId\n\t}\n\tif sessId = r.Form.Get(\"api_token\"); sessId != \"\" {\n\t\treturn sessId\n\t}\n\n\t// Finally try to extract from headers\n\t// Grab the header, and iterate over each instance (http allows multiple headers with the same key)\n\tfor hdr, extractor := range headerExtractors {\n\t\tfor _, v := range r.Header[hdr] {\n\t\t\tif sessId = extractor(hdr, v); sessId != \"\" {\n\t\t\t\tlog.Tracef(\"[Session] Session ID extracted from %s header: '%s'\", hdr, sessId)\n\t\t\t\treturn sessId\n\t\t\t}\n\t\t}\n\t}\n\n\treturn \"\"\n}", "func IsLoggedIn(r *http.Request) (bool, error) {\n\tsession, err := getSession(r)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tt := session.Values[\"accessToken\"]\n\tif t == nil {\n\t\treturn false, nil\n\t}\n\tstoredToken, ok := t.(string)\n\tif !ok {\n\t\treturn false, fmt.Errorf(\"bad type of %q value in session: %v\", \"accessToken\", err)\n\t}\n\tgp := session.Values[\"gplusID\"]\n\tif t == nil {\n\t\treturn false, nil\n\t}\n\tgplusId, ok := gp.(string)\n\tif !ok {\n\t\treturn false, fmt.Errorf(\"bad type of %q value in session: %v\", \"gplusID\", err)\n\t}\n\treturn storedToken != \"\" && isAllowed(gplusId), nil\n}", "func validateSession(f func(http.ResponseWriter, *http.Request)) func(http.ResponseWriter, *http.Request) { \n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\t// get cookie from request\n\t \tc, err := r.Cookie(\"JSESSION_ID\")\n\t\tif err != nil {\n\t\t\tif err == http.ErrNoCookie {\t\t\n\t\t\t\thttp.Redirect(w, r, \"/login?errorM=No session present in request\", http.StatusSeeOther)\n\t\t\t\treturn\t\n\t\t\t}\n\t\t\thttp.Redirect(w, r, \"/login?errorM=Not authorised\", http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\t\t// if no errors, validate the cookie\n\t\tsessToken := c.Value\n\t\tsv := strings.Split(sessToken, \"_\")\t\t\n\t\tif len(sv) != 2 {\t\t\n\t\t\thttp.Redirect(w, r, \"/login?errorM=Invalid cookie format\", http.StatusSeeOther)\n\t\t\treturn\t\n\t\t}\n\t\tuserName := sv[0]\n\t\texpSessToken := getStringFromDB(\"Sessions\", userName)\n\t\tif sv[1] != expSessToken {\n\t\t\thttp.Redirect(w, r, \"/login?errorM=Invalid session\", http.StatusSeeOther)\t\t\n\t\t\treturn\n\t\t}\n\n\t\t// after sucess, refresh the cookie\tto extend the session\n\t\t// Create a new random session token\n\t\tsessionToken, err := uuid.NewUUID()\n\t\tif err != nil {\n\t\t\thttp.Redirect(w, r, \"/login?errorM=Unable to create token\", http.StatusSeeOther)\n\t\t}\n\t\t// Set the token in the db, along with the userName\n\t\tupdateDBString(\"Sessions\", userName, sessionToken.String())\n\n\t\t// set the expiration time\n\t\texpires := time.Now().Add(600 * time.Second)\n\t\tck := http.Cookie{\n\t\t\tName: \"JSESSION_ID\",\n\t\t\tPath: \"/\",\n\t\t\tExpires: expires,\n\t\t\tValue: userName+\"_\"+sessionToken.String(),\n\t\t}\n\n\t\t// write the cookie to response\n\t\thttp.SetCookie(w, &ck)\n\n\t\t// if sucess process the handler\t\t\n\t\tf(w, r)\n\t\t\n\t}\n}", "func (m *RequestValidator) sessionTTL(ctx context.Context, identity tlsca.Identity, r types.AccessRequest) (time.Duration, error) {\n\tttl, err := m.truncateTTL(ctx, identity, r.GetAccessExpiry(), r.GetRoles())\n\tif err != nil {\n\t\treturn 0, trace.BadParameter(\"invalid session TTL: %v\", err)\n\t}\n\n\t// Before returning the TTL, validate that the value requested was smaller\n\t// than the maximum value allowed. Used to return a sensible error to the\n\t// user.\n\trequestedTTL := r.GetAccessExpiry().Sub(m.clock.Now().UTC())\n\tif !r.GetAccessExpiry().IsZero() && requestedTTL > ttl {\n\t\treturn 0, trace.BadParameter(\"invalid session TTL: %v greater than maximum allowed (%v)\", requestedTTL.Round(time.Minute), ttl.Round(time.Minute))\n\t}\n\n\treturn ttl, nil\n}", "func CheckIfAuthenticated(uuid string) bool {\n\t// Make sure a session exists for the extracted UUID\n\t_, sessionFound := sessions[uuid]\n\treturn sessionFound\n}", "func (req *Request) Session() *Session {\n\tif req.session == nil {\n\t\tif err := req.loadSession(); err != nil {\n\t\t\treq.app.Log.Error(\"failed to load session: %v\", err)\n\t\t}\n\t}\n\treturn req.session\n}", "func isAuthenticated(r *http.Request) bool {\n\ts, _ := Store.Get(r, \"sessid\")\n\tval, ok := s.Values[\"authenticated\"].(bool)\n\treturn ok && val\n}", "func validateState(req *http.Request, sess goth.Session) error {\n\trawAuthURL, err := sess.GetAuthURL()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tauthURL, err := url.Parse(rawAuthURL)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treqState := GetState(req)\n\n\toriginalState := authURL.Query().Get(\"state\")\n\tif originalState != \"\" && (originalState != reqState) {\n\t\treturn errors.New(\"state token mismatch\")\n\t}\n\treturn nil\n}", "func sessionKeyFromRequest(req *http.Request) (sessionKey string) {\n\theaderName := Headers[\"session\"]\n\tsessionKey = req.Header.Get(headerName)\n\tif sessionKey == \"\" {\n\t\t// fall back on cookie if we can\n\t\tsessionCookie, err := req.Cookie(headerName)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tsessionKey = sessionCookie.Value\n\t}\n\treturn\n}", "func (s *Service) CheckIfPersonHaveSession(c context.Context, sessionID ulid.ULID, personID ulid.ULID) (*Session, error) {\n\treturn &Session{}, nil\n}", "func session(rp core.Roundtrip) (sess model.Session, err error) {\n\tcookie := rp.GetCookie(\"_cookie\")\n\tif err == nil {\n\t\tsess = model.Session{Uuid: cookie}\n\t\tif ok, _ := sess.Check(); !ok {\n\t\t\terr = errors.New(\"Invalid session\")\n\t\t}\n\t}\n\treturn\n}", "func checkLoggedIn(w http.ResponseWriter, r *http.Request) (*ReqBody, error) {\n\t// get token from request header\n\tReqBody, err := getTknFromReq(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// check if the token is valid or not\n\t// if valid return the user currently logged in\n\tuser := verifyToken(ReqBody.Token)\n\tReqBody.UserDB = user\n\tlog.Println(\"checklogged in\", ReqBody)\n\treturn ReqBody, nil\n}", "func (client *KismetRestClient) ValidConnection() bool {\n\tif newRequest, err := http.NewRequest(\"GET\", client.Url + authCheckPath, strings.NewReader(\"\")) ; err == nil {\n\t\trequest = newRequest\n\t\tdefer request.Body.Close()\n\t} else {\n\t\treturn false\n\t}\n\n\trequest.AddCookie(&client.AuthCookie) // DOH\n\n\tif response, err := httpClient.Do(request) ; err == nil && response.StatusCode == 200 {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func ValidateSession(db *sql.DB, session string) error {\n\n\t// if isrunning and hasfiletarget then OK\n\t// if no rows (session doesn't exist), warning \"ErrNotFound\" -- ignore unless strict (TODO)\n\t// if not autostart and is always on, then OK\n\t// if autostart not isrunning then warning \"ErrNotRunning\" -- error\n\t// if isrunning and not hasfiletarget then warning \"ErrNoFileTarget\" -- error\n\n\tquery := `\n\t\tSELECT\t[session].[name],\n\t\t\t[session].[startup_state] AS [AutoStart],\n\t\t\t--[running].[create_time] AS [StartTime],\n\t\t\tCAST((CASE WHEN ([running].[create_time] IS NULL) THEN 0 ELSE 1 END) AS BIT)AS [IsRunning],\n\t\t\tCAST((CASE WHEN T.[event_session_address] IS NOT NULL THEN 1 ELSE 0 END) AS BIT) AS [HasFileTarget]\n\t\tFROM\t[sys].[server_event_sessions] AS [session]\n\t\tLEFT OUTER JOIN \n\t\t\t[sys].[dm_xe_sessions] AS [running] ON [running].[name] = [session].[name]\n\t\tLEFT OUTER JOIN \n\t\t[sys].[dm_xe_session_targets] T ON T.[event_session_address] = [running].[address]\n\t\t\t\t\t\tAND T.[target_name] = 'event_file'\n\t\tWHERE \t[session].[name] = ?`\n\n\tvar name string\n\tvar autostart, isrunning, hasfiletarget bool\n\terr := db.QueryRow(query, session).Scan(&name, &autostart, &isrunning, &hasfiletarget)\n\tif err == nil && isrunning && hasfiletarget {\n\t\treturn nil\n\t}\n\tif err == sql.ErrNoRows {\n\t\treturn ErrNotFound\n\t}\n\tif autostart && !isrunning {\n\t\treturn ErrNotRunning\n\t}\n\tif name == \"AlwaysOn_health\" && !autostart {\n\t\treturn nil\n\t}\n\tif isrunning && !hasfiletarget {\n\t\treturn ErrNoFileTarget\n\t}\n\tif !autostart && !isrunning {\n\t\treturn nil\n\t}\n\treturn fmt.Errorf(\"autostart: %v; running: %v filetarget: %v\", autostart, isrunning, hasfiletarget)\n}", "func (err *UnauthorizedError) isLoggedIn() bool {\n\treturn err.SessionId != 0 // SessionId is 0 for non-logged in users\n}", "func Validate(next echo.HandlerFunc) echo.HandlerFunc {\n\treturn func(c echo.Context) error {\n\t\ts, _ := CookieStore.Get(c.Request(), CookieName)\n\n\t\tif authn, ok := s.Values[\"authenticated\"].(bool); !ok || !authn {\n\t\t\tglog.V(2).Info(\"Unauthenticated request\")\n\t\t\treturn util.RedirectTo(c, \"login\")\n\t\t}\n\n\t\tif s.Values[\"username\"] == \"\" {\n\t\t\tglog.Error(\"authenticated session, but user not found\")\n\t\t\treturn c.NoContent(http.StatusInternalServerError)\n\t\t}\n\n\t\t// request contains authenticated cookie, call next\n\t\treturn next(c)\n\t}\n}", "func validateSessionTracker(testCtx *TestContext, sessionID string, reason string, invited []string) error {\n\tsessionTracker, err := testCtx.AuthClient.GetSessionTracker(testCtx.Context, sessionID)\n\tif err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tif sessionTracker.GetReason() != reason {\n\t\treturn trace.BadParameter(\"expected reason %q, got %q\", reason, sessionTracker.GetReason())\n\t}\n\tif !reflect.DeepEqual(sessionTracker.GetInvited(), invited) {\n\t\treturn trace.BadParameter(\"expected invited %q, got %q\", invited, sessionTracker.GetInvited())\n\t}\n\treturn nil\n}", "func ParseSessionRequest(request interface{}) (irma.RequestorRequest, error) {\n\tswitch r := request.(type) {\n\tcase irma.RequestorRequest:\n\t\treturn r, nil\n\tcase irma.SessionRequest:\n\t\treturn wrapSessionRequest(r)\n\tcase string:\n\t\treturn ParseSessionRequest([]byte(r))\n\tcase []byte:\n\t\tvar attempts = []irma.Validator{&irma.ServiceProviderRequest{}, &irma.SignatureRequestorRequest{}, &irma.IdentityProviderRequest{}}\n\t\tt, err := tryUnmarshalJson(r, attempts)\n\t\tif err == nil {\n\t\t\treturn t.(irma.RequestorRequest), nil\n\t\t}\n\t\tattempts = []irma.Validator{&irma.DisclosureRequest{}, &irma.SignatureRequest{}, &irma.IssuanceRequest{}}\n\t\tt, err = tryUnmarshalJson(r, attempts)\n\t\tif err == nil {\n\t\t\treturn wrapSessionRequest(t.(irma.SessionRequest))\n\t\t}\n\t\treturn nil, errors.New(\"Failed to JSON unmarshal request bytes\")\n\tdefault:\n\t\treturn nil, errors.New(\"Invalid request type\")\n\t}\n}", "func GetSession(res http.ResponseWriter, req *http.Request) (string, bool) {\n\tvar name string\n\tvar success bool\n\tcorrectlyLogIn := false\n\tuuid, ok := cookiesManager.GetCookieValue(req, CookieName)\n\tif ok {\n\t\tlog.Debugf(\"Found cookie. Value: %s\", uuid)\n\t\tname, success = authClient.GetRequest(uuid)\n\t\tif name != \"\" && success {\n\t\t\tlog.Debugf(\"Found session for key: %s with value: %s\", uuid, name)\n\t\t\tcorrectlyLogIn = true\n\t\t} else {\n\t\t\tlog.Debugf(\"No session found for key; %s\", uuid)\n\t\t\tcookiesManager.RemoveCookie(res, CookieName)\n\t\t}\n\t}\n\n\tlog.Debug(\"Cookie is missing\")\n\treturn name, correctlyLogIn\n}", "func ValidateNewSession(session *Session) error {\n\t// var err error\n\n\t// if session.Options, err = ParseOptions(); err != nil {\n\t// return err\n\t// }\n\n\tif *session.Options.Report != \"\" && filehandler.FileExists(*session.Options.Report) {\n\t\treturn fmt.Errorf(\"file: %s already exists\", *session.Options.Report)\n\t}\n\n\tif *session.Options.Load != \"\" {\n\t\tif !filehandler.FileExists(*session.Options.Load) {\n\t\t\treturn fmt.Errorf(\"session file %s does not exist or is not readable\", *session.Options.Load)\n\t\t}\n\t\tdata, err := ioutil.ReadFile(*session.Options.Load)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := json.Unmarshal(data, &session); err != nil {\n\t\t\treturn fmt.Errorf(\"session file %s is corrupt or not generated this version\", *session.Options.Load)\n\t\t}\n\t}\n\treturn nil\n}", "func IsLogIn(h httprouter.Handle) httprouter.Handle {\n\treturn func(w http.ResponseWriter, r *http.Request, ps httprouter.Params) {\n\n\t\tcookie, err := r.Cookie(\"kRtrima\") //Grab the cookie from the header\n\t\tif err != nil {\n\t\t\tswitch err {\n\t\t\tcase http.ErrNoCookie:\n\t\t\t\tLogger.Println(\"No Cookie was Found with Name kRtrima\")\n\t\t\t\th(w, r, ps)\n\t\t\t\treturn\n\t\t\tdefault:\n\t\t\t\tLogger.Println(\"No Cookie was Found with Name kRtrima\")\n\t\t\t\th(w, r, ps)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tLogger.Println(\"Cookie was Found with Name kRtrima\")\n\n\t\t// Create a BSON ObjectID by passing string to ObjectIDFromHex() method\n\t\tdocID, err := primitive.ObjectIDFromHex(cookie.Value)\n\t\tif err != nil {\n\t\t\tLogger.Printf(\"Cannot Convert %T type to object id\", cookie.Value)\n\t\t\tLogger.Println(err)\n\t\t\th(w, r, ps)\n\t\t\treturn\n\t\t}\n\n\t\tvar SP m.Session\n\t\tif err = m.Sessions.Find(\"_id\", docID, &SP); err != nil {\n\t\t\tLogger.Println(\"Cannot found a valid User Session!!\")\n\t\t\th(w, r, ps)\n\t\t\treturn\n\t\t\t//session is missing, returns with error code 403 Unauthorized\n\t\t}\n\n\t\tLogger.Println(\"Valid User Session was Found!!\")\n\n\t\tvar UP m.LogInUser\n\n\t\terr = m.Users.Find(\"salt\", SP.Salt, &UP)\n\t\tif err != nil {\n\t\t\tLogger.Println(\"Cannot Find user with salt\")\n\t\t\th(w, r, ps)\n\t\t\treturn\n\t\t}\n\n\t\tvar LIP m.LogInUser\n\n\t\terr = m.GetLogInUser(\"User\", &LIP, r)\n\t\tif err != nil {\n\t\t\tm.AddToHeader(\"User\", UP, r)\n\t\t\th(w, r, ps)\n\t\t\treturn\n\t\t} else if UP.Email != LIP.Email {\n\t\t\t//remove the user ID from the session\n\t\t\tr.Header.Del(\"User\")\n\t\t\tm.AddToHeader(\"User\", UP, r)\n\t\t\th(w, r, ps)\n\t\t\treturn\n\t\t}\n\n\t\th(w, r, ps)\n\t}\n}", "func (me TxsdCounterSimpleContentExtensionType) IsSession() bool { return me.String() == \"session\" }", "func (us *UploadSession) IsValid() *AppError {\n\tif !IsValidId(us.Id) {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.id.app_error\", nil, \"\", http.StatusBadRequest)\n\t}\n\n\tif err := us.Type.IsValid(); err != nil {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.type.app_error\", nil, \"\", http.StatusBadRequest).Wrap(err)\n\t}\n\n\tif !IsValidId(us.UserId) && us.UserId != UploadNoUserID {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.user_id.app_error\", nil, \"id=\"+us.Id, http.StatusBadRequest)\n\t}\n\n\tif us.Type == UploadTypeAttachment && !IsValidId(us.ChannelId) {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.channel_id.app_error\", nil, \"id=\"+us.Id, http.StatusBadRequest)\n\t}\n\n\tif us.CreateAt == 0 {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.create_at.app_error\", nil, \"id=\"+us.Id, http.StatusBadRequest)\n\t}\n\n\tif us.Filename == \"\" {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.filename.app_error\", nil, \"id=\"+us.Id, http.StatusBadRequest)\n\t}\n\n\tif us.FileSize <= 0 {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.file_size.app_error\", nil, \"id=\"+us.Id, http.StatusBadRequest)\n\t}\n\n\tif us.FileOffset < 0 || us.FileOffset > us.FileSize {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.file_offset.app_error\", nil, \"id=\"+us.Id, http.StatusBadRequest)\n\t}\n\n\tif us.Path == \"\" {\n\t\treturn NewAppError(\"UploadSession.IsValid\", \"model.upload_session.is_valid.path.app_error\", nil, \"id=\"+us.Id, http.StatusBadRequest)\n\t}\n\n\treturn nil\n}", "func CheckDupeRequest(env databaseConfig, request models.Request) (bool, error) {\n\tctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)\n\tdefer cancel()\n\tcnt, err := getReqCollection(env).CountDocuments(ctx, bson.M{\n\t\t\"singer\": request.Singer,\n\t\t\"song\": request.Song,\n\t\t\"playedtime\": bson.M{\"$exists\": false},\n\t})\n\tif err != nil {\n\t\tenv.GetLog().Printf(\"Failure when checking if request is a duplicate: %v\", err)\n\t\treturn false, err\n\t}\n\treturn cnt > 0, nil\n}", "func (this *Session) Before(txn *Txn) bool {\n\tif txn.Type() != \"html\" {\n\t\t//skip\n\t\treturn true\n\t}\n\t// log.Println(\"SESSION!\")\n\thttpWriter, err := ToHttpWriter(txn)\n\tif err != nil {\n\t\tlog.Printf(\"ERROR in session.before %s\", err)\n\t\treturn true //should we continue with the request?\n\t}\n\n\tcookie, err := httpWriter.HttpRequest.Cookie(\"session_id\")\n\n\tif err != nil {\n\t\t//create new session id\n\t\ttxn.Session.Put(\"session_id\", SessionId())\n\t\t// log.Println(\"Created session!\")\n\t} else {\n\t\t//load the session. \n\t\tsessionId := cookie.Value\n\t\t// log.Printf(\"Found session cookie! %s\", sessionId)\n\t\tbytes, ok := this.cache.Get(sessionId)\n\t\tif !ok {\n\t\t\t//create a new session, since the old one is gone\n\t\t\tsessionId = SessionId()\n\t\t\t// log.Printf(\"Old session expired, setting new one (%s)\", sessionId)\n\t\t} else {\n\t\t\terr = txn.Session.UnmarshalJSON(bytes)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Error unmarshaling json (%s) -> (%s)\", bytes, err)\n\t\t\t}\n\t\t}\n\t\ttxn.Session.Put(\"session_id\", sessionId)\n\t}\n\treturn true\n}", "func (serv *Server) VerifyUser(req *http.Request) (Session, error) {\n jwt := []byte(req.Header.Get(\"Session\"))\n if len(jwt) <= 0 {\n return Session{}, ErrNotLoggedIn\n }\n return UnwrapSessionToken(jwt, serv.macKey[:])\n}", "func (sc *SessionCreate) check() error {\n\tif _, ok := sc.mutation.Created(); !ok {\n\t\treturn &ValidationError{Name: \"created\", err: errors.New(`ent: missing required field \"Session.created\"`)}\n\t}\n\tif _, ok := sc.mutation.Updated(); !ok {\n\t\treturn &ValidationError{Name: \"updated\", err: errors.New(`ent: missing required field \"Session.updated\"`)}\n\t}\n\tif _, ok := sc.mutation.UserID(); !ok {\n\t\treturn &ValidationError{Name: \"user_id\", err: errors.New(`ent: missing required field \"Session.user_id\"`)}\n\t}\n\tif v, ok := sc.mutation.UserID(); ok {\n\t\tif err := session.UserIDValidator(v); err != nil {\n\t\t\treturn &ValidationError{Name: \"user_id\", err: fmt.Errorf(`ent: validator failed for field \"Session.user_id\": %w`, err)}\n\t\t}\n\t}\n\tif _, ok := sc.mutation.PeerID(); !ok {\n\t\treturn &ValidationError{Name: \"peer_id\", err: errors.New(`ent: missing required field \"Session.peer_id\"`)}\n\t}\n\tif v, ok := sc.mutation.PeerID(); ok {\n\t\tif err := session.PeerIDValidator(v); err != nil {\n\t\t\treturn &ValidationError{Name: \"peer_id\", err: fmt.Errorf(`ent: validator failed for field \"Session.peer_id\": %w`, err)}\n\t\t}\n\t}\n\tif _, ok := sc.mutation.SessionType(); !ok {\n\t\treturn &ValidationError{Name: \"session_type\", err: errors.New(`ent: missing required field \"Session.session_type\"`)}\n\t}\n\tif _, ok := sc.mutation.SessionStatus(); !ok {\n\t\treturn &ValidationError{Name: \"session_status\", err: errors.New(`ent: missing required field \"Session.session_status\"`)}\n\t}\n\treturn nil\n}", "func (a *Auth) Session(w http.ResponseWriter, r *http.Request) {\n\tuser := a.userstate.Username(r)\n\temail, err := a.userstate.Email(user)\n\tif err != nil {\n\t\tutil.Error(\"session: %s\", err.Error())\n\t}\n\tutil.OKWith(w, r, map[string]string{\n\t\t\"serialNumber\": user,\n\t\t\"emailAddress\": email,\n\t})\n}", "func (l *RemoteProvider) GetSession(req *http.Request) error {\n\tts, err := l.GetToken(req)\n\tif err != nil || ts == \"\" {\n\t\tlogrus.Infof(\"session not found\")\n\t\treturn err\n\t}\n\n\t_, err = l.VerifyToken(ts)\n\tif err != nil {\n\t\tlogrus.Infof(\"Token validation error : %v\", err.Error())\n\t\tnewts, err := l.refreshToken(ts)\n\t\tif err != nil {\n\t\t\treturn ErrTokenRefresh(err)\n\t\t}\n\t\t_, err = l.VerifyToken(newts)\n\t\tif err != nil {\n\t\t\tlogrus.Errorf(\"Validation of refreshed token failed : %v\", err.Error())\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func MissingSession(w http.ResponseWriter, r *http.Request, h *render.Renderer) {\n\tInternalError(w, r, h, errMissingSession)\n\treturn\n}", "func VerifySession(key string, user int) bool {\n\tvar sessions []Session\n\tdb.Where(\"user is ?\", user).Find(&sessions)\n\tfor i := range sessions {\n\t\tif sessions[i].Key == key {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (r *Request) Session() *Session {\n\treturn r.session\n}", "func (r *Request) Session() *Session {\n\treturn r.session\n}", "func getSession(r *http.Request) (*sessions.Session, error) {\n\tsession, err := store.Get(r, sessionName)\n\tif err != nil {\n\t\tif session.IsNew {\n\t\t\tglog.V(1).Infof(\"ignoring initial session fetch error since session IsNew: %v\\n\", err)\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"error fetching session: %v\", err)\n\t\t}\n\t}\n\treturn session, nil\n}", "func (authSvc *AuthService) validateSessionId(sessionId string) bool {\n\t\n\tvar parts []string = strings.Split(sessionId, \":\")\n\tif len(parts) != 2 {\n\t\tfmt.Println(\"Ill-formatted sessionId:\", sessionId)\n\t\treturn false\n\t}\n\t\n\tvar uniqueNonRandomValue string = parts[0]\n\tvar untrustedHash string = parts[1]\n\tvar empty = []byte{}\n\tvar actualSaltedHashBytes []byte = authSvc.computeHash(uniqueNonRandomValue).Sum(empty)\n\t\n\treturn untrustedHash == fmt.Sprintf(\"%x\", actualSaltedHashBytes)\n}", "func isRequestScopeValid(jwtData *JWTData, requestScope string) bool {\n\n\tif len(requestScope) > 0 {\n\n\t\ttokenScopes := strings.Split(jwtData.Scope, \" \")\n\n\t\tfor _, tokenScope := range tokenScopes {\n\t\t\tif requestScope == tokenScope {\n\t\t\t\treturn true\n\t\t\t}\n\n\t\t}\n\t\tlog.Warnf(\"No matching scopes found!\")\n\t\treturn false\n\t}\n\n\tlog.Infof(\"No scopes defined\")\n\treturn true\n}", "func (req *Request) loadSession() error {\n\treq.session = newSession(req.app.SessionDuration)\n\tcookie, err := req.GetCookie(req.app.SessionName)\n\tif err == http.ErrNoCookie {\n\t\treturn nil\n\t} else if err != nil {\n\t\treturn err\n\t}\n\treturn req.session.unmarshal(cookie.Value, req.app.SessionKey)\n}", "func (c DevSession) IsValid() bool {\n\treturn capnp.Client(c).IsValid()\n}", "func Validate(w http.ResponseWriter, r *http.Request) error {\n\tvar err error\n\tauth := r.Header.Get(\"Authorization\")\n\tif auth == \"\" {\n\t\tw.Header().Set(\"WWW-Authenticate\", `Bearer realm=\"Datadog Agent\"`)\n\t\terr = fmt.Errorf(\"no session token provided\")\n\t\thttp.Error(w, err.Error(), 401)\n\t\treturn err\n\t}\n\n\ttok := strings.Split(auth, \" \")\n\tif tok[0] != \"Bearer\" {\n\t\tw.Header().Set(\"WWW-Authenticate\", `Bearer realm=\"Datadog Agent\"`)\n\t\terr = fmt.Errorf(\"unsupported authorization scheme: %s\", tok[0])\n\t\thttp.Error(w, err.Error(), 401)\n\t\treturn err\n\t}\n\n\tif len(tok) < 2 || tok[1] != GetAuthToken() {\n\t\terr = fmt.Errorf(\"invalid session token\")\n\t\thttp.Error(w, err.Error(), 403)\n\t}\n\n\treturn err\n}", "func (f *Features) canSession(ctx context.Context, adminDB *mongo.Database) {\n\tcmd := bson.D{\n\t\t{\n\t\t\tKey: \"replSetGetStatus\",\n\t\t\tValue: 1,\n\t\t},\n\t}\n\tvar result replInfo\n\terr := adminDB.RunCommand(ctx, cmd).Decode(&result)\n\tif err != nil {\n\t\t// assume we don't have session support on error..\n\t\t// error code 76 will be thrown if replSet is not enabled.\n\t\treturn\n\t}\n\n\tf.HasSessions = result.Ok\n}", "func (m *memoryStorage) sessionChecker() {\n\tfor {\n\t\tm.mutex.Lock()\n\t\tfor sessionID, v := range m.sessions {\n\t\t\tif v.expires < time.Now().Unix() {\n\t\t\t\tdelete(m.sessions, sessionID)\n\t\t\t}\n\t\t}\n\t\tfor nonce, expires := range m.nonces {\n\t\t\tif time.Now().After(expires) {\n\t\t\t\tdelete(m.nonces, nonce)\n\t\t\t}\n\t\t}\n\t\tm.mutex.Unlock()\n\t\ttime.Sleep(sessionCheckInterval * time.Second)\n\t}\n}", "func (r Requests) HasRequest() bool {\n\treturn len(r) > 0\n}", "func isAuthorized(w http.ResponseWriter, r *http.Request) bool {\n\tusername, err := r.Cookie(\"username\")\n\tif err == nil {\n\t\tsessionID, err := r.Cookie(\"sessionID\")\n\t\tif err == nil {\n\t\t\tif sessionID.Value != \"\" && gostuff.SessionManager[username.Value] == sessionID.Value {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\tgostuff.Show404Page(w, r)\n\treturn false\n}", "func (p *hcAutonomywww) isLoggedIn(f http.HandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tlog.Debugf(\"isLoggedIn: %v %v %v %v\", remoteAddr(r), r.Method,\n\t\t\tr.URL, r.Proto)\n\n\t\temail, err := p.getSessionEmail(r)\n\t\tif err != nil {\n\t\t\tutil.RespondWithJSON(w, http.StatusUnauthorized, v1.ErrorReply{\n\t\t\t\tErrorCode: int64(v1.ErrorStatusNotLoggedIn),\n\t\t\t})\n\t\t\treturn\n\t\t}\n\n\t\t// Check if user is authenticated\n\t\tif email == \"\" {\n\t\t\tutil.RespondWithJSON(w, http.StatusUnauthorized, v1.ErrorReply{\n\t\t\t\tErrorCode: int64(v1.ErrorStatusNotLoggedIn),\n\t\t\t})\n\t\t\treturn\n\t\t}\n\n\t\tf(w, r)\n\t}\n}", "func (gs *Service) IsSessionCreated(_ context.Context, req *api.IsSessionCreatedRequest, rsp *api.IsSessionCreatedResponse) error {\n\trsp.IsRunning = 2 // Stands for false.\n\tfor _, value := range gs.sessions {\n\t\tif value.currentGameID == req.GameId {\n\t\t\trsp.IsRunning = 1\n\t\t\trsp.SessionId = value.sessionID\n\t\t\trsp.ChatSocketAddress = value.chatAddress\n\t\t\trsp.SessionSocketAddress = value.sessionAddress\n\t\t\tbreak\n\t\t}\n\t}\n\treturn nil\n}", "func (rp *Provider) SessionExist(ctx context.Context, sid string) (bool, error) {\n\tif ctx == nil {\n\t\tctx = context.TODO()\n\t}\n\treturn rp.c.Contains(ctx, sid), nil\n}" ]
[ "0.68616503", "0.6834911", "0.6676546", "0.66695386", "0.6328408", "0.63170826", "0.6307383", "0.62890023", "0.62826055", "0.6258494", "0.6256342", "0.6208338", "0.61973584", "0.61390716", "0.61384505", "0.61313486", "0.61105573", "0.6034856", "0.6005577", "0.59854245", "0.596214", "0.5949778", "0.5928381", "0.5921608", "0.5916601", "0.5915757", "0.5879217", "0.5872865", "0.5815646", "0.58015025", "0.5794654", "0.5772962", "0.5771158", "0.5770204", "0.5768653", "0.57593775", "0.5741437", "0.57406706", "0.57392275", "0.57150203", "0.5707722", "0.5706875", "0.57028306", "0.5697238", "0.56942636", "0.56816113", "0.56259686", "0.56170315", "0.5616076", "0.5612191", "0.55910563", "0.55756384", "0.5557492", "0.55378836", "0.5533658", "0.5532447", "0.5531924", "0.5529992", "0.5521445", "0.54752916", "0.54587674", "0.5452085", "0.5440852", "0.54360324", "0.54328656", "0.54274213", "0.54129475", "0.541098", "0.53975624", "0.5395334", "0.5357244", "0.53556705", "0.53508246", "0.5348685", "0.53463465", "0.53268725", "0.532088", "0.5307099", "0.5293638", "0.52924865", "0.52898043", "0.5286791", "0.5283404", "0.528212", "0.52738094", "0.5272913", "0.5262747", "0.5262747", "0.5248224", "0.5245479", "0.5239665", "0.5235738", "0.52322227", "0.52209854", "0.52132505", "0.5208596", "0.5208451", "0.52005565", "0.5192323", "0.51837564", "0.51789" ]
0.0
-1
Random returns random string from letters with length
func Random(length int, letters string) string { bytes := make([]byte, length) // A src.Int63() generates 63 random bits, enough for letterIdxMax characters! for index, cache, remain := length-1, randomize.Int63(), letterIdxMax; index >= 0; { if remain == 0 { cache, remain = randomize.Int63(), letterIdxMax } if idx := int(cache & letterIdxMask); idx < len(letters) { bytes[index] = letters[idx] index-- } cache >>= letterIdxBits remain-- } return *(*string)(unsafe.Pointer(&bytes)) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func generateRandomString(length int) string {\n\tletterRunes := []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\")\n\tb := make([]rune, length)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\treturn string(b)\n}", "func generateRandString(length int) string {\n\tb := make([]rune, length)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\treturn string(b)\n}", "func GenerateRandomString(lenght int) string {\n\tvar letterRunes = []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tb := make([]rune, lenght)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\n\treturn string(b)\n}", "func RandomString(length int) string {\n\tvar letters = []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tb := make([]rune, length)\n\trand.Seed(time.Now().UnixNano())\n\tfor i := range b {\n\t\tb[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(b)\n}", "func randomString(length int) string {\n\treturn stringWithCharset(length, charset)\n}", "func getRandomString(length int) (string, error) {\n\tbuf := make([]byte, length)\n\tif _, err := rand.Read(buf); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor i := 0; i < length; {\n\t\tidx := int(buf[i] & letterIdxMask)\n\t\tif idx < letterSize {\n\t\t\tbuf[i] = letters[idx]\n\t\t\ti++\n\t\t} else {\n\t\t\tif _, err := rand.Read(buf[i : i+1]); err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\t\t}\n\t}\n\treturn string(buf), nil\n}", "func RandomString(length int) string {\n\tb := make([]rune, length)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\treturn string(b)\n}", "func RandomString(len int) string {\n\tbytes := make([]byte, len)\n\tfor i := 0; i < len; i++ {\n\t\tbytes[i] = byte(65 + rand.Intn(25)) // A=65 and Z = 65+25\n\t}\n\treturn string(bytes)\n}", "func StrRandom(length int) string {\n\tresult := make([]rune, length)\n\tfor i := range result {\n\t\tresult[i] = alphaNumeric[rand.Intn(len(alphaNumeric))]\n\t}\n\treturn string(result)\n}", "func randomString(prefix string, length int) string {\n\tb := make([]rune, length)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\treturn prefix + string(b)\n}", "func RandomString(strlen int) string {\n rand.Seed(time.Now().UTC().UnixNano())\n const chars = \"abcdefghijklmnopqrstuvwxyz\"\n result := make([]byte, strlen)\n for i := 0; i < strlen; i++ {\n result[i] = chars[rand.Intn(len(chars))]\n }\n return string(result)\n}", "func RandomString(ln int) string {\n\tletters := []rune(\"1234567890abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tb := make([]rune, ln)\n\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\tfor i := range b {\n\t\tb[i] = letters[r.Intn(len(letters))]\n\t}\n\n\treturn string(b)\n}", "func RandomString(ln int) string {\n\tletters := []rune(\"1234567890abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tb := make([]rune, ln)\n\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\tfor i := range b {\n\t\tb[i] = letters[r.Intn(len(letters))]\n\t}\n\n\treturn string(b)\n}", "func randomString() string {\n\tr := make([]rune, 20)\n\tfor i := range r {\n\t\tr[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(r)\n}", "func RandomString(length int) string {\n\tsrc := rand.NewSource(time.Now().UnixNano())\n\tb := make([]byte, length)\n\tfor i, cache, remain := length-1, src.Int63(), letterIndexMax; i >= 0; {\n\t\tif remain == 0 {\n\t\t\tcache, remain = src.Int63(), letterIndexMax\n\t\t}\n\t\tif idx := int(cache & letterIndexMask); idx < len(letterBytes) {\n\t\t\tb[i] = letterBytes[idx]\n\t\t\ti--\n\t\t}\n\t\tcache >>= letterIndexBits\n\t\tremain--\n\t}\n\n\treturn string(b)\n}", "func StringRand(length int) string {\n\treturn StringRandWithCharset(length, CharsetDefault)\n}", "func RandomString(length int) string {\n\tsb := strings.Builder{}\n\n\trand.Seed(time.Now().UnixNano())\n\tfor i := 0; i < length; i++ {\n\t\tsb.WriteByte(_letters[rand.Intn(len(_letters))])\n\t}\n\n\treturn sb.String()\n}", "func RandomString(n int) string {\n\tvar letter = []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\")\n\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letter[rand.Intn(len(letter))]\n\t}\n\treturn string(b)\n}", "func RandomString(n int) string {\n\tvar letter = []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\")\n\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letter[rand.Intn(len(letter))]\n\t}\n\treturn string(b)\n}", "func GenerateRandomString(length int) string {\n\tif length > 0 {\n\t\trand.Seed(time.Now().UnixNano())\n\t\tchars := make([]rune, length)\n\t\tfor i := range chars {\n\t\t\tchars[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t\t}\n\n\t\treturn string(chars)\n\t}\n\n\treturn \"\"\n}", "func RandomString(length uint) (string, error) {\n\tr := make([]byte, length)\n\tbs := int(float64(length) * 1.3)\n\tvar err error\n\tfor i, j, rb := 0, 0, []byte{}; uint(i) < length; j++ {\n\t\tif j%bs == 0 {\n\t\t\trb, err = RandomBytes(uint(bs))\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\t\t}\n\t\tif idx := uint(rb[j%int(length)] & bitmask); idx < uint(len(letters)) {\n\t\t\tr[i] = letters[idx]\n\t\t\ti++\n\t\t}\n\t}\n\n\treturn string(r), nil\n}", "func randomStr(n int) string {\n\tletterRunes := []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\treturn string(b)\n}", "func randomString(len int) string {\n\tbytes := make([]byte, len)\n\tfor i := 0; i < len; i++ {\n\t\tbytes[i] = byte(randomInt(65, 90))\n\t}\n\treturn string(bytes)\n}", "func randomString(len int) string {\n\tbytes := make([]byte, len)\n\tfor i := 0; i < len; i++ {\n\t\tbytes[i] = byte(randomInt(65, 90))\n\t}\n\treturn string(bytes)\n}", "func randomString(len int) string {\n\tbytes := make([]byte, len)\n\tfor i := 0; i < len; i++ {\n\t\tbytes[i] = byte(randomInt(65, 90))\n\t}\n\treturn string(bytes)\n}", "func randomString(len int) string {\n\tbytes := make([]byte, len)\n\tfor i := 0; i < len; i++ {\n\t\tbytes[i] = byte(randomInt(65, 90))\n\t}\n\treturn string(bytes)\n}", "func RandomString(length int) string {\n\trunes := make([]rune, length)\n\tfor i := range runes {\n\t\trunes[i] = allowedCharactersRunes[rand.Intn(len(allowedCharactersRunes))]\n\t}\n\treturn string(runes)\n}", "func RandomString(n int) *string {\n\tvar letter = []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\")\n\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letter[rand.Intn(len(letter))]\n\t}\n\n\ts := string(b)\n\treturn &s\n}", "func randString(length int) string {\n\tb := make([]rune, length)\n\tfor i := range b {\n\t\tb[i] = runes[rand.Intn(len(runes))]\n\t}\n\treturn string(b)\n}", "func RandomString(len int) string {\n\tbytes := make([]byte, len)\n\tfor i := 0; i < len; i++ {\n\t\tbytes[i] = byte(65 + rand.Intn(25))\n\t}\n\treturn string(bytes)\n}", "func RandomString(strlen int) string {\r\n\trand.Seed(time.Now().UTC().UnixNano())\r\n\tconst chars = \"abcdefghijklmnopqrstuvwxyz0123456789\"\r\n\tresult := make([]byte, strlen)\r\n\tfor i := 0; i < strlen; i++ {\r\n\t\tresult[i] = chars[rand.Intn(len(chars))]\r\n\t}\r\n\treturn string(result)\r\n}", "func RandomString(strlen int) string {\r\n\trand.Seed(time.Now().UTC().UnixNano())\r\n\tconst chars = \"abcdefghijklmnopqrstuvwxyz0123456789\"\r\n\tresult := make([]byte, strlen)\r\n\tfor i := 0; i < strlen; i++ {\r\n\t\tresult[i] = chars[rand.Intn(len(chars))]\r\n\t}\r\n\treturn string(result)\r\n}", "func randomString(len int) string {\n\tbytes := make([]byte, len)\n\tfor i := 0; i < len; i++ {\n\t\tbytes[i] = byte(randomInt(97, 122))\n\t}\n\treturn string(bytes)\n}", "func randString(n int) string {\n\n\talpha := make([]rune, n)\n\n\trand.Seed( time.Now().UnixNano() )\n\n\tfor i := 0; i < n; i++ {\n\t\talpha[i] = letters[rand.Intn(len(letters))]\n\t}\n\n\treturn string(alpha)\n}", "func genString(n int) string {\n\tvalidChar := []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\ts := make([]rune, n)\n\tfor i := range s {\n\t\ts[i] = validChar[rand.Intn(len(validChar))]\n\t}\n\treturn string(s)\n}", "func RandomString(n int) string {\n\tb := make([]byte, n)\n\tfor i := range b {\n\t\tb[i] = letterBytes[rand.Intn(len(letterBytes))]\n\t}\n\treturn string(b)\n}", "func RandomString(length int) string {\n\n\tconst charset = \"abcdefghijklmnopqrstuvwxyz\" +\n\t\t\"ABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\" +\n\t\t\"1234567890\"\n\n\tvar seededRand *rand.Rand = rand.New(\n\t\trand.NewSource(time.Now().UnixNano()))\n\n\tb := make([]byte, length)\n\tfor i := range b {\n\t\tb[i] = charset[seededRand.Intn(len(charset))]\n\t}\n\treturn string(b)\n}", "func randomString(n int) string {\n\trand.Seed(time.Now().UnixNano())\n\n\tvar letter = []rune(runeString)\n\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letter[rand.Intn(len(letter))]\n\t}\n\treturn string(b)\n}", "func RandomAlphanumericString(strlen int) []byte {\n\tresult := make([]byte, strlen)\n\tfor i := range result {\n\t\tresult[i] = PasswordCharNumSpace[random.Intn(len(PasswordCharNumSpace))]\n\t}\n\tresult[0] = PasswordCharSpace[random.Intn(len(PasswordCharSpace))]\n\treturn result\n}", "func RandStr(length int) string {\n\tstr := \"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\"\n\treturn RandSomeStr(str, length)\n}", "func RandomString(n int) string {\n\trand.Seed(time.Now().UnixNano())\n\n\tb := make([]rune, n)\n\n\tfor i := range b {\n\t\tb[i] = letters[rand.Intn(len(letters))]\n\t}\n\n\treturn string(b)\n}", "func generateRandomString(n int) string {\n\tvar letter = []rune(\"abcdefghijklmnopqrstuvwxyz0123456789\")\n\n\tb := make([]rune, n)\n\trand.Seed(time.Now().UnixNano())\n\tfor i := range b {\n\t\tb[i] = letter[rand.Intn(len(letter))]\n\t}\n\treturn string(b)\n}", "func randString(length int) string {\n\tcharset := \"abcdefghijklmnopqrstuvwxyz\" +\n\t\t\"ABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\"\n\tvar seededRand *rand.Rand = rand.New(\n\t\trand.NewSource(time.Now().UnixNano()))\n\tb := make([]byte, length)\n\tfor i := range b {\n\t\tb[i] = charset[seededRand.Intn(len(charset))]\n\t}\n\treturn string(b)\n}", "func RandomString(n int) string {\n\tb := make([]byte, n)\n\tfor i := range b {\n\t\tb[i] = letterBytes[rand.Int63()%int64(len(letterBytes))]\n\t}\n\treturn string(b)\n}", "func RandString(n int) string {\n\tletters := []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tlength := len(letters)\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letters[rand.Intn(length)]\n\t}\n\treturn string(b)\n}", "func RandomString(length int, strChars string) string {\n\trand.Seed(time.Now().UnixNano())\n\tchars := []rune(strChars)\n\tfmt.Println(chars)\n\tvar b strings.Builder\n\tfor i := 0; i < length; i++ {\n\t\tb.WriteRune(chars[rand.Intn(len(chars))])\n\t}\n\treturn b.String()\n}", "func GetRandomString(length int) string {\n\tb := make([]byte, length)\n\trnd := rand.New(&source{})\n\n\tfor i := range b {\n\t\tc := rnd.Intn(allowedCharsSize)\n\t\tb[i] = allowedChars[c]\n\t}\n\n\treturn string(b)\n}", "func GetRandomString(length int) string {\n\tstr := \"0123456789abcdefghijklmnopqrstuvwxyz\"\n\tbytes := []byte(str)\n\tresult := []byte{}\n\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\tfor i := 0; i < length; i++ {\n\t\tresult = append(result, bytes[r.Intn(len(bytes))])\n\t}\n\treturn string(result)\n}", "func genString(length int) string {\n\tb := make([]byte, length)\n\tfor i := range b {\n\t\tb[i] = charset[seededRand.Intn(len(charset))]\n\t}\n\treturn string(b)\n}", "func RandomString(n int) string {\n\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\tvar str string\n\tlength := len(alphanum)\n\tfor i := 0; i < n; i++ {\n\t\ta := alphanum[r.Intn(len(alphanum))%length]\n\t\tstr += string(a)\n\t}\n\treturn str\n}", "func Random(length uint) []byte {\n\tvar (\n\t\tr = make([]byte, length)\n\t\tletterBytes = \"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\"\n\t)\n\tif _, err := rand.Reader.Read(r); err != nil {\n\t\tlog.MustNewLogger(&config.DefaultConfig.Log).Warning(err)\n\t\tfor i := range r {\n\t\t\tr[i] = letterBytes[mrand.Intn(len(letterBytes))]\n\t\t}\n\t}\n\treturn r\n}", "func RandString(n int, allowedChars ...[]rune) string {\n\tvar letters []rune\n\tif len(allowedChars) == 0 {\n\t\tletters = defaultLetters\n\t} else {\n\t\tletters = allowedChars[0]\n\t}\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(b)\n}", "func randomString(n int) string {\n\tresult := make([]byte, n)\n\tfor i := range result {\n\t\tresult[i] = charSet[rnd.Intn(len(charSet))]\n\t}\n\treturn string(result)\n}", "func RandString(n int) string {\n\tvar letters = []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(b)\n}", "func RandomString(length int) string {\n\trand.Seed(time.Now().UnixNano())\n\tchars := []rune(\"ABCDEFGHIJKLMNOPQRSTUVWXYZ\" + \"abcdefghijklmnopqrstuvwxyz\" + \"0123456789\")\n\tvar b strings.Builder\n\tfor i := 0; i < length; i++ {\n\t\tb.WriteRune(chars[rand.Intn(len(chars))])\n\t}\n\tstr := b.String()\n\treturn str\n}", "func randString(n int) string {\n\tb := make([]byte, n)\n\t_, err := rand.Read(b)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfor i := range b {\n\t\tb[i] = 'a' + (b[i] % 26)\n\t}\n\treturn string(b)\n}", "func randomString(n int) string {\n\tb := make([]byte, n)\n\t// A src.Int63() generates 63 random bits, enough for letterIdxMax characters!\n\tfor i, cache, remain := n-1, src.Int63(), letterIdxMax; i >= 0; {\n\t\tif remain == 0 {\n\t\t\tcache, remain = src.Int63(), letterIdxMax\n\t\t}\n\t\tif idx := int(cache & letterIdxMask); idx < len(letterBytes) {\n\t\t\tb[i] = letterBytes[idx]\n\t\t\ti--\n\t\t}\n\t\tcache >>= letterIdxBits\n\t\tremain--\n\t}\n\n\treturn string(b)\n}", "func GenerateRandString(n int) string {\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\treturn string(b)\n}", "func randString(strLen int) (randStr string) {\n\tstrChars := \"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ1234567890\"\n\tfor i := 0; i < strLen; i++ {\n\t\trandUint := rand.Uint32()\n\t\tpos := randUint % uint32(len(strChars))\n\t\trandStr += string(strChars[pos])\n\t}\n\treturn\n}", "func generateRandomString(n int) (string, error) {\n\tconst letters = \"0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz-\"\n\tbytes, err := generateRandomBytes(n)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tfor i, b := range bytes {\n\t\tbytes[i] = letters[b%byte(len(letters))]\n\t}\n\treturn string(bytes), nil\n}", "func randomString(n int) string {\n\tvar src = rand.NewSource(time.Now().UnixNano())\n\tb := make([]byte, n)\n\t// A src.Int63() generates 63 random bits, enough for letterIdxMax characters!\n\tfor i, cache, remain := n-1, src.Int63(), letterIdxMax; i >= 0; {\n\t\tif remain == 0 {\n\t\t\tcache, remain = src.Int63(), letterIdxMax\n\t\t}\n\t\tif idx := int(cache & letterIdxMask); idx < len(letterBytes) {\n\t\t\tb[i] = letterBytes[idx]\n\t\t\ti--\n\t\t}\n\t\tcache >>= letterIdxBits\n\t\tremain--\n\t}\n\n\treturn string(b)\n}", "func RandString(n int) string {\n\tb := make([]byte, n)\n\tfor i := range b {\n\t\tb[i] = letterBytes[rand.Intn(len(letterBytes))]\n\t}\n\n\treturn string(b)\n}", "func RandString(n int) string {\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(b)\n}", "func RandString(n int) string {\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(b)\n}", "func RandomString(n int) string {\n\tconst alphanum = \"0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz\"\n\tvar bytes = make([]byte, n)\n\trand.Read(bytes)\n\tfor i, b := range bytes {\n\t\tbytes[i] = alphanum[b%byte(len(alphanum))]\n\t}\n\treturn string(bytes)\n}", "func randString(n int) string {\n\toutput := make([]byte, n)\n\n\t// We will take n bytes, one byte for each character of output.\n\trandomness := make([]byte, n)\n\n\t// read all random\n\t_, err := rand.Read(randomness)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tl := len(letterBytes)\n\t// fill output\n\tfor pos := range output {\n\t\t// get random item\n\t\trandom := uint8(randomness[pos])\n\n\t\t// random % 64\n\t\trandomPos := random % uint8(l)\n\n\t\t// put into output\n\t\toutput[pos] = letterBytes[randomPos]\n\t}\n\n\treturn string(output)\n}", "func (n *N36) Random(l int) string {\n\tmax := len(n.charset)\n\trunes := []rune(n.charset)\n\n\tb := make([]rune, l)\n\tfor i := range b {\n\t\tb[i] = runes[n.seedRand.Intn(max)]\n\t}\n\n\treturn string(b)\n}", "func RandString(length int, letter letter) string {\n\tb := make([]byte, length)\n\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\tfor i := range b {\n\t\tb[i] = letter[r.Int63()%int64(len(letter))]\n\t}\n\treturn string(b)\n}", "func RandomString(n int) string {\n\tb := make([]byte, n)\n\n\tfor i, cache, remain := n-1, rnd.Int63(), letterIdxMax; i >= 0; {\n\t\tif remain == 0 {\n\t\t\tcache, remain = rnd.Int63(), letterIdxMax\n\t\t}\n\t\tif idx := int(cache & letterIdxMask); idx < len(letterBytes) {\n\t\t\tb[i] = letterBytes[idx]\n\t\t\ti--\n\t\t}\n\t\tcache >>= letterIdxBits\n\t\tremain--\n\t}\n\n\treturn string(b)\n}", "func GenerateRandomAlphaNumericString(length int) string {\n\trand.Seed(time.Now().UnixNano())\n\tnumbersAndLetters := []rune(\"0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tif length < 1 {\n\t\treturn \"\"\n\t}\n\tresult := make([]rune, length)\n\tfor i := 0; i < length; i++ {\n\t\tresult[i] = numbersAndLetters[rand.Intn(len(numbersAndLetters))]\n\t}\n\treturn string(result)\n}", "func RandomString(n int) string {\n\tresult := make([]byte, n)\n\tfor i := range result {\n\t\tresult[i] = CharSet[rnd.Intn(len(CharSet))]\n\t}\n\treturn string(result)\n}", "func RandomString(n int) string {\n\tresult := make([]byte, n)\n\tfor i := range result {\n\t\tresult[i] = CharSet[rnd.Intn(len(CharSet))]\n\t}\n\treturn string(result)\n}", "func RandomString(length int) string {\n\tvar seed *rand.Rand = rand.New(rand.NewSource(time.Now().UTC().UnixNano()))\n\n\tb := make([]byte, length)\n\tfor idx := range b {\n\t\tb[idx] = charset[seed.Intn(len(charset))]\n\t}\n\treturn string(b)\n}", "func generateRandomString(n int) (string, error) {\n\tret := make([]byte, n)\n\tfor i := 0; i < n; i++ {\n\t\tnum, err := rand.Int(rand.Reader, big.NewInt(int64(len(LETTERS))))\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tret[i] = LETTERS[num.Int64()]\n\t}\n\n\treturn string(ret), nil\n}", "func RandString(n int) string {\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\treturn string(b)\n}", "func RandString(n int) string {\n\tb := make([]rune, n)\n\tfor i := range b {\n\t\tb[i] = letterRunes[rand.Intn(len(letterRunes))]\n\t}\n\treturn string(b)\n}", "func RandomString(maxlen int, charset string) string {\n\tvar s string\n\tfor i := 0; i < rand.Intn(maxlen)+1; i++ {\n\t\ts = s + string(charset[rand.Intn(len(charset))])\n\t}\n\treturn s\n}", "func RandomChars(n int) string {\n\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\ts := \"\"\n\tfor i := 0; i < n; i++ {\n\t\tindex := r.Int() % (len(alphaNum) - 1)\n\t\ts += alphaNum[index : index+1]\n\t}\n\treturn s\n}", "func TestRandString(t *testing.T) {\n\tassert := asrt.New(t)\n\tstringLengths := []int{2, 4, 8, 16, 23, 47}\n\n\tfor _, stringLength := range stringLengths {\n\t\ttestString := util.RandString(stringLength)\n\t\tassert.Equal(len(testString), stringLength, fmt.Sprintf(\"Generated string is of length %d\", stringLengths))\n\t}\n\tlb := \"a\"\n\tutil.SetLetterBytes(lb)\n\ttestString := util.RandString(1)\n\tassert.Equal(testString, lb)\n}", "func _() string {\n\trand.Seed(time.Now().UnixNano())\n\tchars := []rune(\"ABCDEFGHIJKLMNOPQRSTUVWXYZÅÄÖ\" +\n\t\t\"abcdefghijklmnopqrstuvwxyzåäö\" +\n\t\t\"0123456789\")\n\tlength := 12\n\tvar b strings.Builder\n\tfor i := 0; i < length; i++ {\n\t\tb.WriteRune(chars[rand.Intn(len(chars))])\n\t}\n\treturn b.String()\n}", "func GenerateRandomString() string {\n\tvar letterBytes = []rune(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\n\tbytesToWrite := make([]rune, 1024)\n\tfor i := range bytesToWrite {\n\t\tbytesToWrite[i] = letterBytes[rand.Intn(len(letterBytes))]\n\t}\n\n\treturn string(bytesToWrite)\n}", "func RandomString(rand *rand.Rand, size int) string {\n\tsb := strings.Builder{}\n\tfor sb.Len() <= size {\n\t\tsb.WriteRune(RandomRune(rand, 2, 5))\n\t}\n\tret := sb.String()\n\t_, lastRuneSize := utf8.DecodeLastRuneInString(ret)\n\treturn ret[0 : len(ret)-lastRuneSize]\n}", "func randstring(size int) string {\n\trval := make([]byte,size)\n\trand.Seed(time.Now().UnixNano())\n\tfor i:=0; i<size; i++ {\n\t\trval[i] = byte(rand.Intn(26) + 97)\n\t}\n\treturn string(rval)\n}", "func RandString(n int) string {\n\tconst letterBytes = \"1234567890abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\"\n\tif n < 0 {\n\t\tn = 0\n\t}\n\tb := make([]byte, n)\n\tfor i := range b {\n\t\tb[i] = letterBytes[rand.NewSource(time.Now().UnixNano()).Int63()%int64(len(letterBytes))]\n\t}\n\treturn string(b)\n}", "func randString(n int) string {\n\tconst alphanum = \"0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz\"\n\tvar bytes = make([]byte, n)\n\trand.Read(bytes)\n\tfor i, b := range bytes {\n\t\tbytes[i] = alphanum[b%byte(len(alphanum))]\n\t}\n\treturn string(bytes)\n}", "func RandString(n int) string {\n\tconst letterBytes = \"abcdefghijklmnopqrstuvwxyz\"\n\tb := make([]byte, n)\n\tfor i := range b {\n\t\tb[i] = letterBytes[rand.Intn(len(letterBytes))]\n\t}\n\treturn string(b)\n}", "func RandomString(strSize int) string {\n\n\tdictionary := \"0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz\"\n\n\tvar bytes = make([]byte, strSize)\n\trand.Read(bytes)\n\tfor k, v := range bytes {\n\t\tbytes[k] = dictionary[v%byte(len(dictionary))]\n\t}\n\treturn string(bytes)\n}", "func Random() string {\n\tpassword, _ := RandomLength(4, \"en\")\n\treturn password\n}", "func RandomString() string {\n\tvar letter = []rune(\"ABCDEFGHIJKLMNOPQRSTUVWXYZ\")\n\tvar number = []rune(\"0123456789\")\n\n\tb := make([]rune, 2)\n\tfor i := 0; i < 2; i++ {\n\t\tb[i] = letter[rand.Intn(len(letter))]\n\t}\n\n\tc := make([]rune, 3)\n\tfor i := 0; i < 3; i++ {\n\t\tc[i] = number[rand.Intn(len(number))]\n\t}\n\n\treturn string(append(b, c...))\n}", "func RandomString() string {\n\tsuffix := make([]byte, randSuffixLen)\n\n\tfor i := range suffix {\n\t\tsuffix[i] = letterBytes[rand.Intn(len(letterBytes))]\n\t}\n\treturn string(suffix)\n}", "func getRandString(n int) string {\n\tpool := []rune(\"ABCDEFGHIJKLMNOPQRSTUVWXYZÜÖÄabcdefghijklmnopqrstuvwxyzüöä\")\n\ts := make([]rune, n)\n\n\tfor pos := range s {\n\t\ts[pos] = pool[rand.Intn(len(pool))]\n\t}\n\n\treturn string(s)\n}", "func makeRandomString(bytesLength int) []byte {\n\tbyteVar := make([]byte, bytesLength)\n\tchars := \"abcdefghijklmnopqrstuvwxyz123456789\" // our posibilities\n\tfor i := range byteVar {\n\t\tx := genPseudoRand()\n\t\tbyteVar[i] = chars[x.Intn(len(chars))]\n\t}\n\treturn byteVar\n}", "func RandString(n int) string {\n\trandom := rand.New(rand.NewSource(time.Now().UnixNano()))\n\tb := make([]byte, n)\n\tfor i := range b {\n\t\tb[i] = letterBytes[random.Intn(len(letterBytes))]\n\t}\n\treturn string(b)\n}", "func RandomString(length int) string {\n\trandomString := \"\"\n\n\tfor len(randomString) < length {\n\t\trandomString += strconv.Itoa(rand.Int())\n\t}\n\n\treturn randomString[:length]\n}", "func GenerateRandomString(stringLen int) string {\n\tb := make([]byte, stringLen)\n\tfor i := range b {\n\t\tb[i] = charset[seededRand.Intn(len(charset))]\n\t}\n\treturn string(b)\n}", "func randStringBytesRmndr() string {\n\n\tb := make([]byte, 30)\n\tfor i := range b {\n\t\tb[i] = letterBytes[rand.Int63()%int64(len(letterBytes))]\n\t}\n\treturn string(b)\n}", "func RandString(n int) string {\n\tb := make([]byte, n)\n\tfor i := range b {\n\t\tb[i] = letterBytes[rand.Int63()%int64(len(letterBytes))]\n\t}\n\treturn string(b)\n}", "func randomLetter() string {\n\treturn string('A' + rune(rand.Intn(26)))\n}", "func RandomString(length int) string {\n\tseed := rand.New(rand.NewSource(time.Now().UnixNano()))\n\n\tb := make([]byte, length)\n\tfor i := range b {\n\t\tb[i] = charset[seed.Intn(len(charset))]\n\t}\n\n\treturn string(b)\n}", "func RandString(length int) string {\n\trand.Seed(time.Now().UnixNano())\n\trs := make([]string, length)\n\tfor start := 0; start < length; start++ {\n\t\tt := rand.Intn(3)\n\t\tif t == 0 {\n\t\t\trs = append(rs, strconv.Itoa(rand.Intn(10)))\n\t\t} else if t == 1 {\n\t\t\trs = append(rs, string(rand.Intn(26)+65))\n\t\t} else {\n\t\t\trs = append(rs, string(rand.Intn(26)+97))\n\t\t}\n\t}\n\treturn strings.Join(rs, \"\")\n}" ]
[ "0.83070576", "0.8210682", "0.8164084", "0.8048164", "0.802901", "0.80230534", "0.8022366", "0.8022267", "0.7973639", "0.79713494", "0.796581", "0.79470444", "0.79470444", "0.7938094", "0.79303855", "0.7924201", "0.79141414", "0.79013175", "0.79013175", "0.7890087", "0.7862105", "0.7859868", "0.78406173", "0.78406173", "0.78406173", "0.78406173", "0.78329283", "0.7813987", "0.7810244", "0.7802205", "0.77975094", "0.77975094", "0.7790116", "0.77616936", "0.77492464", "0.77461797", "0.77264696", "0.7713609", "0.7683104", "0.7681767", "0.76797754", "0.76686656", "0.7652634", "0.7630214", "0.7618677", "0.76135653", "0.7613087", "0.7607383", "0.7595004", "0.75752467", "0.7565157", "0.7556467", "0.7555839", "0.7552847", "0.75401795", "0.7540014", "0.75369453", "0.7520699", "0.75195044", "0.7518491", "0.7515", "0.7513931", "0.75011003", "0.75011003", "0.75003284", "0.7495379", "0.7493945", "0.7485497", "0.747867", "0.7477244", "0.74759835", "0.74759835", "0.74689287", "0.7464549", "0.7459885", "0.7459885", "0.745506", "0.7453166", "0.7448064", "0.7447753", "0.74407893", "0.74384314", "0.74356127", "0.74275523", "0.74221736", "0.7419652", "0.7418431", "0.74183387", "0.7418106", "0.7403864", "0.74035084", "0.7399802", "0.7399285", "0.73837346", "0.7381847", "0.7377817", "0.7364146", "0.7354461", "0.73469776", "0.73452735" ]
0.8115874
3
Fetch the updated stats
func (s *StatFS) Fetch() error { return syscall.Statfs(s.path, &s.stat) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *StatsCollector) fetchStats() (map[string]string, error) {\n\n\t// TODO measure rpc time\n\t//timer := prometheus.NewTimer(rpc_request_duration)\n\t//defer timer.ObserveDuration()\n\n\t// establish connection to Kamailio server\n\tvar err error\n\tvar conn net.Conn\n\tif c.kamailioHost == \"\" {\n\t\tlog.Debug(\"Requesting stats from kamailio via domain socket \", c.socketPath)\n\t\tconn, err = net.Dial(\"unix\", c.socketPath)\n\t} else {\n\t\taddress := fmt.Sprintf(\"%s:%d\", c.kamailioHost, c.kamailioPort)\n\t\tlog.Debug(\"Requesting stats from kamailio via binrpc \", address)\n\t\tconn, err = net.Dial(\"tcp\", address)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer conn.Close()\n\t// TODO\n\t// c.conn.SetDeadline(time.Now().Add(c.Timeout))\n\n\t// WritePacket returns the cookie generated\n\tcookie, err := binrpc.WritePacket(conn, \"stats.fetch\", \"all\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// the cookie is passed again for verification\n\t// we receive records in response\n\trecords, err := binrpc.ReadPacket(conn, cookie)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// convert the structure into a simple key=>value map\n\titems, _ := records[0].StructItems()\n\tresult := make(map[string]string)\n\tfor _, item := range items {\n\t\tvalue, _ := item.Value.String()\n\t\tresult[item.Key] = value\n\t}\n\n\treturn result, nil\n}", "func (st *httpStats) updateStats(r *http.Request, w *httpResponseRecorder) {\n\t// A successful request has a 2xx response code\n\tsuccessReq := (w.respStatusCode >= 200 && w.respStatusCode < 300)\n\t// Update stats according to method verb\n\tswitch r.Method {\n\tcase \"HEAD\":\n\t\tst.totalHEADs.Inc(1)\n\t\tif successReq {\n\t\t\tst.successHEADs.Inc(1)\n\t\t}\n\tcase \"GET\":\n\t\tst.totalGETs.Inc(1)\n\t\tif successReq {\n\t\t\tst.successGETs.Inc(1)\n\t\t}\n\tcase \"PUT\":\n\t\tst.totalPUTs.Inc(1)\n\t\tif successReq {\n\t\t\tst.successPUTs.Inc(1)\n\t\t}\n\tcase \"POST\":\n\t\tst.totalPOSTs.Inc(1)\n\t\tif successReq {\n\t\t\tst.successPOSTs.Inc(1)\n\t\t}\n\tcase \"DELETE\":\n\t\tst.totalDELETEs.Inc(1)\n\t\tif successReq {\n\t\t\tst.successDELETEs.Inc(1)\n\t\t}\n\t}\n}", "func (fM *FeslManager) UpdateStats(event GameSpy.EventClientTLSCommand) {\n\tif !event.Client.IsActive {\n\t\tlog.Noteln(\"Client left\")\n\t\treturn\n\t}\n\n\tanswer := event.Command.Message\n\tanswer[\"TXN\"] = \"UpdateStats\"\n\n\tuserId := event.Client.RedisState.Get(\"uID\")\n\n\tusers, _ := strconv.Atoi(event.Command.Message[\"u.[]\"])\n\n\tif users == 0 {\n\t\tlog.Warning(\"No u.[], defaulting to 1\")\n\t\tusers = 1\n\t}\n\n\tfor i := 0; i < users; i++ {\n\t\towner, ok := event.Command.Message[\"u.\"+strconv.Itoa(i)+\".o\"]\n\t\tif event.Client.RedisState.Get(\"clientType\") == \"server\" {\n\n\t\t\tvar id, userIDhero, heroName, online string\n\t\t\terr := fM.stmtGetHeroeByID.QueryRow(owner).Scan(&id, &userIDhero, &heroName, &online)\n\t\t\tif err != nil {\n\t\t\t\tlog.Noteln(\"Persona not worthy!\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tuserId = userIDhero\n\t\t\tlog.Noteln(\"Server updating stats\")\n\t\t}\n\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\n\t\tstats := make(map[string]*stat)\n\n\t\t// Get current stats from DB\n\t\t// Generate our argument list for the statement -> heroID, userID, key1, key2, key3, ...\n\t\tvar argsGet []interface{}\n\t\tstatsKeys := make(map[string]string)\n\t\targsGet = append(argsGet, owner)\n\t\targsGet = append(argsGet, userId)\n\t\tkeys, _ := strconv.Atoi(event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.[]\"])\n\t\tfor j := 0; j < keys; j++ {\n\t\t\targsGet = append(argsGet, event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".k\"])\n\t\t\tstatsKeys[event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".k\"]] = strconv.Itoa(j)\n\t\t}\n\n\t\trows, err := fM.getStatsStatement(keys).Query(argsGet...)\n\t\tif err != nil {\n\t\t\tlog.Errorln(\"Failed gettings stats for hero \"+owner, err.Error())\n\t\t}\n\n\t\tcount := 0\n\t\tfor rows.Next() {\n\t\t\tvar userID, heroID, statsKey, statsValue string\n\t\t\terr := rows.Scan(&userID, &heroID, &statsKey, &statsValue)\n\t\t\tif err != nil {\n\t\t\t\tlog.Errorln(\"Issue with database:\", err.Error())\n\t\t\t}\n\n\t\t\tintValue, err := strconv.ParseFloat(statsValue, 64)\n\t\t\tif err != nil {\n\t\t\t\tintValue = 0\n\t\t\t}\n\t\t\tstats[statsKey] = &stat{\n\t\t\t\ttext: statsValue,\n\t\t\t\tvalue: intValue,\n\t\t\t}\n\n\t\t\tdelete(statsKeys, statsKey)\n\t\t\tcount++\n\t\t}\n\n\t\t// Send stats not found with default value of \"\"\n\t\tfor key := range statsKeys {\n\t\t\tstats[key] = &stat{\n\t\t\t\ttext: \"\",\n\t\t\t\tvalue: 0,\n\t\t\t}\n\n\t\t\tcount++\n\t\t}\n\t\t// end Get current stats from DB\n\n\t\t// Generate our argument list for the statement -> userId, owner, key1, value1, userId, owner, key2, value2, userId, owner, ...\n\t\tvar args []interface{}\n\t\tkeys, _ = strconv.Atoi(event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.[]\"])\n\t\tfor j := 0; j < keys; j++ {\n\n\t\t\tif event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".ut\"] != \"3\" {\n\t\t\t\tlog.Noteln(\"Update new Type:\", event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".k\"], event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".t\"], event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".ut\"], event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".v\"], event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".pt\"])\n\t\t\t}\n\n\t\t\tkey := event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".k\"]\n\t\t\tvalue := event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".t\"]\n\n\t\t\tif value == \"\" {\n\t\t\t\tlog.Noteln(\"Updating stat\", key+\":\", event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".v\"], \"+\", stats[key].value)\n\t\t\t\t// We are dealing with a number\n\t\t\t\tvalue = event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".v\"]\n\n\t\t\t\t// ut seems to be 3 when we need to add up (xp has ut 0 when you level'ed up, otherwise 3)\n\t\t\t\tif event.Command.Message[\"u.\"+strconv.Itoa(i)+\".s.\"+strconv.Itoa(j)+\".ut\"] == \"3\" {\n\t\t\t\t\tintValue, err := strconv.ParseFloat(value, 64)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t// Couldn't transfer it to a number, skip updating this stat\n\t\t\t\t\t\tlog.Errorln(\"Skipping stat \"+key, err)\n\n\t\t\t\t\t\tanswer := make(map[string]string)\n\t\t\t\t\t\tanswer[\"TXN\"] = \"UpdateStats\"\n\n\t\t\t\t\t\tevent.Client.WriteFESL(event.Command.Query, answer, event.Command.PayloadID)\n\t\t\t\t\t\tfM.logAnswer(event.Command.Query, answer, event.Command.PayloadID)\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\tif intValue <= 0 || event.Client.RedisState.Get(\"clientType\") == \"server\" || key == \"c_ltp\" || key == \"c_sln\" || key == \"c_ltm\" || key == \"c_slm\" || key == \"c_wmid0\" || key == \"c_wmid1\" || key == \"c_tut\" || key == \"c_wmid2\" {\n\t\t\t\t\t\t// Only allow increasing numbers (like HeroPoints) by the server for now\n\t\t\t\t\t\tnewValue := stats[key].value + intValue\n\n\t\t\t\t\t\tif key == \"c_wallet_hero\" && newValue < 0 {\n\t\t\t\t\t\t\tlog.Errorln(\"Not allowed to process stat. c_wallet_hero lower than 0\", key)\n\t\t\t\t\t\t\tanswer := make(map[string]string)\n\t\t\t\t\t\t\tanswer[\"TXN\"] = \"UpdateStats\"\n\t\t\t\t\t\t\tevent.Client.WriteFESL(event.Command.Query, answer, event.Command.PayloadID)\n\t\t\t\t\t\t\tfM.logAnswer(event.Command.Query, answer, event.Command.PayloadID)\n\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tvalue = strconv.FormatFloat(newValue, 'f', 4, 64)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tlog.Errorln(\"Not allowed to process stat\", key)\n\t\t\t\t\t\tanswer := make(map[string]string)\n\t\t\t\t\t\tanswer[\"TXN\"] = \"UpdateStats\"\n\t\t\t\t\t\tevent.Client.WriteFESL(event.Command.Query, answer, event.Command.PayloadID)\n\t\t\t\t\t\tfM.logAnswer(event.Command.Query, answer, event.Command.PayloadID)\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// We need to append 3 values for each insert/update,\n\t\t\t// owner, key and value\n\t\t\tlog.Noteln(\"Updating stats:\", userId, owner, key, value)\n\t\t\targs = append(args, userId)\n\t\t\targs = append(args, owner)\n\t\t\targs = append(args, key)\n\t\t\targs = append(args, value)\n\t\t}\n\n\t\t_, err = fM.setStatsStatement(keys).Exec(args...)\n\t\tif err != nil {\n\t\t\tlog.Errorln(\"Failed setting stats for hero \"+owner, err.Error())\n\t\t}\n\t}\n\n\tevent.Client.WriteFESL(event.Command.Query, answer, event.Command.PayloadID)\n\tfM.logAnswer(event.Command.Query, answer, event.Command.PayloadID)\n}", "func (s *DevStat) UpdateSnmpGetStats(g int64, p int64, e int64) {\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\ts.Counters[SnmpOIDGetAll] = s.Counters[SnmpOIDGetAll].(int) + int(g)\n\ts.Counters[SnmpOIDGetProcessed] = s.Counters[SnmpOIDGetProcessed].(int) + int(p)\n\ts.Counters[SnmpOIDGetErrors] = s.Counters[SnmpOIDGetErrors].(int) + int(e)\n}", "func (s SolrPlugin) FetchMetrics() (map[string]interface{}, error) {\n\tstat := make(map[string]interface{})\n\tfor core, stats := range s.Stats {\n\t\tfor k, v := range stats {\n\t\t\tstat[core+\"_\"+k] = v\n\t\t}\n\t}\n\treturn stat, nil\n}", "func (c *Client) getStatistics() *AllStats {\n\n\tvar status Status\n\tstatusURL := fmt.Sprintf(statusURLPattern, c.protocol, c.hostname, c.port)\n\tbody := c.MakeRequest(statusURL)\n\terr := json.Unmarshal(body, &status)\n\tif err != nil {\n\t\tlog.Println(\"Unable to unmarshal Adguard log statistics to log statistics struct model\", err)\n\t}\n\n\tvar stats Stats\n\tstatsURL := fmt.Sprintf(statsURLPattern, c.protocol, c.hostname, c.port)\n\tbody = c.MakeRequest(statsURL)\n\terr = json.Unmarshal(body, &stats)\n\tif err != nil {\n\t\tlog.Println(\"Unable to unmarshal Adguard statistics to statistics struct model\", err)\n\t}\n\n\tvar logstats LogStats\n\tlogstatsURL := fmt.Sprintf(logstatsURLPattern, c.protocol, c.hostname, c.port, c.logLimit)\n\tbody = c.MakeRequest(logstatsURL)\n\terr = json.Unmarshal(body, &logstats)\n\tif err != nil {\n\t\tlog.Println(\"Unable to unmarshal Adguard log statistics to log statistics struct model\", err)\n\t}\n\n\tvar allstats AllStats\n\tallstats.status = &status\n\tallstats.stats = &stats\n\tallstats.logStats = &logstats\n\n\treturn &allstats\n}", "func (m VarnishPlugin) FetchMetrics() (map[string]interface{}, error) {\n\tvar out []byte\n\tvar err error\n\n\tif m.VarnishName == \"\" {\n\t\tout, err = exec.Command(m.VarnishStatPath, \"-1\").CombinedOutput()\n\t} else {\n\t\tout, err = exec.Command(m.VarnishStatPath, \"-1\", \"-n\", m.VarnishName).CombinedOutput()\n\t}\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%s: %s\", err, out)\n\t}\n\n\tlineexp := regexp.MustCompile(`^([^ ]+) +(\\d+)`)\n\tsmaexp := regexp.MustCompile(`^SMA\\.([^\\.]+)\\.(.+)$`)\n\n\tstat := map[string]interface{}{\n\t\t\"requests\": float64(0),\n\t}\n\n\tvar tmpv float64\n\tfor _, line := range strings.Split(string(out), \"\\n\") {\n\t\tmatch := lineexp.FindStringSubmatch(line)\n\t\tif match == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\ttmpv, err = strconv.ParseFloat(match[2], 64)\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tswitch match[1] {\n\t\tcase \"cache_hit\", \"MAIN.cache_hit\":\n\t\t\tstat[\"cache_hits\"] = tmpv\n\t\t\tstat[\"requests\"] = stat[\"requests\"].(float64) + tmpv\n\t\tcase \"cache_miss\", \"MAIN.cache_miss\":\n\t\t\tstat[\"requests\"] = stat[\"requests\"].(float64) + tmpv\n\t\tcase \"cache_hitpass\", \"MAIN.cache_hitpass\":\n\t\t\tstat[\"requests\"] = stat[\"requests\"].(float64) + tmpv\n\t\tcase \"MAIN.backend_req\":\n\t\t\tstat[\"backend_req\"] = tmpv\n\t\tcase \"MAIN.backend_conn\":\n\t\t\tstat[\"backend_conn\"] = tmpv\n\t\tcase \"MAIN.backend_fail\":\n\t\t\tstat[\"backend_fail\"] = tmpv\n\t\tcase \"MAIN.backend_reuse\":\n\t\t\tstat[\"backend_reuse\"] = tmpv\n\t\tcase \"MAIN.backend_recycle\":\n\t\t\tstat[\"backend_recycle\"] = tmpv\n\t\tcase \"MAIN.n_object\":\n\t\t\tstat[\"n_object\"] = tmpv\n\t\tcase \"MAIN.n_objectcore\":\n\t\t\tstat[\"n_objectcore\"] = tmpv\n\t\tcase \"MAIN.n_expired\":\n\t\t\tstat[\"n_expired\"] = tmpv\n\t\tcase \"MAIN.n_objecthead\":\n\t\t\tstat[\"n_objecthead\"] = tmpv\n\t\tcase \"MAIN.busy_sleep\":\n\t\t\tstat[\"busy_sleep\"] = tmpv\n\t\tcase \"MAIN.busy_wakeup\":\n\t\t\tstat[\"busy_wakeup\"] = tmpv\n\t\tdefault:\n\t\t\tsmamatch := smaexp.FindStringSubmatch(match[1])\n\t\t\tif smamatch == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif smamatch[2] == \"g_alloc\" {\n\t\t\t\tstat[\"varnish.sma.g_alloc.\"+smamatch[1]+\".g_alloc\"] = tmpv\n\t\t\t} else if smamatch[2] == \"g_bytes\" {\n\t\t\t\tstat[\"varnish.sma.memory.\"+smamatch[1]+\".allocated\"] = tmpv\n\t\t\t} else if smamatch[2] == \"g_space\" {\n\t\t\t\tstat[\"varnish.sma.memory.\"+smamatch[1]+\".available\"] = tmpv\n\t\t\t}\n\t\t}\n\t}\n\n\treturn stat, err\n}", "func (s *session) fetchMaildropStats() error {\n\tmsgCount, err := s.handler.GetMessageCount()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor i := uint64(0); i < msgCount; i++ {\n\t\tmSize, err := s.handler.GetMessageSize(i + 1)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\ts.msgSizes[i+1] = mSize\n\t}\n\treturn nil\n}", "func (fM *FeslManager) GetStats(event GameSpy.EventClientTLSCommand) {\n\tif !event.Client.IsActive {\n\t\tlog.Noteln(\"Client left\")\n\t\treturn\n\t}\n\n\tloginPacket := make(map[string]string)\n\tloginPacket[\"TXN\"] = \"GetStats\"\n\n\towner := event.Command.Message[\"owner\"]\n\n\tlog.Noteln(event.Command.Message[\"owner\"])\n\n\t// TODO\n\t// Check for mysql injection\n\tvar query string\n\tkeys, _ := strconv.Atoi(event.Command.Message[\"keys.[]\"])\n\tfor i := 0; i < keys; i++ {\n\t\tquery += event.Command.Message[\"keys.\"+strconv.Itoa(i)+\"\"] + \", \"\n\t}\n\n\t// Result is your slice string.\n\trawResult := make([][]byte, keys+1)\n\tresult := make([]string, keys+1)\n\n\tdest := make([]interface{}, keys+1) // A temporary interface{} slice\n\tfor i := range rawResult {\n\t\tdest[i] = &rawResult[i] // Put pointers to each string in the interface slice\n\t}\n\n\t// Owner==0 is for accounts-stats.\n\t// Otherwise hero-stats\n\tif owner == \"0\" || owner == event.Client.RedisState.Get(\"uID\") {\n\t\tstmt, err := fM.db.Prepare(\"SELECT \" + query + \"uid FROM west_heroes_accounts WHERE uid = ?\")\n\t\tlog.Debugln(stmt)\n\t\tdefer stmt.Close()\n\t\tif err != nil {\n\t\t\tlog.Errorln(err)\n\n\t\t\t// DEV CODE; REMOVE BEFORE TAKING LIVE!!!!!\n\t\t\t// Creates a missing column\n\n\t\t\tvar columns []string\n\t\t\tkeys, _ = strconv.Atoi(event.Command.Message[\"keys.[]\"])\n\t\t\tfor i := 0; i < keys; i++ {\n\t\t\t\tcolumns = append(columns, event.Command.Message[\"keys.\"+strconv.Itoa(i)+\"\"])\n\t\t\t}\n\n\t\t\tfor _, column := range columns {\n\t\t\t\tlog.Debugln(\"Checking column \" + column)\n\t\t\t\tstmt2, err := fM.db.Prepare(\"SELECT count(*) FROM INFORMATION_SCHEMA.COLUMNS WHERE TABLE_SCHEMA = DATABASE() AND TABLE_NAME = \\\"west_heroes_accounts\\\" AND COLUMN_NAME = \\\"\" + column + \"\\\"\")\n\t\t\t\tdefer stmt2.Close()\n\t\t\t\tif err != nil {\n\t\t\t\t}\n\t\t\t\tvar count int\n\t\t\t\terr = stmt2.QueryRow().Scan(&count)\n\t\t\t\tif err != nil {\n\t\t\t\t}\n\n\t\t\t\tif count == 0 {\n\t\t\t\t\tlog.Debugln(\"Creating column \" + column)\n\t\t\t\t\t// If we land here, the column doesn't exist, so create it\n\n\t\t\t\t\t_, err := fM.db.Exec(\"ALTER TABLE `west_heroes_accounts` ADD COLUMN `\" + column + \"` TEXT NULL\")\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t//DEV CODE; REMOVE BEFORE TAKING LIVE!!!!!\n\t\t\treturn\n\t\t}\n\n\t\terr = stmt.QueryRow(event.Client.RedisState.Get(\"uID\")).Scan(dest...)\n\t\tif err != nil {\n\t\t\tlog.Debugln(err)\n\t\t\treturn\n\t\t}\n\t\tif err != nil {\n\t\t\tlog.Errorln(err)\n\t\t\treturn\n\t\t}\n\n\t\tfor i, raw := range rawResult {\n\t\t\tif raw == nil {\n\t\t\t\tresult[i] = \"\\\\N\"\n\t\t\t} else {\n\t\t\t\tresult[i] = string(raw)\n\t\t\t}\n\t\t}\n\n\t\tloginPacket[\"ownerId\"] = result[len(result)-1]\n\t\tloginPacket[\"ownerType\"] = \"1\"\n\t\tloginPacket[\"stats.[]\"] = event.Command.Message[\"keys.[]\"]\n\t\tfor i := 0; i < keys; i++ {\n\t\t\tloginPacket[\"stats.\"+strconv.Itoa(i)+\".key\"] = event.Command.Message[\"keys.\"+strconv.Itoa(i)+\"\"]\n\t\t\tloginPacket[\"stats.\"+strconv.Itoa(i)+\".value\"] = result[i]\n\t\t}\n\n\t\tevent.Client.WriteFESL(event.Command.Query, loginPacket, event.Command.PayloadID)\n\t\tfM.logAnswer(event.Command.Query, loginPacket, event.Command.PayloadID)\n\n\t\treturn\n\t}\n\n\t// DO the same as above but for hero-stats instead of hero-account\n\n\tstmt, err := fM.db.Prepare(\"SELECT \" + query + \"pid FROM west_heroes_stats WHERE pid = ?\")\n\n\tdefer stmt.Close()\n\tif err != nil {\n\t\tlog.Errorln(err)\n\n\t\t// DEV CODE; REMOVE BEFORE TAKING LIVE!!!!!\n\t\t// Creates a missing column\n\n\t\tvar columns []string\n\t\tkeys, _ = strconv.Atoi(event.Command.Message[\"keys.[]\"])\n\t\tfor i := 0; i < keys; i++ {\n\t\t\tcolumns = append(columns, event.Command.Message[\"keys.\"+strconv.Itoa(i)+\"\"])\n\t\t}\n\n\t\tfor _, column := range columns {\n\t\t\tlog.Debugln(\"Checking column \" + column)\n\t\t\tstmt2, err := fM.db.Prepare(\"SELECT count(*) FROM INFORMATION_SCHEMA.COLUMNS WHERE TABLE_SCHEMA = DATABASE() AND TABLE_NAME = \\\"west_heroes_stats\\\" AND COLUMN_NAME = \\\"\" + column + \"\\\"\")\n\t\t\tdefer stmt2.Close()\n\t\t\tif err != nil {\n\t\t\t}\n\t\t\tvar count int\n\t\t\terr = stmt2.QueryRow().Scan(&count)\n\t\t\tif err != nil {\n\t\t\t}\n\n\t\t\tif count == 0 {\n\t\t\t\tlog.Debugln(\"Creating column \" + column)\n\t\t\t\t// If we land here, the column doesn't exist, so create it\n\n\t\t\t\tsql := \"ALTER TABLE `west_heroes_stats` ADD COLUMN `\" + column + \"` TEXT NULL\"\n\t\t\t\t_, err := fM.db.Exec(sql)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Errorln(sql)\n\t\t\t\t\tlog.Errorln(err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t//DEV CODE; REMOVE BEFORE TAKING LIVE!!!!!\n\t\t//return\n\t}\n\tsql := \"UPDATE `west_heroes_stats` SET c_wallet_hero = round(level/2-0.1,0) WHERE pid = \" + owner + \"\"\n\t_, err = fM.db.Exec(sql)\n\tif err != nil {\n\t\tlog.Errorln(err)\n\t}\n\n\terr = stmt.QueryRow(owner).Scan(dest...)\n\tif err != nil {\n\t\tlog.Debugln(err)\n\t\treturn\n\t}\n\tif err != nil {\n\t\tlog.Errorln(err)\n\t\treturn\n\t}\n\n\tfor i, raw := range rawResult {\n\t\tif raw == nil {\n\t\t\tresult[i] = \"\\\\N\"\n\t\t} else {\n\t\t\tresult[i] = string(raw)\n\t\t}\n\t}\n\n\tloginPacket[\"ownerId\"] = result[len(result)-1]\n\tloginPacket[\"ownerType\"] = \"1\"\n\tloginPacket[\"stats.[]\"] = event.Command.Message[\"keys.[]\"]\n\tfor i := 0; i < keys; i++ {\n\t\tloginPacket[\"stats.\"+strconv.Itoa(i)+\".key\"] = event.Command.Message[\"keys.\"+strconv.Itoa(i)+\"\"]\n\t\tloginPacket[\"stats.\"+strconv.Itoa(i)+\".value\"] = result[i]\n\t}\n\n\tevent.Client.WriteFESL(event.Command.Query, loginPacket, event.Command.PayloadID)\n\tfM.logAnswer(event.Command.Query, loginPacket, event.Command.PayloadID)\n\n}", "func (s *Service) Stats(r *http.Request, args *StatsArgs, result *StatsResponse) error {\n\tif args.UserID == \"\" {\n\t\tresult.Error = uidMissing\n\t\treturn nil\n\t}\n\tresult.Whole = -1\n\tresult.Bookmarks = -1\n\tresult.Pim = -1\n\tresult.Org = -1\n\tcoll := s.Session.DB(MentatDatabase).C(args.UserID)\n\twholeCount, err := coll.Count()\n\tif err != nil {\n\t\tresult.Error = fmt.Sprintf(\"failed getting stats/whole count: %s\", err)\n\t\treturn nil\n\t}\n\tresult.Whole = wholeCount\n\tif args.Detailed {\n\t\tvar entries []Entry\n\t\terr := coll.Find(bson.M{\"type\": \"bookmark\"}).All(&entries)\n\t\tif err != nil {\n\t\t\tresult.Error = fmt.Sprintf(\"failed getting stats/bookmarks count: %s\", err)\n\t\t\treturn nil\n\t\t}\n\t\tresult.Bookmarks = len(entries)\n\t\terr = coll.Find(bson.M{\"type\": \"pim\"}).All(&entries)\n\t\tif err != nil {\n\t\t\tresult.Error = fmt.Sprintf(\"failed getting stats/pim count: %s\", err)\n\t\t\treturn nil\n\t\t}\n\t\tresult.Pim = len(entries)\n\t\terr = coll.Find(bson.M{\"type\": \"org\"}).All(&entries)\n\t\tif err != nil {\n\t\t\tresult.Error = fmt.Sprintf(\"failed getting stats/org count: %s\", err)\n\t\t\treturn nil\n\t\t}\n\t\tresult.Org = len(entries)\n\t}\n\treturn nil\n}", "func (m *Monitor) Stats(ctx *context.Context) {\n\tctx.JSON(m.Holder.GetStats())\n}", "func (m Plugin) FetchMetrics() (map[string]float64, error) {\n\tresp, err := http.Get(fmt.Sprintf(\"http://%s/stats\", m.Target))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\tstats := struct {\n\t\tConnections float64 `json:\"connections\"`\n\t\tTotalConnections float64 `json:\"total_connections\"`\n\t\tTotalMessages float64 `json:\"total_messages\"`\n\t\tConnectErrors float64 `json:\"connect_errors\"`\n\t\tMessageErrors float64 `json:\"message_errors\"`\n\t\tClosingConnections float64 `json:\"closing_connections\"`\n\t}{}\n\tif err := json.NewDecoder(resp.Body).Decode(&stats); err != nil {\n\t\treturn nil, err\n\t}\n\tret := make(map[string]float64, 6)\n\tret[\"conn_current\"] = stats.Connections\n\tret[\"conn_total\"] = stats.TotalConnections\n\tret[\"conn_errors\"] = stats.ConnectErrors\n\tret[\"conn_closing\"] = stats.ClosingConnections\n\tret[\"messages_total\"] = stats.TotalMessages\n\tret[\"messages_errors\"] = stats.MessageErrors\n\n\treturn ret, nil\n}", "func (sr *ServicedStatsReporter) updateStats() {\n\t// Stats for host.\n\tsr.updateHostStats()\n\t// Stats for the containers.\n\tstates, err := zkservice.GetHostStates(sr.conn, \"\", sr.hostID)\n\tif err != nil {\n\t\tplog.WithFields(logrus.Fields{\n\t\t\t\"conn\": sr.conn,\n\t\t\t\"hostID\": sr.hostID,\n\t\t}).WithError(err).Error(\"Could not get host states from Zookeeper\")\n\t}\n\n\tfor _, rs := range states {\n\t\tif rs.ContainerID != \"\" {\n\n\t\t\tcontainerRegistry := sr.getOrCreateContainerRegistry(rs.ServiceID, rs.InstanceID)\n\t\t\tstats, err := sr.docker.GetContainerStats(rs.ContainerID, 30*time.Second)\n\t\t\tif err != nil || stats == nil { //stats may be nil if service is shutting down\n\t\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\t\"serviceID\": rs.ServiceID,\n\t\t\t\t\t\"instanceID\": rs.InstanceID,\n\t\t\t\t}).WithError(err).Warn(\"Couldn't get stats from docker\")\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Check to see if we have the previous stats for this running instance\n\t\t\tusePreviousStats := true\n\t\t\tkey := rs.ContainerID\n\t\t\tif _, found := sr.previousStats[key]; !found {\n\t\t\t\tsr.previousStats[key] = make(map[string]uint64)\n\t\t\t\tusePreviousStats = false\n\t\t\t}\n\n\t\t\t// CPU Stats\n\t\t\t// TODO: Consolidate this into a single object that both ISVCS and non-ISVCS can use\n\t\t\tvar (\n\t\t\t\tkernelCPUPercent float64\n\t\t\t\tuserCPUPercent float64\n\t\t\t\ttotalCPUChange uint64\n\t\t\t)\n\n\t\t\tkernelCPU := stats.CPUStats.CPUUsage.UsageInKernelmode\n\t\t\tuserCPU := stats.CPUStats.CPUUsage.UsageInUsermode\n\t\t\ttotalCPU := stats.CPUStats.SystemCPUUsage\n\n\t\t\t// Total CPU Cycles\n\t\t\tpreviousTotalCPU, found := sr.previousStats[key][\"totalCPU\"]\n\t\t\tif found {\n\t\t\t\tif totalCPU <= previousTotalCPU {\n\t\t\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\t\t\"totalCPU\": totalCPU,\n\t\t\t\t\t\t\"previousTotalCPU\": previousTotalCPU,\n\t\t\t\t\t}).Debug(\"Change in total CPU usage was nonpositive, skipping CPU stats update\")\n\t\t\t\t\tusePreviousStats = false\n\t\t\t\t} else {\n\t\t\t\t\ttotalCPUChange = totalCPU - previousTotalCPU\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tusePreviousStats = false\n\t\t\t}\n\t\t\tsr.previousStats[key][\"totalCPU\"] = totalCPU\n\n\t\t\t// CPU Cycles in Kernel mode\n\t\t\tif previousKernelCPU, found := sr.previousStats[key][\"kernelCPU\"]; found && usePreviousStats {\n\t\t\t\tkernelCPUChange := kernelCPU - previousKernelCPU\n\t\t\t\tkernelCPUPercent = (float64(kernelCPUChange) / float64(totalCPUChange)) * float64(len(stats.CPUStats.CPUUsage.PercpuUsage)) * 100.0\n\t\t\t} else {\n\t\t\t\tusePreviousStats = false\n\t\t\t}\n\t\t\tsr.previousStats[key][\"kernelCPU\"] = kernelCPU\n\n\t\t\t// CPU Cycles in User mode\n\t\t\tif previousUserCPU, found := sr.previousStats[key][\"userCPU\"]; found && usePreviousStats {\n\t\t\t\tuserCPUChange := userCPU - previousUserCPU\n\t\t\t\tuserCPUPercent = (float64(userCPUChange) / float64(totalCPUChange)) * float64(len(stats.CPUStats.CPUUsage.PercpuUsage)) * 100.0\n\t\t\t} else {\n\t\t\t\tusePreviousStats = false\n\t\t\t}\n\t\t\tsr.previousStats[key][\"userCPU\"] = userCPU\n\n\t\t\t// Update CPU metrics\n\t\t\tif usePreviousStats {\n\t\t\t\tmetrics.GetOrRegisterGaugeFloat64(\"docker.usageinkernelmode\", containerRegistry).Update(kernelCPUPercent)\n\t\t\t\tmetrics.GetOrRegisterGaugeFloat64(\"docker.usageinusermode\", containerRegistry).Update(userCPUPercent)\n\t\t\t} else {\n\t\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\t\"serviceID\": rs.ServiceID,\n\t\t\t\t\t\"instanceID\": rs.InstanceID,\n\t\t\t\t}).Debug(\"Skipping CPU stats, no previous values to compare to\")\n\t\t\t}\n\n\t\t\t// Memory Stats\n\t\t\tpgFault := int64(stats.MemoryStats.Stats.Pgfault)\n\t\t\ttotalRSS := int64(stats.MemoryStats.Stats.TotalRss)\n\t\t\tcache := int64(stats.MemoryStats.Stats.Cache)\n\t\t\tif pgFault < 0 || totalRSS < 0 || cache < 0 {\n\t\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\t\"serviceID\": rs.ServiceID,\n\t\t\t\t\t\"instanceID\": rs.InstanceID,\n\t\t\t\t}).Debug(\"Memory metric value too big for int64\")\n\t\t\t}\n\t\t\tmetrics.GetOrRegisterGauge(\"cgroup.memory.pgmajfault\", containerRegistry).Update(pgFault)\n\t\t\tmetrics.GetOrRegisterGauge(\"cgroup.memory.totalrss\", containerRegistry).Update(totalRSS)\n\t\t\tmetrics.GetOrRegisterGauge(\"cgroup.memory.cache\", containerRegistry).Update(cache)\n\n\t\t} else {\n\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\"serviceID\": rs.ServiceID,\n\t\t\t\t\"instanceID\": rs.InstanceID,\n\t\t\t}).Debug(\"Skipping stats update, no container ID exists\")\n\t\t}\n\t}\n\t// Clean out old container registries\n\tsr.removeStaleRegistries(states)\n}", "func (c *Cache) updateStats() error {\n\tvar newUsed int64\n\terr := c.walk(func(osPath string, fi os.FileInfo, name string) error {\n\t\tif !fi.IsDir() {\n\t\t\t// Update the atime with that of the file\n\t\t\tatime := times.Get(fi).AccessTime()\n\t\t\tc.updateStat(name, atime, fi.Size())\n\t\t\tnewUsed += fi.Size()\n\t\t} else {\n\t\t\tc.cacheDir(name)\n\t\t}\n\t\treturn nil\n\t})\n\tc.itemMu.Lock()\n\tc.used = newUsed\n\tc.itemMu.Unlock()\n\treturn err\n}", "func getStats(url string) *grequests.Response {\n\tresp, err := grequests.Get(url, requestOptions)\n\t// you can modify the request by passing an optional RequestOptions struct\n\tif err != nil {\n\t\tlog.Fatalln(\"Unable to make request: \", err)\n\t}\n\treturn resp\n}", "func (m *Monitor) fetchStats(container dockerContainer, labelMap map[string]string, envMap map[string]string, enhancedMetricsConfig EnhancedMetricsConfig) {\n\tctx, cancel := context.WithTimeout(m.ctx, m.timeout)\n\tstats, err := m.client.ContainerStats(ctx, container.ID, false)\n\tif err != nil {\n\t\tcancel()\n\t\tif isContainerNotFound(err) {\n\t\t\tm.logger.Debugf(\"container %s is not found in cache\", container.ID)\n\t\t\treturn\n\t\t}\n\t\tm.logger.WithError(err).Errorf(\"Could not fetch docker stats for container id %s\", container.ID)\n\t\treturn\n\t}\n\n\tvar parsed dtypes.StatsJSON\n\terr = json.NewDecoder(stats.Body).Decode(&parsed)\n\tstats.Body.Close()\n\tif err != nil {\n\t\tcancel()\n\t\t// EOF means that there aren't any stats, perhaps because the container\n\t\t// is gone. Just return nothing and no error.\n\t\tif err == io.EOF {\n\t\t\treturn\n\t\t}\n\t\tm.logger.WithError(err).Errorf(\"Could not parse docker stats for container id %s\", container.ID)\n\t\treturn\n\t}\n\n\tdps, err := ConvertStatsToMetrics(container.ContainerJSON, &parsed, enhancedMetricsConfig)\n\tcancel()\n\tif err != nil {\n\t\tm.logger.WithError(err).Errorf(\"Could not convert docker stats for container id %s\", container.ID)\n\t\treturn\n\t}\n\n\tfor i := range dps {\n\t\tfor k, dimName := range envMap {\n\t\t\tif v := container.EnvMap[k]; v != \"\" {\n\t\t\t\tdps[i].Dimensions[dimName] = v\n\t\t\t}\n\t\t}\n\t\tfor k, dimName := range labelMap {\n\t\t\tif v := container.Config.Labels[k]; v != \"\" {\n\t\t\t\tdps[i].Dimensions[dimName] = v\n\t\t\t}\n\t\t}\n\t}\n\tm.Output.SendDatapoints(dps...)\n}", "func (p *PoolShard) getStats() (stats PoolStats) {\n\tstats.Shard = p.server\n\tstats.NumActive = int(atomic.LoadInt32(&p.active))\n\tif atomic.LoadUint32(&p.available) == 1 {\n\t\tstats.Available = true\n\t} else {\n\t\tstats.Available = false\n\t}\n\n\tstats.NumGet = atomic.SwapUint64(&p.stats.NumGet, 0)\n\tstats.NumPut = atomic.SwapUint64(&p.stats.NumPut, 0)\n\tstats.NumBroken = atomic.SwapUint64(&p.stats.NumBroken, 0)\n\tstats.NumClose = atomic.SwapUint64(&p.stats.NumClose, 0)\n\tstats.NumDial = atomic.SwapUint64(&p.stats.NumDial, 0)\n\tstats.NumDialError = atomic.SwapUint64(&p.stats.NumDialError, 0)\n\tstats.NumEvict = atomic.SwapUint64(&p.stats.NumEvict, 0)\n\treturn\n}", "func (h *StatsHandlers) getStats(c *gin.Context) {\n\tdb, ok := c.MustGet(\"databaseConn\").(*gorm.DB)\n\tif !ok {\n\t\treturn\n\t}\n\n\tvar stats []models.Stats\n\tdb.Limit(60 * 5).Order(\"created_date desc\").Find(&stats)\n\n\tSuccess(c, \"stats\", gin.H{\n\t\t\"title\": \"Stats\",\n\t\t\"stats\": stats})\n\n}", "func (lc *LocalCache) Stats() (stats Stats) {\n\tlc.RLock()\n\tdefer lc.RUnlock()\n\n\tstats.Count = len(lc.Data)\n\tnow := time.Now().UTC()\n\tfor _, item := range lc.Data {\n\t\tage := now.Sub(item.Timestamp)\n\t\tif stats.MaxAge < age {\n\t\t\tstats.MaxAge = age\n\t\t}\n\t\tstats.SizeBytes += int(unsafe.Sizeof(item))\n\t}\n\treturn\n}", "func (c networkDependencyCollector) Update(ch chan<- prometheus.Metric) error {\n\ttraffic := darkstat.Get()\n\tserverProcesses, upstreams, downstreams := socketstat.Get()\n\tlocalInventory := inventory.GetLocalInventory()\n\n\tfor _, m := range traffic {\n\t\tch <- prometheus.MustNewConstMetric(c.traffic, prometheus.GaugeValue, m.Bandwidth,\n\t\t\tm.LocalHostgroup, m.Direction, m.RemoteHostgroup, m.RemoteIPAddr, m.LocalDomain, m.RemoteDomain)\n\t}\n\tfor _, m := range upstreams {\n\t\tch <- prometheus.MustNewConstMetric(c.upstream, prometheus.GaugeValue, 1,\n\t\t\tm.LocalHostgroup, m.RemoteHostgroup, m.LocalAddress, m.RemoteAddress, m.Port, m.Protocol, m.ProcessName)\n\t}\n\tfor _, m := range downstreams {\n\t\tch <- prometheus.MustNewConstMetric(c.downstream, prometheus.GaugeValue, 1,\n\t\t\tm.LocalHostgroup, m.RemoteHostgroup, m.LocalAddress, m.RemoteAddress, m.Port, m.Protocol, m.ProcessName)\n\t}\n\tfor _, m := range serverProcesses {\n\t\tch <- prometheus.MustNewConstMetric(c.serverProcesses, prometheus.GaugeValue, 1,\n\t\t\tlocalInventory.Hostgroup, m.Bind, m.Name, m.Port)\n\t}\n\n\treturn nil\n}", "func getLoadStats(id uint16, cfd *os.File, conn *Connection) (info.LoadStats, error) {\n\tmsg := prepareCmdMessage(id, cfd.Fd())\n\terr := conn.WriteMessage(msg.toRawMsg())\n\tif err != nil {\n\t\treturn info.LoadStats{}, err\n\t}\n\n\tresp, err := conn.ReadMessage()\n\tif err != nil {\n\t\treturn info.LoadStats{}, err\n\t}\n\n\tparsedmsg, err := parseLoadStatsResp(resp)\n\tif err != nil {\n\t\treturn info.LoadStats{}, err\n\t}\n\treturn parsedmsg.Stats, nil\n}", "func (ms *MetricSet) Fetch(report mb.ReporterV2) {\n\tneedsStateUpdate := time.Since(ms.lastState) > ms.config.effectiveStatePeriod()\n\tif needsStateUpdate || ms.cache.IsEmpty() {\n\t\tms.log.Debugf(\"State update needed (needsStateUpdate=%v, cache.IsEmpty()=%v)\", needsStateUpdate, ms.cache.IsEmpty())\n\t\terr := ms.reportState(report)\n\t\tif err != nil {\n\t\t\tms.log.Error(err)\n\t\t\treport.Error(err)\n\t\t}\n\t\tms.log.Debugf(\"Next state update by %v\", ms.lastState.Add(ms.config.effectiveStatePeriod()))\n\t}\n\n\terr := ms.reportChanges(report)\n\tif err != nil {\n\t\tms.log.Error(err)\n\t\treport.Error(err)\n\t}\n}", "func (c *softnetCollector) Update(ch chan<- prometheus.Metric) error {\n\tvar cpu string\n\n\tstats, err := c.fs.NetSoftnetStat()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not get softnet statistics: %w\", err)\n\t}\n\n\tfor _, cpuStats := range stats {\n\t\tcpu = strconv.FormatUint(uint64(cpuStats.Index), 10)\n\n\t\tch <- prometheus.MustNewConstMetric(\n\t\t\tc.processed,\n\t\t\tprometheus.CounterValue,\n\t\t\tfloat64(cpuStats.Processed),\n\t\t\tcpu,\n\t\t)\n\t\tch <- prometheus.MustNewConstMetric(\n\t\t\tc.dropped,\n\t\t\tprometheus.CounterValue,\n\t\t\tfloat64(cpuStats.Dropped),\n\t\t\tcpu,\n\t\t)\n\t\tch <- prometheus.MustNewConstMetric(\n\t\t\tc.timeSqueezed,\n\t\t\tprometheus.CounterValue,\n\t\t\tfloat64(cpuStats.TimeSqueezed),\n\t\t\tcpu,\n\t\t)\n\t\tch <- prometheus.MustNewConstMetric(\n\t\t\tc.cpuCollision,\n\t\t\tprometheus.CounterValue,\n\t\t\tfloat64(cpuStats.CPUCollision),\n\t\t\tcpu,\n\t\t)\n\t\tch <- prometheus.MustNewConstMetric(\n\t\t\tc.receivedRps,\n\t\t\tprometheus.CounterValue,\n\t\t\tfloat64(cpuStats.ReceivedRps),\n\t\t\tcpu,\n\t\t)\n\t\tch <- prometheus.MustNewConstMetric(\n\t\t\tc.flowLimitCount,\n\t\t\tprometheus.CounterValue,\n\t\t\tfloat64(cpuStats.FlowLimitCount),\n\t\t\tcpu,\n\t\t)\n\t\tch <- prometheus.MustNewConstMetric(\n\t\t\tc.softnetBacklogLen,\n\t\t\tprometheus.GaugeValue,\n\t\t\tfloat64(cpuStats.SoftnetBacklogLen),\n\t\t\tcpu,\n\t\t)\n\t}\n\n\treturn nil\n}", "func (u RackStatsPlugin) FetchMetrics() (stats map[string]interface{}, err error) {\n\tstats, err = u.parseStats()\n\treturn stats, err\n}", "func (c *PrometheusCollector) Update(r metricCollector.MetricResult) {\n\tc.RWMutex.Lock()\n\tdefer c.RWMutex.Unlock()\n\n\t// check circuit open\n\tif r.Successes > 0 {\n\t\tgauge := c.gauges[metricCircuitOpen]\n\t\tgauge.Set(0)\n\n\t\tcounter := c.counters[metricSuccesses]\n\t\tcounter.Add(r.Successes)\n\t}\n\tif r.ShortCircuits > 0 {\n\t\tgauge := c.gauges[metricCircuitOpen]\n\t\tgauge.Set(1)\n\n\t\tcounter := c.counters[metricShortCircuits]\n\t\tcounter.Add(r.ShortCircuits)\n\t}\n\t// update metric\n\tif r.Attempts > 0 {\n\t\tcounter := c.counters[metricAttempts]\n\t\tcounter.Add(r.Attempts)\n\t}\n\tif r.Errors > 0 {\n\t\tcounter := c.counters[metricErrors]\n\t\tcounter.Add(r.Errors)\n\t}\n\tif r.Failures > 0 {\n\t\tcounter := c.counters[metricFailures]\n\t\tcounter.Add(r.Failures)\n\t}\n\tif r.Rejects > 0 {\n\t\tcounter := c.counters[metricRejects]\n\t\tcounter.Add(r.Rejects)\n\t}\n\tif r.Timeouts > 0 {\n\t\tcounter := c.counters[metricTimeouts]\n\t\tcounter.Add(r.Timeouts)\n\t}\n\tif r.FallbackSuccesses > 0 {\n\t\tcounter := c.counters[metricFallbackSuccesses]\n\t\tcounter.Add(r.FallbackSuccesses)\n\t}\n\tif r.FallbackFailures > 0 {\n\t\tcounter := c.counters[metricFallbackFailures]\n\t\tcounter.Add(r.FallbackFailures)\n\t}\n\n\tgauge := c.gauges[metricTotalDuration]\n\tgauge.Set(r.TotalDuration.Seconds())\n\n\tgauge = c.gauges[metricRunDuration]\n\tgauge.Set(r.RunDuration.Seconds())\n\n\tgauge = c.gauges[metricConcurrencyInUse]\n\tgauge.Set(r.ConcurrencyInUse)\n}", "func (e *Exporter) ovnMetricsUpdate() {\n\tfor {\n\t\te.exportOvnStatusGauge()\n\t\te.exportOvnLogFileSizeGauge()\n\t\te.exportOvnDBFileSizeGauge()\n\t\te.exportOvnRequestErrorGauge()\n\t\te.exportOvnDBStatusGauge()\n\n\t\te.exportOvnChassisGauge()\n\t\te.exportLogicalSwitchGauge()\n\t\te.exportLogicalSwitchPortGauge()\n\n\t\te.exportOvnClusterEnableGauge()\n\t\tif isClusterEnabled {\n\t\t\te.exportOvnClusterInfoGauge()\n\t\t}\n\n\t\ttime.Sleep(time.Duration(e.pollInterval) * time.Second)\n\t}\n}", "func (m *MetricSet) Fetch(reporter mb.ReporterV2) {\n\tdbs, err := m.getDbsNames()\n\tif err != nil {\n\t\treporter.Error(err)\n\t\treturn\n\t}\n\n\tfor _, db := range dbs {\n\t\tmoduleFields := common.MapStr{\n\t\t\t\"database\": common.MapStr{\n\t\t\t\t\"id\": db.id,\n\t\t\t\t\"name\": db.name,\n\t\t\t},\n\t\t}\n\t\tmetricsetFields := common.MapStr{}\n\n\t\tspaceUsage, err := m.getLogSpaceUsageForDb(db.name)\n\t\tif err != nil {\n\t\t\treporter.Error(err)\n\t\t} else {\n\t\t\tmetricsetFields[\"space_usage\"] = spaceUsage\n\t\t}\n\n\t\tstats, err := m.getLogStats(db)\n\t\tif err != nil {\n\t\t\treporter.Error(err)\n\t\t} else {\n\t\t\tmetricsetFields[\"stats\"] = stats\n\t\t}\n\n\t\tif len(metricsetFields) == 0 {\n\t\t\tm.log.Debug(\"no data to report\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// Both log space and log size are available, so report both\n\t\tif isReported := reporter.Event(mb.Event{\n\t\t\tModuleFields: moduleFields,\n\t\t\tMetricSetFields: metricsetFields,\n\t\t}); !isReported {\n\t\t\tm.log.Debug(\"event not reported\")\n\t\t}\n\n\t}\n}", "func (p RekognitionPlugin) FetchMetrics() (map[string]float64, error) {\n\tstat := make(map[string]float64)\n\n\tfor _, met := range [...]string{\n\t\t\"SuccessfulRequestCount\",\n\t\t\"ThrottledCount\",\n\t\t\"ResponseTime\",\n\t\t\"DetectedFaceCount\",\n\t\t\"DetectedLabelCount\",\n\t\t\"ServerErrorCount\",\n\t\t\"UserErrorCount\",\n\t} {\n\t\tv, err := p.getLastPoint(met)\n\t\tif err == nil {\n\t\t\tstat[met] = v\n\t\t} else {\n\t\t\tlog.Printf(\"%s: %s\", met, err)\n\t\t}\n\t}\n\n\treturn stat, nil\n}", "func GetStats() Stats {\r\n\r\n\treturn stats\r\n}", "func (d *Daemon) fetchData() {\n\td.Data.Health.Sensu = make(map[string]structs.SensuHealth, len(*d.Datacenters))\n\n\tfor _, datacenter := range *d.Datacenters {\n\t\tlogger.Infof(\"Updating the datacenter %s\", datacenter.Name)\n\n\t\t// set default health status\n\t\td.Data.Health.Sensu[datacenter.Name] = structs.SensuHealth{Output: datacenterErrorString, Status: 2}\n\t\td.Data.Health.Uchiwa = \"ok\"\n\n\t\t// fetch sensu data from the datacenter\n\t\tstashes, err := datacenter.GetStashes()\n\t\tif err != nil {\n\t\t\tlogger.Warningf(\"Connection failed to the datacenter %s\", datacenter.Name)\n\t\t\tcontinue\n\t\t}\n\t\tsilenced, err := datacenter.GetSilenced()\n\t\tif err != nil {\n\t\t\tlogger.Warningf(\"Connection failed to the datacenter %s.\", datacenter.Name)\n\t\t\tcontinue\n\t\t}\n\t\tchecks, err := datacenter.GetChecks()\n\t\tif err != nil {\n\t\t\tlogger.Warningf(\"Connection failed to the datacenter %s\", datacenter.Name)\n\t\t\tcontinue\n\t\t}\n\t\tclients, err := datacenter.GetClients()\n\t\tif err != nil {\n\t\t\tlogger.Warningf(\"Connection failed to the datacenter %s\", datacenter.Name)\n\t\t\tcontinue\n\t\t}\n\t\tevents, err := datacenter.GetEvents()\n\t\tif err != nil {\n\t\t\tlogger.Warningf(\"Connection failed to the datacenter %s\", datacenter.Name)\n\t\t\tcontinue\n\t\t}\n\t\tinfo, err := datacenter.GetInfo()\n\t\tif err != nil {\n\t\t\tlogger.Warningf(\"Connection failed to the datacenter %s\", datacenter.Name)\n\t\t\tcontinue\n\t\t}\n\t\taggregates, err := datacenter.GetAggregates()\n\t\tif err != nil {\n\t\t\tlogger.Warningf(\"Connection failed to the datacenter %s\", datacenter.Name)\n\t\t\tcontinue\n\t\t}\n\n\t\tif d.Enterprise {\n\t\t\td.Data.SERawMetrics = *getEnterpriseMetrics(&datacenter, &d.Data.SERawMetrics)\n\t\t}\n\n\t\t// Determine the status of the datacenter\n\t\tif !info.Redis.Connected {\n\t\t\td.Data.Health.Sensu[datacenter.Name] = structs.SensuHealth{Output: \"Not connected to Redis\", Status: 1}\n\t\t} else if !info.Transport.Connected {\n\t\t\td.Data.Health.Sensu[datacenter.Name] = structs.SensuHealth{Output: \"Not connected to the transport\", Status: 1}\n\t\t} else {\n\t\t\td.Data.Health.Sensu[datacenter.Name] = structs.SensuHealth{Output: \"ok\", Status: 0}\n\t\t}\n\n\t\t// add fetched data into d.Data interface\n\t\tfor _, v := range stashes {\n\t\t\tsetDc(v, datacenter.Name)\n\t\t\td.Data.Stashes = append(d.Data.Stashes, v)\n\t\t}\n\t\tfor _, v := range silenced {\n\t\t\tsetDc(v, datacenter.Name)\n\t\t\td.Data.Silenced = append(d.Data.Silenced, v)\n\t\t}\n\t\tfor _, v := range checks {\n\t\t\tsetDc(v, datacenter.Name)\n\t\t\td.Data.Checks = append(d.Data.Checks, v)\n\t\t}\n\t\tfor _, v := range clients {\n\t\t\tsetDc(v, datacenter.Name)\n\t\t\td.Data.Clients = append(d.Data.Clients, v)\n\t\t}\n\t\tfor _, v := range events {\n\t\t\tsetDc(v, datacenter.Name)\n\t\t\td.Data.Events = append(d.Data.Events, v)\n\t\t}\n\t\tfor _, v := range aggregates {\n\t\t\tsetDc(v, datacenter.Name)\n\t\t\td.Data.Aggregates = append(d.Data.Aggregates, v)\n\t\t}\n\n\t\t// build datacenter\n\t\tdc := d.buildDatacenter(&datacenter.Name, info)\n\t\tdc.Stats[\"aggregates\"] = len(aggregates)\n\t\tdc.Stats[\"checks\"] = len(checks)\n\t\tdc.Stats[\"clients\"] = len(clients)\n\t\tdc.Stats[\"events\"] = len(events)\n\t\tdc.Stats[\"silenced\"] = len(silenced)\n\t\tdc.Stats[\"stashes\"] = len(stashes)\n\t\td.Data.Dc = append(d.Data.Dc, dc)\n\t}\n}", "func (s *Simulator) Stats() *Stats {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\telapsed := time.Since(s.now).Seconds()\n\tpThrough := float64(s.writtenN) / elapsed\n\trespMean := 0\n\tif len(s.latencyHistory) > 0 {\n\t\trespMean = int(s.totalLatency) / len(s.latencyHistory) / int(time.Millisecond)\n\t}\n\tstats := &Stats{\n\t\tTime: time.Unix(0, int64(time.Since(s.now))),\n\t\tTags: s.ReportTags,\n\t\tFields: models.Fields(map[string]interface{}{\n\t\t\t\"T\": int(elapsed),\n\t\t\t\"points_written\": s.writtenN,\n\t\t\t\"values_written\": s.writtenN * s.FieldsPerPoint,\n\t\t\t\"points_ps\": pThrough,\n\t\t\t\"values_ps\": pThrough * float64(s.FieldsPerPoint),\n\t\t\t\"write_error\": s.currentErrors,\n\t\t\t\"resp_wma\": int(s.wmaLatency),\n\t\t\t\"resp_mean\": respMean,\n\t\t\t\"resp_90\": int(s.quartileResponse(0.9) / time.Millisecond),\n\t\t\t\"resp_95\": int(s.quartileResponse(0.95) / time.Millisecond),\n\t\t\t\"resp_99\": int(s.quartileResponse(0.99) / time.Millisecond),\n\t\t}),\n\t}\n\n\tvar isCreating bool\n\tif s.writtenN < s.SeriesN() {\n\t\tisCreating = true\n\t}\n\tstats.Tags[\"creating_series\"] = fmt.Sprint(isCreating)\n\n\t// Reset error count for next reporting.\n\ts.currentErrors = 0\n\n\t// Add runtime stats for the remote instance.\n\tvar vars Vars\n\tresp, err := http.Get(strings.TrimSuffix(s.Host, \"/\") + \"/debug/vars\")\n\tif err != nil {\n\t\t// Don't log error as it can get spammy.\n\t\treturn stats\n\t}\n\tdefer resp.Body.Close()\n\n\tif err := json.NewDecoder(resp.Body).Decode(&vars); err != nil {\n\t\tfmt.Fprintln(s.Stderr, err)\n\t\treturn stats\n\t}\n\n\tstats.Fields[\"heap_alloc\"] = vars.Memstats.HeapAlloc\n\tstats.Fields[\"heap_in_use\"] = vars.Memstats.HeapInUse\n\tstats.Fields[\"heap_objects\"] = vars.Memstats.HeapObjects\n\treturn stats\n}", "func loadStats(db *sql.DB) error {\n\tvar err error\n\tvar s model.Stats\n\n\terr = db.QueryRow(`SELECT COUNT(users.id) FROM users`).Scan(&s.UsersCount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = db.QueryRow(`SELECT COUNT(repositories.id) FROM repositories`).Scan(\n\t\t&s.RepositoriesCount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = db.QueryRow(`SELECT COUNT(commits.id) FROM commits`).Scan(\n\t\t&s.CommitsCount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = db.QueryRow(`SELECT COUNT(commit_diff_deltas.id) FROM commit_diff_deltas`).Scan(\n\t\t&s.CommitDeltasCount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = db.QueryRow(`SELECT COUNT(features.id) FROM features`).Scan(\n\t\t&s.FeaturesCount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = db.QueryRow(`SELECT COUNT(gh_users.id) FROM gh_users`).Scan(\n\t\t&s.GhUsersCount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = db.QueryRow(`SELECT COUNT(gh_organizations.id) FROM gh_organizations`).Scan(\n\t\t&s.GhOrganizationsCount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = db.QueryRow(`SELECT COUNT(gh_repositories.id) FROM gh_repositories`).Scan(\n\t\t&s.GhRepositoriesCount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tstats = &s\n\n\treturn nil\n}", "func (client *NginxClient) GetStats() (*Stats, error) {\n\tinfo, err := client.GetNginxInfo()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tcaches, err := client.GetCaches()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tprocesses, err := client.GetProcesses()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tslabs, err := client.GetSlabs()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tcons, err := client.GetConnections()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\trequests, err := client.GetHTTPRequests()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tssl, err := client.GetSSL()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tzones, err := client.GetServerZones()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tupstreams, err := client.GetUpstreams()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tstreamZones, err := client.GetStreamServerZones()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tstreamUpstreams, err := client.GetStreamUpstreams()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tstreamZoneSync, err := client.GetStreamZoneSync()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tlocationZones, err := client.GetLocationZones()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tresolvers, err := client.GetResolvers()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tlimitReqs, err := client.GetHTTPLimitReqs()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tlimitConnsHTTP, err := client.GetHTTPConnectionsLimit()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tlimitConnsStream, err := client.GetStreamConnectionsLimit()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\tworkers, err := client.GetWorkers()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get stats: %w\", err)\n\t}\n\n\treturn &Stats{\n\t\tNginxInfo: *info,\n\t\tCaches: *caches,\n\t\tProcesses: *processes,\n\t\tSlabs: *slabs,\n\t\tConnections: *cons,\n\t\tHTTPRequests: *requests,\n\t\tSSL: *ssl,\n\t\tServerZones: *zones,\n\t\tStreamServerZones: *streamZones,\n\t\tUpstreams: *upstreams,\n\t\tStreamUpstreams: *streamUpstreams,\n\t\tStreamZoneSync: streamZoneSync,\n\t\tLocationZones: *locationZones,\n\t\tResolvers: *resolvers,\n\t\tHTTPLimitRequests: *limitReqs,\n\t\tHTTPLimitConnections: *limitConnsHTTP,\n\t\tStreamLimitConnections: *limitConnsStream,\n\t\tWorkers: workers,\n\t}, nil\n}", "func (s *server) getStats(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t_, err := s.getUserFromContextToken(r.Context())\n\tif err != nil {\n\t\twriteErrorResponse(w, fmt.Sprintf(\"unable to get stats: %v\", err))\n\t\treturn\n\t}\n\n\tsite, ok := r.URL.Query()[\"site\"]\n\tif !ok {\n\t\twriteErrorResponse(w, \"site id is not defined\")\n\t\treturn\n\t}\n\n\tparsedSite, err := strconv.ParseInt(site[0], 10, 64)\n\tif err != nil {\n\t\twriteErrorResponse(w, err.Error())\n\t\treturn\n\t}\n\n\tresp, err := s.hand.GetStats(&pb.GetStatsRequest{\n\t\tSiteID: parsedSite,\n\t})\n\tif err != nil {\n\t\twriteErrorResponse(w, err.Error())\n\t\treturn\n\t}\n\tw.WriteHeader(http.StatusOK)\n\tdata, err := json.Marshal(resp)\n\tif err != nil {\n\t\twriteErrorResponse(w, err.Error())\n\t\treturn\n\t}\n\tfmt.Fprintf(w, string(data))\n}", "func (s *searcher) refreshData(initialDir string) (*downloadStats, error) {\n\tif s.logger != nil {\n\t\ts.logger.Log(\"download\", \"Starting refresh of data\")\n\n\t\tif initialDir != \"\" {\n\t\t\ts.logger.Log(\"download\", fmt.Sprintf(\"reading files from %s\", initialDir))\n\t\t}\n\t}\n\n\tlastDataRefreshFailure.WithLabelValues(\"SDNs\").Set(float64(time.Now().Unix()))\n\n\tresults, err := ofacRecords(s.logger, initialDir)\n\tif err != nil {\n\t\tlastDataRefreshFailure.WithLabelValues(\"SDNs\").Set(float64(time.Now().Unix()))\n\n\t\treturn nil, fmt.Errorf(\"OFAC records: %v\", err)\n\t}\n\n\tsdns := precomputeSDNs(results.SDNs, results.Addresses, s.pipe)\n\tadds := precomputeAddresses(results.Addresses)\n\talts := precomputeAlts(results.AlternateIdentities)\n\n\tdeniedPersons, err := dplRecords(s.logger, initialDir)\n\tif err != nil {\n\t\tlastDataRefreshFailure.WithLabelValues(\"DPs\").Set(float64(time.Now().Unix()))\n\n\t\treturn nil, fmt.Errorf(\"DPL records: %v\", err)\n\t}\n\tdps := precomputeDPs(deniedPersons, s.pipe)\n\n\tconsolidatedLists, err := cslRecords(s.logger, initialDir)\n\tif err != nil {\n\t\tlastDataRefreshFailure.WithLabelValues(\"CSL\").Set(float64(time.Now().Unix()))\n\n\t\treturn nil, fmt.Errorf(\"CSL records: %v\", err)\n\t}\n\tssis := precomputeSSIs(consolidatedLists.SSIs, s.pipe)\n\tels := precomputeBISEntities(consolidatedLists.ELs, s.pipe)\n\n\tstats := &downloadStats{\n\t\t// OFAC\n\t\tSDNs: len(sdns),\n\t\tAlts: len(alts),\n\t\tAddresses: len(adds),\n\t\tSectoralSanctions: len(ssis),\n\t\t// BIS\n\t\tBISEntities: len(els),\n\t\tDeniedPersons: len(dps),\n\t}\n\tstats.RefreshedAt = lastRefresh(initialDir)\n\n\t// record prometheus metrics\n\tlastDataRefreshCount.WithLabelValues(\"SDNs\").Set(float64(len(sdns)))\n\tlastDataRefreshCount.WithLabelValues(\"SSIs\").Set(float64(len(ssis)))\n\tlastDataRefreshCount.WithLabelValues(\"BISEntities\").Set(float64(len(els)))\n\tlastDataRefreshCount.WithLabelValues(\"DPs\").Set(float64(len(dps)))\n\n\t// Set new records after precomputation (to minimize lock contention)\n\ts.Lock()\n\t// OFAC\n\ts.SDNs = sdns\n\ts.Addresses = adds\n\ts.Alts = alts\n\ts.SSIs = ssis\n\t// BIS\n\ts.DPs = dps\n\ts.BISEntities = els\n\t// metadata\n\ts.lastRefreshedAt = stats.RefreshedAt\n\ts.Unlock()\n\n\tif s.logger != nil {\n\t\ts.logger.Log(\"download\", \"Finished refresh of data\")\n\t}\n\n\t// record successful data refresh\n\tlastDataRefreshSuccess.WithLabelValues().Set(float64(time.Now().Unix()))\n\n\treturn stats, nil\n}", "func (s *Stats) Update(c *gin.Context, last time.Time) {\n\tStatsWith(c, s.update(last))\n}", "func (cache *Cache) GetStats(reset bool) Stats {\n\tcache.mu.Lock()\n\tdefer cache.mu.Unlock()\n\n\ts := Stats{\n\t\tReceived: cache.received,\n\t\tTotalReceived: cache.totalReceived + cache.received,\n\t\tExpected: cache.expected,\n\t\tTotalExpected: cache.totalExpected + cache.expected,\n\t\tESeqno: uint32(cache.cycle)<<16 | uint32(cache.last),\n\t}\n\n\tif reset {\n\t\tcache.totalExpected += cache.expected\n\t\tcache.expected = 0\n\t\tcache.totalReceived += cache.received\n\t\tcache.received = 0\n\t}\n\treturn s\n}", "func (s *Stats) Update(code int) {\n\ts.Lock()\n\ts.StatusCodes[code]++\n\ts.Unlock()\n}", "func (db *Database) UpdateStats(ctx context.Context, codes ...*VerificationCode) {\n\tissued := len(codes)\n\tif issued == 0 {\n\t\treturn\n\t}\n\tlogger := logging.FromContext(ctx).Named(\"issueapi.recordStats\")\n\tv := codes[0]\n\tdate := timeutils.UTCMidnight(v.CreatedAt)\n\n\t// If the issuer was a user, update the user stats for the day.\n\tif v.IssuingUserID != 0 {\n\t\tsql := `\n\t\t\tINSERT INTO user_stats (date, realm_id, user_id, codes_issued)\n\t\t\t\tVALUES ($1, $2, $3, $4)\n\t\t\tON CONFLICT (date, realm_id, user_id) DO UPDATE\n\t\t\t\tSET codes_issued = user_stats.codes_issued + $4`\n\n\t\tif err := db.db.Exec(sql, date, v.RealmID, v.IssuingUserID, issued).Error; err != nil {\n\t\t\tlogger.Warnw(\"failed to update user stats\", \"error\", err)\n\t\t}\n\t}\n\n\t// If the request was an API request, we might have an external issuer ID.\n\tif len(v.IssuingExternalID) != 0 {\n\t\tsql := `\n\t\t\tINSERT INTO external_issuer_stats (date, realm_id, issuer_id, codes_issued)\n\t\t\t\tVALUES ($1, $2, $3, $4)\n\t\t\tON CONFLICT (date, realm_id, issuer_id) DO UPDATE\n\t\t\t\tSET codes_issued = external_issuer_stats.codes_issued + $4\n\t\t`\n\n\t\tif err := db.db.Exec(sql, date, v.RealmID, v.IssuingExternalID, issued).Error; err != nil {\n\t\t\tlogger.Warnw(\"failed to update external-issuer stats\", \"error\", err)\n\t\t}\n\t}\n\n\t// If the issuer was a app, update the app stats for the day.\n\tif v.IssuingAppID != 0 {\n\t\tsql := `\n\t\t\tINSERT INTO authorized_app_stats (date, authorized_app_id, codes_issued)\n\t\t\t\tVALUES ($1, $2, $3)\n\t\t\tON CONFLICT (date, authorized_app_id) DO UPDATE\n\t\t\t\tSET codes_issued = authorized_app_stats.codes_issued + $3\n\t\t`\n\n\t\tif err := db.db.Exec(sql, date, v.IssuingAppID, issued).Error; err != nil {\n\t\t\tlogger.Warnw(\"failed to update authorized app stats\", \"error\", err)\n\t\t}\n\t}\n\n\tif v.RealmID != 0 {\n\t\t// Count the number of user initiated reports\n\t\tuserReports := 0\n\t\tfor _, vc := range codes {\n\t\t\tif vc.TestType == verifyapi.ReportTypeSelfReport {\n\t\t\t\tuserReports++\n\t\t\t}\n\t\t}\n\n\t\tsql := `\n\t\t\tINSERT INTO realm_stats(date, realm_id, codes_issued, user_reports_issued)\n\t\t\t\tVALUES ($1, $2, $3, $4)\n\t\t\tON CONFLICT (date, realm_id) DO UPDATE\n\t\t\t\tSET codes_issued = realm_stats.codes_issued + $3,\n\t\t\t\tuser_reports_issued = realm_stats.user_reports_issued + $4`\n\n\t\tif err := db.db.Exec(sql, date, v.RealmID, issued, userReports).Error; err != nil {\n\t\t\tlogger.Warnw(\"failed to update realm stats\", \"error\", err)\n\t\t}\n\t}\n}", "func (q AptCheckPlugin) FetchMetrics() (map[string]interface{}, error) {\n\tres, err := q.invokeAptCheck()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn map[string]interface{}{\n\t\t\"updates\": uint64(res.NumOfUpdates),\n\t\t\"security_updates\": uint64(res.NumOfSecurityUpdates),\n\t}, nil\n}", "func manualRefreshHandler(logger log.Logger, searcher *searcher, updates chan *DownloadStats, downloadRepo downloadRepository) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tlogger.Log(\"admin: refreshing data\")\n\n\t\tif stats, err := searcher.refreshData(\"\"); err != nil {\n\t\t\tlogger.LogErrorf(\"ERROR: admin: problem refreshing data: %v\", err)\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t} else {\n\t\t\tif err := downloadRepo.recordStats(stats); err != nil {\n\t\t\t\tmoovhttp.Problem(w, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tgo func() {\n\t\t\t\tupdates <- stats\n\t\t\t}()\n\n\t\t\tlogger.Info().With(log.Fields{\n\t\t\t\t\"SDNs\": log.Int(stats.SDNs),\n\t\t\t\t\"AltNames\": log.Int(stats.Alts),\n\t\t\t\t\"Addresses\": log.Int(stats.Addresses),\n\t\t\t\t\"SSI\": log.Int(stats.SectoralSanctions),\n\t\t\t\t\"DPL\": log.Int(stats.DeniedPersons),\n\t\t\t\t\"BISEntities\": log.Int(stats.BISEntities),\n\t\t\t\t\"UVL\": log.Int(stats.Unverified),\n\t\t\t\t\"ISN\": log.Int(stats.NonProliferationSanctions),\n\t\t\t\t\"FSE\": log.Int(stats.ForeignSanctionsEvaders),\n\t\t\t\t\"PLC\": log.Int(stats.PalestinianLegislativeCouncil),\n\t\t\t\t\"CAP\": log.Int(stats.CAPTA),\n\t\t\t\t\"DTC\": log.Int(stats.ITARDebarred),\n\t\t\t\t\"CMIC\": log.Int(stats.ChineseMilitaryIndustrialComplex),\n\t\t\t\t\"NS_MBS\": log.Int(stats.NonSDNMenuBasedSanctions),\n\t\t\t\t\"EUCSL\": log.Int(stats.EUCSL),\n\t\t\t\t\"UKCSL\": log.Int(stats.UKCSL),\n\t\t\t\t\"UKSanctionsList\": log.Int(stats.UKSanctionsList),\n\t\t\t}).Logf(\"admin: finished data refresh %v ago\", time.Since(stats.RefreshedAt))\n\n\t\t\tjson.NewEncoder(w).Encode(stats)\n\t\t}\n\t}\n}", "func (s *TXPoolServer) getStats() []uint64 {\n\ts.stats.RLock()\n\tdefer s.stats.RUnlock()\n\tret := make([]uint64, 0, len(s.stats.count))\n\tfor _, v := range s.stats.count {\n\t\tret = append(ret, v)\n\t}\n\treturn ret\n}", "func (c *tabletStatsCache) StatsUpdate(stats *discovery.TabletStats) {\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tkeyspace := stats.Target.Keyspace\n\tshard := stats.Target.Shard\n\tcell := stats.Tablet.Alias.Cell\n\ttabletType := stats.Target.TabletType\n\n\taliasKey := tabletToMapKey(stats)\n\tts, ok := c.statusesByAlias[aliasKey]\n\tif !stats.Up {\n\t\tif !ok {\n\t\t\t// Tablet doesn't exist and was recently deleted or changed its type. Panic as this is unexpected behavior.\n\t\t\tpanic(fmt.Sprintf(\"BUG: tablet (%v) doesn't exist\", aliasKey))\n\t\t}\n\t\t// The tablet still exists in our cache but was recently deleted or changed its type. Delete it now.\n\t\tc.statuses[keyspace][shard][cell][tabletType] = remove(c.statuses[keyspace][shard][cell][tabletType], stats.Tablet.Alias)\n\t\tdelete(c.statusesByAlias, aliasKey)\n\t\tc.tabletCountsByCell[cell]--\n\t\tif c.tabletCountsByCell[cell] == 0 {\n\t\t\tdelete(c.tabletCountsByCell, cell)\n\t\t}\n\t\treturn\n\t}\n\n\tif !ok {\n\t\t// Tablet isn't tracked yet so just add it.\n\t\tshards, ok := c.statuses[keyspace]\n\t\tif !ok {\n\t\t\tshards = make(map[string]map[string]map[topodatapb.TabletType][]*discovery.TabletStats)\n\t\t\tc.statuses[keyspace] = shards\n\t\t}\n\n\t\tcells, ok := c.statuses[keyspace][shard]\n\t\tif !ok {\n\t\t\tcells = make(map[string]map[topodatapb.TabletType][]*discovery.TabletStats)\n\t\t\tc.statuses[keyspace][shard] = cells\n\t\t}\n\n\t\ttypes, ok := c.statuses[keyspace][shard][cell]\n\t\tif !ok {\n\t\t\ttypes = make(map[topodatapb.TabletType][]*discovery.TabletStats)\n\t\t\tc.statuses[keyspace][shard][cell] = types\n\t\t}\n\n\t\ttablets, ok := c.statuses[keyspace][shard][cell][tabletType]\n\t\tif !ok {\n\t\t\ttablets = make([]*discovery.TabletStats, 0)\n\t\t\tc.statuses[keyspace][shard][cell][tabletType] = tablets\n\t\t}\n\n\t\tc.statuses[keyspace][shard][cell][tabletType] = append(c.statuses[keyspace][shard][cell][tabletType], stats)\n\t\tsort.Sort(byTabletUID(c.statuses[keyspace][shard][cell][tabletType]))\n\t\tc.statusesByAlias[aliasKey] = stats\n\t\tc.tabletCountsByCell[cell]++\n\t\treturn\n\t}\n\n\t// Tablet already exists so just update it in the cache.\n\t*ts = *stats\n}", "func (r IpvsPlugin) FetchMetrics() (map[string]float64, error) {\n file, err := os.Open(r.Target)\n if err != nil {\n return nil, err\n }\n defer file.Close()\n\n return Parse(file)\n}", "func (r *RollingStoreStats) Observe(stats *pdpb.StoreStats) {\n\tstatInterval := stats.GetInterval()\n\tinterval := time.Duration(statInterval.GetEndTimestamp()-statInterval.GetStartTimestamp()) * time.Second\n\tlog.Debug(\"update store stats\",\n\t\tzap.Uint64(\"key-write\", stats.KeysWritten),\n\t\tzap.Uint64(\"bytes-write\", stats.BytesWritten),\n\t\tzap.Uint64(\"key-read\", stats.KeysRead),\n\t\tzap.Uint64(\"bytes-read\", stats.BytesRead),\n\t\tzap.Uint64(\"query-write\", core.GetWriteQueryNum(stats.QueryStats)),\n\t\tzap.Uint64(\"query-read\", core.GetReadQueryNum(stats.QueryStats)),\n\t\tzap.Duration(\"interval\", interval),\n\t\tzap.Uint64(\"store-id\", stats.GetStoreId()))\n\tr.Lock()\n\tdefer r.Unlock()\n\treadQueryNum, writeQueryNum := core.GetReadQueryNum(stats.QueryStats), core.GetWriteQueryNum(stats.QueryStats)\n\tr.timeMedians[utils.StoreWriteBytes].Add(float64(stats.BytesWritten), interval)\n\tr.timeMedians[utils.StoreWriteKeys].Add(float64(stats.KeysWritten), interval)\n\tr.timeMedians[utils.StoreWriteQuery].Add(float64(writeQueryNum), interval)\n\tr.timeMedians[utils.StoreReadBytes].Add(float64(stats.BytesRead), interval)\n\tr.timeMedians[utils.StoreReadKeys].Add(float64(stats.KeysRead), interval)\n\tr.timeMedians[utils.StoreReadQuery].Add(float64(readQueryNum), interval)\n\n\t// Updates the cpu usages and disk rw rates of store.\n\tr.movingAvgs[utils.StoreCPUUsage].Add(collect(stats.GetCpuUsages()))\n\tr.movingAvgs[utils.StoreDiskReadRate].Add(collect(stats.GetReadIoRates()))\n\tr.movingAvgs[utils.StoreDiskWriteRate].Add(collect(stats.GetWriteIoRates()))\n}", "func (a *app) gatherStat() {\n\tabout, err := a.srv.About.Get().Fields(\"storageQuota\").Do()\n\tif err != nil {\n\t\tlog.Fatalf(\"Unable to execute an about request: %v\", err)\n\t}\n\n\ta.sq = about.StorageQuota\n}", "func (r *Refresher) maybeRefreshStats(\n\tctx context.Context,\n\tstopper *stop.Stopper,\n\ttableID descpb.ID,\n\trowsAffected int64,\n\tasOf time.Duration,\n) {\n\ttableStats, err := r.cache.GetTableStats(ctx, tableID)\n\tif err != nil {\n\t\tlog.Errorf(ctx, \"failed to get table statistics: %v\", err)\n\t\treturn\n\t}\n\n\tvar rowCount float64\n\tmustRefresh := false\n\tif stat := mostRecentAutomaticStat(tableStats); stat != nil {\n\t\t// Check if too much time has passed since the last refresh.\n\t\t// This check is in place to corral statistical outliers and avoid a\n\t\t// case where a significant portion of the data in a table has changed but\n\t\t// the stats haven't been refreshed. Randomly add some extra time to the\n\t\t// limit check to avoid having multiple nodes trying to create stats at\n\t\t// the same time.\n\t\t//\n\t\t// Note that this can cause some unnecessary runs of CREATE STATISTICS\n\t\t// in the case where there is a heavy write load followed by a very light\n\t\t// load. For example, suppose the average refresh time is 1 hour during\n\t\t// the period of heavy writes, and the average refresh time should be 1\n\t\t// week during the period of light load. It could take ~16 refreshes over\n\t\t// 3-4 weeks before the average settles at around 1 week. (Assuming the\n\t\t// refresh happens at exactly 2x the current average, and the average\n\t\t// refresh time is calculated from the most recent 4 refreshes. See the\n\t\t// comment in stats/delete_stats.go.)\n\t\tmaxTimeBetweenRefreshes := stat.CreatedAt.Add(2*avgRefreshTime(tableStats) + r.extraTime)\n\t\tif timeutil.Now().After(maxTimeBetweenRefreshes) {\n\t\t\tmustRefresh = true\n\t\t}\n\t\trowCount = float64(stat.RowCount)\n\t} else {\n\t\t// If there are no statistics available on this table, we must perform a\n\t\t// refresh.\n\t\tmustRefresh = true\n\t}\n\n\ttargetRows := int64(rowCount*AutomaticStatisticsFractionStaleRows.Get(&r.st.SV)) +\n\t\tAutomaticStatisticsMinStaleRows.Get(&r.st.SV)\n\tif !mustRefresh && rowsAffected < math.MaxInt32 && r.randGen.randInt(targetRows) >= rowsAffected {\n\t\t// No refresh is happening this time.\n\t\treturn\n\t}\n\n\tif err := r.refreshStats(ctx, tableID, asOf); err != nil {\n\t\tif errors.Is(err, ConcurrentCreateStatsError) {\n\t\t\t// Another stats job was already running. Attempt to reschedule this\n\t\t\t// refresh.\n\t\t\tif mustRefresh {\n\t\t\t\t// For the cases where mustRefresh=true (stats don't yet exist or it\n\t\t\t\t// has been 2x the average time since a refresh), we want to make sure\n\t\t\t\t// that maybeRefreshStats is called on this table during the next\n\t\t\t\t// cycle so that we have another chance to trigger a refresh. We pass\n\t\t\t\t// rowsAffected=0 so that we don't force a refresh if another node has\n\t\t\t\t// already done it.\n\t\t\t\tr.mutations <- mutation{tableID: tableID, rowsAffected: 0}\n\t\t\t} else {\n\t\t\t\t// If this refresh was caused by a \"dice roll\", we want to make sure\n\t\t\t\t// that the refresh is rescheduled so that we adhere to the\n\t\t\t\t// AutomaticStatisticsFractionStaleRows statistical ideal. We\n\t\t\t\t// ensure that the refresh is triggered during the next cycle by\n\t\t\t\t// passing a very large number for rowsAffected.\n\t\t\t\tr.mutations <- mutation{tableID: tableID, rowsAffected: math.MaxInt32}\n\t\t\t}\n\t\t\treturn\n\t\t}\n\n\t\t// Log other errors but don't automatically reschedule the refresh, since\n\t\t// that could lead to endless retries.\n\t\tlog.Warningf(ctx, \"failed to create statistics on table %d: %v\", tableID, err)\n\t\treturn\n\t}\n}", "func (api *API) getGroupUpdatesStats(group *Group) (*UpdatesStats, error) {\n\tvar updatesStats UpdatesStats\n\n\tpackageVersion := \"\"\n\tif group.Channel != nil && group.Channel.Package != nil {\n\t\tpackageVersion = group.Channel.Package.Version\n\t}\n\tquery, _, err := goqu.From(\"instance_application\").Select(\n\t\tgoqu.COUNT(\"*\").As(\"total_instances\"),\n\t\tgoqu.COALESCE(goqu.SUM(goqu.L(\"case when last_update_version = ? then 1 else 0 end\", packageVersion)), 0).As(\"updates_to_current_version_granted\"),\n\t\tgoqu.COALESCE(goqu.SUM(goqu.L(\"case when update_in_progress = 'false' and last_update_version = ? then 1 else 0 end\", packageVersion)), 0).As(\"updates_to_current_version_attempted\"),\n\t\tgoqu.COALESCE(goqu.SUM(goqu.L(\"case when update_in_progress = 'false' and last_update_version = ? and last_update_version = version then 1 else 0 end\", packageVersion)), 0).As(\"updates_to_current_version_succeeded\"),\n\t\tgoqu.COALESCE(goqu.SUM(goqu.L(\"case when update_in_progress = 'false' and last_update_version = ? and last_update_version != version then 1 else 0 end\", packageVersion)), 0).As(\"updates_to_current_version_failed\"),\n\t\tgoqu.COALESCE(goqu.SUM(goqu.L(\"case when last_update_granted_ts > now() at time zone 'utc' - interval ? then 1 else 0 end\", group.PolicyPeriodInterval)), 0).As(\"updates_granted_in_last_period\"),\n\t\tgoqu.COALESCE(goqu.SUM(goqu.L(\"case when update_in_progress = 'true' and now() at time zone 'utc' - last_update_granted_ts <= interval ? then 1 else 0 end\", group.PolicyUpdateTimeout)), 0).As(\"updates_in_progress\"),\n\t\tgoqu.COALESCE(goqu.SUM(goqu.L(\"case when update_in_progress = 'true' and now() at time zone 'utc' - last_update_granted_ts > interval ? then 1 else 0 end\", group.PolicyUpdateTimeout)), 0).As(\"updates_timed_out\"),\n\t).Where(goqu.C(\"group_id\").Eq(group.ID), goqu.L(\"last_check_for_updates > now() at time zone 'utc' - interval ?\", validityInterval),\n\t\tgoqu.L(ignoreFakeInstanceCondition(\"instance_id\")),\n\t).ToSQL()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = api.db.QueryRowx(query).StructScan(&updatesStats)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &updatesStats, nil\n}", "func (s *Service) updateCache(st *api.Stat) (err error) {\n\tvar (\n\t\tstat3 = &api.Stat{\n\t\t\tAid: st.Aid,\n\t\t\tView: int32(st.View),\n\t\t\tDanmaku: int32(st.Danmaku),\n\t\t\tReply: int32(st.Reply),\n\t\t\tFav: int32(st.Fav),\n\t\t\tCoin: int32(st.Coin),\n\t\t\tShare: int32(st.Share),\n\t\t\tNowRank: int32(st.NowRank),\n\t\t\tHisRank: int32(st.HisRank),\n\t\t\tLike: int32(st.Like),\n\t\t\tDisLike: 0,\n\t\t}\n\t\tclick *archive.Click3\n\t\tupclick = true\n\t)\n\tif click, err = s.dao.Click(context.TODO(), st.Aid); err != nil {\n\t\tupclick = false\n\t}\n\tif click == nil {\n\t\tclick = &archive.Click3{}\n\t}\n\tfor _, mc := range s.memcaches {\n\t\tvar c = context.TODO()\n\t\tconn := mc.Get(c)\n\t\tif err = conn.Set(&memcache.Item{Key: statPBKey(stat3.Aid), Object: stat3, Flags: memcache.FlagProtobuf, Expiration: 0}); err != nil {\n\t\t\tlog.Error(\"conn1.Set(%s, %+v) error(%v)\", statPBKey(stat3.Aid), stat3, err)\n\t\t}\n\t\tif upclick {\n\t\t\tif err = conn.Set(&memcache.Item{Key: clickPBKey(stat3.Aid), Object: click, Flags: memcache.FlagProtobuf, Expiration: 0}); err != nil {\n\t\t\t\tlog.Error(\"conn1.Set(%s, %+v) error(%v)\", clickPBKey(stat3.Aid), click, err)\n\t\t\t}\n\t\t}\n\t\tif err == nil {\n\t\t\tlog.Info(\"update cache aid(%d) stat(%+v) success\", st.Aid, stat3)\n\t\t\tlog.Info(\"update cache aid(%d) click(%+v) success\", st.Aid, click)\n\t\t}\n\t\tconn.Close()\n\t}\n\treturn\n}", "func UpdateStatistics(settings *playfab.Settings, postData *UpdateStatisticsRequestModel, entityToken string) (*UpdateStatisticsResponseModel, error) {\n if entityToken == \"\" {\n return nil, playfab.NewCustomError(\"entityToken should not be an empty string\", playfab.ErrorGeneric)\n }\n b, errMarshal := json.Marshal(postData)\n if errMarshal != nil {\n return nil, playfab.NewCustomError(errMarshal.Error(), playfab.ErrorMarshal)\n }\n\n sourceMap, err := playfab.Request(settings, b, \"/Statistic/UpdateStatistics\", \"X-EntityToken\", entityToken)\n if err != nil {\n return nil, err\n }\n \n result := &UpdateStatisticsResponseModel{}\n\n config := mapstructure.DecoderConfig{\n DecodeHook: playfab.StringToDateTimeHook,\n Result: result,\n }\n \n decoder, errDecoding := mapstructure.NewDecoder(&config)\n if errDecoding != nil {\n return nil, playfab.NewCustomError(errDecoding.Error(), playfab.ErrorDecoding)\n }\n \n errDecoding = decoder.Decode(sourceMap)\n if errDecoding != nil {\n return nil, playfab.NewCustomError(errDecoding.Error(), playfab.ErrorDecoding)\n }\n\n return result, nil\n}", "func (p S3RequestsPlugin) FetchMetrics() (map[string]float64, error) {\n\tstats := make(map[string]float64)\n\n\tfor _, met := range s3RequestMetricsGroup {\n\t\tv, err := getLastPointFromCloudWatch(p.CloudWatch, p.BucketName, p.FilterID, met)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"%s: %s\", met, err)\n\t\t} else if v != nil {\n\t\t\tstats = mergeStatsFromDatapoint(stats, v, met)\n\t\t}\n\t}\n\treturn stats, nil\n}", "func (s *Server) statproc() {\n\tvar (\n\t\tv *volume.Volume\n\t\tolds *stat.Stats\n\t\tnews = new(stat.Stats)\n\t)\n\tfor {\n\t\tolds = s.info.Stats\n\t\t*news = *olds\n\t\ts.info.Stats = news // use news instead, for current display\n\t\tolds.Reset()\n\t\tfor _, v = range s.store.Volumes {\n\t\t\tv.Stats.Calc()\n\t\t\tolds.Merge(v.Stats)\n\t\t}\n\t\tolds.Calc()\n\t\ts.info.Stats = olds\n\t\ttime.Sleep(statDuration)\n\t}\n}", "func (this *ReceiverHolder) stats(c *gin.Context) {\n\n\tflightData := this.receiver.GetInFlightRavens()\n\n\tdeadBoxData := this.receiver.GetDeadBoxCount()\n\tboxes := make([]string, 0)\n\tfor _, box := range this.receiver.msgReceivers {\n\t\tboxes = append(boxes, box.id)\n\t}\n\n\tdata := gin.H{\n\t\t\"Queue\": this.receiver.source.GetName(),\n\t\t\"IsReliable\": this.receiver.options.isReliable,\n\t\t\"Boxes\": boxes,\n\t\t\"Inflight\": flightData,\n\t\t\"DeadBox\": deadBoxData,\n\t}\n\tc.JSON(200, data)\n}", "func (d *hdbDriver) Stats() *Stats { return d.metrics.stats() }", "func FetchNodeStats(http *helper.HTTP, host *url.URL) (interface{}, error) {\n\turi := strings.TrimSuffix(host.String(), \"/\") + NODES_LOCAL_STATS_PATH\n\tnodeBody := NodeStatsBody{}\n\n\tbody, err := http.FetchContent(uri)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = json.Unmarshal(body, &nodeBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar nodeInfo NodeStatsInfo\n\tfor _, node := range nodeBody.Nodes {\n\t\tnodeInfo = node\n\t\tnodeInfo.ClusterName = nodeBody.ClusterName\n\t\tbreak\n\t}\n\n\treturn &nodeInfo, nil\n}", "func (s *SingleRuntime) GetStats() *models.NativeStatsCollection {\n\trAPI := \"\"\n\tif s.worker != 0 {\n\t\trAPI = fmt.Sprintf(\"%s@%v\", s.socketPath, s.worker)\n\t} else {\n\t\trAPI = s.socketPath\n\t}\n\tresult := &models.NativeStatsCollection{RuntimeAPI: rAPI}\n\trawdata, err := s.ExecuteRaw(\"show stat\")\n\tif err != nil {\n\t\tresult.Error = err.Error()\n\t\treturn result\n\t}\n\tlines := strings.Split(rawdata[2:], \"\\n\")\n\tstats := []*models.NativeStat{}\n\tkeys := strings.Split(lines[0], \",\")\n\t//data := []map[string]string{}\n\tfor i := 1; i < len(lines); i++ {\n\t\tdata := map[string]string{}\n\t\tline := strings.Split(lines[i], \",\")\n\t\tif len(line) < len(keys) {\n\t\t\tcontinue\n\t\t}\n\t\tfor index, key := range keys {\n\t\t\tif len(line[index]) > 0 {\n\t\t\t\tdata[key] = line[index]\n\t\t\t}\n\t\t}\n\t\toneLineData := &models.NativeStat{}\n\t\ttString := strings.ToLower(line[1])\n\t\tif tString == \"backend\" || tString == \"frontend\" {\n\t\t\toneLineData.Name = line[0]\n\t\t\toneLineData.Type = tString\n\t\t} else {\n\t\t\toneLineData.Name = tString\n\t\t\toneLineData.Type = \"server\"\n\t\t\toneLineData.BackendName = line[0]\n\t\t}\n\n\t\tvar st models.NativeStatStats\n\t\tdecoder, err := mapstructure.NewDecoder(&mapstructure.DecoderConfig{\n\t\t\tResult: &st,\n\t\t\tWeaklyTypedInput: true,\n\t\t\tTagName: \"json\",\n\t\t})\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\terr = decoder.Decode(data)\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\t\toneLineData.Stats = &st\n\n\t\tstats = append(stats, oneLineData)\n\t}\n\tresult.Stats = stats\n\treturn result\n}", "func (ghs *Service) GetStats() gclient.RepoStats {\n\tclient, err := gclient.NewInstallationClient(ghs.AppID, ghs.InstallationID, ghs.PEMFile, &ghs.logger)\n\n\tif err != nil {\n\t\tghs.logger.Error().Err(err).Msg(\"NewInstallationClient\")\n\t\treturn gclient.RepoStats{}\n\t}\n\n\treturn client.GetStats(\"ThundeRatz\")\n}", "func (s *Server) GetStats(w rest.ResponseWriter, r *rest.Request) {\n\tresponse := models.BaseResponse{}\n\tresponse.Init(w)\n\n\tcurrentUser, err := s.LoginProcess(response, r)\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tuserID, err := s.GetUserIDFromParams(r)\n\n\tif err != nil {\n\t\tresponse.SendError(err.Error())\n\t\treturn\n\t}\n\n\tif userID == 0 {\n\t\tuserID = currentUser.ID\n\t}\n\n\tb := new(badgecontroller.Badge)\n\n\tstats, err := b.List(s.Db, userID)\n\n\tif err != nil {\n\t\tresponse.SendError(err.Error())\n\t\treturn\n\t}\n\n\tresponse.SendSuccess(stats)\n}", "func (m *MetricSet) Fetch(r mb.ReporterV2) {\n\tvar uptime sigar.Uptime\n\tif err := uptime.Get(); err != nil {\n\t\tr.Error(errors.Wrap(err, \"failed to get uptime\"))\n\t\treturn\n\t}\n\n\tr.Event(mb.Event{\n\t\tMetricSetFields: common.MapStr{\n\t\t\t\"duration\": common.MapStr{\n\t\t\t\t\"ms\": int64(uptime.Length * 1000),\n\t\t\t},\n\t\t},\n\t})\n}", "func (tc *LegacyTabletStatsCache) StatsUpdate(ts *LegacyTabletStats) {\n\tif ts.Target.TabletType != topodatapb.TabletType_MASTER &&\n\t\tts.Tablet.Alias.Cell != tc.cell &&\n\t\ttc.getAliasByCell(ts.Tablet.Alias.Cell) != tc.getAliasByCell(tc.cell) {\n\t\t// this is for a non-master tablet in a different cell and a different alias, drop it\n\t\treturn\n\t}\n\n\te := tc.getOrCreateEntry(ts.Target)\n\te.mu.Lock()\n\tdefer e.mu.Unlock()\n\n\t// Update our full map.\n\ttrivialNonMasterUpdate := false\n\tif existing, ok := e.all[ts.Key]; ok {\n\t\tif ts.Up {\n\t\t\t// We have an existing entry, and a new entry.\n\t\t\t// Remember if they are both good (most common case).\n\t\t\ttrivialNonMasterUpdate = existing.LastError == nil && existing.Serving && ts.LastError == nil &&\n\t\t\t\tts.Serving && ts.Target.TabletType != topodatapb.TabletType_MASTER && existing.TrivialStatsUpdate(ts)\n\n\t\t\t// We already have the entry, update the\n\t\t\t// values if necessary. (will update both\n\t\t\t// 'all' and 'healthy' as they use pointers).\n\t\t\tif !trivialNonMasterUpdate {\n\t\t\t\t*existing = *ts\n\t\t\t}\n\t\t} else {\n\t\t\t// We have an entry which we shouldn't. Remove it.\n\t\t\tdelete(e.all, ts.Key)\n\t\t}\n\t} else {\n\t\tif ts.Up {\n\t\t\t// Add the entry.\n\t\t\te.all[ts.Key] = ts\n\t\t} else {\n\t\t\t// We were told to remove an entry which we\n\t\t\t// didn't have anyway, nothing should happen.\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Update our healthy list.\n\tvar allArray []*LegacyTabletStats\n\tif ts.Target.TabletType == topodatapb.TabletType_MASTER {\n\t\t// The healthy list is different for TabletType_MASTER: we\n\t\t// only keep the most recent one.\n\t\te.updateHealthyMapForMaster(ts)\n\t} else {\n\t\t// For non-master, if it is a trivial update,\n\t\t// we just skip everything else. We don't even update the\n\t\t// aggregate stats.\n\t\tif trivialNonMasterUpdate {\n\t\t\treturn\n\t\t}\n\n\t\t// Now we need to do some work. Recompute our healthy list.\n\t\tallArray = make([]*LegacyTabletStats, 0, len(e.all))\n\t\tfor _, s := range e.all {\n\t\t\tallArray = append(allArray, s)\n\t\t}\n\t\te.healthy = FilterLegacyStatsByReplicationLag(allArray)\n\t}\n}", "func (nsc *NilConsumerStatsCollector) UpdateGetRecordsDuration(time.Duration) {}", "func (t *Tracker) GetStats() *TrackerReport {\n\tlogrus.Info(\"Fetching stats\")\n\tr := NewReport()\n\tt.Query(func(t *Tracker) {\n\t\tr.Total = t.total.get()\n\t\tfor k, v := range t.hosts {\n\t\t\tr.ByHost[k] = v.get()\n\t\t}\n\t\tfor k, v := range t.services {\n\t\t\tr.ByService[k] = v.get()\n\t\t}\n\t\tfor k, v := range t.subServices {\n\t\t\tr.BySubService[k] = v.get()\n\t\t}\n\t\tfor k, v := range t.hostTimes {\n\t\t\tr.LastSeenByHost[k] = v.Unix()\n\t\t}\n\t})\n\n\treturn r\n}", "func (d *Dao) Stat(c context.Context, mid int64, ip string) (st *growup.Stat, err error) {\n\tparams := url.Values{}\n\tparams.Set(\"mid\", strconv.FormatInt(mid, 10))\n\tparams.Set(\"ts\", strconv.FormatInt(time.Now().UnixNano()/1000000, 10))\n\tvar res struct {\n\t\tCode int `json:\"code\"`\n\t\tData *growup.Stat\n\t\tMessage string `json:\"message\"`\n\t}\n\tif err = d.client.Get(c, d.statURL, ip, params, &res); err != nil {\n\t\tlog.Error(\"growup url(%s) response(%v) error(%v)\", d.statURL+\"?\"+params.Encode(), res, err)\n\t\terr = ecode.CreativeOrderAPIErr\n\t\treturn\n\t}\n\tif res.Code != 0 {\n\t\tlog.Error(\"growup url(%s) res(%v)\", d.statURL, res)\n\t\terr = ecode.CreativeOrderAPIErr\n\t\treturn\n\t}\n\tst = res.Data\n\treturn\n}", "func UpdateMetrics(result *Results) {\n\n\t// Publish system variables\n\tupTimeGauge.Set(float64(result.SysMonitorInfo.Uptime))\n\tcpuUsageGauge.Set(float64(result.SysMonitorInfo.CpuUsagePercent))\n\n\t// Memory\n\tmemUsagePercentGauge.Set(result.SysMonitorInfo.MemUsagePercent)\n\tmemTotalGauge.Set(float64(result.SysMonitorInfo.MemTotal))\n\tmemAvailableGauge.Set(float64(result.SysMonitorInfo.MemAvailable))\n\n\t// Bandwidth\n\tbandwidthUsageTotalGauge.Set(float64(result.SysMonitorInfo.BandwidthUsageTotal))\n\tbandwidthUsageSentGauge.Set(float64(result.SysMonitorInfo.BandwidthUsageSent))\n\tbandwidthUsageRecvGauge.Set(float64(result.SysMonitorInfo.BandwidthUsageRecv))\n\n\tfor _, driveUsage := range result.SysMonitorInfo.DriveUsage {\n\t\t// \"drive_path\", \"available\", \"growth_rate\", \"full_in\", \"physical_drive\"\n\n\t\tdays := strconv.FormatFloat(driveUsage.DaysTillFull, 'f', 3, 64)\n\n\t\tif math.IsInf(driveUsage.DaysTillFull, 0) {\n\t\t\tdays = \"10 years\"\n\t\t}\n\n\t\tdriveSpace.WithLabelValues(driveUsage.Path,\n\t\t\tstrconv.FormatFloat(driveUsage.PercentUsed, 'f', 3, 64),\n\t\t\tstrconv.FormatUint(driveUsage.GrowthPerDayBytes, 10),\n\t\t\tdays,\n\t\t\tdriveUsage.VolumeName).Set(driveUsage.PercentUsed)\n\t}\n\n\t// Publish endpoints being monitored\n\tfor _, uptimeResponse := range result.UptimeList {\n\n\t\tif uptimeResponse.ResponseCode == 200 {\n\t\t\tendpointAvailable.WithLabelValues(uptimeResponse.Endpoint).Set(1)\n\t\t} else {\n\t\t\tendpointAvailable.WithLabelValues(uptimeResponse.Endpoint).Set(0)\n\t\t}\n\n\t\tendpointDuration.WithLabelValues(uptimeResponse.Endpoint).Set(uptimeResponse.ResponseTime.Seconds())\n\t}\n\n\tfor _, backupInfo := range result.BackupInfoList {\n\n\t\t/*\n\t\t\tif backupInfo.WasBackedUp {\n\t\t\t\tbackupsDone.WithLabelValues(backupInfo.Folder).Set(1)\n\t\t\t} else {\n\t\t\t\tbackupsDone.WithLabelValues(backupInfo.Folder).Set(0)\n\t\t\t}\n\t\t*/\n\n\t\t// {\"backup_directory\", \"backup_in_last_24_hours\", \"last_backup_size\", \"last_backup_date\", \"last_backup_time\"})\n\n\t\t// backupsSize.WithLabelValues(backupInfo.Folder).Set(float64(backupInfo.BackupFileSize))\n\n\t\tbackupInfoGauge.WithLabelValues(backupInfo.Folder,\n\t\t\tbtoa(backupInfo.WasBackedUp),\n\t\t\titoa(backupInfo.LastBackupSize),\n\t\t\tttoa(backupInfo.LastBackupTime),\n\t\t\tbackupInfo.LastBackupFile).Set(btof(backupInfo.WasBackedUp))\n\t}\n\n\t// TODO: This loop is not needed, you can build the summary on the first loop\n\tvar too_many_lines = 500\n\tfor _, logLine := range result.LoglineList {\n\n\t\tsummary, ok := result.LogSummary[logLine.LogPath]\n\n\t\tif ok == false {\n\t\t\tsummary = LogSummary{}\n\t\t\tsummary.StatusCount = make(map[string]int64)\n\t\t\tsummary.SeverityLevelCount = make(map[string]int64)\n\t\t}\n\n\t\tsummary.StatusCount[logLine.StatusCode] = summary.StatusCount[logLine.StatusCode] + 1\n\n\t\tif len(logLine.Severity) > 0 {\n\t\t\tsummary.SeverityLevelCount[logLine.Severity] = summary.SeverityLevelCount[logLine.Severity] + 1\n\t\t}\n\n\t\tresult.LogSummary[logLine.LogPath] = summary\n\n\t\tif too_many_lines <= 0 {\n\t\t\t// Pending a better solution, let's not allow the processing\n\t\t\t// of too many lines, to not kill the server\n\t\t\tlLog.Print(\"Too many lines for a single tick to process\")\n\t\t\tbreak\n\t\t}\n\n\t}\n\n\t// Set the values for the logs. We use two labels (logpath, code)\n\tfor logFilePath, logSummary := range result.LogSummary {\n\n\t\tfor s, value := range logSummary.StatusCount {\n\t\t\tstatusCodes.WithLabelValues(logFilePath, s).Set(float64(value))\n\t\t}\n\n\t\tfor s, value := range logSummary.SeverityLevelCount {\n\t\t\tseverity.WithLabelValues(logFilePath, s).Set(float64(value))\n\t\t}\n\n\t}\n}", "func (s *Stats) Get() Stats {\n\n\tout := Stats{\n\t\tEventsTotal: atomic.LoadUint64(&s.EventsTotal),\n\t\tEventsUpdate: atomic.LoadUint64(&s.EventsUpdate),\n\t\tEventsDestroy: atomic.LoadUint64(&s.EventsDestroy),\n\t}\n\n\t// Get Update source stats if present.\n\tif s.UpdateSourceStats != nil {\n\t\ts := s.UpdateSourceStats.Get()\n\t\tout.UpdateSourceStats = &s\n\t}\n\n\t// Get Destroy source stats if present.\n\tif s.DestroySourceStats != nil {\n\t\ts := s.DestroySourceStats.Get()\n\t\tout.DestroySourceStats = &s\n\t}\n\n\treturn out\n}", "func (s *shard) loadStats(stats *CacheStats) {\n\t//get\n\tstats.Hits += atomic.LoadUint64(&s.hits)\n\tstats.Misses += atomic.LoadUint64(&s.misses)\n\n\tstats.Collisions += atomic.LoadUint64(&s.collisions)\n\n\t// del\n\tstats.DelHits += atomic.LoadUint64(&s.delHits)\n\tstats.DelMisses += atomic.LoadUint64(&s.delMisses)\n\n\ts.rwMutex.RLock()\n\tstats.EntriesCount += uint64(len(s.entryIndexes))\n\tstats.CacheBytes += s.ring.Cap()\n\ts.rwMutex.RUnlock()\n}", "func (s *BatchProvidingSystem) Stat(ctx context.Context) (BatchedProviderStats, error) {\n\t// TODO: Does it matter that there is no locking around the total+average values?\n\treturn BatchedProviderStats{\n\t\tTotalProvides: s.totalProvides,\n\t\tLastReprovideBatchSize: s.lastReprovideBatchSize,\n\t\tAvgProvideDuration: s.avgProvideDuration,\n\t\tLastReprovideDuration: s.lastReprovideDuration,\n\t}, nil\n}", "func (pool *Pool) RefreshStats() (err error) {\n\tif 0 != C.refresh_stats(pool.list) {\n\t\treturn errors.New(\"error refreshing stats\")\n\t}\n\treturn nil\n}", "func (r *Repository) Stats(context.Context) (*adagio.Stats, error) {\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\n\tnodeCounts := &adagio.Stats_NodeCounts{}\n\n\tfor _, runState := range r.runs {\n\t\tfor _, node := range runState.lookup {\n\t\t\tswitch node.Status {\n\t\t\tcase adagio.Node_WAITING:\n\t\t\t\tnodeCounts.WaitingCount++\n\t\t\tcase adagio.Node_READY:\n\t\t\t\tnodeCounts.ReadyCount++\n\t\t\tcase adagio.Node_RUNNING:\n\t\t\t\tnodeCounts.RunningCount++\n\t\t\tcase adagio.Node_COMPLETED:\n\t\t\t\tnodeCounts.CompletedCount++\n\t\t\t}\n\t\t}\n\t}\n\n\treturn &adagio.Stats{\n\t\tRunCount: int64(len(r.runs)),\n\t\tNodeCounts: nodeCounts,\n\t}, nil\n}", "func (d *Dao) Stats(c context.Context, aids []int64, ip string) (a map[int64]*api.Stat, err error) {\n\tvar arg = &archive.ArgAids2{Aids: aids, RealIP: ip}\n\tif a, err = d.arc.Stats3(c, arg); err != nil {\n\t\tlog.Error(\"rpc Stats (%v) error(%v)\", aids, err)\n\t\terr = ecode.CreativeArcServiceErr\n\t}\n\treturn\n}", "func (h *Handlers) GetStats(w http.ResponseWriter, r *http.Request) {\n\tdataJSON, err := h.pkgManager.GetStatsJSON(r.Context())\n\tif err != nil {\n\t\th.logger.Error().Err(err).Str(\"method\", \"GetStats\").Send()\n\t\thttp.Error(w, \"\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\thelpers.RenderJSON(w, dataJSON, helpers.DefaultAPICacheMaxAge)\n}", "func (s *service) Stats() Stats {\n\ts.m.Lock()\n\tdefer s.m.Unlock()\n\n\tstats := Stats{\n\t\tServiceIdentity: s.serviceIdentity(),\n\t\tEndpoints: make([]*EndpointStats, 0),\n\t\tType: StatsResponseType,\n\t\tStarted: s.started,\n\t}\n\tfor _, endpoint := range s.endpoints {\n\t\tendpointStats := &EndpointStats{\n\t\t\tName: endpoint.stats.Name,\n\t\t\tSubject: endpoint.stats.Subject,\n\t\t\tNumRequests: endpoint.stats.NumRequests,\n\t\t\tNumErrors: endpoint.stats.NumErrors,\n\t\t\tLastError: endpoint.stats.LastError,\n\t\t\tProcessingTime: endpoint.stats.ProcessingTime,\n\t\t\tAverageProcessingTime: endpoint.stats.AverageProcessingTime,\n\t\t}\n\t\tif s.StatsHandler != nil {\n\t\t\tdata, _ := json.Marshal(s.StatsHandler(endpoint))\n\t\t\tendpointStats.Data = data\n\t\t}\n\t\tstats.Endpoints = append(stats.Endpoints, endpointStats)\n\t}\n\treturn stats\n}", "func (m *Monitor) ReadStat() pb.Stat {\n\n\tcpu := make(chan float32, 1)\n\tmemory := make(chan float32, 1)\n\tbandwidthR := make(chan float32, 1)\n\tbandwidthW := make(chan float32, 1)\n\n\tgo func() {\n\t\tm.Stat.CPUStats()\n\t\ttime.Sleep(time.Second)\n\t\tcpus := m.Stat.CPUStats()\n\n\t\t//logger.LogDebugf(\"CPU Idle %f %f %f %f %f %f\\n\", cpus.User, cpus.Kernel, cpus.Idle, cpus.IOWait, cpus.Swap, cpus.Nice)\n\t\tcpu <- (100 - float32(cpus.Idle)) / 100 // range from 0 ~ 1\n\n\t}()\n\n\tgo func() {\n\n\t\tm.Stat.MemStats()\n\t\ttime.Sleep(time.Second)\n\t\tmemorys := m.Stat.MemStats()\n\n\t\tmemory <- (float32(memorys.Used) / float32(memorys.Total)) //range from 0 ~ 1\n\n\t}()\n\n\tgo func() {\n\n\t\tm.Stat.NetIOStats()\n\t\ttime.Sleep(time.Second)\n\t\tio := m.Stat.NetIOStats()\n\n\t\tif m.adapterIdx == -1 {\n\t\t\tm.adapterIdx = m.GetAdapterIndex(io)\n\t\t\tif m.adapterIdx == -1 {\n\t\t\t\tlogger.LogErr(\"Cannot find the adapter name\")\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\ttotalbandwidth := config.CfgWorker.LoadReport.MaxNetworkBandwidth\n\n\t\tbandwidthR <- (float32(io[m.adapterIdx].RX) / 1024 / 1024 * 8) / totalbandwidth // range from 0 ~ 1\n\t\tbandwidthW <- (float32(io[m.adapterIdx].TX) / 1024 / 1024 * 8) / totalbandwidth // range from 0 ~ 1\n\n\t}()\n\n\treturn pb.Stat{CpuUsageRate: <-cpu, MemoryUsageRate: <-memory, ReadNetworkIOUsageRate: <-bandwidthR, WriteNetworkIOUsageRate: <-bandwidthW}\n\n}", "func (runner *McRunner) updateStatus() {\n\trunner.WaitGroup.Add(1)\n\tdefer runner.WaitGroup.Done()\n\tfor {\n\t\tselect {\n\t\tcase <-runner.StatusRequestChannel:\n\t\t\tif runner.State != Running {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tstatus := new(Status)\n\t\t\tstatus.Name = runner.Settings.Name\n\t\t\tstatus.PlayerMax = runner.Settings.MaxPlayers\n\t\t\tswitch runner.State {\n\t\t\tcase NotRunning:\n\t\t\t\tstatus.Status = \"Not Running\"\n\t\t\tcase Starting:\n\t\t\t\tstatus.Status = \"Starting\"\n\t\t\tcase Running:\n\t\t\t\tstatus.Status = \"Running\"\n\t\t\t}\n\t\t\tstatus.ActiveTime = int(time.Since(runner.startTime).Seconds())\n\n\t\t\tproc, _ := process.NewProcess(int32(runner.cmd.Process.Pid))\n\t\t\tmemInfo, _ := proc.MemoryInfo()\n\t\t\tstatus.MemoryMax = runner.Settings.MaxRAM\n\t\t\tstatus.Memory = int(memInfo.RSS / (1024 * 1024))\n\n\t\t\tworldPath := filepath.Join(McServerPath(), \"world\")\n\t\t\tusage, _ := disk.Usage(worldPath)\n\t\t\tstatus.Storage = usage.Used / (1024 * 1024)\n\t\t\tstatus.StorageMax = usage.Total / (1024 * 1024)\n\n\t\t\trunner.executeCommand(\"list\")\n\t\t\tstatus.PlayerCount = <-runner.playerChannel\n\n\t\t\ttpsMap := make(map[int]float32)\n\t\t\trunner.executeCommand(\"forge tps\")\n\t\tloop:\n\t\t\tfor {\n\t\t\t\tselect {\n\t\t\t\tcase m := <-runner.tpsChannel:\n\t\t\t\t\tfor k, v := range m {\n\t\t\t\t\t\ttpsMap[k] = v\n\t\t\t\t\t}\n\t\t\t\tcase <-time.After(1 * time.Second):\n\t\t\t\t\tbreak loop\n\t\t\t\t}\n\t\t\t}\n\t\t\tvar tpsStrBuilder strings.Builder\n\t\t\ttpsStrBuilder.WriteString(\"{ \")\n\t\t\tfor k, v := range tpsMap {\n\t\t\t\ttpsStrBuilder.WriteString(fmt.Sprintf(\"\\\"%d\\\": %f, \", k, v))\n\t\t\t}\n\t\t\ttpsStr := tpsStrBuilder.String()[:tpsStrBuilder.Len()-3]\n\t\t\ttpsStrBuilder.Reset()\n\t\t\ttpsStrBuilder.WriteString(tpsStr)\n\t\t\ttpsStrBuilder.WriteString(\"}\")\n\t\t\ttpsStr = tpsStrBuilder.String()\n\t\t\tstatus.TPS = []byte(tpsStr)\n\n\t\t\trunner.StatusChannel <- status\n\t\tcase <-runner.killChannel:\n\t\t\treturn\n\t\t}\n\t}\n}", "func Stats(w http.ResponseWriter, r *http.Request) {\n\tAuthorised(r)\n\tif Wow.Authorised == false {\n\t\treturn\n\t}\n\tErrorHandler(w, r, nil, 0)\n\tif r.Method == \"GET\" {\n\t\tErrorHandler(w, r, errors.New(\"page for ajax\"), 2)\n\t} else {\n\t\tvar x StatsTH\n\t\terr := json.NewDecoder(r.Body).Decode(&x)\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t}\n\t\tID, err := strconv.Atoi(x.ThreadID)\n\t\tErrorHandler(w, r, err, 1)\n\t\tdata.AddNewValueToThread(Wow.ID, ID, x.Value)\n\t\tThread, err := data.GetThreadByID(ID)\n\t\tx.Likes = strconv.Itoa(Thread.Likes)\n\t\tx.Dislikes = strconv.Itoa(Thread.Dislikes)\n\t\tx.Liked = strconv.Itoa(data.CheckUserLikedThread(Wow.ID, ID))\n\t\ta, err := json.Marshal(x)\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t}\n\t\tw.Write(a)\n\t}\n}", "func (stat *NetworkStat) Update() {\n\tfor i, node := range stat.bn.nodeIndex {\n\t\tassignment := node.GetAssignment()\n\t\tif assignment == \"F\" {\n\t\t\tcontinue\n\t\t}\n\t\tstat.count[i] += 1\n\t}\n\tstat.total += 1\n}", "func (client *Client) Stats(key string) ([]StatValue, error) {\n\trv := make([]StatValue, 0, 128)\n\n\treq := &gomemcached.MCRequest{\n\t\tOpcode: gomemcached.STAT,\n\t\tVBucket: 0,\n\t\tKey: []byte(key),\n\t\tCas: 0,\n\t\tOpaque: 918494,\n\t\tExtras: []byte{}}\n\n\terr := transmitRequest(client.conn, req)\n\tif err != nil {\n\t\treturn rv, err\n\t}\n\n\tfor {\n\t\tres, err := client.getResponse()\n\t\tif err != nil {\n\t\t\treturn rv, err\n\t\t}\n\t\tk := string(res.Key)\n\t\tif k == \"\" {\n\t\t\tbreak\n\t\t}\n\t\trv = append(rv, StatValue{\n\t\t\tKey: k,\n\t\t\tVal: string(res.Body),\n\t\t})\n\t}\n\n\treturn rv, nil\n}", "func (h *Handler) GetStats(req *pb.GetStatsRequest) ([]*models.PingData, error) {\n\tresp, err := h.Storage.GetStats(req)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"unable to search stats\")\n\t}\n\treturn resp, nil\n}", "func (m *MetricSet) Fetch(reporter mb.ReporterV2) {\n\tvar err error\n\tvar rows *sql.Rows\n\trows, err = m.db.Query(`SELECT object_name, \n counter_name, \n instance_name, \n cntr_value \nFROM sys.dm_os_performance_counters \nWHERE counter_name = 'SQL Compilations/sec' \n OR counter_name = 'SQL Re-Compilations/sec' \n OR counter_name = 'User Connections' \n OR counter_name = 'Page splits/sec' \n OR ( counter_name = 'Lock Waits/sec' \n AND instance_name = '_Total' ) \n OR counter_name = 'Page splits/sec' \n OR ( object_name = 'SQLServer:Buffer Manager' \n AND counter_name = 'Page life expectancy' ) \n OR counter_name = 'Batch Requests/sec' \n OR ( counter_name = 'Buffer cache hit ratio' \n AND object_name = 'SQLServer:Buffer Manager' ) \n OR ( counter_name = 'Target pages' \n AND object_name = 'SQLServer:Buffer Manager' ) \n OR ( counter_name = 'Database pages' \n AND object_name = 'SQLServer:Buffer Manager' ) \n OR ( counter_name = 'Checkpoint pages/sec' \n AND object_name = 'SQLServer:Buffer Manager' ) \n OR ( counter_name = 'Lock Waits/sec' \n AND instance_name = '_Total' ) \n OR ( counter_name = 'Transactions' \n AND object_name = 'SQLServer:General Statistics' ) \n OR ( counter_name = 'Logins/sec' \n AND object_name = 'SQLServer:General Statistics' ) \n OR ( counter_name = 'Logouts/sec' \n AND object_name = 'SQLServer:General Statistics' ) \n OR ( counter_name = 'Connection Reset/sec' \n AND object_name = 'SQLServer:General Statistics' ) \n OR ( counter_name = 'Active Temp Tables' \n AND object_name = 'SQLServer:General Statistics' )`)\n\tif err != nil {\n\t\treporter.Error(errors.Wrapf(err, \"error closing rows\"))\n\t\treturn\n\t}\n\tdefer func() {\n\t\tif err := rows.Close(); err != nil {\n\t\t\tm.log.Error(\"error closing rows: %s\", err.Error())\n\t\t}\n\t}()\n\n\tmapStr := common.MapStr{}\n\tfor rows.Next() {\n\t\tvar row performanceCounter\n\t\tif err = rows.Scan(&row.objectName, &row.counterName, &row.instanceName, &row.counterValue); err != nil {\n\t\t\treporter.Error(errors.Wrap(err, \"error scanning rows\"))\n\t\t\tcontinue\n\t\t}\n\n\t\t//cell values contains spaces at the beginning and at the end of the 'actual' value. They must be removed.\n\t\trow.counterName = strings.TrimSpace(row.counterName)\n\t\trow.instanceName = strings.TrimSpace(row.instanceName)\n\t\trow.objectName = strings.TrimSpace(row.objectName)\n\n\t\tif row.counterName == \"Buffer cache hit ratio\" {\n\t\t\tmapStr[row.counterName] = fmt.Sprintf(\"%v\", float64(*row.counterValue)/100)\n\t\t} else {\n\t\t\tmapStr[row.counterName] = fmt.Sprintf(\"%v\", *row.counterValue)\n\t\t}\n\t}\n\n\tres, err := schema.Apply(mapStr)\n\tif err != nil {\n\t\tm.log.Error(errors.Wrap(err, \"error applying schema\"))\n\t\treturn\n\t}\n\n\tif isReported := reporter.Event(mb.Event{\n\t\tMetricSetFields: res,\n\t}); !isReported {\n\t\tm.log.Debug(\"event not reported\")\n\t}\n}", "func (sc statsCache) update(tables []*statistics.Table, deletedIDs []int64, newVersion uint64, opts ...TableStatsOpt) statsCache {\n\toption := &tableStatsOption{}\n\tfor _, opt := range opts {\n\t\topt(option)\n\t}\n\tnewCache := sc.copy()\n\tif newVersion == newCache.version {\n\t\tnewCache.minorVersion += uint64(1)\n\t} else {\n\t\tnewCache.version = newVersion\n\t\tnewCache.minorVersion = uint64(0)\n\t}\n\tfor _, tbl := range tables {\n\t\tid := tbl.PhysicalID\n\t\tif option.byQuery {\n\t\t\tnewCache.PutByQuery(id, tbl)\n\t\t} else {\n\t\t\tnewCache.Put(id, tbl)\n\t\t}\n\t}\n\tfor _, id := range deletedIDs {\n\t\tnewCache.Del(id)\n\t}\n\treturn newCache\n}", "func (c *Client) Stats(indexList []string, extraArgs url.Values) (*Response, error) {\n\tr := Request{\n\t\tIndexList: indexList,\n\t\tExtraArgs: extraArgs,\n\t\tMethod: \"GET\",\n\t\tAPI: \"_stats\",\n\t}\n\n\treturn c.Do(&r)\n}", "func metricsUpdate() {\n metricsXml()\n}", "func (p ECachePlugin) FetchMetrics() (map[string]float64, error) {\n\tsess, err := session.NewSession()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tconfig := aws.NewConfig()\n\tif p.AccessKeyID != \"\" && p.SecretAccessKey != \"\" {\n\t\tconfig = config.WithCredentials(credentials.NewStaticCredentials(p.AccessKeyID, p.SecretAccessKey, \"\"))\n\t}\n\tif p.Region != \"\" {\n\t\tconfig = config.WithRegion(p.Region)\n\t}\n\n\tcloudWatch := cloudwatch.New(sess, config)\n\n\tstat := make(map[string]float64)\n\n\tperInstances := []*cloudwatch.Dimension{\n\t\t{\n\t\t\tName: aws.String(\"CacheClusterId\"),\n\t\t\tValue: aws.String(p.CacheClusterID),\n\t\t},\n\t\t{\n\t\t\tName: aws.String(\"CacheNodeId\"),\n\t\t\tValue: aws.String(p.CacheNodeID),\n\t\t},\n\t}\n\n\tfor _, met := range p.CacheMetrics {\n\t\tv, err := getLastPoint(cloudWatch, perInstances, met)\n\t\tif err == nil {\n\t\t\tstat[met] = v\n\t\t} else {\n\t\t\tlog.Printf(\"%s: %s\", met, err)\n\t\t}\n\t}\n\n\treturn stat, nil\n}", "func (c *GorCache) Statistic() Stats {\n\tgetter := &statItem{ //TODO make pool for this\n\t\tresponce: make(chan Stats, 1),\n\t}\n\tc.statsChan <- getter\n\treturn <-getter.responce\n}", "func (d *Data)GetStats() (int, int64) {\n total := len(d.hashMap)\n var averageMilli int64\n if total > 0 {\n averageNano := d.totalHashNanoseconds / int64(total)\n averageMilli = averageNano / 1000\n }\n return total, averageMilli\n}", "func (c *Conn) Stats(cmd *Cmd) int {\n\tc.statsMut.RLock()\n\tdefer c.statsMut.RUnlock()\n\n\treturn c.stats[cmd.hash()]\n}", "func getUpdatesAPI(c *ConfUpdates) (header string, content string, err error) {\n\tvar r api.Response\n\treqURL := \"http://con/api?updates\"\n\tif c.Every != \"\" {\n\t\treqURL += fmt.Sprintf(\"&refresh&immediate&every=%s\", c.Every)\n\t}\n\theader, r, err = getUpdatesResponse(c.Address, reqURL, c.padL, c.padR)\n\tif err != nil {\n\t\treturn\n\t}\n\tif r.Error != \"\" {\n\t\tlog.Warnf(\"[updates] response contains error %s\", r.Error)\n\t\tcontent = fmt.Sprintf(\"%s\\n\", utils.Warn(r.Error))\n\t}\n\tif r.Queued != nil && *r.Queued == true {\n\t\tif r.Data == nil {\n\t\t\theader = fmt.Sprintf(\"%s: No data, refreshing\\n\", utils.Wrap(\"Updates\", c.padL, c.padR))\n\t\t} else {\n\t\t\theader = fmt.Sprintf(\"%s: %d pending, refreshing\\n\", utils.Wrap(\"Updates\", c.padL, c.padR), len(r.Data.Updates))\n\t\t}\n\t} else {\n\t\tif r.Data == nil {\n\t\t\theader = fmt.Sprintf(\"%s: No data\\n\", utils.Wrap(\"Updates\", c.padL, c.padR))\n\t\t\treturn\n\t\t}\n\t\tt, err := time.Parse(time.RFC3339, r.Data.Checked)\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"[updates] cannot parse timestamp %s: %v\", r.Data.Checked, err)\n\t\t\theader = fmt.Sprintf(\"%s: %d pending, cannot parse timestamp\\n\", utils.Wrap(\"Updates\", c.padL, c.padR), len(r.Data.Updates))\n\t\t}\n\t\tvar timeElapsed = time.Since(t)\n\t\theader = fmt.Sprintf(\"%s: %d pending, checked %s ago\\n\",\n\t\t\tutils.Wrap(\"Updates\", c.padL, c.padR), len(r.Data.Updates), timeStr(timeElapsed, 2, c.ShortNames))\n\t}\n\tif r.Data == nil || c.Show == nil || *c.Show == false {\n\t\treturn\n\t}\n\tcontent += fmt.Sprint(utils.Wrap(r.Data.String(), c.padL, c.padR))\n\treturn\n}", "func (f *lightFetcher) checkUpdateStats(p *peer, newEntry *updateStatsEntry) {\n\tnow := mclock.Now()\n\tfp := f.peers[p]\n\tif fp == nil {\n\t\tp.Log().Debug(\"Unknown peer to check update stats\")\n\t\treturn\n\t}\n\n\tif newEntry != nil && fp.firstUpdateStats == nil {\n\t\tfp.firstUpdateStats = newEntry\n\t}\n\tfor fp.firstUpdateStats != nil && fp.firstUpdateStats.time <= now-mclock.AbsTime(blockDelayTimeout) {\n\t\tf.handler.backend.serverPool.adjustBlockDelay(p.poolEntry, blockDelayTimeout)\n\t\tfp.firstUpdateStats = fp.firstUpdateStats.next\n\t}\n\tif fp.confirmedTd != nil {\n\t\tfor fp.firstUpdateStats != nil && fp.firstUpdateStats.td.Cmp(fp.confirmedTd) <= 0 {\n\t\t\tf.handler.backend.serverPool.adjustBlockDelay(p.poolEntry, time.Duration(now-fp.firstUpdateStats.time))\n\t\t\tfp.firstUpdateStats = fp.firstUpdateStats.next\n\t\t}\n\t}\n}", "func (c *HBComp) updateMetricMoverStats() {\n\tvt := models.ValueType{\n\t\tKind: \"STRING\",\n\t\tValue: c.app.MetricMover.Status().String(),\n\t}\n\tc.app.Service.SetServiceAttribute(com.ServiceAttrMetricMoverStatus, vt)\n}", "func (e EntityRepository) GetStats() (models.EntityStats, error) {\n\treturn repositoryHandler.entityRepository.GetStats()\n}", "func (l *channelLink) Stats() (uint64, lnwire.MilliSatoshi, lnwire.MilliSatoshi) {\n\tsnapshot := l.channel.StateSnapshot()\n\n\treturn snapshot.NumUpdates,\n\t\tsnapshot.TotalMilliSatoshisSent,\n\t\tsnapshot.TotalMilliSatoshisReceived\n}", "func (u UptimePlugin) FetchMetrics() (map[string]float64, error) {\n\tut, err := uptime.Get()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to fetch uptime metrics: %s\", err)\n\t}\n\treturn map[string]float64{\"seconds\": ut.Seconds()}, nil\n}", "func (a *PeriodicalAnnouncer) Stats() Stats {\n\tvar stats Stats\n\treq := statsRequest{Response: make(chan Stats, 1)}\n\tselect {\n\tcase a.statsCommandC <- req:\n\tcase <-a.closeC:\n\t}\n\tselect {\n\tcase stats = <-req.Response:\n\tcase <-a.closeC:\n\t}\n\treturn stats\n}", "func (ds *DiscoveryService) GetCacheStats(_ *restful.Request, response *restful.Response) {\n\tstats := make(map[string]*discoveryCacheStatEntry)\n\tfor k, v := range ds.sdsCache.stats() {\n\t\tstats[k] = v\n\t}\n\tfor k, v := range ds.cdsCache.stats() {\n\t\tstats[k] = v\n\t}\n\tfor k, v := range ds.rdsCache.stats() {\n\t\tstats[k] = v\n\t}\n\tif err := response.WriteEntity(discoveryCacheStats{stats}); err != nil {\n\t\tglog.Warning(err)\n\t}\n}", "func (client *Client) Stats(key string) ([]StatValue, error) {\n\trv := make([]StatValue, 0, 128)\n\n\treq := &gomemcached.MCRequest{\n\t\tOpcode: gomemcached.STAT,\n\t\tVBucket: 0,\n\t\tKey: []byte(key),\n\t\tCas: 0,\n\t\tOpaque: 918494,\n\t\tExtras: []byte{}}\n\n\terr := transmitRequest(client.conn, req)\n\tif err != nil {\n\t\treturn rv, err\n\t}\n\n\tfor {\n\t\tres, err := getResponse(client.conn, client.hdrBuf)\n\t\tif err != nil {\n\t\t\treturn rv, err\n\t\t}\n\t\tk := string(res.Key)\n\t\tif k == \"\" {\n\t\t\tbreak\n\t\t}\n\t\trv = append(rv, StatValue{\n\t\t\tKey: k,\n\t\t\tVal: string(res.Body),\n\t\t})\n\t}\n\n\treturn rv, nil\n}", "func (s *Storage) Stats() (blobsFetched, bytesFetched int64) {\n\treturn atomic.LoadInt64(&s.blobsFetched), atomic.LoadInt64(&s.bytesFetched)\n}", "func getStats() string {\n\tvar mutex = &sync.Mutex{}\n\tvar stat []models.ActionOutput\n\t//using mutex to protect critical section and prevent race conditions.\n\tmutex.Lock()\n\tfor action, actionCounter := range actionMap {\n\t\tavgTime := actionCounter.TotalTime / (float64)(actionCounter.Counter)\n\t\tao := models.ActionOutput{\n\t\t\tAction: action,\n\t\t\tAvg: avgTime,\n\t\t}\n\t\tstat = append(stat, ao)\n\t}\n\tactionStats, _ := json.Marshal(stat)\n\tmutex.Unlock()\n\n\t//convert to serialized json string array\n\treturn string(actionStats)\n\n}", "func (c *HBComp) updateCrudeStatistics() {\n\tvt := models.ValueType{\n\t\tKind: \"STRING\",\n\t\tValue: c.app.CrudeOps.GetStats().String(),\n\t}\n\tc.app.Service.SetServiceAttribute(com.ServiceAttrCrudeStatistics, vt)\n}", "func (t *Tortoise) Updates() []result.Layer {\n\tt.mu.Lock()\n\tdefer t.mu.Unlock()\n\tif t.trtl.pending == 0 {\n\t\treturn nil\n\t}\n\trst, err := t.results(t.trtl.pending, t.trtl.processed)\n\tif err != nil {\n\t\tt.logger.Panic(\"unexpected error\",\n\t\t\tzap.Uint32(\"pending\", t.trtl.pending.Uint32()),\n\t\t\tzap.Uint32(\"processed\", t.trtl.pending.Uint32()),\n\t\t\tzap.Error(err),\n\t\t)\n\t}\n\tt.trtl.pending = 0\n\tif t.tracer != nil {\n\t\tt.tracer.On(&UpdatesTrace{ResultsTrace{\n\t\t\tFrom: t.trtl.pending, To: t.trtl.processed,\n\t\t\tResults: rst,\n\t\t}})\n\t}\n\treturn rst\n}", "func (bg *wormhole) stats() {\n\tfor {\n\t\tfor k := range bg.destinationList {\n\t\t\tbg.destinationList[k].Stat(bg.checkCycle / 2)\n\t\t}\n\t\tif bg.algo.Type() == CLOSEST {\n\t\t\tsort.Sort(svrList(bg.destinationList))\n\t\t}\n\t\t/*\n\t\t\tfor _, be := range bg.destinationList {\n\t\t\t\tfmt.Printf(\"Stat: %s => %v\\n\", be.name, be.connectTime)\n\t\t\t}*/\n\t\ttime.Sleep(time.Second * time.Duration(bg.checkCycle))\n\t}\n}", "func (q *QueryBenchmarker) processStats(telemetrySink chan *report.Point) {\n\n\tq.statMapping = StatsMap{\n\t\tAllQueriesLabel: &StatGroup{},\n\t}\n\n\tlastRefresh := time.Time{}\n\ti := uint64(0)\n\tfor stat := range q.statChan {\n\t\tq.isBurnIn = i < q.burnIn\n\t\tif q.isBurnIn {\n\t\t\ti++\n\t\t\tq.statPool.Put(stat)\n\t\t\tcontinue\n\t\t} else if i == q.burnIn && q.burnIn > 0 {\n\t\t\tlog.Printf(\"burn-in complete after %d queries with %d workers\\n\", q.burnIn, q.workers)\n\t\t}\n\n\t\tif _, ok := q.statMapping[string(stat.Label)]; !ok {\n\t\t\tq.statMapping[string(stat.Label)] = &StatGroup{}\n\t\t}\n\n\t\tnow := time.Now()\n\n\t\tif stat.IsActual {\n\t\t\tq.movingAverageStat.Push(now, stat.Value)\n\t\t\tq.statMapping[AllQueriesLabel].Push(stat.Value)\n\t\t\tq.statMapping[string(stat.Label)].Push(stat.Value)\n\t\t\ti++\n\t\t}\n\n\t\tq.statPool.Put(stat)\n\n\t\tif lastRefresh.Nanosecond() == 0 || now.Sub(lastRefresh).Seconds() >= 1.0 {\n\t\t\tq.movingAverageStat.UpdateAvg(now, q.workers)\n\t\t\tlastRefresh = now\n\t\t\t// Report telemetry, if applicable:\n\t\t\tif telemetrySink != nil {\n\t\t\t\tp := report.GetPointFromGlobalPool()\n\t\t\t\tp.Init(\"benchmarks_telemetry\", now.UnixNano())\n\t\t\t\tfor _, tagpair := range q.reportTags {\n\t\t\t\t\tp.AddTag(tagpair[0], tagpair[1])\n\t\t\t\t}\n\t\t\t\tp.AddTag(\"client_type\", \"query\")\n\t\t\t\tp.AddFloat64Field(\"query_response_time_mean\", q.statMapping[AllQueriesLabel].Mean)\n\t\t\t\tp.AddFloat64Field(\"query_response_time_moving_mean\", q.movingAverageStat.Avg())\n\t\t\t\tp.AddIntField(\"query_workers\", q.workers)\n\t\t\t\tp.AddInt64Field(\"queries\", int64(i))\n\t\t\t\ttelemetrySink <- p\n\t\t\t}\n\t\t}\n\t\t// print stats to stderr (if printInterval is greater than zero):\n\t\tif q.printInterval > 0 && i > 0 && i%q.printInterval == 0 && (int64(i) < q.limit || q.limit < 0) {\n\t\t\tlog.Printf(\"%s: after %d queries with %d workers:\\n\", time.Now().String(), i-q.burnIn, q.workers)\n\t\t\tfprintStats(os.Stderr, q)\n\t\t\tlog.Printf(\"\\n\")\n\t\t}\n\n\t}\n\n\tlog.Printf(\"run complete after %d queries with %d workers:\\n\", i-q.burnIn, q.workers)\n\tq.totalQueries = int(i)\n\tq.statGroup.Done()\n}" ]
[ "0.6630799", "0.63485396", "0.626949", "0.62368155", "0.6143968", "0.5994762", "0.59772927", "0.5966089", "0.59394723", "0.59333026", "0.59260935", "0.58925277", "0.5889321", "0.58758885", "0.5873812", "0.58506167", "0.5810288", "0.5777145", "0.5775306", "0.5769908", "0.57540727", "0.57468426", "0.5739778", "0.57328534", "0.5717082", "0.57154363", "0.5709203", "0.56998867", "0.5688009", "0.56842244", "0.5683041", "0.56716484", "0.5670035", "0.56688064", "0.5654618", "0.56533694", "0.5652776", "0.5646741", "0.5644648", "0.56372297", "0.56354195", "0.5634313", "0.56325173", "0.5628553", "0.5620131", "0.5617138", "0.559768", "0.5592947", "0.55905926", "0.55858517", "0.5583749", "0.55623955", "0.554946", "0.554624", "0.55377525", "0.5525844", "0.5523541", "0.5520874", "0.5507654", "0.5499668", "0.5498312", "0.5495121", "0.54888475", "0.5484624", "0.54834807", "0.54807144", "0.5478931", "0.54780716", "0.5473282", "0.5471595", "0.5467948", "0.54677737", "0.5464609", "0.5462061", "0.5460986", "0.5460927", "0.54548764", "0.54427856", "0.54369897", "0.5435024", "0.54275", "0.54225296", "0.5404861", "0.54040647", "0.54037297", "0.5403599", "0.5402667", "0.5399343", "0.5398264", "0.5396703", "0.53928584", "0.5389571", "0.53888476", "0.5381632", "0.53805196", "0.5380095", "0.53780043", "0.5377848", "0.53742343", "0.53740746", "0.53736156" ]
0.0
-1
MaxFiles that can be written based on filesize and free blocks
func (s *StatFS) MaxFiles(size int) int { return int(s.stat.Bfree / (uint64(size) / uint64(s.stat.Bsize))) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func DefineFiles(tsize uint64, hilimit uint64, flS *FileCollection) error {\n\tvar nfiles, remain uint64\n\ttfs, err := flS.totalFileSize() \n\tif err != nil {\n\t\tlog.Printf(\"DefineFiles(): Error computing total file size: %s\", err.Error())\n\t\treturn err\n\t}\n\tif tsize > tfs && tsize > hilimit { //Trying to add files and the total size exceeds the limit\n\t\treturn fmt.Errorf(\"Size requested is over the limit: requested %d bytes, limit: %d bytes.\", tsize, hilimit)\n\t}\n\tfor index, fsize := range flS.fileSizes {\n\t\tnfiles = tsize / fsize\n\t\tremain = tsize % fsize\n\t\tif nfiles > limitFiles { //Use all files of this size, keep adding more files of higher capacities\n\t\t\ttsize -= limitFiles * fsize\n\t\t\tflS.fileAmmount[index] = limitFiles\n\t\t} else if nfiles == 0 {\n\t\t\tflS.fileAmmount[index] = 0\n\t\t} else {\n\t\t\ttsize -= nfiles * fsize\n\t\t\tflS.fileAmmount[index] = nfiles\n\t\t}\n\t}\n\tif tsize > flS.fileSizes[len(flS.fileSizes)-1] { //The remaining size to allocate is bigger than the biggest file sezie, Add more parts of the maximum size\n\t\tnfiles = tsize / flS.fileSizes[len(flS.fileSizes)-1]\n\t\tremain = tsize % flS.fileSizes[len(flS.fileSizes)-1]\n\t\tflS.fileAmmount[len(flS.fileAmmount)-1] += nfiles\n\t}\n\tif remain > 0 { //The remain must be smaller than the bigger file size.\n\t\tfor index, fsize := range flS.fileSizes {\n\t\t\tif remain <= 3*fsize {\n\t\t\t\tsignRemain := int(remain)\n\t\t\t\tfor signRemain > 0 {\n\t\t\t\t\tflS.fileAmmount[index]++\n\t\t\t\t\tsignRemain -= int(fsize)\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func MaxOpenFiles() int { return maxOpenFiles() }", "func setMaxFiles(log logger.Logger) {\n\tvar limits syscall.Rlimit\n\n\terr := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &limits)\n\tif err != nil {\n\t\tlog.Println(\"Failed to get file limit:\", err)\n\t\treturn\n\t}\n\n\tlog.Verbosef(\"Current file limits: %d soft, %d hard\", limits.Cur, limits.Max)\n\tif limits.Cur == limits.Max {\n\t\treturn\n\t}\n\n\tlimits.Cur = limits.Max\n\terr = syscall.Setrlimit(syscall.RLIMIT_NOFILE, &limits)\n\tif err != nil {\n\t\tlog.Println(\"Failed to increase file limit:\", err)\n\t}\n}", "func (s *StatFS) BlockFiles(size int) int {\n\treturn size / int(s.stat.Bsize)\n}", "func GetFileLimit() int {\n\treturn 10000\n}", "func (g *Group) MaxFiles() int64 {\n\treturn g.Max\n}", "func (opts *FIFOCompactionOptions) GetMaxTableFilesSize() uint64 {\n\treturn uint64(C.rocksdb_fifo_compaction_options_get_max_table_files_size(opts.c))\n}", "func fixedCopy(ctx *context, files []os.FileInfo) (written int64, err error) {\n\tvar wholesize int64\n\tctx.filecount = uint64(len(files))\n\tif *ctx.verbose {\n\t\ttimeremaining := time.Duration(ctx.estimatesize/ctx.limit) * time.Second\n\t\tfmt.Printf(\"Total size: %s\\nFiles: %d\\nEstimated time: %v\\n\",\n\t\t\thumanize.Bytes(ctx.estimatesize),\n\t\t\tctx.filecount,\n\t\t\ttimeremaining)\n\t\tctx.starttime = time.Now()\n\t\tcolor.Set(color.FgYellow)\n\t\tfmt.Printf(\"**START** (%v)\\n\", ctx.starttime)\n\t\tcolor.Unset()\n\t\tdefer func() { ctx.endtime = time.Now() }()\n\t}\n\tfor _, file := range files {\n\t\tbytesco, err := copyOneFile(ctx, file)\n\t\tif err != nil {\n\t\t\treturn wholesize, err\n\t\t}\n\t\tctx.filecopied++\n\t\twholesize += bytesco\n\t}\n\treturn wholesize, nil\n}", "func FileSize(min, max uint64) MatcherFunc { return SizeRange(min, max) }", "func maxItemsPerCachedPart() uint64 {\n\tmem := memory.Remaining()\n\t// Production data shows that each item occupies ~4 bytes in the compressed part.\n\t// It is expected no more than defaultPartsToMerge/2 parts exist\n\t// in the OS page cache before they are merged into bigger part.\n\t// Halft of the remaining RAM must be left for lib/storage parts,\n\t// so the maxItems is calculated using the below code:\n\tmaxItems := uint64(mem) / (4 * defaultPartsToMerge)\n\tif maxItems < 1e6 {\n\t\tmaxItems = 1e6\n\t}\n\treturn maxItems\n}", "func (c *DirentCache) setMaxSize(max uint64) {\n\tc.mu.Lock()\n\tc.maxSize = max\n\tc.maybeShrink()\n\tc.mu.Unlock()\n}", "func (opts *FIFOCompactionOptions) SetMaxTableFilesSize(value uint64) {\n\tC.rocksdb_fifo_compaction_options_set_max_table_files_size(opts.c, C.uint64_t(value))\n}", "func TestMaxTransfer(t *testing.T) {\n\tctx := context.Background()\n\tctx, ci := fs.AddConfig(ctx)\n\tci.MaxTransfer = 3 * 1024\n\tci.Transfers = 1\n\tci.Checkers = 1\n\tci.CutoffMode = fs.CutoffModeHard\n\n\ttest := func(t *testing.T, cutoff fs.CutoffMode) {\n\t\tr := fstest.NewRun(t)\n\t\tci.CutoffMode = cutoff\n\n\t\tif r.Fremote.Name() != \"local\" {\n\t\t\tt.Skip(\"This test only runs on local\")\n\t\t}\n\n\t\t// Create file on source\n\t\tfile1 := r.WriteFile(\"file1\", string(make([]byte, 5*1024)), t1)\n\t\tfile2 := r.WriteFile(\"file2\", string(make([]byte, 2*1024)), t1)\n\t\tfile3 := r.WriteFile(\"file3\", string(make([]byte, 3*1024)), t1)\n\t\tr.CheckLocalItems(t, file1, file2, file3)\n\t\tr.CheckRemoteItems(t)\n\n\t\taccounting.GlobalStats().ResetCounters()\n\n\t\terr := Sync(ctx, r.Fremote, r.Flocal, false)\n\t\texpectedErr := fserrors.FsError(accounting.ErrorMaxTransferLimitReachedFatal)\n\t\tif cutoff != fs.CutoffModeHard {\n\t\t\texpectedErr = accounting.ErrorMaxTransferLimitReachedGraceful\n\t\t}\n\t\tfserrors.Count(expectedErr)\n\t\tassert.Equal(t, expectedErr, err)\n\t}\n\n\tt.Run(\"Hard\", func(t *testing.T) { test(t, fs.CutoffModeHard) })\n\tt.Run(\"Soft\", func(t *testing.T) { test(t, fs.CutoffModeSoft) })\n\tt.Run(\"Cautious\", func(t *testing.T) { test(t, fs.CutoffModeCautious) })\n}", "func CreditToFileUploadSize(credit float64) (bytes int) {\n\tbytes = freeFileUploadBytes\n\tfor {\n\t\tif credit > 0 {\n\t\t\tbytes += freeFileUploadBytes\n\t\t\tcredit -= creditSteps\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn\n}", "func MaxDataBytes(maxBytes, evidenceBytes int64, valsCount int) int64 {\n\tmaxDataBytes := maxBytes -\n\t\tMaxOverheadForBlock -\n\t\tMaxHeaderBytes -\n\t\tMaxCommitBytes(valsCount) -\n\t\tevidenceBytes\n\n\tif maxDataBytes < 0 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"Negative MaxDataBytes. Block.MaxBytes=%d is too small to accommodate header&lastCommit&evidence=%d\",\n\t\t\tmaxBytes,\n\t\t\t-(maxDataBytes - maxBytes),\n\t\t))\n\t}\n\n\treturn maxDataBytes\n}", "func (r *Receiver) countBytes(length uint64, update bool) error {\n\tif length > MaxTransfer {\n\t\treturn fmt.Errorf(\"file too large, %d\", length)\n\t}\n\tif r.byteLimit != 0 && r.totalBytes > uint64(r.byteLimit)-length {\n\t\treturn fmt.Errorf(\"file too large, %d\", length)\n\t}\n\tif update {\n\t\tr.totalBytes += length\n\t}\n\treturn nil\n}", "func (w *Writer) writeFile(fil *fileHolder, write io.WriterAt, startOffset int64) (endOffset int64, err error) {\n\tendOffset = startOffset\n\tvar sizes []uint32\n\tif fil.fragIndex != -1 {\n\t\tsizes = fil.blockSizes[:len(fil.blockSizes)-1]\n\t} else {\n\t\tsizes = fil.blockSizes\n\t}\n\tif rdrAt, ok := fil.reader.(io.ReaderAt); ok {\n\t\ttype writeReturn struct {\n\t\t\terr error\n\t\t\tbyts []byte\n\t\t\ti int\n\t\t}\n\t\tout := make(chan *writeReturn)\n\t\tvar filOffset int64\n\t\tvar sync sync.WaitGroup\n\t\tsync.Add(len(sizes))\n\t\tfor i, size := range sizes {\n\t\t\tgo func(offset int64, size uint32, i int) {\n\t\t\t\tvar ret writeReturn\n\t\t\t\tret.i = i\n\t\t\t\tdefer func() {\n\t\t\t\t\tout <- &ret\n\t\t\t\t}()\n\t\t\t\tret.byts = make([]byte, size)\n\t\t\t\t_, ret.err = rdrAt.ReadAt(ret.byts, offset)\n\t\t\t\tif ret.err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tret.byts, ret.err = w.compressData(ret.byts)\n\t\t\t\tsync.Done()\n\t\t\t}(filOffset, size, i)\n\t\t\tfilOffset += int64(size)\n\t\t}\n\t\tvar curInd int\n\t\tvar holdingArea []*writeReturn\n\t\tfor curInd < len(sizes) {\n\t\t\tvar tmp *writeReturn\n\t\t\tfor _, ret := range holdingArea {\n\t\t\t\tif ret.i == curInd {\n\t\t\t\t\ttmp = ret\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif tmp == nil {\n\t\t\t\ttmp = <-out\n\t\t\t\tif tmp.err != nil {\n\t\t\t\t\tsync.Wait()\n\t\t\t\t\treturn endOffset, tmp.err\n\t\t\t\t}\n\t\t\t\tif tmp.i != curInd {\n\t\t\t\t\tholdingArea = append(holdingArea, tmp)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\tfil.blockSizes[curInd] = uint32(len(tmp.byts))\n\t\t\tif len(tmp.byts) == int(w.BlockSize) {\n\t\t\t\t//set uncompressed bit if not compressed\n\t\t\t\tfil.blockSizes[curInd] |= (1 << 24)\n\t\t\t}\n\t\t\tvar n int\n\t\t\tn, err = write.WriteAt(tmp.byts, endOffset)\n\t\t\tendOffset += int64(n)\n\t\t\tif err != nil {\n\t\t\t\tsync.Wait()\n\t\t\t\treturn\n\t\t\t}\n\t\t\tcurInd++\n\t\t}\n\t\treturn\n\t}\n\tvar byts []byte\n\tfor i, size := range sizes {\n\t\tbyts = make([]byte, size)\n\t\t_, err = fil.reader.Read(byts)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tbyts, err = w.compressData(byts)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tfil.blockSizes[i] = uint32(len(byts))\n\t\tif len(byts) == int(w.BlockSize) {\n\t\t\t//set uncompressed bit if not compressed\n\t\t\tfil.blockSizes[i] |= (1 << 24)\n\t\t}\n\t\tvar n int\n\t\tn, err = write.WriteAt(byts, endOffset)\n\t\tendOffset += int64(n)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func adrefiles(fS *FileCollection) error {\n\tfor index,value := range fS.fileSizes {\n\t\tdirectory := fmt.Sprintf(\"%s/d-%d\",fS.frandi,value)\n\t\t//Create a list of files in directory\n\t\tfileList,err := getFilesInDir(directory)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"adrefiles(): Error listing directory: %s\",directory)\n\t\t\treturn err\n\t\t} \n\t\t//Sort the list of files\n\t\tsort.Slice(fileList, func(i,j int) bool {\n\t\t\ts1 := strings.TrimLeft(fileList[i].Name(),\"f-\")\n\t\t\ts2 := strings.TrimLeft(fileList[j].Name(),\"f-\")\n\t\t\tn1,_ := strconv.ParseInt(s1,10,32)\n\t\t\tn2,_ := strconv.ParseInt(s2,10,32)\n\t\t\treturn n1 < n2\n\t\t})\n\t\t//Get the number of the last file created, 0 if none has been\n\t\tvar lastfnum uint64\n\t\tif len(fileList) > 0 {\n\t\t\tlastfnum,_ = strconv.ParseUint(strings.TrimLeft(fileList[len(fileList)-1].Name(),\"f-\"),10,32)\n\t\t} else {\n\t\t\tlastfnum = 0 \n\t\t}\n\t\tlog.Printf(\"Last file number: %d\",lastfnum)\n\t\t\n\t\t//Get the total size in bytes consumed by the files\n\t\tvar tfsize,rqsize,deltasize,fdelta uint64\n\t\tfor _,v := range fileList {\n\t\t\ttfsize += uint64(v.Size())\n\t\t\t//log.Printf(\"File: %s - Size: %d\",v.Name(),v.Size())\n\t\t}\n\t\tlog.Printf(\"Total file size in dir %s: %d\",directory,tfsize)\n\t\trqsize = fS.fileAmmount[index]*value\n\t\tlog.Printf(\"Requested size: %d\",rqsize)\n\t\tif tfsize > rqsize { //Need to remove files\n\t\t\tdeltasize = tfsize - rqsize\n\t\t\tfdelta = deltasize / value\n\t\t\tlog.Printf(\"- Need to remove %d bytes, %d files of size %d\",deltasize,fdelta,value)\n\t\t\tfor n:=0;n<int(fdelta);n++{\n\t\t\t\tfilename := fmt.Sprintf(\"%s/d-%d/f-%d\",fS.frandi,value,int(lastfnum)-n)\n\t\t\t\terr = os.Remove(filename)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Printf(\"adrefiles(): error deleting file %s:\",filename)\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t} else if tfsize < rqsize { //Need to create files\n\t\t\tdeltasize = rqsize - tfsize\n\t\t\tfdelta = deltasize / value\n\t\t\tlog.Printf(\"+ Need to add %d bytes, %d files of size %d\",deltasize,fdelta,value)\n\t\t\tfor n:=1;n<=int(fdelta);n++ {\n\t\t\t\tfilename := fmt.Sprintf(\"%s/d-%d/f-%d\",fS.frandi,value,n+int(lastfnum))\n\t\t\t\terr = newFile(filename,value)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Printf(\"adrefiles(): error creating file %s:\",filename)\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t} else { //No need to add or remove anything \n\t\t\tlog.Printf(\"= No need to add or remove any files\")\n\t\t}\n\t}\n\treturn nil\n}", "func sequentialWriteDisk(blockSize, totalSize int) {\n var message string\n for i := 0; i < blockSize * KiloByte / 10; i++ {\n message += \"1234567890\"\n }\n message += \"\\n\"\n\n // create file\n f, err := os.OpenFile(\"server\", os.O_CREATE|os.O_APPEND|os.O_RDWR, 0666)\n if err != nil {\n fmt.Println(err)\n return\n }\n // clean file\n defer f.Close()\n defer os.Remove(\"server\")\n\n writer := bufio.NewWriter(f)\n round := totalSize * MByte / (blockSize * KiloByte)\n startTime := time.Now()\n for i := 0; i < round; i++ {\n writer.WriteString(message)\n }\n writer.Flush()\n\n // print statistic info\n timeElapsed := float64(time.Now().Sub(startTime).Nanoseconds()) / (1000 * 1000)\n writeSpeed := int64(float64(totalSize) / timeElapsed * 1000)\n fmt.Printf(\"| %vMB \\t| %vKB \\t | %vms \\t| %vMB/s \\t|\\n\", totalSize, blockSize, timeElapsed, writeSpeed)\n}", "func (r *FileSizeRotator) reachLimit(n int) bool {\n\tatomic.AddUint64(&r.currSize, uint64(n))\n\tif r.currSize > r.limitSize {\n\t\treturn true\n\t}\n\treturn false\n}", "func TestMountMaxWrite(t *testing.T) {\n\topts := []MountOptions{\n\t\t{MaxWrite: 0}, // go-fuse default\n\t\t{MaxWrite: 1},\n\t\t{MaxWrite: 123},\n\t\t{MaxWrite: 1 * 1024},\n\t\t{MaxWrite: 4 * 1024},\n\t\t{MaxWrite: 8 * 1024},\n\t\t{MaxWrite: 64 * 1024}, // go-fuse default\n\t\t{MaxWrite: 128 * 1024}, // limit in Linux v4.19 and older\n\t\t{MaxWrite: 999 * 1024},\n\t\t{MaxWrite: 1024 * 1024}, // limit in Linux v4.20+\n\t}\n\tfor _, o := range opts {\n\t\tname := fmt.Sprintf(\"MaxWrite%d\", o.MaxWrite)\n\t\tt.Run(name, func(t *testing.T) {\n\t\t\tmnt, err := ioutil.TempDir(\"\", name)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\tfs := NewDefaultRawFileSystem()\n\t\t\tsrv, err := NewServer(fs, mnt, &o)\n\t\t\tif err != nil {\n\t\t\t\tt.Error(err)\n\t\t\t} else {\n\t\t\t\tgo srv.Serve()\n\t\t\t\tsrv.Unmount()\n\t\t\t}\n\t\t})\n\t}\n}", "func GetFileLimit() int {\n\tlimit := 50000\n\n\tvar lim syscall.Rlimit\n\tif err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &lim); err == nil {\n\t\tlim.Cur = lim.Max\n\t\tsyscall.Setrlimit(syscall.RLIMIT_NOFILE, &lim)\n\t}\n\n\tif err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &lim); err == nil {\n\t\tif cur := int(lim.Cur); cur < limit {\n\t\t\tlimit = cur\n\t\t}\n\t}\n\treturn limit\n}", "func MaxBlockLen(ct CompressionType) uint64 {\n\tif ct == Snappy {\n\t\t// https://github.com/golang/snappy/blob/2a8bb927dd31d8daada140a5d09578521ce5c36a/encode.go#L76\n\t\treturn 6 * (0xffffffff - 32) / 7\n\t}\n\treturn math.MaxUint64\n}", "func BufferedByteLimit(n int) LoggerOption { return bufferedByteLimit(n) }", "func rangesMIMESize(ranges []httpRange, contentType string, contentSize int64) (encSize int64) {\n var w countingWriter\n mw := multipart.NewWriter(&w)\n for _, ra := range ranges {\n mw.CreatePart(ra.mimeHeader(contentType, contentSize))\n encSize += ra.length\n }\n mw.Close()\n encSize += int64(w)\n return\n}", "func (m *wasiSnapshotPreview1Impl) fdFilestatSetSize(pfd wasiFd, psize wasiFilesize) (err wasiErrno) {\n\tf, err := m.files.getFile(pfd, wasiRightsFdRead)\n\tif err != wasiErrnoSuccess {\n\t\treturn err\n\t}\n\n\tif ferr := f.SetSize(psize); ferr != nil {\n\t\treturn fileErrno(ferr)\n\t}\n\treturn wasiErrnoSuccess\n}", "func TestFileLimit(t *testing.T) {\n\tcfg := &guerrilla.AppConfig{LogFile: log.OutputOff.String()}\n\tsc := guerrilla.ServerConfig{\n\t\tListenInterface: \"127.0.0.1:2526\",\n\t\tIsEnabled: true,\n\t\tMaxClients: 1000,\n\t}\n\tcfg.Servers = append(cfg.Servers, sc)\n\td := guerrilla.Daemon{Config: cfg}\n\tif ok, maxClients, fileLimit := guerrilla.CheckFileLimit(d.Config); !ok {\n\t\tt.Errorf(\"Combined max clients for all servers (%d) is greater than open file limit (%d). \"+\n\t\t\t\"Please increase your open file limit. Please check your OS docs for how to increase the limit.\", maxClients, fileLimit)\n\t}\n}", "func MaxDataBytes(maxBytes int64, keyType crypto.KeyType, evidenceBytes int64, valsCount int) int64 {\n\tmaxDataBytes := maxBytes -\n\t\tMaxOverheadForBlock -\n\t\tMaxHeaderBytes -\n\t\tMaxCoreChainLockSize -\n\t\tMaxCommitOverheadBytes -\n\t\tevidenceBytes\n\n\tif maxDataBytes < 0 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"Negative MaxDataBytes. Block.MaxBytes=%d is too small to accommodate header&lastCommit&evidence=%d\",\n\t\t\tmaxBytes,\n\t\t\t-(maxDataBytes - maxBytes),\n\t\t))\n\t}\n\n\treturn maxDataBytes\n}", "func Merge_Files(file_name, sortAlg string, elemen_size, max_size, number_of_processors int, \n\treadData util.ReadData, cmp util.Compare, fragment util.Fragment_files, writeData util.WriteData) {\n\tinit_time := time.Now()\t\t\t\t\t\t\n\t\n\tconst size_unity = 1000000\t\t\t\t\t\t// unidade de medida para max_size\n\t\n\tmax_size = max_size * size_unity\t\t\t\t// define a quantidade maxima de memoria\n\t\n\tcond_files = sync.NewCond(queueLock)\t\t\t//Inicializa a variavel condicao\n\t\n\tfiles_queue = util.NewList()\t\t\t\t\t//Inicializa a fila que vai conter os arquivos ja ordenados\n\tcount_files = 0\n\n\tctx := context.Background()\t\t\t\t\t\t//Contexto da thread principal\n\n\t//semaforo que controla as threads da read and sort\n\tsem_RAS = semaphore.NewWeighted(int64(number_of_processors))\n\n\t// Obtem uma lista com os files descriptors necessarios\n\tfds, size_fd := fragment(file_name, number_of_processors, elemen_size, int64(max_size))\n\tfds_qtd := len(fds)\n\n\t// Controla a barra de progresso\n\tgeneral_pbar = pb.New((fds_qtd * 2) - 1)\n\tgeneral_pbar.Prefix(\"Total\")\n\tgeneral_pbar.ShowSpeed = false\n\tpPool = util.NewPBar(general_pbar)\n\n\t//fragmenta e ordena os arquivos\n\tvar i int\n\tfor i = 0; i < fds_qtd; i++ {\n\t\twg.Add(1)\n\t\tsem_RAS.Acquire(ctx, 1) // pega uma permissao do sem\n\t\tgo read_And_Sort(sortAlg, i, int64(size_fd[i]), fds[i], readData, writeData, cmp)\n\t}\n\n\tvar output_name string\n\t//controla a mesclagem de arquivos\n\n\tfor count := 0; count < fds_qtd-1; count += 1 {\n\t\tsem_RAS.Acquire(ctx, 1)\t\t\t\t\t\t\t\t// Garante que o numero de threads esteja dentro do permitido\n\t\t\n\t\tcond_files.L.Lock()\t\t\t\t\t\t\t\t\t// Da inicio a regiao de exclusao mutua\n\n\t\tfor count_files < 2 {\t\t\t\t\t\t\t\t// Caso o numero de arquivos prontos seja menor que 2, \n\t\t\tcond_files.Wait()\t\t\t\t\t\t\t\t// Aguarda na variavel condicao\n\t\t}\t\t\n\t\t\n\t\tfile1_name := (files_queue.Pop_front()).(string)\t// Se chegou ate aqui, eh porque existem ao menos dois arquivos prontos\n\t\tfile2_name := (files_queue.Pop_front()).(string)\n\t\tcount_files -= 2\t\t\t\t\t\t\t\t\t// Subtrai 2 da quantidade total de arquivos\n\n\t\tcond_files.L.Unlock()\t\t\t\t\t\t\t\t// Da fim a regiao de exclusao mutua\n\n\t\toutput_name = \"out\" + strconv.Itoa(i)\n\t\ti++\n\t\twg.Add(1)\n\t\tgo merge_arrays(file1_name, file2_name, output_name, elemen_size, int64(max_size), readData, writeData, cmp)\n\t}\n\n\twg.Wait() //Espera todo mundo terminar\n\n\tif fds_qtd == 1 {\n\t\toutput_name = \"out\" + strconv.Itoa(i-1)\n\t}\n\n\t//Renomeia o arquivo, move ele pra raiz e deleta a temp\n\tos.Rename(\"temp\"+string(os.PathSeparator)+output_name+\".bin\", \".\"+string(os.PathSeparator)+\"Sorted\"+\".bin\")\n\tos.Remove(\"temp\")\n\n\tpPool.End()\n\tfmt.Println(fds_qtd)\n\tsince := time.Since(init_time)\n\tfmt.Printf(\"Tempo decorrido: %dm%ds\\n\", int(since.Minutes()), (int(since.Seconds()))%60)\n}", "func merge(path string, size int64, in chan *chunk, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\tvar total int64\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\t// make sure we have exactly the amount of space we need\n\tif err = f.Truncate(size); err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\tdefer f.Close()\n\n\tfor ch := range in {\n\t\t_, err = f.Seek(ch.off, 0)\n\t\tif err != nil { // downloaded too much?\n\t\t\tlog.Println(err)\n\t\t\tcontinue\n\t\t}\n\t\tn, err := f.Write(ch.data)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t\tcontinue\n\t\t}\n\t\ttotal += int64(n)\n\t\tfmt.Print(\"\\r\", total, \"/\", size)\n\t}\n}", "func (mm *BytesMonitor) MaximumBytes() int64 {\n\tmm.mu.Lock()\n\tdefer mm.mu.Unlock()\n\treturn mm.mu.maxAllocated\n}", "func (r *SizeRotator) Write(bytes []byte) (n int, err error) {\r\n\tr.mutex.Lock()\r\n\tdefer r.mutex.Unlock()\r\n\r\n\tif r.file == nil {\r\n\t\t// Check file existence\r\n\t\tstat, _ := os.Lstat(r.path)\r\n\t\tif stat != nil {\r\n\t\t\t// Update initial size by file size\r\n\t\t\tr.totalSize = stat.Size()\r\n\t\t}\r\n\t}\r\n\t// Do rotate when size exceeded\r\n\tif r.totalSize+int64(len(bytes)) > r.RotationSize {\r\n\t\t// Get available file name to be rotated\r\n\t\tfor i := 1; i <= r.MaxRotation; i++ {\r\n\t\t\tfext := filepath.Ext(r.path)\r\n\t\t\trenamedPath := \"\"\r\n\t\t\tif len(fext) != 0 {\r\n\t\t\t\trenamedPath = strings.TrimRight(r.path, fext) + \"_\" + LeftPad2Len(strconv.Itoa(i), \"0\", len(strconv.Itoa(r.MaxRotation))) + fext\r\n\t\t\t} else {\r\n\t\t\t\trenamedPath = r.path + \"_\" + LeftPad2Len(strconv.Itoa(i), \"0\", len(strconv.Itoa(r.MaxRotation)))\r\n\t\t\t}\r\n\t\t\tstat, _ := os.Lstat(renamedPath)\r\n\t\t\tif stat == nil {\r\n\t\t\t\tif r.file != nil {\r\n\t\t\t\t\t// reset file reference\r\n\t\t\t\t\tr.file.Close()\r\n\t\t\t\t\tr.file = nil\r\n\t\t\t\t}\r\n\t\t\t\terr := os.Rename(r.path, renamedPath)\r\n\t\t\t\tif err != nil {\r\n\t\t\t\t\treturn 0, err\r\n\t\t\t\t}\r\n\t\t\t\tbreak\r\n\t\t\t}\r\n\t\t\tif i == r.MaxRotation {\r\n\t\t\t\treturn 0, errors.New(\"Rotation count has been exceeded\")\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n\tif r.file == nil {\r\n\t\tr.file, err = os.OpenFile(r.path, os.O_APPEND|os.O_WRONLY|os.O_CREATE, 0644)\r\n\t\tif err != nil {\r\n\t\t\treturn 0, err\r\n\t\t}\r\n\t\t// Switch current date\r\n\t\tr.totalSize = 0\r\n\t}\r\n\tn, err = r.file.Write(bytes)\r\n\tr.totalSize += int64(n)\r\n\treturn n, err\r\n}", "func TestDeleteCleanerBytesBelowLimit(t *testing.T) {\n\topts := deleteCleanerOptions{Name: \"foo\", Logger: noopLogger()}\n\topts.Retention.Bytes = 50\n\tcleaner := newDeleteCleaner(opts)\n\tdir := tempDir(t)\n\tdefer remove(t, dir)\n\n\texpected := make([]*segment, 5)\n\tfor i := 0; i < 5; i++ {\n\t\texpected[i] = createSegment(t, dir, int64(i), 20)\n\t}\n\tactual, err := cleaner.Clean(expected)\n\trequire.NoError(t, err)\n\trequire.Equal(t, expected, actual)\n}", "func (f *FileRotator) nextFile() error {\n\tnextFileIdx := f.logFileIdx\n\tfor {\n\t\tnextFileIdx += 1\n\t\tlogFileName := filepath.Join(f.path, fmt.Sprintf(\"%s.%d\", f.baseFileName, nextFileIdx))\n\t\tif fi, err := os.Stat(logFileName); err == nil {\n\t\t\tif fi.IsDir() || fi.Size() >= f.FileSize {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tf.logFileIdx = nextFileIdx\n\t\tif err := f.createFile(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tbreak\n\t}\n\t// Purge old files if we have more files than MaxFiles\n\tf.fileLock.Lock()\n\tdefer f.fileLock.Unlock()\n\tif f.logFileIdx-f.oldestLogFileIdx >= f.MaxFiles && !f.closed {\n\t\tselect {\n\t\tcase f.purgeCh <- struct{}{}:\n\t\tdefault:\n\t\t}\n\t}\n\treturn nil\n}", "func (s *settings) GetMaxWriteSize() uint {\n\treturn s.wMaxSize\n}", "func (z *zpaqWriter) writeFile(w *writer, b []byte) (int, error) {\n\tc1 := z.c1\n\n\tfor i, c := range b {\n\t\tsplit := false\n\t\tv := sigmap[c]\n\t\tif len(v) > 0 && i < len(b)-6 {\n\t\t\tfor _, s := range v {\n\t\t\t\tsplit = true\n\t\t\t\tfor j, expect := range s {\n\t\t\t\t\tif b[j+1] != expect {\n\t\t\t\t\t\tsplit = false\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif c == z.o1[c1] {\n\t\t\tz.h = (z.h + uint32(c) + 1) * 314159265\n\t\t} else {\n\t\t\tz.h = (z.h + uint32(c) + 1) * 271828182\n\t\t}\n\t\tz.o1[c1] = c\n\t\tc1 = c\n\t\tw.cur[w.off] = c\n\t\tw.off++\n\n\t\t// Filled the buffer? Send it off!\n\t\tif w.off >= z.minFragment && (z.h < z.maxHash || split || w.off >= z.maxFragment) {\n\t\t\tb := <-w.buffers\n\t\t\t// Swap block with current\n\t\t\tw.cur, b.data = b.data[:w.maxSize], w.cur[:w.off]\n\t\t\tb.N = w.nblocks\n\n\t\t\tw.input <- b\n\t\t\tw.write <- b\n\t\t\tw.nblocks++\n\t\t\tw.off = 0\n\t\t\tz.h = 0\n\t\t\tc1 = 0\n\t\t}\n\t}\n\tz.c1 = c1\n\treturn len(b), nil\n}", "func increaseFDRlimit() error {\n\tvar l syscall.Rlimit\n\tif err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &l); err != nil {\n\t\treturn err\n\t}\n\tif l.Cur == l.Max {\n\t\t// Already at soft max, nothing to do\n\t\treturn nil\n\t}\n\tl.Cur = l.Max\n\n\t// The following is a workaround for this issue:\n\t// https://github.com/golang/go/issues/30401\n\tif runtime.GOOS == \"darwin\" && l.Cur > 24576 {\n\t\t// The max file limit is 24576, even though the max returned by\n\t\t// Getrlimit is 1<<63-1.\n\t\tl.Cur = 24576\n\t}\n\n\treturn syscall.Setrlimit(syscall.RLIMIT_NOFILE, &l)\n}", "func GetMaxBlockSize() int64 {\r\n\treturn converter.StrToInt64(SysString(MaxBlockSize))\r\n}", "func MaxUploads(value int) Option {\n\treturn addParam(\"max-uploads\", strconv.Itoa(value))\n}", "func (c *Config) MaxSize(stream string) (uint, error) {\n\tkey, err := keyName(stream, \"maxsize\")\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn c.v.GetSizeInBytes(key), nil\n}", "func MaxBytesHandler(h Handler, n int64) Handler {\n\treturn HandlerFunc(func(w ResponseWriter, r *Request) {\n\t\tr2 := *r\n\t\tr2.Body = MaxBytesReader(w, r.Body, n)\n\t\th.ServeHTTP(w, &r2)\n\t})\n}", "func (db *DB) grow(sz int) error {\n\t// Ignore if the new size is less than available file size.\n\tif sz <= db.filesz {\n\t\treturn nil\n\t}\n\n\t// If the data is smaller than the alloc size then only allocate what's needed.\n\t// Once it goes over the allocation size then allocate in chunks.\n\tif db.datasz <= db.AllocSize {\n\t\tsz = db.datasz\n\t} else {\n\t\tsz += db.AllocSize\n\t}\n\n\t// Truncate and fsync to ensure file size metadata is flushed.\n\t// https://github.com/boltdb/bolt/issues/284\n\tif !db.NoGrowSync && !db.readOnly {\n\t\tif runtime.GOOS != \"windows\" {\n\t\t\tif err := db.file.Truncate(int64(sz)); err != nil {\n\t\t\t\treturn fmt.Errorf(\"file resize error: %s\", err)\n\t\t\t}\n\t\t}\n\t\tif err := db.file.Sync(); err != nil {\n\t\t\treturn fmt.Errorf(\"file sync error: %s\", err)\n\t\t}\n\t\tif db.Mlock {\n\t\t\t// unlock old file and lock new one\n\t\t\tif err := db.mrelock(db.filesz, sz); err != nil {\n\t\t\t\treturn fmt.Errorf(\"mlock/munlock error: %s\", err)\n\t\t\t}\n\t\t}\n\t}\n\n\tdb.filesz = sz\n\treturn nil\n}", "func GetFileLimit() int {\n\tvar lim syscall.Rlimit\n\n\tlimit := 100000\n\tif err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &lim); err == nil {\n\t\tlim.Cur = lim.Max\n\t\tif lim.Cur == unix.RLIM_INFINITY || lim.Cur > 100000 {\n\t\t\tlim.Cur = 100000\n\t\t}\n\t\tsyscall.Setrlimit(syscall.RLIMIT_NOFILE, &lim)\n\t}\n\n\tif err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &lim); err == nil {\n\t\tlimit = int(lim.Cur)\n\t}\n\treturn limit\n}", "func (f *LogFile) bytes() int {\n\tvar b int\n\tb += 24 // mu RWMutex is 24 bytes\n\tb += 16 // wg WaitGroup is 16 bytes\n\tb += int(unsafe.Sizeof(f.id))\n\t// Do not include f.data because it is mmap'd\n\t// TODO(jacobmarble): Uncomment when we are using go >= 1.10.0\n\t//b += int(unsafe.Sizeof(f.w)) + f.w.Size()\n\tb += int(unsafe.Sizeof(f.buf)) + len(f.buf)\n\tb += int(unsafe.Sizeof(f.keyBuf)) + len(f.keyBuf)\n\t// Do not count SeriesFile because it belongs to the code that constructed this Index.\n\tb += int(unsafe.Sizeof(f.size))\n\tb += int(unsafe.Sizeof(f.modTime))\n\tb += int(unsafe.Sizeof(f.seriesIDSet)) + f.seriesIDSet.Bytes()\n\tb += int(unsafe.Sizeof(f.tombstoneSeriesIDSet)) + f.tombstoneSeriesIDSet.Bytes()\n\tb += int(unsafe.Sizeof(f.mms)) + f.mms.bytes()\n\tb += int(unsafe.Sizeof(f.path)) + len(f.path)\n\treturn b\n}", "func calculateBufferSize(blocks uint64) uint64 {\n\tif nbb := NBufferBlocks; blocks < nbb {\n\t\treturn blocks\n\t} else {\n\t\treturn nbb\n\t}\n}", "func makeThumbnails6(filenames <-chan string) int64{\n\tsizes := make(chan int64)\n\tvar wg sync.WaitGroup // number of working goroutines\n\tfor f := range filenames{\n\t\twg.Add(1)\n\t\t//worker\n\t\tgo func(f string){\n\t\t\tdefer wg.Done()\n\t\t\tthumb, err := thumbnail.ImageFile(f)\n\t\t\tif err != nil{\n\t\t\t\tlog.Println(err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tinfo, _ :=os.Stat(thumb) // OK to ingore error\n\t\t\tsizes <- info.Size()\n\t\t}(f)\n\t}\n\n\n\t//closer\n\tgo func(){\n\t\twg.Wait()\n\t\tclose(sizes)\n\t}()\n\n\tvar total int64\n\tfor size := range sizes{\n\t\ttotal += size\n\t}\n\treturn total\n}", "func MaxBytes(m int64) optionSetter {\n\treturn func(o *options) error {\n\t\to.maxBytes = m\n\t\treturn nil\n\t}\n}", "func SetMaxFileDescriptors(requestedSoftFDLimit uint64) (uint64, error) {\n\n\tvar limits syscall.Rlimit\n\n\tif err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &limits); err != nil {\n\t\treturn 0, err\n\t}\n\n\trequiresUpdate, recommendedSoftFDLimit := getSoftFDLimit(\n\t\trequestedSoftFDLimit,\n\t\tlimits,\n\t)\n\n\t// No call to Setrlimit required, because the requested soft limit is lower than current soft limit\n\tif !requiresUpdate {\n\t\treturn 0, nil\n\t}\n\n\t// Update the soft limit (but don't bother updating the hard limit, since only root can do that,\n\t// and it's assumed that this process is not running as root)\n\tlimits.Cur = recommendedSoftFDLimit\n\terr := syscall.Setrlimit(syscall.RLIMIT_NOFILE, &limits)\n\n\tif err == nil {\n\t\tLogf(\"Configured process to allow %d open file descriptors\", recommendedSoftFDLimit)\n\t}\n\n\treturn recommendedSoftFDLimit, err\n\n}", "func fwrite(tls TLS, ptr uintptr, size, nmemb size_t, stream uintptr) size_t {\n\thi, lo := mathutil.MulUint128_64(uint64(size), uint64(nmemb))\n\tif hi != 0 || lo > uint64(len(rawmem{})) {\n\t\ttls.setErrno(errno.XE2BIG)\n\t\treturn 0\n\t}\n\n\tn, err := files.writer(stream).Write((*rawmem)(unsafe.Pointer(ptr))[:lo])\n\tif err != nil {\n\t\ttls.setErrno(errno.XEIO)\n\t}\n\treturn size_t(n) / size\n}", "func (db *DB) grow(sz int) error {\n\t// Ignore if the new size is less than available file size.\n\tif sz <= db.filesz {\n\t\treturn nil\n\t}\n\n\t// If the data is smaller than the alloc size then only allocate what's needed.\n\t// Once it goes over the allocation size then allocate in chunks.\n\tif db.datasz < db.AllocSize {\n\t\tsz = db.datasz\n\t} else {\n\t\tsz += db.AllocSize\n\t}\n\n\t// Truncate and fsync to ensure file size metadata is flushed.\n\t// https://github.com/boltdb/bolt/issues/284\n\tif !db.NoGrowSync && !db.readOnly {\n\t\tif runtime.GOOS != \"windows\" {\n\t\t\tif err := db.file.Truncate(int64(sz)); err != nil {\n\t\t\t\treturn fmt.Errorf(\"file resize error: %s\", err)\n\t\t\t}\n\t\t}\n\t\tif err := db.file.Sync(); err != nil {\n\t\t\treturn fmt.Errorf(\"file sync error: %s\", err)\n\t\t}\n\t}\n\n\tdb.filesz = sz\n\treturn nil\n}", "func (kvStore *KvStore) Write(key string, value []byte) (err error) {\n\t// So here we want to check the size of the file and if it's > max size we should create\n\t// a new one\n\t// The consideration that we have to think about is that the latest file could already be in the process\n\t// of being written to. Could a RWMutex help us here..? As long as we haven't hit a crucial file size we\n\t// we can allow as many processes as are needed\n\t//\n\tcount := len(kvStore.files)\n\terr = kvStore.files[count-1].Write(key, value)\n\tsize, _ := kvStore.files[count-1].Size()\n\t// Just use 100 for the moment\n\tif size > 100 {\n\t\tnewFile, err2 := gklogfile.Open(fmt.Sprintf(\"c:\\\\devwork\\\\go\\\\gokave_data\\\\%s\\\\%d.gkv\", kvStore.storeName, time.Now().UTC().UnixNano()))\n\t\tif err != nil {\n\t\t\treturn err2\n\t\t}\n\t\t// This needs to be in a write mutex when we update the current file. All operations apart from this are 'read'\n\t\tkvStore.files = append(kvStore.files, newFile)\n\t}\n\tfmt.Printf(\"File size: %d\\n\", size)\n\treturn\n\n}", "func maxBodySize(sz int64) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tr.Body = http.MaxBytesReader(w, r.Body, sz)\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func writeFragmentBlocks(fileList []*finalizeFileInfo, f util.File, ws string, blocksize int, options FinalizeOptions, location int64) ([]fragmentBlock, int64, error) {\n\tcompressor := options.Compression\n\tif options.NoCompressFragments {\n\t\tcompressor = nil\n\t}\n\tfragmentData := make([]byte, 0)\n\tvar (\n\t\tallWritten int64\n\t\tfragmentBlockIndex uint32\n\t\tfragmentBlocks []fragmentBlock\n\t)\n\tfileCloseList := make([]*os.File, 0)\n\tdefer func() {\n\t\tfor _, f := range fileCloseList {\n\t\t\tf.Close()\n\t\t}\n\t}()\n\tfor _, e := range fileList {\n\t\t// only copy data for regular files\n\t\tif e.fileType != fileRegular {\n\t\t\tcontinue\n\t\t}\n\t\tvar (\n\t\t\twritten int64\n\t\t\terr error\n\t\t)\n\n\t\t// how much is there to put in a fragment?\n\t\tremainder := e.Size() % int64(blocksize)\n\t\tif remainder == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// would adding this data cause us to write?\n\t\tif len(fragmentData)+int(remainder) > blocksize {\n\t\t\twritten, compressed, err := finalizeFragment(fragmentData, f, location, compressor)\n\t\t\tif err != nil {\n\t\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"error writing fragment block %d: %v\", fragmentBlockIndex, err)\n\t\t\t}\n\t\t\tfragmentBlocks = append(fragmentBlocks, fragmentBlock{\n\t\t\t\tsize: uint32(written),\n\t\t\t\tcompressed: compressed,\n\t\t\t\tlocation: location,\n\t\t\t})\n\t\t\t// increment as all writes will be to next block block\n\t\t\tfragmentBlockIndex++\n\t\t\tfragmentData = fragmentData[:blocksize]\n\t\t}\n\n\t\te.fragment = &fragmentRef{\n\t\t\tblock: fragmentBlockIndex,\n\t\t\toffset: uint32(len(fragmentData)),\n\t\t}\n\t\t// save the fragment data from the file\n\n\t\tfrom, err := os.Open(path.Join(ws, e.path))\n\t\tif err != nil {\n\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"failed to open file for reading %s: %v\", e.path, err)\n\t\t}\n\t\tfileCloseList = append(fileCloseList, from)\n\t\tbuf := make([]byte, remainder)\n\t\tn, err := from.ReadAt(buf, e.Size()-remainder)\n\t\tif err != nil && err != io.EOF {\n\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"error reading final %d bytes from file %s: %v\", remainder, e.Name(), err)\n\t\t}\n\t\tif n != len(buf) {\n\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"failed reading final %d bytes from file %s, only read %d\", remainder, e.Name(), n)\n\t\t}\n\t\tfrom.Close()\n\t\tfragmentData = append(fragmentData, buf...)\n\n\t\tallWritten += written\n\t\tif written > 0 {\n\t\t\tfragmentBlockIndex++\n\t\t}\n\t}\n\n\t// write remaining fragment data\n\tif len(fragmentData) > 0 {\n\t\twritten, compressed, err := finalizeFragment(fragmentData, f, location, compressor)\n\t\tif err != nil {\n\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"error writing fragment block %d: %v\", fragmentBlockIndex, err)\n\t\t}\n\t\tfragmentBlocks = append(fragmentBlocks, fragmentBlock{\n\t\t\tsize: uint32(written),\n\t\t\tcompressed: compressed,\n\t\t\tlocation: location,\n\t\t})\n\t\t// increment as all writes will be to next block block\n\t\tallWritten += int64(written)\n\t}\n\treturn fragmentBlocks, allWritten, nil\n}", "func addMaxRSS(w io.Writer) {\n}", "func MaxDataBytesNoEvidence(maxBytes int64, valsCount int) int64 {\n\tmaxDataBytes := maxBytes -\n\t\tMaxOverheadForBlock -\n\t\tMaxHeaderBytes -\n\t\tMaxCommitBytes(valsCount)\n\n\tif maxDataBytes < 0 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"Negative MaxDataBytesUnknownEvidence. Block.MaxBytes=%d is too small to accommodate header&lastCommit&evidence=%d\",\n\t\t\tmaxBytes,\n\t\t\t-(maxDataBytes - maxBytes),\n\t\t))\n\t}\n\n\treturn maxDataBytes\n}", "func getMaxUploadSizeMiddleware(maxSize int64, typesAllowed []string) middlewares.MiddlewareFunc {\n\treturn middlewares.NewMaxUploadSize(&middlewares.MaxUploadSizeOptions{\n\t\tAllowedFileType: typesAllowed,\n\t\tSize: maxSize,\n\t})\n}", "func (f *fileScorer) adjustSaveBytes(n int64) bool {\n\tif int64(f.saveBytes)+n <= 0 {\n\t\tf.saveBytes = 0\n\t\tf.trimQueue()\n\t\treturn false\n\t}\n\tif n < 0 {\n\t\tf.saveBytes -= ^uint64(n - 1)\n\t} else {\n\t\tf.saveBytes += uint64(n)\n\t}\n\tif f.saveBytes == 0 {\n\t\tf.queue.Init()\n\t\tf.saveBytes = 0\n\t\treturn false\n\t}\n\tif n < 0 {\n\t\tf.trimQueue()\n\t}\n\treturn true\n}", "func (s *settings) SetMaxWriteSize(size uint) {\n\ts.wMaxSize = size\n}", "func (fc FileCollection) totalFileSize() (uint64,error)\t{\n\tvar tfsize uint64\n\tfor _,fsize := range fc.fileSizes {\n\t\tdirectory := fmt.Sprintf(\"%s/d-%d\",fc.frandi,fsize)\n\t\tfileList,err := getFilesInDir(directory)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"totalSizeFiles(): Error listing directory: %s\\n%s\",directory,err.Error())\n\t\t\treturn 0,err\n\t\t}\n\t\tfor _,v := range fileList {\n\t\t\ttfsize += uint64(v.Size())\n\t\t}\n\t}\n\treturn tfsize,nil\n}", "func limit(n int64) int {\n\tif n < 0 || maxio < n {\n\t\tFatal(\"bad io size:\", n)\n\t}\n\treturn int(n)\n}", "func raiseFileLimit(n uint64) {\n\tsoftLimit, hardLimit, err := getFileLimits()\n\tif err != nil {\n\t\tlog.Warnf(`unable to increase file limit, the current value could`+\n\t\t\t` not be retrieved. If you are using an init script to start`+\n\t\t\t` the New Relic Daemon trying adding \"ulimit -n %d\" to your`+\n\t\t\t` init script. The error was %v.`, n, err)\n\t\treturn\n\t}\n\n\tif n <= softLimit {\n\t\treturn\n\t}\n\n\t// Maintain the following invariant: softLimit <= hardLimit\n\t// Failure to abide makes The Dude sad. Also causes EINVAL.\n\n\tif n > hardLimit {\n\t\t// The hard limit also needs to be raised. Try to raise the soft and\n\t\t// hard limits at the same time.\n\t\terr := setFileLimits(n, n)\n\t\tif err == nil {\n\t\t\tlog.Infof(\"increased file limit to %d\", n)\n\t\t\treturn\n\t\t}\n\n\t\t// Couldn't raise the hard limit. Log the failure and fall through\n\t\t// below to raise the soft limit as high as we can.\n\t\tif err == syscall.EPERM {\n\t\t\tlog.Warnf(\"unable to increase file hard limit from %d to %d.\"+\n\t\t\t\t\" Raising the hard limit requires super-user privileges,\"+\n\t\t\t\t\" please contact your system administrator for assistance.\"+\n\t\t\t\t\" An attempt will be made to raise the soft limit to %[1]d.\",\n\t\t\t\thardLimit, n)\n\t\t} else {\n\t\t\tlog.Warnf(\"unable to increase file hard limit from %d to %d.\"+\n\t\t\t\t\" The error was %v. An attempt will be made to raise the\"+\n\t\t\t\t\" soft limit to %[1]d.\", hardLimit, n, err)\n\t\t}\n\t}\n\n\tsoftLimit = n\n\n\t// Ensure softLimit = min(n, hardLimit). We may have failed to raise\n\t// the hard limit to be greater than or equal to n above.\n\tif softLimit > hardLimit {\n\t\tsoftLimit = hardLimit\n\t}\n\n\terr = setFileLimits(softLimit, hardLimit)\n\tif err != nil {\n\t\tlog.Warnf(\"unable to increase file limit: %v\", err)\n\t\treturn\n\t}\n\n\tlog.Infof(\"increased file limit to %d\", softLimit)\n}", "func WithMaxFileSZ(sz int) Option {\n\treturn func(opts *options) {\n\t\topts.maxFileSZ = sz\n\t}\n}", "func (o *OSS) ReturnSize(groupID int64) error {\n\n partLine := partLine()\n\n totalData := map[string]map[string]int{}\n wg := &sync.WaitGroup{}\n ch := make(chan base.BaseInfo, 1000)\n wg.Add(2)\n go register(groupID, ch, wg)\n go fileCalc(groupID, ch, wg, o, totalData)\n\n time.Sleep(2 * time.Second)\n wg.Wait()\n\n for t := range totalData {\n ts := strconv.Itoa(totalData[t][\"totalSize\"])\n\n write.CreateFile(t, partLine + \"\\n\")\n write.CreateFile(t, fmt.Sprintf(\"Total: RecordCount: %d ; FileCount: %d ; FileSize: %s .\\n\",totalData[t][\"RecordCount\"],totalData[t][\"totalCount\"], utils.FormatSize(ts) ))\n }\n return nil\n}", "func testSyncWithMaxDuration(t *testing.T, cutoffMode fs.CutoffMode) {\n\tctx := context.Background()\n\tctx, ci := fs.AddConfig(ctx)\n\tif *fstest.RemoteName != \"\" {\n\t\tt.Skip(\"Skipping test on non local remote\")\n\t}\n\tr := fstest.NewRun(t)\n\n\tmaxDuration := 250 * time.Millisecond\n\tci.MaxDuration = maxDuration\n\tci.CutoffMode = cutoffMode\n\tci.CheckFirst = true\n\tci.OrderBy = \"size\"\n\tci.Transfers = 1\n\tci.Checkers = 1\n\tbytesPerSecond := 10 * 1024\n\taccounting.TokenBucket.SetBwLimit(fs.BwPair{Tx: fs.SizeSuffix(bytesPerSecond), Rx: fs.SizeSuffix(bytesPerSecond)})\n\tdefer accounting.TokenBucket.SetBwLimit(fs.BwPair{Tx: -1, Rx: -1})\n\n\t// write one small file which we expect to transfer and one big one which we don't\n\tfile1 := r.WriteFile(\"file1\", string(make([]byte, 16)), t1)\n\tfile2 := r.WriteFile(\"file2\", string(make([]byte, 50*1024)), t1)\n\tr.CheckLocalItems(t, file1, file2)\n\tr.CheckRemoteItems(t)\n\n\taccounting.GlobalStats().ResetCounters()\n\tstartTime := time.Now()\n\terr := Sync(ctx, r.Fremote, r.Flocal, false)\n\trequire.True(t, errors.Is(err, ErrorMaxDurationReached))\n\n\tif cutoffMode == fs.CutoffModeHard {\n\t\tr.CheckRemoteItems(t, file1)\n\t\tassert.Equal(t, int64(1), accounting.GlobalStats().GetTransfers())\n\t} else {\n\t\tr.CheckRemoteItems(t, file1, file2)\n\t\tassert.Equal(t, int64(2), accounting.GlobalStats().GetTransfers())\n\t}\n\n\telapsed := time.Since(startTime)\n\tconst maxTransferTime = 20 * time.Second\n\n\twhat := fmt.Sprintf(\"expecting elapsed time %v between %v and %v\", elapsed, maxDuration, maxTransferTime)\n\tassert.True(t, elapsed >= maxDuration, what)\n\tassert.True(t, elapsed < maxTransferTime, what)\n}", "func (c *Cache) SizeMaxBytes() int {\n\tn := 0\n\tfor _, shard := range c.shards {\n\t\tn += shard.SizeMaxBytes()\n\t}\n\treturn n\n}", "func BoostedTreesCreateQuantileStreamResourceMaxElements(value int64) BoostedTreesCreateQuantileStreamResourceAttr {\n\treturn func(m optionalAttr) {\n\t\tm[\"max_elements\"] = value\n\t}\n}", "func (l *fileSink) initializeNewOutputFile(\n\tfile *os.File, now time.Time,\n) (newWriter *bufio.Writer, nbytes int64, err error) {\n\t// bufferSize sizes the buffer associated with each log file. It's large\n\t// so that log records can accumulate without the logging thread blocking\n\t// on disk I/O. The flushDaemon will block instead.\n\tconst bufferSize = 256 * 1024\n\n\tnewWriter = bufio.NewWriterSize(file, bufferSize)\n\n\tif l.getStartLines != nil {\n\t\tbufs := l.getStartLines(now)\n\t\tfor _, buf := range bufs {\n\t\t\tvar n int\n\t\t\tvar thisErr error\n\t\t\tn, thisErr = file.Write(buf.Bytes())\n\t\t\tnbytes += int64(n)\n\t\t\t// Note: we combine the errors, instead of stopping at the first\n\t\t\t// error encountered, to ensure that all the buffers get\n\t\t\t// released back to the pool.\n\t\t\terr = errors.CombineErrors(err, thisErr)\n\t\t\tputBuffer(buf)\n\t\t}\n\t}\n\n\tif err != nil {\n\t\treturn nil, nbytes, err\n\t}\n\n\treturn newWriter, nbytes, nil\n}", "func (p *Policy) setMaxBlockSize(ic *interop.Context, args []stackitem.Item) stackitem.Item {\n\tvalue := uint32(toBigInt(args[0]).Int64())\n\tif value > payload.MaxSize {\n\t\tpanic(fmt.Errorf(\"MaxBlockSize cannot be more than the maximum payload size = %d\", payload.MaxSize))\n\t}\n\tok, err := p.checkValidators(ic)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif !ok {\n\t\treturn stackitem.NewBool(false)\n\t}\n\tp.lock.Lock()\n\tdefer p.lock.Unlock()\n\terr = p.setUint32WithKey(ic.DAO, maxBlockSizeKey, value)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tp.isValid = false\n\treturn stackitem.NewBool(true)\n}", "func trackDynamicBandwidthLimit(ctx context.Context, waitGroup *sync.WaitGroup, fileBackedDevice *FileBackedDevice) {\n\twaitGroup.Add(1)\n\tdefer waitGroup.Done()\n\tvar codCount int\n\tfor !fileBackedDevice.IsFullySynced() && !isCancelSignaled(ctx) {\n\t\tcodProcesses, err := FindMaxProcessesCounts(LocalFs{}, logrus.StandardLogger(), fileBackedDevice.processFiles)\n\t\tif err != nil {\n\t\t\tfileBackedDevice.log.Warnf(\"Could not read progress file(s): %v\", fileBackedDevice.processFiles)\n\t\t} else {\n\t\t\tcodCount = codProcesses\n\t\t}\n\t\tupdateBandwidth(codCount, fileBackedDevice)\n\t\ttime.Sleep(5 * time.Second)\n\t}\n\treturn\n}", "func EntryByteLimit(n int) LoggerOption { return entryByteLimit(n) }", "func fileSplitOnly(w *writer, b []byte) (int, error) {\n\tfor i, c := range b {\n\t\tsplit := false\n\t\tv := sigmap[c]\n\t\tif len(v) > 0 && i < len(b)-6 {\n\t\t\tfor _, s := range v {\n\t\t\t\tsplit = true\n\t\t\t\tfor j, expect := range s {\n\t\t\t\t\tif b[j+1] != expect {\n\t\t\t\t\t\tsplit = false\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tw.cur[w.off] = c\n\t\tw.off++\n\n\t\t// Filled the buffer? Send it off!\n\t\tif split || w.off >= w.maxSize {\n\t\t\tb := <-w.buffers\n\t\t\t// Swap block with current\n\t\t\tw.cur, b.data = b.data[:w.maxSize], w.cur[:w.off]\n\t\t\tb.N = w.nblocks\n\n\t\t\tw.input <- b\n\t\t\tw.write <- b\n\t\t\tw.nblocks++\n\t\t\tw.off = 0\n\t\t}\n\t}\n\treturn len(b), nil\n}", "func ConcurrentWriteLimit(n int) LoggerOption { return concurrentWriteLimit(n) }", "func (self File) TailBytes(limitSize int) ([]byte, error) {\n\tvar limitBytes []byte\n\tfile := self.Open()\n\n\treadBytes, err := io.ReadAtLeast(file, limitBytes, limitSize)\n\tif readBytes != limitSize {\n\t\treturn limitBytes, fmt.Errorf(\"error: failed to complete read: read \", readBytes, \" out of \", limitSize, \"bytes\")\n\t} else {\n\t\treturn limitBytes, err\n\t}\n}", "func SplitFile(file *os.File, size int) {\n\tfinfo, err := file.Stat()\n\tif err != nil {\n\t\tfmt.Println(\"get file info failed:\", file, size)\n\t}\n\n\tfmt.Println(finfo, size)\n\n\t//每次最多拷贝1m\n\tbufsize := 1024 * 1024\n\tif size < bufsize {\n\t\tbufsize = size\n\t}\n\n\tbuf := make([]byte, bufsize)\n\n\tnum := (int(finfo.Size()) + size - 1) / size\n\tfmt.Println(num, len(buf))\n\n\tfor i := 0; i < num; i++ {\n\t\tcopylen := 0\n\t\tnewfilename := finfo.Name() + strconv.Itoa(i)\n\t\tnewfile, err1 := os.Create(newfilename)\n\t\tif err1 != nil {\n\t\t\tfmt.Println(\"failed to create file\", newfilename)\n\t\t} else {\n\t\t\tfmt.Println(\"create file:\", newfilename)\n\t\t}\n\n\t\tfor copylen < size {\n\t\t\tn, err2 := file.Read(buf)\n\t\t\tif err2 != nil && err2 != io.EOF {\n\t\t\t\tfmt.Println(err2, \"failed to read from:\", file)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif n <= 0 {\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\t//fmt.Println(n, len(buf))\n\n\t\t\t//写文件\n\t\t\tw_buf := buf[:n]\n\t\t\tnewfile.Write(w_buf)\n\t\t\tcopylen += n\n\t\t}\n\t}\n\n\treturn\n}", "func FileSize(s int64) string {\n\tsizes := []string{\"B\", \"KB\", \"MB\", \"GB\", \"TB\", \"PB\", \"EB\"}\n\treturn humanateBytes(uint64(s), 1024, sizes)\n}", "func FileSize(s int64) string {\n\tsizes := []string{\"B\", \"KB\", \"MB\", \"GB\", \"TB\", \"PB\", \"EB\"}\n\treturn humanateBytes(uint64(s), 1024, sizes)\n}", "func (p *Policy) getMaxBlockSize(ic *interop.Context, _ []stackitem.Item) stackitem.Item {\n\treturn stackitem.NewBigInteger(big.NewInt(int64(p.GetMaxBlockSizeInternal(ic.DAO))))\n}", "func makeThumbnails6(filenames []string) int64 {\n\tsizes := make(chan int64)\n\tvar wg sync.WaitGroup // counter. number of working goroutines\n\tfor _, f := range filenames {\n\t\twg.Add(1)\n\t\tgo func(f string) {\n\t\t\tdefer wg.Done()\n\t\t\tthumb, err := ImageFile(f)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tinfo, _ := os.Stat(thumb)\n\t\t\tsizes <- info.Size()\n\t\t}(f)\n\t}\n\n\tgo func() {\n\t\twg.Wait()\n\t\tclose(sizes)\n\t}()\n\n\tvar total int64\n\tfor size := range sizes {\n\t\ttotal += size\n\t}\n\treturn total\n}", "func (p *btrfsPool) Limit(size uint64) error {\n\treturn fmt.Errorf(\"not implemented\")\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigPtrOutput) FsFileMax() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.FsFileMax\n\t}).(pulumi.IntPtrOutput)\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigOutput) FsFileMax() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int { return v.FsFileMax }).(pulumi.IntPtrOutput)\n}", "func (s *ServerStat) bytes(capacity int64) int64 {\n\treturn int64(s.limit_maxbytes) - capacity\n}", "func makeBuf(max int) []byte {\n\tif max > BufferSize {\n\t\tmax = BufferSize\n\t}\n\treturn make([]byte, max)\n}", "func copyFileContents(size int64, src, dst string, bwlimit uint64) (written int64, err error) {\n\tif *contexte.verbose {\n\t\tfmt.Printf(\"%s -> %s (%s)\", src, dst, humanize.Bytes(uint64(size)))\n\t}\n\tif !*contexte.verbose {\n\t\tfmt.Print(\".\")\n\t}\n\n\tpool := iothrottler.NewIOThrottlerPool(iothrottler.BytesPerSecond * iothrottler.Bandwidth(bwlimit))\n\tdefer pool.ReleasePool()\n\n\tfile, err := os.Open(src)\n\tif err != nil {\n\t\t// fmt.Println(\"Error:\", err) // handle error\n\t\treturn 0, err\n\t}\n\tdefer func() {\n\t\tfile.Close()\n\t\tif err != nil {\n\t\t\tcolor.Set(color.FgRed)\n\t\t\tif *contexte.verbose {\n\t\t\t\tfmt.Print(\" KO\\n\")\n\t\t\t}\n\t\t\tif !*contexte.verbose {\n\t\t\t\tfmt.Print(\".\")\n\t\t\t}\n\t\t\tcolor.Unset()\n\t\t\treturn\n\t\t}\n\t\tcolor.Set(color.FgGreen)\n\t\tif *contexte.verbose {\n\t\t\tfmt.Print(\" OK\\n\")\n\t\t}\n\t\tif !*contexte.verbose {\n\t\t\tfmt.Print(\".\")\n\t\t}\n\t\tcolor.Unset()\n\t}()\n\n\tthrottledFile, err := pool.AddReader(file)\n\tif err != nil {\n\t\t// fmt.Println(\"Error:\", err) // handle error\n\t\t// handle error\n\t\treturn 0, err\n\t}\n\n\tout, err := os.Create(dst)\n\tif err != nil {\n\t\t// fmt.Println(\"Error:\", err) // handle error\n\t\treturn 0, err\n\t}\n\tdefer func() {\n\t\tcerr := out.Close()\n\t\tif err == nil {\n\t\t\terr = cerr\n\t\t}\n\t}()\n\tbytesw, err := io.Copy(out, throttledFile)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\terr = out.Sync()\n\treturn bytesw, err\n}", "func (s *stagingSink) Write(data []byte) (int, error) {\n\t// Watch for size violations.\n\tif s.maximumSize != 0 && (s.maximumSize-s.currentSize) < uint64(len(data)) {\n\t\treturn 0, errors.New(\"maximum file size reached\")\n\t}\n\n\t// Write to the underlying storage.\n\tn, err := s.storage.Write(data)\n\n\t// Write as much to the digester as we wrote to the underlying storage. This\n\t// can't fail.\n\ts.digester.Write(data[:n])\n\n\t// Update the current size. We needn't worry about this overflowing, because\n\t// if maximumSize is 0, then this isn't even used, and if maximumSize is\n\t// non-zero, then the check above is sufficient to know that this amount of\n\t// data won't overflow the maximum uint64 value.\n\ts.currentSize += uint64(n)\n\n\t// Done.\n\treturn n, err\n}", "func (info *Info) writeFiles(w io.Writer, open func(fi FileInfo) (io.ReadCloser, error)) error {\n\tfor _, fi := range info.UpvertedFiles() {\n\t\tr, err := open(fi)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error opening %v: %s\", fi, err)\n\t\t}\n\t\twn, err := io.CopyN(w, r, fi.Length)\n\t\tr.Close()\n\t\tif wn != fi.Length {\n\t\t\treturn fmt.Errorf(\"error copying %v: %s\", fi, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (b *BlockSplitterSimple) SetMaxDirEntriesPerBlockForTesting(newMax int) {\n\tb.maxDirEntriesPerBlock = newMax\n}", "func (g *Gauges) BuffersMaxWrittenClean() prometheus.Gauge {\n\treturn g.new(\n\t\tprometheus.GaugeOpts{\n\t\t\tName: \"postgresql_buffers_maxwritten_clean\",\n\t\t\tHelp: \"Number of times the background writer stopped a cleaning scan because it had written too many buffers\",\n\t\t\tConstLabels: g.labels,\n\t\t},\n\t\t\"SELECT maxwritten_clean FROM pg_stat_bgwriter\",\n\t)\n}", "func WithMaxBytes(maxBytes int) LimitedWriterOption {\n\tbytesWritten := 0\n\treturn func(w io.WriteCloser) io.WriteCloser {\n\t\tpreCheck := NewPreWriteCallbacks(w, func(p []byte) error {\n\t\t\tif bytesWritten+len(p) > maxBytes {\n\t\t\t\tif err := w.Close(); err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"failed to close WriteCloser writing maxBytes; Close error was: %w\", err)\n\t\t\t\t}\n\t\t\t\treturn ErrTooLargeWrite\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\n\t\treturn NewPostWriteCallbacks(preCheck, func(p []byte, n int, err error) {\n\t\t\tbytesWritten += n\n\t\t})\n\t}\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigOutput) FsFileMax() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int { return v.FsFileMax }).(pulumi.IntPtrOutput)\n}", "func getFreeLog(file *os.File, index int64, limit int64) int64 {\n\t//Obtenemos la bitacora inicial\n\tbita, pr := getBitacora(file, index, 0)\n\tfor i := 1; bita.TransactionDate != [19]byte{}; i++ {\n\t\t//Obtenemos la siguiente bitacora\n\t\tbita, pr = getBitacora(file, index, int64(i))\n\t}\n\tlimit = limit*int64(binary.Size(bita)) + index - int64(binary.Size(bita))\n\tif pr > limit {\n\t\treturn -1 \n\t}\n\treturn pr\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigPtrOutput) FsFileMax() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.FsFileMax\n\t}).(pulumi.IntPtrOutput)\n}", "func CalculateMaxSize() uint64 {\n\tif logging.logFile != \"\" {\n\t\tif logging.logFileMaxSizeMB == 0 {\n\t\t\t// If logFileMaxSizeMB is zero, we don't have limitations on the log size.\n\t\t\treturn math.MaxUint64\n\t\t}\n\t\t// Flag logFileMaxSizeMB is in MB for user convenience.\n\t\treturn logging.logFileMaxSizeMB * 1024 * 1024\n\t}\n\t// If \"log_file\" flag is not specified, the target file (sb.file) will be cleaned up when reaches a fixed size.\n\treturn MaxSize\n}", "func (s *server) bufferSize(digest *pb.Digest) int {\n\tif digest.SizeBytes < googleapi.DefaultUploadChunkSize {\n\t\treturn int(digest.SizeBytes)\n\t}\n\treturn googleapi.DefaultUploadChunkSize\n}", "func BySize(left, right *FileInfoPath) bool {\n\treturn left.Mode().IsRegular() &&\n\t\tright.Mode().IsRegular() &&\n\t\t(left.Size() < right.Size())\n}", "func increaseLimit() {\n\tvar rlimit syscall.Rlimit\n\tif err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &rlimit); err != nil {\n\t\tpanic(err)\n\t}\n\trlimit.Cur = rlimit.Max\n\tif err := syscall.Setrlimit(syscall.RLIMIT_NOFILE, &rlimit); err != nil {\n\t\tpanic(err)\n\t}\n\tlog.Printf(\"rlimit.Max = %d\\n\", rlimit.Max)\n\tlog.Printf(\"rlimit.Cur = %d\\n\", rlimit.Cur)\n}", "func increaseLimit() {\n\tvar rlimit syscall.Rlimit\n\tif err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &rlimit); err != nil {\n\t\tpanic(err)\n\t}\n\trlimit.Cur = rlimit.Max\n\tif err := syscall.Setrlimit(syscall.RLIMIT_NOFILE, &rlimit); err != nil {\n\t\tpanic(err)\n\t}\n\tlog.Printf(\"rlimit.Max = %d\\n\", rlimit.Max)\n\tlog.Printf(\"rlimit.Cur = %d\\n\", rlimit.Cur)\n}", "func usedSize(size, offset, chunkSize int) int {\n\tsize += offset % chunkSize\n\tusedSize := (size / chunkSize) * chunkSize\n\tif size%chunkSize > 0 {\n\t\tusedSize += chunkSize\n\t}\n\n\treturn usedSize\n}", "func createFiles(t *testing.T, dir, prefix string, n int, d time.Duration) int {\n\tt.Helper()\n\n\tif d == 0 {\n\t\td = 9 * time.Minute\n\t}\n\n\tfmtNum := func(n int) string {\n\t\ts := fmt.Sprintf(\"%09d\", n)\n\t\treturn s[:3] + \"_\" + s[3:6] + \"_\" + s[6:]\n\t}\n\n\tvar (\n\t\tmax = time.After(d)\n\t\tcreated int\n\t)\n\tfor i := 0; i < n; i++ {\n\t\tselect {\n\t\tcase <-max:\n\t\t\tt.Logf(\"createFiles: stopped at %s files because it took longer than %s\", fmtNum(created), d)\n\t\t\treturn created\n\t\tdefault:\n\t\t\tpath := join(dir, prefix+fmtNum(i))\n\t\t\tfp, err := os.Create(path)\n\t\t\tif err != nil {\n\t\t\t\tt.Errorf(\"create failed for %s: %s\", fmtNum(i), err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif err := fp.Close(); err != nil {\n\t\t\t\tt.Errorf(\"close failed for %s: %s\", fmtNum(i), err)\n\t\t\t}\n\t\t\tif err := os.Remove(path); err != nil {\n\t\t\t\tt.Errorf(\"remove failed for %s: %s\", fmtNum(i), err)\n\t\t\t}\n\t\t\tif i%10_000 == 0 {\n\t\t\t\tt.Logf(\"createFiles: %s\", fmtNum(i))\n\t\t\t}\n\t\t\tcreated++\n\t\t}\n\t}\n\treturn created\n}", "func (c *stream) writeChunks(b []byte) (int, error) {\n\ttotalN := 0\n\tfor {\n\t\ttoWrite := b\n\t\tlast := true\n\t\tif len(b) > MaxDataLen {\n\t\t\ttoWrite = b[:MaxDataLen]\n\t\t\tb = b[MaxDataLen:]\n\t\t\tlast = false\n\t\t}\n\t\tn, err := c.Write(toWrite)\n\t\ttotalN += n\n\t\tif last || err != nil {\n\t\t\treturn totalN, err\n\t\t}\n\t}\n}" ]
[ "0.67073935", "0.6590335", "0.65184194", "0.61902106", "0.61621016", "0.6069341", "0.59757304", "0.57250816", "0.56450105", "0.55433017", "0.5528977", "0.55281764", "0.54828227", "0.54323435", "0.5431975", "0.5417879", "0.5390343", "0.5368576", "0.5367149", "0.5358498", "0.5344472", "0.5338559", "0.5324176", "0.529943", "0.52974796", "0.5291434", "0.5277663", "0.5266775", "0.5263233", "0.5227803", "0.5220164", "0.52121705", "0.5210397", "0.520673", "0.52018464", "0.5199863", "0.5167477", "0.5165549", "0.5161617", "0.5158298", "0.51582193", "0.5157805", "0.51508576", "0.5150644", "0.51402444", "0.51371026", "0.5134302", "0.51312107", "0.51311463", "0.51249593", "0.5121283", "0.51053053", "0.5104977", "0.5087115", "0.5075603", "0.5071938", "0.5071663", "0.50656104", "0.5057426", "0.505624", "0.50538844", "0.50481623", "0.5046322", "0.504005", "0.5035556", "0.50330603", "0.5032591", "0.5030644", "0.5024338", "0.5022754", "0.5019938", "0.50090146", "0.50052905", "0.49974966", "0.49911118", "0.49911118", "0.49845004", "0.49818936", "0.49742627", "0.4954281", "0.49463648", "0.4943454", "0.4942384", "0.49373958", "0.49200976", "0.49123594", "0.49098447", "0.4902034", "0.49011397", "0.4895748", "0.489209", "0.489113", "0.48901334", "0.4889947", "0.48861772", "0.48805696", "0.48805696", "0.4878748", "0.4878253", "0.48565623" ]
0.6947793
0
BlockFiles calculates the amount of blocks needes for a file based on filesize and blocksize
func (s *StatFS) BlockFiles(size int) int { return size / int(s.stat.Bsize) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func blocks(n int) int {\n\tb := align(n+2, blocksize)\n\treturn b / blocksize\n}", "func (d *digest) BlockSize() int { return 1 }", "func calculateBlocks(size uint64, isDecrypt bool) (blocks uint64) {\n\tblocks = size / BlockSize\n\tif !isDecrypt {\n\t\tblocks++\n\t}\n\treturn\n}", "func (sm3 *SM3) BlockSize() int { return 64 }", "func (h *Hash) BlockSize() int { return len(h.buf) }", "func calculateBufferSize(blocks uint64) uint64 {\n\tif nbb := NBufferBlocks; blocks < nbb {\n\t\treturn blocks\n\t} else {\n\t\treturn nbb\n\t}\n}", "func BlockSizeECB(f EncryptFunc) int {\n\tconst (\n\t\tbufLen = 1024\n\t\tminBlockSize = 4\n\t\tmaxBlockSize = bufLen / 4\n\t)\n\n\tenc := f(A(bufLen))\n\n\tfor bs := minBlockSize; bs <= maxBlockSize; bs++ {\n\t\tnumNeeded := bufLen/bs - 2 // first or last may be misaligned\n\t\tvar prevBlock []byte // last block that was seen\n\t\tblockCount := 0 // consecutive occurrences of prevBlock\n\t\tfor start := 0; start+bs < len(enc); start += bs {\n\t\t\tbl := enc[start : start+bs]\n\t\t\tif prevBlock == nil || !bytes.Equal(bl, prevBlock) {\n\t\t\t\tprevBlock = bl\n\t\t\t\tblockCount = 0\n\t\t\t} else {\n\t\t\t\tblockCount++\n\t\t\t\tif blockCount >= numNeeded {\n\t\t\t\t\treturn bs\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tpanic(\"couldn't find block size\")\n}", "func (*digest) BlockSize() int {\n\treturn int(BlockSize)\n}", "func (*digest) BlockSize() int {\n\treturn int(BlockSize)\n}", "func (d Avx512Digest) BlockSize() int { return BlockSize }", "func (d *RabinKarp64) BlockSize() int { return 1 }", "func (lf *ListFile) NumBytes() int64 {\n\t// NOTE: here we don't use IsClosed() because\n\t// it uses the mutex; Size() is used in noMutexIterateLines\n\t// which is called after another mutex is locked,\n\t// making IsClosed() wait forever for the mutex unlock.\n\tif lf.isClosed {\n\t\treturn 0\n\t}\n\n\terr := lf.file.Sync()\n\tif err != nil {\n\t\t// TODO: not panic??\n\t\tpanic(err)\n\t}\n\n\tinfo, err := lf.file.Stat()\n\tif err != nil {\n\t\t// TODO: not panic??\n\t\tpanic(err)\n\t}\n\n\treturn info.Size()\n}", "func calculateChunks(state *State, segmentnum uint) []*chunk {\n\tcount := int64(segmentnum)\n\n\tpieceLength := int64(state.BitfieldPieceLength)\n\n\t// calculate the chunks of a resumable file.\n\tif state.Bitfield.Count() != 0 {\n\t\tvar chunks []*chunk\n\t\tvar idx uint32\n\t\tfor {\n\t\t\tstart, ok := state.Bitfield.FirstClear(idx)\n\t\t\tif !ok {\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tend, ok := state.Bitfield.FirstSet(start)\n\t\t\tif !ok {\n\t\t\t\tchunks = append(chunks, &chunk{\n\t\t\t\t\toffset: int64(start) * pieceLength,\n\t\t\t\t\tlength: state.FileLength - int64(start)*pieceLength,\n\t\t\t\t})\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tchunks = append(chunks, &chunk{\n\t\t\t\toffset: int64(start) * pieceLength,\n\t\t\t\tlength: int64(end-start) * pieceLength,\n\t\t\t})\n\n\t\t\tidx = end\n\t\t}\n\t\treturn chunks\n\t}\n\n\t// calculate the chunks of a fresh new file.\n\n\tfilesize := state.FileLength\n\t// don't even consider smaller files\n\tif filesize <= pieceLength || count <= 1 {\n\t\treturn []*chunk{{offset: 0, length: filesize}}\n\t}\n\n\t// how many blocks fit perfectly on a filesize\n\tblockCount := filesize / pieceLength\n\t// how many bytes are left out\n\texcessBytes := filesize % pieceLength\n\n\t// If there are no blocks available for the given blocksize, we're gonna\n\t// reduce the count to the max available block count.\n\tif blockCount < count {\n\t\tcount = blockCount\n\t}\n\n\tblocksPerUnit := blockCount / count\n\texcessBlocks := blockCount % count\n\n\tvar chunks []*chunk\n\tfor i := int64(0); i < count; i++ {\n\t\tchunks = append(chunks, &chunk{\n\t\t\toffset: i * blocksPerUnit * pieceLength,\n\t\t\tlength: blocksPerUnit * pieceLength,\n\t\t})\n\t}\n\n\tif excessBlocks > 0 {\n\t\toffset := count * blocksPerUnit * pieceLength\n\t\tlength := excessBlocks * pieceLength\n\t\tchunks = append(chunks, &chunk{\n\t\t\toffset: offset,\n\t\t\tlength: length,\n\t\t})\n\t}\n\n\t// append excess bytes to the last chunk\n\tif excessBytes > 0 {\n\t\tc := chunks[len(chunks)-1]\n\t\tc.length += excessBytes\n\t}\n\n\treturn chunks\n}", "func (digest *sm3Digest) BlockSize() int {\n\treturn BlockSize\n}", "func getBlkSize(p string) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(p, &fs)\n\tif err != nil {\n\t\tlog.Fatal(\"get block size error with:\", err)\n\t}\n\tBlockSize = uint64(fs.Bsize)\n}", "func (bdi *Info) CountBlocks() int {\n\treturn int(C.spdk_bdev_get_num_blocks(bdi.ptr()))\n}", "func (tfb *TempFileBlock) Size() (int64, error) {\n\tf, err := os.Open(tfb.TempFile)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"Error opening file block at \\\"\"+tfb.TempFile+\"\\\" for reading\")\n\t}\n\tfi, err := f.Stat()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"Error getting file info for file block at \\\"\"+tfb.TempFile+\"\\\"\")\n\t}\n\treturn fi.Size(), nil\n}", "func (x *ecbDecrypter) BlockSize() int { return x.blockSize }", "func BlockSize() int {\n\treturn config.Record.BlockSize\n}", "func (xxh *xxHash) BlockSize() int {\n\treturn 1\n}", "func (x *ecbEncrypter) BlockSize() int { return x.blockSize }", "func (d *state) BlockSize() int { return d.rate }", "func (h *ihash) BlockSize() int { return h.blockSize }", "func (cipher CipherFunction) BlockSize() int {\n\tswitch cipher {\n\tcase TripleDES:\n\t\treturn des.BlockSize\n\tcase CAST5:\n\t\treturn 8\n\tcase AES128, AES192, AES256:\n\t\treturn 16\n\t}\n\treturn 0\n}", "func (r *Rollsum32) BlockSize() int {\n\treturn int(r.blockSize)\n}", "func (db *DB) FileSize() (int64, error) {\n\tvar size int64\n\tfiles, err := db.opts.FileSystem.ReadDir(\".\")\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tfor _, file := range files {\n\t\tsize += file.Size()\n\t}\n\treturn size, nil\n}", "func (s *State) BlockSize() int { return 8 }", "func (s *State) BlockSize() int { return 2 }", "func (fc FileCollection) totalFileSize() (uint64,error)\t{\n\tvar tfsize uint64\n\tfor _,fsize := range fc.fileSizes {\n\t\tdirectory := fmt.Sprintf(\"%s/d-%d\",fc.frandi,fsize)\n\t\tfileList,err := getFilesInDir(directory)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"totalSizeFiles(): Error listing directory: %s\\n%s\",directory,err.Error())\n\t\t\treturn 0,err\n\t\t}\n\t\tfor _,v := range fileList {\n\t\t\ttfsize += uint64(v.Size())\n\t\t}\n\t}\n\treturn tfsize,nil\n}", "func blockSize(rate int) int {\n\tconst playbackBufferedTimeUs = 5000\n\treturn rate * playbackBufferedTimeUs / 1000000\n}", "func CreditToFileUploadSize(credit float64) (bytes int) {\n\tbytes = freeFileUploadBytes\n\tfor {\n\t\tif credit > 0 {\n\t\t\tbytes += freeFileUploadBytes\n\t\t\tcredit -= creditSteps\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn\n}", "func (s *State) BlockSize() int { return 1 }", "func BlockCount() int {\n\tc := cryptoutil.NewCipher(EncryptedCipherText)\n\treturn len(c) / BlockSize\n}", "func usedSize(size, offset, chunkSize int) int {\n\tsize += offset % chunkSize\n\tusedSize := (size / chunkSize) * chunkSize\n\tif size%chunkSize > 0 {\n\t\tusedSize += chunkSize\n\t}\n\n\treturn usedSize\n}", "func (i *identity) BlockSize() int {\n\treturn 0\n}", "func (f *LogFile) bytes() int {\n\tvar b int\n\tb += 24 // mu RWMutex is 24 bytes\n\tb += 16 // wg WaitGroup is 16 bytes\n\tb += int(unsafe.Sizeof(f.id))\n\t// Do not include f.data because it is mmap'd\n\t// TODO(jacobmarble): Uncomment when we are using go >= 1.10.0\n\t//b += int(unsafe.Sizeof(f.w)) + f.w.Size()\n\tb += int(unsafe.Sizeof(f.buf)) + len(f.buf)\n\tb += int(unsafe.Sizeof(f.keyBuf)) + len(f.keyBuf)\n\t// Do not count SeriesFile because it belongs to the code that constructed this Index.\n\tb += int(unsafe.Sizeof(f.size))\n\tb += int(unsafe.Sizeof(f.modTime))\n\tb += int(unsafe.Sizeof(f.seriesIDSet)) + f.seriesIDSet.Bytes()\n\tb += int(unsafe.Sizeof(f.tombstoneSeriesIDSet)) + f.tombstoneSeriesIDSet.Bytes()\n\tb += int(unsafe.Sizeof(f.mms)) + f.mms.bytes()\n\tb += int(unsafe.Sizeof(f.path)) + len(f.path)\n\treturn b\n}", "func newBlockfileMgr(id string, conf *Conf, indexConfig *blkstorage.IndexConfig, indexStore *leveldbhelper.DBHandle) *blockfileMgr {\n\tlogger.Debugf(\"newBlockfileMgr() initializing file-based block storage for ledger: %s \", id)\n\tvar rwMutexs []*sync.RWMutex\n\n\t//Determine the root directory for the blockfile storage, if it does not exist create it\n\trootDir := conf.getLedgerBlockDir(id)\n\t_, err := util.CreateDirIfMissing(rootDir)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Error: %s\", err))\n\t}\n\t// Instantiate the manager, i.e. blockFileMgr structure\n\tmgr := &blockfileMgr{rootDir: rootDir, conf: conf, db: indexStore, rwMutexs: rwMutexs}\n\n\t// cp = checkpointInfo, retrieve from the database the file suffix or number of where blocks were stored.\n\t// It also retrieves the current size of that file and the last block number that was written to that file.\n\t// At init checkpointInfo:latestFileChunkSuffixNum=[0], latestFileChunksize=[0], lastBlockNumber=[0]\n\tcpInfo, err := mgr.loadCurrentInfo()\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not get block file info for current block file from db: %s\", err))\n\t}\n\tif cpInfo == nil {\n\t\tlogger.Info(`Getting block information from block storage`)\n\t\tif cpInfo, err = constructCheckpointInfoFromBlockFiles(rootDir); err != nil {\n\t\t\tpanic(fmt.Sprintf(\"Could not build checkpoint info from block files: %s\", err))\n\t\t}\n\t\tlogger.Debugf(\"Info constructed by scanning the blocks dir = %s\", spew.Sdump(cpInfo))\n\t} else {\n\t\tlogger.Debug(`Synching block information from block storage (if needed)`)\n\t\tsyncCPInfoFromFS(rootDir, cpInfo)\n\t}\n\terr = mgr.saveCurrentInfo(cpInfo, true)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not save next block file info to db: %s\", err))\n\t}\n\n\tmgr.oldestFileChunkSuffixNum = syncOldestFileNum(rootDir)\n\t//If start up is a restart of an existing storage,new the rwMutex for the files\n\tif conf.dumpConf.Enabled {\n\t\tfor i := 0; i <= cpInfo.latestFileChunkSuffixNum; i++ {\n\t\t\trwMutex := new(sync.RWMutex)\n\t\t\tmgr.rwMutexs = append(mgr.rwMutexs, rwMutex)\n\t\t}\n\t}\n\tmgr.dumpMutex = new(sync.Mutex)\n\n\t//Open a writer to the file identified by the number and truncate it to only contain the latest block\n\t// that was completely saved (file system, index, cpinfo, etc)\n\tcurrentFileWriter, err := newBlockfileWriter(deriveBlockfilePath(rootDir, cpInfo.latestFileChunkSuffixNum))\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not open writer to current file: %s\", err))\n\t}\n\t//Truncate the file to remove excess past last block\n\terr = currentFileWriter.truncateFile(cpInfo.latestFileChunksize)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not truncate current file to known size in db: %s\", err))\n\t}\n\n\t// Create a new KeyValue store database handler for the blocks index in the keyvalue database\n\tmgr.index = newBlockIndex(indexConfig, indexStore)\n\n\t// Update the manager with the checkpoint info and the file writer\n\tmgr.cpInfo = cpInfo\n\tmgr.currentFileWriter = currentFileWriter\n\t// Create a checkpoint condition (event) variable, for the goroutine waiting for\n\t// or announcing the occurrence of an event.\n\tmgr.cpInfoCond = sync.NewCond(&sync.Mutex{})\n\n\t// init BlockchainInfo for external API's\n\tbcInfo := &common.BlockchainInfo{\n\t\tHeight: 0,\n\t\tCurrentBlockHash: nil,\n\t\tPreviousBlockHash: nil}\n\n\tif !cpInfo.isChainEmpty {\n\t\t//If start up is a restart of an existing storage, sync the index from block storage and update BlockchainInfo for external API's\n\t\tmgr.syncIndex()\n\t\tlastBlockHeader, err := mgr.retrieveBlockHeaderByNumber(cpInfo.lastBlockNumber)\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"Could not retrieve header of the last block form file: %s\", err))\n\t\t}\n\t\tlastBlockHash := lastBlockHeader.Hash()\n\t\tpreviousBlockHash := lastBlockHeader.PreviousHash\n\t\tbcInfo = &common.BlockchainInfo{\n\t\t\tHeight: cpInfo.lastBlockNumber + 1,\n\t\t\tCurrentBlockHash: lastBlockHash,\n\t\t\tPreviousBlockHash: previousBlockHash}\n\t}\n\tmgr.bcInfo.Store(bcInfo)\n\treturn mgr\n}", "func (layout Layout) hashesPerBlock() int64 {\n\treturn layout.blockSize / layout.digestSize\n}", "func (b *block) uncompressedSizeBytes() uint64 {\n\trowsCount := uint64(b.Len())\n\n\t// Take into account timestamps\n\tn := rowsCount * uint64(len(time.RFC3339Nano))\n\n\t// Take into account columns\n\tcs := b.columns\n\tfor i := range cs {\n\t\tc := &cs[i]\n\t\tnameLen := uint64(len(c.name))\n\t\tif nameLen == 0 {\n\t\t\tnameLen = uint64(len(\"_msg\"))\n\t\t}\n\t\tfor _, v := range c.values {\n\t\t\tif len(v) > 0 {\n\t\t\t\tn += nameLen + 2 + uint64(len(v))\n\t\t\t}\n\t\t}\n\t}\n\n\t// Take into account constColumns\n\tccs := b.constColumns\n\tfor i := range ccs {\n\t\tcc := &ccs[i]\n\t\tnameLen := uint64(len(cc.Name))\n\t\tif nameLen == 0 {\n\t\t\tnameLen = uint64(len(\"_msg\"))\n\t\t}\n\t\tn += rowsCount * (2 + nameLen + uint64(len(cc.Value)))\n\t}\n\n\treturn n\n}", "func (c *BlockCache) blockLength(height int) int {\n\tindex := height - c.firstBlock\n\treturn int(c.starts[index+1] - c.starts[index] - 8)\n}", "func (bdi *Info) BlockSize() int {\n\treturn int(C.spdk_bdev_get_block_size(bdi.ptr()))\n}", "func calculateBuffers(blocks uint64) (buffers uint64) {\n\tbuffers = blocks / NBufferBlocks\n\tif blocks%NBufferBlocks != 0 {\n\t\tbuffers++\n\t}\n\treturn\n}", "func getBlockSize(inputsize uint64) uint32 {\n\t// Substract version and block size in available size\n\tmaxSize := MAX_SIG_SIZE - 4 - 4\n\n\tmaxBlocks := math.Ceil(float64(maxSize) / float64(ROLLSUM_SIZE+sha1.Size))\n\treturn uint32(math.Ceil(float64(inputsize) / maxBlocks))\n}", "func (d *Dataset) BlockCount() int {\n\tif len(d.blocks) == 0 {\n\t\td.loadBlocks()\n\t}\n\treturn len(d.blocks)\n}", "func (s *StatFS) MaxFiles(size int) int {\n\treturn int(s.stat.Bfree / (uint64(size) / uint64(s.stat.Bsize)))\n}", "func (s *Server) UploadBlocks(stream service.DirSync_UploadBlocksServer) error {\n\tmeta, err := extractUploadMeta(stream.Context())\n\tif err != nil {\n\t\treturn err\n\t}\n\tpath := meta.Path\n\tblockSize := meta.BlockSize\n\n\ttmpfile, err := ioutil.TempFile(\"\", \"dirsync\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer tmpfile.Close()\n\n\tabsPath := filepath.Join(s.absDir, fsutil.CleanPath(path))\n\n\terr = os.MkdirAll(filepath.Dir(absPath), 0755)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfile, err := os.OpenFile(absPath, os.O_CREATE, 0755)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\tfileHash := sha256.New()\n\tstartTime := time.Now()\n\tfor {\n\t\tblock, err := stream.Recv()\n\t\tif err != nil {\n\t\t\tif err == io.EOF {\n\t\t\t\t// Atomically move tmp file to directory.\n\t\t\t\terr := os.Rename(tmpfile.Name(), absPath)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tfileChecksum := hex.EncodeToString(fileHash.Sum(nil))\n\t\t\t\ts.updateChecksumMapping(absPath, fileChecksum)\n\t\t\t\tlog.Printf(\"uploading file %s, elapsed: %s\\n\", path, time.Since(startTime))\n\t\t\t\treturn stream.SendAndClose(&service.UploadResponse{})\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t\tif block.GetReference() {\n\t\t\t_, err := file.Seek(int64(block.GetNumber())*int64(blockSize), 0)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tbuf := make([]byte, blockSize)\n\t\t\tn, _ := io.ReadFull(file, buf)\n\t\t\t_, err = tmpfile.Write(buf[:n])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t_, err = fileHash.Write(buf[:n])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\t_, err := tmpfile.Write(block.GetPayload())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t_, err = fileHash.Write(block.GetPayload())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n}", "func (c *cipher256) BlockSize() int { return blockSize256 }", "func FileSize(s int64) string {\n\tsizes := []string{\"B\", \"KB\", \"MB\", \"GB\", \"TB\", \"PB\", \"EB\"}\n\treturn humanateBytes(uint64(s), 1024, sizes)\n}", "func FileSize(s int64) string {\n\tsizes := []string{\"B\", \"KB\", \"MB\", \"GB\", \"TB\", \"PB\", \"EB\"}\n\treturn humanateBytes(uint64(s), 1024, sizes)\n}", "func (k Keeper) BlockByteSize(ctx sdk.Ctx) (res int64) {\n\tk.Paramstore.Get(ctx, types.KeyBlockByteSize, &res)\n\treturn\n}", "func BlockCount(xml []byte) int {\n\tr := regexp.MustCompile(`(<block)`)\n\tres := r.FindAllStringSubmatch(string(xml), -1)\n\treturn len(res)\n}", "func (f FormatHeader) BlockSize() uint16 {\n\treturn (f.BitsPerSample / 8) * f.NumChannels\n}", "func (s *Source) BlockSize() (n uint) {\n\treturn s.blockSize\n}", "func GetFileSplits(filePath string) ([]Split, error) {\n\t// open file and get size\n\tf, err := os.Open(filePath)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Error opening %s for read : %s\", filePath, err)\n\t}\n\tdefer f.Close()\n\tstat, err := os.Stat(filePath)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Error getting stat for file %s : %s\", filePath, err)\n\t}\n\tfileSize := stat.Size()\n\n\t// list blocks\n\tblocLocBytes := make([]byte, 64*1024, 64*1024)\n\tbytesLen, err := syscall.Getxattr(filePath, \"user.mfs.blockLocs\", blocLocBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tblocLocStr := string(blocLocBytes[:bytesLen])\n\tlines := strings.Split(blocLocStr, \"\\n\")\n\n\tret := make([]Split, 0, len(lines))\n\t// find first linebreak after each to make splits\n\tfor _, line := range lines {\n\t\tif line == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tlineSplits := strings.Split(line, \"\\t\")\n\t\tif len(lineSplits) < 3 {\n\t\t\treturn nil, fmt.Errorf(\"Error, improperly formatted blockLocs line (less than 3 elements): %s\", line)\n\t\t}\n\t\tblockStartPos, err := strconv.ParseInt(lineSplits[0], 10, 64)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Error, blockStartPos not a number! %s : %s\", err, line)\n\t\t}\n\t\tblockEndPos, err := strconv.ParseInt(lineSplits[1], 10, 64)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Error, blockEndPos not a number! %s : %s\", err, line)\n\t\t}\n\t\tif blockStartPos > 0 {\n\t\t\tblockStartPos, err = nextLineBreak(f, blockStartPos)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t\tif blockEndPos < fileSize {\n\t\t\tblockEndPos, err = nextLineBreak(f, blockEndPos)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t\tblockHosts := strings.Split(lineSplits[2], \",\")\n\t\tif len(blockHosts) != 0 {\n\t\t\tret = append(ret, Split{filePath, blockHosts[0], blockStartPos, blockEndPos})\n\t\t}\n\n\t}\n\treturn ret, nil\n}", "func (f *DynamicDiskBlockFactory) GetBlockCount() int64 {\n\treturn int64(f.params.BlockAllocationTable.BATEntriesCount)\n}", "func DefineFiles(tsize uint64, hilimit uint64, flS *FileCollection) error {\n\tvar nfiles, remain uint64\n\ttfs, err := flS.totalFileSize() \n\tif err != nil {\n\t\tlog.Printf(\"DefineFiles(): Error computing total file size: %s\", err.Error())\n\t\treturn err\n\t}\n\tif tsize > tfs && tsize > hilimit { //Trying to add files and the total size exceeds the limit\n\t\treturn fmt.Errorf(\"Size requested is over the limit: requested %d bytes, limit: %d bytes.\", tsize, hilimit)\n\t}\n\tfor index, fsize := range flS.fileSizes {\n\t\tnfiles = tsize / fsize\n\t\tremain = tsize % fsize\n\t\tif nfiles > limitFiles { //Use all files of this size, keep adding more files of higher capacities\n\t\t\ttsize -= limitFiles * fsize\n\t\t\tflS.fileAmmount[index] = limitFiles\n\t\t} else if nfiles == 0 {\n\t\t\tflS.fileAmmount[index] = 0\n\t\t} else {\n\t\t\ttsize -= nfiles * fsize\n\t\t\tflS.fileAmmount[index] = nfiles\n\t\t}\n\t}\n\tif tsize > flS.fileSizes[len(flS.fileSizes)-1] { //The remaining size to allocate is bigger than the biggest file sezie, Add more parts of the maximum size\n\t\tnfiles = tsize / flS.fileSizes[len(flS.fileSizes)-1]\n\t\tremain = tsize % flS.fileSizes[len(flS.fileSizes)-1]\n\t\tflS.fileAmmount[len(flS.fileAmmount)-1] += nfiles\n\t}\n\tif remain > 0 { //The remain must be smaller than the bigger file size.\n\t\tfor index, fsize := range flS.fileSizes {\n\t\t\tif remain <= 3*fsize {\n\t\t\t\tsignRemain := int(remain)\n\t\t\t\tfor signRemain > 0 {\n\t\t\t\t\tflS.fileAmmount[index]++\n\t\t\t\t\tsignRemain -= int(fsize)\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (cipher *Simon96Cipher) BlockSize() int {\n\treturn 12\n}", "func (superBlock SuperBlock) InodeCount() Size {\n\treturn Size(superBlock.ClusterBitmapStartAddress-superBlock.InodeStartAddress) / Size(unsafe.Sizeof(PseudoInode{}))\n}", "func NumChunks(capacity int, size int64) int {\n\treturn int(math.Ceil(float64(size*8) / 5 / float64(capacity)))\n}", "func (f File) Size() int {\n\treturn (2 + len(f.BencodedPath) + 1) * 8;\n}", "func (m *Mode) NBlocks() uint64 {\n\treturn m.blocks\n}", "func (r *Receiver) countBytes(length uint64, update bool) error {\n\tif length > MaxTransfer {\n\t\treturn fmt.Errorf(\"file too large, %d\", length)\n\t}\n\tif r.byteLimit != 0 && r.totalBytes > uint64(r.byteLimit)-length {\n\t\treturn fmt.Errorf(\"file too large, %d\", length)\n\t}\n\tif update {\n\t\tr.totalBytes += length\n\t}\n\treturn nil\n}", "func ReadBlock(filePth string, bufSize int64, skipbitlen int64) ([]byte, error) {\n\tvar merchant int64 = skipbitlen / 8\n\tvar remainder int64 = skipbitlen % 8\n\tvar blockbitsize int64\n\tvar iter int64\n\tif (bufSize % 8) > 0 {\n\t\tblockbitsize = (bufSize/8 + 1)\n\t} else {\n\t\tblockbitsize = (bufSize / 8)\n\t}\n\tvar datalen = blockbitsize + 1\n\t// File\n\tfile, err := os.Open(filePth)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer file.Close()\n\tdata := make([]byte, datalen)\n\tfor iter = 0; iter < datalen; iter++ {\n\t\tdata[iter] = 0\n\t}\n\tretdata := make([]byte, blockbitsize)\n\tfor iter = 0; iter < blockbitsize; iter++ {\n\t\tretdata[iter] = 0\n\t}\n\t_, err = file.ReadAt(data, merchant)\n\t//fmt.Println(\"count is \",count)\n\t//datasignStr := fmt.Sprintf(\"%x\", data)\n\t//fmt.Println(\"data is \", datasignStr)\n\n\tif remainder > 0 {\n\t\tfor iter = 0; iter < blockbitsize; iter++ {\n\t\t\tretdata[iter] = (data[iter] << remainder) ^ (data[iter+1] >> (8 - remainder))\n\t\t}\n\t} else {\n\t\tfor iter = 0; iter < blockbitsize; iter++ {\n\t\t\tretdata[iter] = data[iter]\n\t\t}\n\t}\n\treturn retdata, fmt.Errorf(\"\")\n}", "func (w *writer) chunkCount(b []byte) int {\n\tlenB := len(b)\n\tif lenB <= w.chunkSize {\n\t\treturn 1\n\t}\n\n\treturn len(b)/w.chunkDataSize + 1\n}", "func HashFile(strFilepath string) []BlockHash {\n\tvar c, startWindowPosition, index, cmatch, lenCurr int = 0, 0, 0, 0, -1\n\tvar hash uint64 = 0\n\tvar currByte byte\n\tvar window WindowBytes\n\tvar hashBlock [16]byte\n\tvar arrBlockHash []BlockHash\n\n\t//\n\tfmt.Println(\"Start hash of file \", strFilepath)\n\n\t// Check if file exists\n\tif _, err := os.Stat(strFilepath); os.IsNotExist(err) {\n\t\treturn arrBlockHash\n\t}\n\n\twindow.init(HASH_WINDOW_SIZE)\n\n\t// Read file\n\tf, err := os.Open(strFilepath)\n\tif err != nil {\n\t\tfmt.Println(\"Err in opening file\", err)\n\t\treturn arrBlockHash\n\t}\n\tdefer func() {\n\t\tif err := f.Close(); err != nil {\n\t\t\tfmt.Println(\"Err in closing file\", err)\n\t\t\treturn\n\t\t}\n\t}()\n\treader := bufio.NewReader(f)\n\n\t// Reset the read window, we'll slide from there\n\tlenCurr, err = window.readFull(reader)\n\tif err != nil && lenCurr <= 0 {\n\t\tfmt.Println(\"Err in reading file\", err)\n\t\treturn arrBlockHash\n\t}\n\tc += lenCurr\n\t// Calculate window hash (first time)\n\tfor index, currByte = range window.currBytes {\n\t\thash += uint64(currByte) * math.Pow(HASH_PRIME_ROOT, HASH_WINDOW_SIZE-index-1)\n\t}\n\n\tfor {\n\t\t// Check if we fit the match, and at least a certain amount of bytes\n\t\tif (hash | HASH_MASK) == hash {\n\n\t\t\t// New match, md5 it\n\t\t\tcmatch++\n\t\t\thashBlock = md5.Sum(window.currBlock)\n\t\t\tarrBlockHash = append(arrBlockHash, BlockHash{Length: lenCurr, Hash: hashBlock, PositionInFile: startWindowPosition})\n\n\t\t\t// Reset the read window, we'll slide from there\n\t\t\tlenCurr, err = window.readFull(reader)\n\t\t\tif err != nil && lenCurr <= 0 {\n\t\t\t\tfmt.Println(\"Error in hashfile\", err)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tstartWindowPosition = c\n\t\t\tc += lenCurr\n\t\t\t// Calculate next window hash\n\t\t\thash = 0\n\t\t\tfor index, currByte = range window.currBytes {\n\t\t\t\thash += uint64(currByte) * math.Pow(HASH_PRIME_ROOT, HASH_WINDOW_SIZE-index-1)\n\t\t\t}\n\n\t\t} else {\n\t\t\t// No fit, we keep going for this block\n\t\t\tcurrByte, err = reader.ReadByte()\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(\"Error in hashfile2\", err)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\t// Magic hash\n\t\t\thash -= uint64(window.getFirstByte()) * math.Pow(HASH_PRIME_ROOT, HASH_WINDOW_SIZE-1)\n\t\t\thash *= HASH_PRIME_ROOT\n\t\t\thash += uint64(currByte)\n\n\t\t\t// Add new byte read\n\t\t\twindow.addByte(currByte)\n\t\t\tc++\n\t\t\tlenCurr++\n\t\t}\n\t}\n\n\t// Last block, if not empty\n\tif lenCurr > 0 {\n\t\thashBlock = md5.Sum(window.currBlock)\n\t\tarrBlockHash = append(arrBlockHash, BlockHash{Length: lenCurr, Hash: hashBlock, PositionInFile: startWindowPosition})\n\t}\n\n\tfmt.Printf(\"Found %d matches!\\n\", cmatch)\n\tfmt.Printf(\"Went through %d bytes!\\n\", c)\n\n\treturn arrBlockHash\n}", "func numChunks(size int64) int32 {\n\td := float64(size) / float64(1024*1024)\n\tn := int(math.Ceil(d))\n\treturn int32(n)\n}", "func fileSize(file string) int64 {\n\tf, e := os.Stat(file)\n\tif e != nil {\n\t\tfmt.Println(e.Error())\n\t\treturn 0\n\t}\n\n\treturn f.Size()\n}", "func (msg *Block) SerializeSize() int {\n\t// Block header bytes + Serialized varint size for the number of transactions.\n\tn := blockHeaderLen + VarIntSerializeSize(uint64(len(msg.Transactions)))\n\tfor _, tx := range msg.Transactions {\n\t\tn += tx.SerializeSize()\n\t}\n\treturn n\n}", "func main() {\n file, err := os.Open(os.Args[1])\n\n fi, err := file.Stat();\n\n //each split file size\n size := fi.Size()\n b := make([]byte, size);\n _, err = file.Read(b);\n file.Close()\n\n fmt.Println(string(b))\n}", "func DetectBlockSize(oracle func([]byte) []byte) uint {\n\n\tsize := len(oracle(make([]byte, 0)))\n\tdata := make([]byte, 1)\n\tfor true {\n\t\tdif := len(oracle(data)) - size\n\t\tif dif > 0 {\n\t\t\treturn uint(dif)\n\t\t} else {\n\t\t\tdata = append(data, byte(0))\n\t\t}\n\t}\n\treturn 0 // no valid block size found\n}", "func (f FixedFileinfo) Size() int { return binary.Size(f) }", "func (o *StoragePhysicalDiskAllOf) GetNumBlocks() string {\n\tif o == nil || o.NumBlocks == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.NumBlocks\n}", "func getSize(n string) [4]byte {\n\tfilesL.RLock()\n\tdefer filesL.RUnlock()\n\n\t/* Get hold of the file */\n\tf, ok := files[n]\n\tif !ok {\n\t\tlog.Panicf(\"no file %q for size\", n)\n\t}\n\n\t/* Get the file size */\n\tvar a [4]byte\n\ta = f.size\n\treturn a\n}", "func (tf tFiles) size() (sum int64) {\n\tfor _, t := range tf {\n\t\tsum += t.size\n\t}\n\treturn sum\n}", "func (f *FileList) NFiles() int {\n\t_, nfiles, _ := f.NTotalDirsAndFile()\n\treturn nfiles\n}", "func (f *File) Size() (uint32, error) {\n\tif err := f.unlocker.Ensure(); err != nil {\n\t\treturn 0, err\n\t}\n\t// file chunks include an embedded length field at the start\n\tbinlength, _, err := f.t.client.Read(f.chunk, 0, 4)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn binary.LittleEndian.Uint32(binlength), nil\n}", "func nbSectors(bs *BlockDriverState) (int64, error) {\n\tdrv := bs.Drv\n\n\tif drv == nil {\n\t\treturn 0, ENOMEDIUM\n\t}\n\n\tif drv.hasVariableLength {\n\t\terr := refreshTotalSectors(bs, bs.TotalSectors)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\treturn bs.TotalSectors, nil\n}", "func fileSize(file string) (int64, error) {\n\tst := syscall.Stat_t{}\n\n\terr := syscall.Stat(file, &st)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn st.Size, nil\n}", "func fileSplitOnly(w *writer, b []byte) (int, error) {\n\tfor i, c := range b {\n\t\tsplit := false\n\t\tv := sigmap[c]\n\t\tif len(v) > 0 && i < len(b)-6 {\n\t\t\tfor _, s := range v {\n\t\t\t\tsplit = true\n\t\t\t\tfor j, expect := range s {\n\t\t\t\t\tif b[j+1] != expect {\n\t\t\t\t\t\tsplit = false\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tw.cur[w.off] = c\n\t\tw.off++\n\n\t\t// Filled the buffer? Send it off!\n\t\tif split || w.off >= w.maxSize {\n\t\t\tb := <-w.buffers\n\t\t\t// Swap block with current\n\t\t\tw.cur, b.data = b.data[:w.maxSize], w.cur[:w.off]\n\t\t\tb.N = w.nblocks\n\n\t\t\tw.input <- b\n\t\t\tw.write <- b\n\t\t\tw.nblocks++\n\t\t\tw.off = 0\n\t\t}\n\t}\n\treturn len(b), nil\n}", "func (storage *B2Storage) ListFiles(threadIndex int, dir string) (files []string, sizes []int64, err error) {\n for len(dir) > 0 && dir[len(dir) - 1] == '/' {\n dir = dir[:len(dir) - 1]\n }\n length := len(dir) + 1\n\n includeVersions := false\n if dir == \"chunks\" {\n includeVersions = true\n }\n\n entries, err := storage.clients[threadIndex].ListFileNames(dir, false, includeVersions)\n if err != nil {\n return nil, nil, err\n }\n\n if dir == \"snapshots\" {\n\n subDirs := make(map[string]bool)\n\n for _, entry := range entries {\n name := entry.FileName[length:]\n subDir := strings.Split(name, \"/\")[0]\n subDirs[subDir + \"/\"] = true\n }\n\n for subDir, _ := range subDirs {\n files = append(files, subDir)\n }\n } else if dir == \"chunks\" {\n lastFile := \"\"\n for _, entry := range entries {\n if entry.FileName == lastFile {\n continue\n }\n lastFile = entry.FileName\n if entry.Action == \"hide\" {\n files = append(files, entry.FileName[length:] + \".fsl\")\n } else {\n files = append(files, entry.FileName[length:])\n }\n sizes = append(sizes, entry.Size)\n }\n } else {\n for _, entry := range entries {\n files = append(files, entry.FileName[length:])\n }\n }\n\n return files, sizes, nil\n}", "func MaxBlockLen(ct CompressionType) uint64 {\n\tif ct == Snappy {\n\t\t// https://github.com/golang/snappy/blob/2a8bb927dd31d8daada140a5d09578521ce5c36a/encode.go#L76\n\t\treturn 6 * (0xffffffff - 32) / 7\n\t}\n\treturn math.MaxUint64\n}", "func (x *gcm) BlockSize() int { return x.blockSize }", "func writeFragmentBlocks(fileList []*finalizeFileInfo, f util.File, ws string, blocksize int, options FinalizeOptions, location int64) ([]fragmentBlock, int64, error) {\n\tcompressor := options.Compression\n\tif options.NoCompressFragments {\n\t\tcompressor = nil\n\t}\n\tfragmentData := make([]byte, 0)\n\tvar (\n\t\tallWritten int64\n\t\tfragmentBlockIndex uint32\n\t\tfragmentBlocks []fragmentBlock\n\t)\n\tfileCloseList := make([]*os.File, 0)\n\tdefer func() {\n\t\tfor _, f := range fileCloseList {\n\t\t\tf.Close()\n\t\t}\n\t}()\n\tfor _, e := range fileList {\n\t\t// only copy data for regular files\n\t\tif e.fileType != fileRegular {\n\t\t\tcontinue\n\t\t}\n\t\tvar (\n\t\t\twritten int64\n\t\t\terr error\n\t\t)\n\n\t\t// how much is there to put in a fragment?\n\t\tremainder := e.Size() % int64(blocksize)\n\t\tif remainder == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// would adding this data cause us to write?\n\t\tif len(fragmentData)+int(remainder) > blocksize {\n\t\t\twritten, compressed, err := finalizeFragment(fragmentData, f, location, compressor)\n\t\t\tif err != nil {\n\t\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"error writing fragment block %d: %v\", fragmentBlockIndex, err)\n\t\t\t}\n\t\t\tfragmentBlocks = append(fragmentBlocks, fragmentBlock{\n\t\t\t\tsize: uint32(written),\n\t\t\t\tcompressed: compressed,\n\t\t\t\tlocation: location,\n\t\t\t})\n\t\t\t// increment as all writes will be to next block block\n\t\t\tfragmentBlockIndex++\n\t\t\tfragmentData = fragmentData[:blocksize]\n\t\t}\n\n\t\te.fragment = &fragmentRef{\n\t\t\tblock: fragmentBlockIndex,\n\t\t\toffset: uint32(len(fragmentData)),\n\t\t}\n\t\t// save the fragment data from the file\n\n\t\tfrom, err := os.Open(path.Join(ws, e.path))\n\t\tif err != nil {\n\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"failed to open file for reading %s: %v\", e.path, err)\n\t\t}\n\t\tfileCloseList = append(fileCloseList, from)\n\t\tbuf := make([]byte, remainder)\n\t\tn, err := from.ReadAt(buf, e.Size()-remainder)\n\t\tif err != nil && err != io.EOF {\n\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"error reading final %d bytes from file %s: %v\", remainder, e.Name(), err)\n\t\t}\n\t\tif n != len(buf) {\n\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"failed reading final %d bytes from file %s, only read %d\", remainder, e.Name(), n)\n\t\t}\n\t\tfrom.Close()\n\t\tfragmentData = append(fragmentData, buf...)\n\n\t\tallWritten += written\n\t\tif written > 0 {\n\t\t\tfragmentBlockIndex++\n\t\t}\n\t}\n\n\t// write remaining fragment data\n\tif len(fragmentData) > 0 {\n\t\twritten, compressed, err := finalizeFragment(fragmentData, f, location, compressor)\n\t\tif err != nil {\n\t\t\treturn fragmentBlocks, 0, fmt.Errorf(\"error writing fragment block %d: %v\", fragmentBlockIndex, err)\n\t\t}\n\t\tfragmentBlocks = append(fragmentBlocks, fragmentBlock{\n\t\t\tsize: uint32(written),\n\t\t\tcompressed: compressed,\n\t\t\tlocation: location,\n\t\t})\n\t\t// increment as all writes will be to next block block\n\t\tallWritten += int64(written)\n\t}\n\treturn fragmentBlocks, allWritten, nil\n}", "func (n NZB) NumFiles() int {\n\treturn len(n.Files)\n}", "func BlockFile(h int) string {\n\treturn fmt.Sprintf(\"./blocks_gen/%d.dat\", h)\n}", "func DirSizeByte(path string) uint64 {\n\tdirSize = 0\n\tfilepath.Walk(path, readSize)\n\treturn dirSize\n}", "func (p *pbkdf2) getHashLengthBitBlockCount() int {\n\treturn int(math.Ceil(float64(p.size) / float64(p.pseudoRandomFunction.Size())))\n}", "func (b *logEventBuffer) blockNumberIndex(bn int64) int {\n\treturn int(bn-1) % b.bufferSize()\n}", "func HashFile(ctx context.Context, fs fs.Filesystem, path string, blockSize int, counter Counter, useWeakHashes bool) ([]protocol.BlockInfo, error) {\n\tfd, err := fs.Open(path)\n\tif err != nil {\n\t\tl.Debugln(\"open:\", err)\n\t\treturn nil, err\n\t}\n\tdefer fd.Close()\n\n\t// Get the size and modtime of the file before we start hashing it.\n\n\tfi, err := fd.Stat()\n\tif err != nil {\n\t\tl.Debugln(\"stat before:\", err)\n\t\treturn nil, err\n\t}\n\tsize := fi.Size()\n\tmodTime := fi.ModTime()\n\n\t// Hash the file. This may take a while for large files.\n\n\tblocks, err := Blocks(ctx, fd, blockSize, size, counter, useWeakHashes)\n\tif err != nil {\n\t\tl.Debugln(\"blocks:\", err)\n\t\treturn nil, err\n\t}\n\n\t// Recheck the size and modtime again. If they differ, the file changed\n\t// while we were reading it and our hash results are invalid.\n\n\tfi, err = fd.Stat()\n\tif err != nil {\n\t\tl.Debugln(\"stat after:\", err)\n\t\treturn nil, err\n\t}\n\tif size != fi.Size() || !modTime.Equal(fi.ModTime()) {\n\t\treturn nil, errors.New(\"file changed during hashing\")\n\t}\n\n\treturn blocks, nil\n}", "func (m *MockHash64) BlockSize() int {\n\tret := m.ctrl.Call(m, \"BlockSize\")\n\tret0, _ := ret[0].(int)\n\treturn ret0\n}", "func FileSizeFormating(b int64) string {\n\tconst unit = 1024\n\tif b < unit {\n\t\treturn fmt.Sprintf(\"%d B\", b)\n\t}\n\tdiv, exp := int64(unit), 0\n\tfor n := b / unit; n >= unit; n /= unit {\n\t\tdiv *= unit\n\t\texp++\n\t}\n\treturn fmt.Sprintf(\"%.1f %cB\",\n\t\tfloat64(b)/float64(div), \"kMGTPE\"[exp])\n}", "func fetchBatchesPerBlock(ctx context.Context, client *ethclient.Client, number uint64, signer types.Signer, config Config) (uint64, uint64, error) {\n\tvalidBatchCount := uint64(0)\n\tinvalidBatchCount := uint64(0)\n\tctx, cancel := context.WithTimeout(ctx, 10*time.Second)\n\tdefer cancel()\n\tblock, err := client.BlockByNumber(ctx, new(big.Int).SetUint64(number))\n\tif err != nil {\n\t\treturn 0, 0, err\n\t}\n\tfmt.Println(\"Fetched block: \", number)\n\tfor i, tx := range block.Transactions() {\n\t\tif tx.To() != nil && *tx.To() == config.BatchInbox {\n\t\t\tsender, err := signer.Sender(tx)\n\t\t\tif err != nil {\n\t\t\t\treturn 0, 0, err\n\t\t\t}\n\t\t\tvalidSender := true\n\t\t\tif _, ok := config.BatchSenders[sender]; !ok {\n\t\t\t\tfmt.Printf(\"Found a transaction (%s) from an invalid sender (%s)\\n\", tx.Hash().String(), sender.String())\n\t\t\t\tinvalidBatchCount += 1\n\t\t\t\tvalidSender = false\n\t\t\t}\n\n\t\t\tvalidFrames := true\n\t\t\tframeError := \"\"\n\t\t\tframes, err := derive.ParseFrames(tx.Data())\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Found a transaction (%s) with invalid data: %v\\n\", tx.Hash().String(), err)\n\t\t\t\tvalidFrames = false\n\t\t\t\tframeError = err.Error()\n\t\t\t}\n\n\t\t\tif validSender && validFrames {\n\t\t\t\tvalidBatchCount += 1\n\t\t\t} else {\n\t\t\t\tinvalidBatchCount += 1\n\t\t\t}\n\n\t\t\ttxm := &TransactionWithMetadata{\n\t\t\t\tTx: tx,\n\t\t\t\tSender: sender,\n\t\t\t\tValidSender: validSender,\n\t\t\t\tTxIndex: uint64(i),\n\t\t\t\tBlockNumber: block.NumberU64(),\n\t\t\t\tBlockHash: block.Hash(),\n\t\t\t\tBlockTime: block.Time(),\n\t\t\t\tChainId: config.ChainID.Uint64(),\n\t\t\t\tInboxAddr: config.BatchInbox,\n\t\t\t\tFrames: frames,\n\t\t\t\tFrameErr: frameError,\n\t\t\t\tValidFrames: validFrames,\n\t\t\t}\n\t\t\tfilename := path.Join(config.OutDirectory, fmt.Sprintf(\"%s.json\", tx.Hash().String()))\n\t\t\tfile, err := os.Create(filename)\n\t\t\tif err != nil {\n\t\t\t\treturn 0, 0, err\n\t\t\t}\n\t\t\tdefer file.Close()\n\t\t\tenc := json.NewEncoder(file)\n\t\t\tif err := enc.Encode(txm); err != nil {\n\t\t\t\treturn 0, 0, err\n\t\t\t}\n\t\t}\n\t}\n\treturn validBatchCount, invalidBatchCount, nil\n}", "func (s *FileSet) TotalFileSize() int64 {\n\tvar result int64\n\ts.Walk(func(f string) {\n\t\tinfo, err := os.Stat(f)\n\t\tif err != nil {\n\t\t\tlog.Error().Str(\"file\", f).Msg(\"Could not stat\")\n\t\t}\n\t\tresult += info.Size()\n\t})\n\treturn result\n}", "func TestSizeBasedSplit(t *testing.T) {\n\tif testing.Short() {\n\t\tt.SkipNow()\n\t}\n\n\ttestFileConsistency(t, 32*512, 512)\n\ttestFileConsistency(t, 32*4096, 4096)\n\n\t// Uneven offset\n\ttestFileConsistency(t, 31*4095, 4096)\n}", "func (f *DynamicDiskBlockFactory) GetBlockSize() int64 {\n\treturn int64(f.params.VhdHeader.BlockSize)\n}", "func (h *MsgBlockHeader) NumParentBlocks() byte {\n\tnumParents := len(h.ParentHashes)\n\tif numParents > math.MaxUint8 {\n\t\tpanic(errors.Errorf(\"number of parents is %d, which is more than one byte can fit\", numParents))\n\t}\n\treturn byte(numParents)\n}", "func (m *MockHash) BlockSize() int {\n\tret := m.ctrl.Call(m, \"BlockSize\")\n\tret0, _ := ret[0].(int)\n\treturn ret0\n}", "func (p *SimplePipeline) BlockSize() uint {\n\treturn p.source.BlockSize()\n}", "func (b *Block) Size() int {\n\tpbb, err := b.ToProto()\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn pbb.Size()\n}", "func (b *Block) Size() int {\n\tpbb, err := b.ToProto()\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn pbb.Size()\n}" ]
[ "0.6831651", "0.64359915", "0.6351425", "0.61531454", "0.60735196", "0.6053935", "0.59970754", "0.59917593", "0.59917593", "0.59467095", "0.59163296", "0.5890694", "0.5873502", "0.5858289", "0.58537126", "0.5849925", "0.58035946", "0.5725327", "0.57226384", "0.5721372", "0.5704868", "0.5704112", "0.5686368", "0.5659873", "0.56383026", "0.56335264", "0.56248045", "0.56163067", "0.56133986", "0.5609269", "0.5572325", "0.5541451", "0.5535294", "0.55218124", "0.55167896", "0.5507961", "0.5499317", "0.5494869", "0.5491738", "0.5475532", "0.54718965", "0.54496354", "0.54432434", "0.5438084", "0.5428672", "0.5418755", "0.5407061", "0.5404045", "0.5404045", "0.5400038", "0.5393645", "0.53917", "0.5383651", "0.5380635", "0.5365647", "0.5351255", "0.534785", "0.53466564", "0.5339704", "0.53371704", "0.5336222", "0.5334116", "0.53215575", "0.531843", "0.53176373", "0.5314984", "0.5253755", "0.5252306", "0.5249448", "0.5235407", "0.5218566", "0.5213241", "0.5208575", "0.5206149", "0.518752", "0.5186624", "0.5168998", "0.5168638", "0.5151566", "0.5151349", "0.51484257", "0.51159865", "0.51132774", "0.5101668", "0.5100717", "0.50997853", "0.5097142", "0.5092485", "0.50893515", "0.5083652", "0.5070499", "0.5061514", "0.50593144", "0.5053091", "0.5048236", "0.50421035", "0.50383514", "0.50325125", "0.5028545", "0.5028545" ]
0.78074825
0
Returns a quic.Session object with a .OpenStreamSync method to send streams
func DialAsQuicClient(ctx context.Context, hostname string, port int) (quic.Session, error) { // TODO(jophba): Change InsecureSkipVerify tlsConfig := &tls.Config{InsecureSkipVerify: true} addr := fmt.Sprintf("%s:%d", hostname, port) session, err := quic.DialAddrContext(ctx, addr, tlsConfig, nil) return session, err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Session) Open() (str *Stream, err error) {\n\t// Wait for acks\n\tselect {\n\tcase s.opened <- struct{}{}:\n\tcase <-s.done:\n\t\treturn nil, ErrSessionClosed\n\t}\n\tsid := s.generateStreamID()\n\tif sid == 0 {\n\t\treturn nil, ErrMaxOpenStreams\n\t}\n\tstr = newStream(s, sid)\n\tif _, err = str.SendWindowUpdate(flagSYN, 0); err != nil {\n\t\treturn nil, err\n\t}\n\tif err = s.putStream(str); err != nil {\n\t\treturn nil, ErrSessionClosed\n\t}\n\treturn str, nil\n}", "func (s *Yamux) OpenStream() (net.Conn, error) {\n\treturn s.session.OpenStream()\n}", "func (t testConn) NewStream(ctx context.Context) (network.Stream, error) { return nil, nil }", "func (s *JudgePlayServerCallStub) SendStream() interface {\n\tSend(item JudgeAction) error\n} {\n\treturn implJudgePlayServerCallSend{s}\n}", "func Open(oauthClient *oauth.Client, accessToken *oauth.Credentials, urlStr string, params url.Values) (*Stream, error) {\n\tts := new(Stream)\n\n\tu, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\taddr := u.Host\n\tif strings.LastIndex(addr, \":\") <= strings.LastIndex(addr, \"]\") {\n\t\tif u.Scheme == \"http\" {\n\t\t\taddr = addr + \":80\"\n\t\t} else {\n\t\t\taddr = addr + \":443\"\n\t\t}\n\t}\n\n\tif u.Scheme == \"http\" {\n\t\tts.conn, err = net.Dial(\"tcp\", addr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tts.conn, err = tls.Dial(\"tcp\", addr, nil)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif err = ts.conn.(*tls.Conn).VerifyHostname(addr[:strings.LastIndex(addr, \":\")]); err != nil {\n\t\t\treturn nil, ts.fatal(err)\n\t\t}\n\t}\n\n\t// Setup request body.\n\tpcopy := url.Values{}\n\tfor key, values := range params {\n\t\tpcopy[key] = values\n\t}\n\toauthClient.SignParam(accessToken, \"POST\", urlStr, pcopy)\n\tbody := pcopy.Encode()\n\n\tvar req bytes.Buffer\n\treq.WriteString(\"POST \")\n\treq.WriteString(u.RequestURI())\n\treq.WriteString(\" HTTP/1.1\")\n\treq.WriteString(\"\\r\\nHost: \")\n\treq.WriteString(u.Host)\n\treq.WriteString(\"\\r\\nContent-Type: application/x-www-form-urlencoded\")\n\treq.WriteString(\"\\r\\nContent-Length: \")\n\treq.WriteString(strconv.Itoa(len(body)))\n\treq.WriteString(\"\\r\\n\\r\\n\")\n\treq.WriteString(body)\n\t_, err = ts.conn.Write(req.Bytes())\n\tif err != nil {\n\t\treturn nil, ts.fatal(err)\n\t}\n\n\t// Must connect in 60 seconds.\n\terr = ts.conn.SetReadDeadline(time.Now().Add(60 * time.Second))\n\tif err != nil {\n\t\treturn nil, ts.fatal(err)\n\t}\n\n\tts.r = bufio.NewReaderSize(ts.conn, 8192)\n\tp, err := ts.r.ReadSlice('\\n')\n\tif err != nil {\n\t\treturn nil, ts.fatal(err)\n\t}\n\n\tm := responseLineRegexp.FindSubmatch(p)\n\tif m == nil {\n\t\treturn nil, ts.fatal(errors.New(\"twitterstream: bad http response line\"))\n\t}\n\n\t// Skip headers\n\tfor {\n\t\tp, err = ts.r.ReadSlice('\\n')\n\t\tif err != nil {\n\t\t\treturn nil, ts.fatal(err)\n\t\t}\n\t\tif len(p) <= 2 {\n\t\t\tbreak\n\t\t}\n\t}\n\n\tstatusCode, _ := strconv.Atoi(string(m[1]))\n\tif statusCode != 200 {\n\t\tp, _ := ioutil.ReadAll(ts.r)\n\t\treturn nil, HTTPStatusError{statusCode, string(p)}\n\t}\n\n\tts.chunkState = stateStart\n\treturn ts, nil\n}", "func (s *Session) OpenWriteStream(sType SessionType) (WriteStream, error) {\n\tswitch sType {\n\tcase SessionWriteCapture:\n\t\ts.wp.Add()\n\t\tparStream := newSessionStream(s, s.dbo, s.schema, s.wp)\n\t\tws, err := newWriteCapStream(parStream, s.cancel)\n\t\tif err != nil {\n\t\t\ts.wp.Done()\n\t\t}\n\t\treturn ws, err\n\tcase SessionWriteEntity:\n\t\tparStream := newSessionStream(s, s.dbo, s.schema, s.wp)\n\t\tws, err := newWriteEntityStream(parStream, s.cancel)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ts.wp.Add()\n\t\treturn ws, err\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unsupported write stream type\")\n\t}\n}", "func (conn *Conn) Open(port int) (stream net.Conn, err error) {\n\tstream, err = conn.sess.OpenStream()\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to open smux stream\")\n\t}\n\n\tvar portData [8]byte\n\tbinary.BigEndian.PutUint64(portData[:], uint64(port))\n\n\t_, err = stream.Write(portData[:])\n\tif err != nil {\n\t\tstream.Close()\n\t\treturn nil, errors.Wrapf(err, \"failed to write port to smux stream\")\n\t}\n\n\tlog.WithField(\"peer\", conn.peerPublicKey).\n\t\tWithField(\"port\", port).\n\t\tInfo(\"opened connection\")\n\n\treturn stream, nil\n}", "func (sc *SoundCloud) Stream(track string) (io.ReadCloser, error) {\n\t// Get the HTTP Stream\n\trsp, err := http.Get(sc.streamUrl(track).String())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Createa http stream buffer\n\tbuff := buffer.HTTPBuffer(rsp)\n\tgo buff.Buffer() // Start buffering\n\tscs := &SoundCloudStream{\n\t\tbuffer: buff,\n\t\tdecoder: &mpa.Reader{Decoder: &mpa.Decoder{Input: buff}},\n\t}\n\treturn scs, nil\n}", "func (c *webSocketWillSendHandshakeRequestClient) GetStream() rpcc.Stream { return c.Stream }", "func (s *SessionSRTCP) OpenWriteStream() (*WriteStream, error) {\n\treturn s.writeStream, nil\n}", "func (pc *PeerConnection) Stream() error {\n\t// Wait for ICE agent to establish a connection.\n\ttimeoutCtx, _ := context.WithTimeout(pc.ctx, connectTimeout)\n\tdataStream, err := pc.iceAgent.GetDataStream(timeoutCtx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer dataStream.Close()\n\n\t// Instantiate a new net.Conn multiplexer\n\tdataMux := mux.NewMux(dataStream, 8192)\n\tdefer dataMux.Close()\n\n\t// Instantiate a new endpoint for DTLS from multiplexer\n\tdtlsEndpoint := dataMux.NewEndpoint(mux.MatchDTLS)\n\n\t// Instantiate a new endpoint for SRTP from multiplexer\n\tsrtpEndpoint := dataMux.NewEndpoint(func(b []byte) bool {\n\t\t// First byte looks like 10??????, representing RTP version 2.\n\t\treturn b[0]&0xb0 == 0x80\n\t})\n\n\t// Configuration for DTLS handshake, namely certificate and private key\n\tconfig := &dtls.Config{Certificate: pc.certificate, PrivateKey: pc.privateKey}\n\n\t// Initiate a DTLS handshake as a client\n\tdtlsConn, err := dtls.Client(dtlsEndpoint, config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Create SRTP keys from DTLS handshake (see RFC5764 Section 4.2)\n\tkeys, err := dtlsConn.ExportKeyingMaterial(\"EXTRACTOR-dtls_srtp\", nil, 2*keyLen+2*saltLen)\n\tif err != nil {\n\t\treturn err\n\t}\n\tkeyReader := bytes.NewBuffer(keys)\n\twriteKey := keyReader.Next(keyLen)\n\treadKey := keyReader.Next(keyLen)\n\twriteSalt := keyReader.Next(saltLen)\n\treadSalt := keyReader.Next(saltLen)\n\n\trtpSession := rtp.NewSession(rtp.SessionOptions{\n\t\tMuxConn: srtpEndpoint, // rtcp-mux assumed\n\t\tReadKey: readKey,\n\t\tReadSalt: readSalt,\n\t\tWriteKey: writeKey,\n\t\tWriteSalt: writeSalt,\n\t})\n\n\tvideoStreamOpts := rtp.StreamOptions{\n\t\tDirection: \"sendonly\",\n\t}\n\tfor _, m := range pc.localDescription.Media {\n\t\tif m.Type == \"video\" {\n\t\t\tfmt.Sscanf(m.GetAttr(\"ssrc\"), \"%d cname:%s\", &videoStreamOpts.LocalSSRC, &videoStreamOpts.LocalCNAME)\n\t\t\tbreak\n\t\t}\n\t}\n\tfor _, m := range pc.remoteDescription.Media {\n\t\tif m.Type == \"video\" {\n\t\t\tfmt.Sscanf(m.GetAttr(\"ssrc\"), \"%d cname:%s\", &videoStreamOpts.RemoteSSRC, &videoStreamOpts.RemoteCNAME)\n\t\t\tbreak\n\t\t}\n\t}\n\n\tvideoStream := rtpSession.AddStream(videoStreamOpts)\n\tgo videoStream.SendVideo(pc.ctx.Done(), pc.DynamicType, pc.localVideo)\n\n\t//rtpSession, err := rtp.NewSecureSession(rtpEndpoint, readKey, readSalt, writeKey, writeSalt)\n\t//go streamH264(pc.ctx, pc.localVideoTrack, rtpSession.NewH264Stream(ssrc, cname))\n\n\t// Start goroutine for processing incoming SRTCP packets\n\t//go srtcpReaderRunloop(dataMux, readKey, readSalt)\n\n\t// Begin a new SRTP session\n\t//srtpSession, err := srtp.NewSession(srtpEndpoint, pc.DynamicType, writeKey, writeSalt)\n\t//if err != nil {\n\t//\treturn err\n\t//}\n\n\t// There are two termination conditions that we need to deal with here:\n\t// 1. Context cancellation. If Close() is called explicitly, or if the\n\t// parent context is canceled, we should terminate cleanly.\n\t// 2. Connection timeout. If the remote peer disconnects unexpectedly, the\n\t// read loop on the underlying net.UDPConn will time out. The associated\n\t// ice.DataStream will then be marked dead, which we check for here.\n\tselect {\n\tcase <-pc.ctx.Done():\n\t\treturn nil\n\tcase <-dataStream.Done():\n\t\treturn dataStream.Err()\n\t}\n}", "func (s *MessengerDiffServerCallStub) SendStream() interface {\n\tSend(item []bool) error\n} {\n\treturn implMessengerDiffServerCallSend{s}\n}", "func NewSession(peerID string, stream MultiplexStream, bufferSize uint32, sendTimeout time.Duration) Session {\n\ts := &session{\n\t\tpeerID: peerID,\n\t\terr: make(chan error),\n\t\tserverStream: &kdsServerStream{\n\t\t\tctx: stream.Context(),\n\t\t\tbufferStream: newBufferStream(bufferSize),\n\t\t},\n\t\tclientStream: &kdsClientStream{\n\t\t\tctx: stream.Context(),\n\t\t\tbufferStream: newBufferStream(bufferSize),\n\t\t},\n\t}\n\tgo func() {\n\t\ts.handleSend(stream, sendTimeout)\n\t}()\n\tgo func() {\n\t\ts.handleRecv(stream)\n\t}()\n\treturn s\n}", "func newStream(session *Session, id uint32, state streamState) *Stream {\n\ts := &Stream{\n\t\tid: id,\n\t\tsession: session,\n\t\tstate: state,\n\t\tcontrolHdr: header(make([]byte, headerSize)),\n\t\tcontrolErr: make(chan error, 1),\n\t\tsendHdr: header(make([]byte, headerSize)),\n\t\tsendErr: make(chan error, 1),\n\t\trecvWindow: initialStreamWindow,\n\t\tsendWindow: initialStreamWindow,\n\t\trecvNotifyCh: make(chan struct{}, 1),\n\t\tsendNotifyCh: make(chan struct{}, 1),\n\t\testablishCh: make(chan struct{}, 1),\n\t}\n\ts.readDeadline.Store(time.Time{})\n\ts.writeDeadline.Store(time.Time{})\n\treturn s\n}", "func (s *SessionSRTP) OpenWriteStream() (*WriteStream, error) {\n\treturn s.writeStream, nil\n}", "func (c *webSocketCreatedClient) GetStream() rpcc.Stream { return c.Stream }", "func (s *Session) OpenStream() (*Stream, error) {\n\tif s.IsClosed() {\n\t\treturn nil, ErrSessionShutdown\n\t}\n\nGET_ID:\n\t// Get an ID, and check for stream exhaustion\n\tid := atomic.LoadUint32(&s.nextStreamID)\n\tif id >= math.MaxUint32-1 {\n\t\treturn nil, ErrStreamsExhausted\n\t}\n\tif !atomic.CompareAndSwapUint32(&s.nextStreamID, id, id+2) {\n\t\tgoto GET_ID\n\t}\n\n\t// Register the stream\n\tstream := newStream(s, id)\n\ts.streams.Store(id, stream)\n\tatomic.AddInt32(&s.streamsCounter, 1)\n\tvar timeout <-chan time.Time\n\terr := s.writeFrame(newLenFrame(flagSYN, id, 0, nil), timeout)\n\tif nil != err {\n\t\treturn nil, err\n\t}\n\treturn stream, nil\n}", "func (s *session) OpenStream() (Stream, error) {\n\treturn s.streamsMap.OpenStream()\n}", "func (s *Sync) Stream() <-chan Execer {\n\treturn s.s.ExecStream(s.a.Events())\n}", "func (c *webSocketFrameSentClient) GetStream() rpcc.Stream { return c.Stream }", "func (s *Stream) Session() *Session {\n\treturn s.session\n}", "func (s *Session) Stream() error {\n\t// In parallel read from client, send to broker\n\t// and read from broker, send to client.\n\terrs := make(chan error, 2)\n\n\tgo s.stream(up, s.inbound, s.outbound, errs)\n\tgo s.stream(down, s.outbound, s.inbound, errs)\n\n\t// Handle whichever error happens first.\n\t// The other routine won't be blocked when writing\n\t// to the errors channel because it is buffered.\n\terr := <-errs\n\n\ts.handler.Disconnect(&s.Client)\n\treturn err\n}", "func (m *Market) Stream(timerCh chan<- cq.TimerMsg) error {\n\terr := connectWS(m, timerCh)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Client) OpenStream(name string) (*Stream, error) {\n\tid := c.defines.allocate(name)\n\tif id.isNumeric() {\n\t\tif err := c.define(name, uint8(id.numericID())); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\ts := &Stream{\n\t\tclient: c,\n\t\tstreamID: id,\n\t}\n\treturn s, nil\n}", "func (c *Client) StartSession(ctx context.Context) (*Session, error) {\n\tstreamClient, err := c.client.Update(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Session{\n\t\tclosed: make(chan struct{}),\n\t\tstreamClient: streamClient,\n\t}, nil\n}", "func New(conn io.ReadWriteCloser, client bool, config *Config) (*Session, error) {\n\tif config == nil {\n\t\tconfig = DefaultConfig()\n\t} else if err := config.Verify(); err != nil {\n\t\treturn nil, err\n\t}\n\ts := Session{\n\t\tdone: make(chan struct{}),\n\t\tconfig: config,\n\t\tconn: conn,\n\t\tstreams: make(map[uint32]*Stream),\n\t}\n\tif client {\n\t\ts.nextID = 1\n\t\ts.client = 1\n\t} else {\n\t\ts.nextID = 2\n\t}\n\ts.accept = make(chan *Stream, s.config.AcceptBacklog)\n\ts.opened = make(chan struct{}, s.config.AcceptBacklog)\n\t// Set up i/o\n\t{\n\t\tw := timeoutWriter(conn, &s)\n\t\ts.send = bufio.NewWriterSize(w, s.config.SendBufferSize)\n\t}\n\t{\n\t\tr := timeoutReader(conn, &s)\n\t\ts.recv = bufio.NewReaderSize(r, s.config.RecvBufferSize)\n\t}\n\tgo func() {\n\t\tif err := s.recvLoop(); err != nil {\n\t\t\ts.closeWithError(err)\n\t\t}\n\t}()\n\tif s.config.KeepAliveEnabled {\n\t\tgo s.keepAlive()\n\t}\n\treturn &s, nil\n}", "func (e *streamExecutor) Stream(stdin io.Reader, stdout, stderr io.Writer, tty bool) error {\n\tsupportedProtocols := []string{StreamProtocolV2Name, StreamProtocolV1Name}\n\tconn, protocol, err := e.Dial(supportedProtocols...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer conn.Close()\n\n\tvar streamer streamProtocolHandler\n\n\tswitch protocol {\n\tcase StreamProtocolV2Name:\n\t\tstreamer = &streamProtocolV2{\n\t\t\tstdin: stdin,\n\t\t\tstdout: stdout,\n\t\t\tstderr: stderr,\n\t\t\ttty: tty,\n\t\t}\n\tcase \"\":\n\t\tglog.V(4).Infof(\"The server did not negotiate a streaming protocol version. Falling back to %s\", StreamProtocolV1Name)\n\t\tfallthrough\n\tcase StreamProtocolV1Name:\n\t\tstreamer = &streamProtocolV1{\n\t\t\tstdin: stdin,\n\t\t\tstdout: stdout,\n\t\t\tstderr: stderr,\n\t\t\ttty: tty,\n\t\t}\n\t}\n\n\treturn streamer.stream(conn)\n}", "func (fss *StreamingService) Stream(wg *sync.WaitGroup) error { return nil }", "func (c *webTransportConnectionEstablishedClient) GetStream() rpcc.Stream { return c.Stream }", "func (s *MessageRepositoryPullServerCallStub) SendStream() interface {\n\tSend(item []byte) error\n} {\n\treturn implMessageRepositoryPullServerCallSend{s}\n}", "func (bs *BlockStream) Session(ctx context.Context, token access.Token, autosave bool, peers ...peer.ID) (*Session, error) {\n\tses := newSession(ctx)\n\n\tfor _, p := range peers {\n\t\ts, err := bs.Host.NewStream(ctx, p, Protocol)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\terr = giveHand(s, token)\n\t\tif err != nil {\n\t\t\ts.Reset()\n\t\t\treturn nil, err\n\t\t}\n\n\t\tses.addProvider(s, func(f func() error) {\n\t\t\tbs.wg.Add(1)\n\t\t\tdefer bs.wg.Done()\n\n\t\t\tif err := f(); err != nil {\n\t\t\t\tlog.Error(err)\n\t\t\t\ts.Reset()\n\t\t\t\tses.removeProvider()\n\n\t\t\t\tif ses.getProviders() == 0 {\n\t\t\t\t\tlog.Error(\"Closing session: \", ErrStreamsReset)\n\t\t\t\t\tses.err = ErrStreamsReset\n\t\t\t\t\tses.cancel()\n\t\t\t\t}\n\t\t\t}\n\t\t})\n\t}\n\n\treturn ses, nil\n}", "func (e *GetSmscd) Stream(ctx context.Context, req *GetSmsCD.StreamingRequest, stream GetSmsCD.GetSmscd_StreamStream) error {\n\tlog.Logf(\"Received GetSmscd.Stream request with count: %d\", req.Count)\n\n\tfor i := 0; i < int(req.Count); i++ {\n\t\tlog.Logf(\"Responding: %d\", i)\n\t\tif err := stream.Send(&GetSmsCD.StreamingResponse{\n\t\t\tCount: int64(i),\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func newStream(id uint32, frameSize int, sess *Session) *Stream {\n\ts := new(Stream)\n\ts.id = id\n\ts.chReadEvent = make(chan struct{}, 1)\n\ts.frameSize = frameSize\n\ts.sess = sess\n\ts.die = make(chan struct{})\n\treturn s\n}", "func NewStream(ctx context.Context, method string, conn *Conn) (Stream, error) {\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\n\treturn newStreamClient(ctx, method, conn)\n}", "func (sc *serveconn) StreamRequest(cmd Cmd, flags FrameFlag, payload []byte) (StreamWriter, Response, error) {\n\n\tflags = flags.ToStream()\n\t_, resp, writer, err := sc.writeFirstFrame(cmd, flags, payload)\n\tif err != nil {\n\t\tl.Error(\"StreamRequest writeFirstFrame\", zap.Error(err))\n\t\treturn nil, nil, err\n\t}\n\n\twriter.checkExist = true\n\treturn (*defaultStreamWriter)(writer), resp, nil\n}", "func (stream *MAMWriteStream) Open() (trinary.Trytes, error) {\n\tchannelID, err := stream.m.ChannelCreate(5)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tstream.currentChannelID = channelID\n\treturn channelID, nil\n}", "func (c *webTransportCreatedClient) GetStream() rpcc.Stream { return c.Stream }", "func (v Vehicle) Stream() (chan *StreamEvent, chan error, error) {\n\turl := StreamURL + \"/stream/\" + strconv.Itoa(v.VehicleID) + \"/?values=\" + StreamParams\n\treq, _ := http.NewRequest(\"GET\", url, nil)\n\treq.SetBasicAuth(ActiveClient.Auth.Email, v.Tokens[0])\n\tresp, err := ActiveClient.HTTP.Do(req)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\teventChan := make(chan *StreamEvent)\n\terrChan := make(chan error)\n\tgo readStream(resp, eventChan, errChan)\n\n\treturn eventChan, errChan, nil\n}", "func (st *state) connectStream(path string, attrs url.Values, extraHeaders http.Header) (base.Stream, error) {\n\ttarget := url.URL{\n\t\tScheme: \"wss\",\n\t\tHost: st.addr,\n\t\tPath: path,\n\t\tRawQuery: attrs.Encode(),\n\t}\n\t// TODO(macgreagoir) IPv6. Ubuntu still always provides IPv4 loopback,\n\t// and when/if this changes localhost should resolve to IPv6 loopback\n\t// in any case (lp:1644009). Review.\n\tcfg, err := websocket.NewConfig(target.String(), \"http://localhost/\")\n\t// Add any cookies because they will not be sent to websocket\n\t// connections by default.\n\tfor header, values := range extraHeaders {\n\t\tfor _, value := range values {\n\t\t\tcfg.Header.Add(header, value)\n\t\t}\n\t}\n\n\tconnection, err := websocketDialConfig(cfg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := readInitialStreamError(connection); err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\treturn connection, nil\n}", "func Send(name string, options SendOptions) (io.ReadCloser, error) {\n\tcmd := &Cmd{}\n\n\tr, w, err := os.Pipe()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\toptions.Fd = int32(w.Fd())\n\n\tstream := sendStream{\n\t\terrorChan: make(chan error, 1),\n\t\tr: r,\n\t}\n\n\tgo func() {\n\t\terr := NvlistIoctl(zfsHandle.Fd(), ZFS_IOC_SEND_NEW, name, cmd, options, &struct{}{}, nil)\n\t\tstream.errorChan <- err\n\t\tw.Close()\n\t}()\n\n\tbuf := make([]byte, 1) // We want at least 1 byte of output to enter streaming mode\n\n\t_, err = stream.peek(buf)\n\tif err != nil {\n\t\tr.Close()\n\t\tw.Close()\n\t\treturn nil, err\n\t}\n\n\treturn &stream, nil\n}", "func (s *MessageRepositoryManifestServerCallStub) SendStream() interface {\n\tSend(item Message) error\n} {\n\treturn implMessageRepositoryManifestServerCallSend{s}\n}", "func (ws *WrappedStream) Stream() net.Stream {\n\treturn ws.stream\n}", "func (c *webSocketClosedClient) GetStream() rpcc.Stream { return c.Stream }", "func (hs *Handshake) OpenedStream(s p2p.Stream) {\n\n}", "func (s *Conn) Stream(id uint64) (*Stream, error) {\n\treturn s.getOrCreateStream(id, true)\n}", "func newMockStream() *mockStream {\n\t// just need a new context specific to this stream,\n\t// easiest way to get it.\n\tctx, _ := context.WithCancel(context.TODO())\n\n\treturn &mockStream{\n\t\tsendChan: make(chan *Response, 100),\n\t\trecvChan: make(chan *Request, 100),\n\t\tctx: ctx,\n\t}\n}", "func newStream(id common.StreamId, hostStr string, handler MutationHandler) (*Stream, error) {\n\n\t// TODO: use constant\n\tmutch := make(chan interface{}, 1000)\n\tstopch := make(chan bool)\n\n\ts := &Stream{id: id,\n\t\thostStr: hostStr,\n\t\thandler: handler,\n\t\tmutch: mutch,\n\t\tstopch: stopch,\n\t\tstatus: false}\n\n\treturn s, nil\n}", "func NewStream(\n\tURI string,\n\tstoringDirectory string,\n\tkeepFiles bool,\n\taudio bool,\n\tloggingOpts ProcessLoggingOpts,\n\twaitTimeOut time.Duration,\n) (*Stream, string) {\n\tid := uuid.New().String()\n\tpath := fmt.Sprintf(\"%s/%s\", storingDirectory, id)\n\terr := os.MkdirAll(path, os.ModePerm)\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\treturn nil, \"\"\n\t}\n\tprocess := NewProcess(keepFiles, audio)\n\tcmd := process.Spawn(path, URI)\n\n\t// Create nil pointer in case logging is not enabled\n\tcmdLogger := (*lumberjack.Logger)(nil)\n\t// Create logger otherwise\n\tif loggingOpts.Enabled {\n\t\tcmdLogger = &lumberjack.Logger{\n\t\t\tFilename: fmt.Sprintf(\"%s/%s.log\", loggingOpts.Directory, id),\n\t\t\tMaxSize: loggingOpts.MaxSize,\n\t\t\tMaxBackups: loggingOpts.MaxBackups,\n\t\t\tMaxAge: loggingOpts.MaxAge,\n\t\t\tCompress: loggingOpts.Compress,\n\t\t}\n\t\tcmd.Stderr = cmdLogger\n\t\tcmd.Stdout = cmdLogger\n\t}\n\tstream := Stream{\n\t\tID: id,\n\t\tCMD: cmd,\n\t\tProcess: process,\n\t\tMux: &sync.Mutex{},\n\t\tPath: fmt.Sprintf(\"/%s/index.m3u8\", filepath.Join(\"stream\", id)),\n\t\tStorePath: path,\n\t\tStreak: hotstreak.New(hotstreak.Config{\n\t\t\tLimit: 10,\n\t\t\tHotWait: time.Minute * 2,\n\t\t\tActiveWait: time.Minute * 4,\n\t\t}).Activate(),\n\t\tOriginalURI: URI,\n\t\tKeepFiles: keepFiles,\n\t\tLoggingOpts: &loggingOpts,\n\t\tLogger: cmdLogger,\n\t\tRunning: false,\n\t\tWaitTimeOut: waitTimeOut,\n\t}\n\tlogrus.Debugf(\"%s store path created | Stream\", stream.StorePath)\n\treturn &stream, id\n}", "func (c *Client) Stream(t *soundcloud.Track) (*StreamOutput, error) {\n\tvar (\n\t\terr error\n\t\tresp *http.Response\n\t\treq *http.Request\n\t\tout StreamOutput\n\t)\n\n\tif req, err = sling.New().\n\t\tGet(fmt.Sprintf(\"https://api.soundcloud.com/tracks/%d/streams\", t.Id)).\n\t\tQueryStruct(StreamParam{ClientID: c.clientID}).\n\t\tRequest(); err != nil {\n\t\treturn nil, fmt.Errorf(\"create query: %w\", err)\n\t}\n\n\tif resp, err = c.hc.Do(req); err != nil {\n\t\treturn nil, fmt.Errorf(\"execute query: %w\", err)\n\t}\n\tdefer resp.Body.Close()\n\n\to, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"read body: %w\", err)\n\t}\n\n\tif err = json.Unmarshal(o, &out); err != nil {\n\t\treturn nil, fmt.Errorf(\"decode body: %w\", err)\n\t}\n\n\treturn &out, nil\n}", "func (d *Discord) Open() (<-chan Message, error) {\n\tshards := d.Shards\n\tif shards < 1 {\n\t\tshards = 1\n\t}\n\n\td.Sessions = make([]*discordgo.Session, shards)\n\n\tfor i := 0; i < shards; i++ {\n\t\tsession, err := discordgo.New(d.args...)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tsession.State.TrackPresences = false\n\t\tsession.ShardCount = shards\n\t\tsession.ShardID = i\n\t\tsession.AddHandler(d.onMessageCreate)\n\t\tsession.AddHandler(d.onMessageUpdate)\n\t\tsession.AddHandler(d.onMessageDelete)\n\n\t\td.Sessions[i] = session\n\t}\n\n\td.Session = d.Sessions[0]\n\n\tfor i := 0; i < len(d.Sessions); i++ {\n\t\td.Sessions[i].Open()\n\t}\n\n\treturn d.messageChan, nil\n}", "func (e *GetSmsCd) Stream(ctx context.Context, req *pb.StreamingRequest, stream pb.GetSmsCd_StreamStream) error {\n\tlog.Infof(\"Received GetSmsCd.Stream request with count: %d\", req.Count)\n\n\tfor i := 0; i < int(req.Count); i++ {\n\t\tlog.Infof(\"Responding: %d\", i)\n\t\tif err := stream.Send(&pb.StreamingResponse{\n\t\t\tCount: int64(i),\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (d *Discord) Open() (<-chan Message, error) {\n\tgateway, err := discordgo.New(d.args[0].(string))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ts, err := gateway.GatewayBot()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\td.Sessions = make([]*discordgo.Session, s.Shards)\n\n\tlog.Printf(\"%s opening with %d shards\\n\", d.Name(), s.Shards)\n\twg := sync.WaitGroup{}\n\tfor i := 0; i < s.Shards; i++ {\n\t\tlog.Printf(\"%s opening shard %d\\n\", d.Name(), i+1)\n\t\tsession, err := discordgo.New(d.args[0].(string))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif d.Session == nil {\n\t\t\td.Session = session\n\t\t}\n\t\td.Sessions[i] = session\n\t\tsession.ShardCount = s.Shards\n\t\tsession.ShardID = i\n\t\tsession.State.TrackPresences = false\n\t\twg.Add(1)\n\t\tgo func(session *discordgo.Session) {\n\t\t\tdefer wg.Done()\n\t\t\terr := session.Open()\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"error opening shard %s\", err)\n\t\t\t}\n\t\t}(d.Sessions[i])\n\t}\n\twg.Wait()\n\n\tfor _, session := range d.Sessions {\n\t\tsession.AddHandler(d.onMessageCreate)\n\t\tsession.AddHandler(d.onMessageUpdate)\n\t\tsession.AddHandler(d.onMessageDelete)\n\t}\n\n\treturn d.messageChan, nil\n}", "func GetStream(w http.ResponseWriter, r *http.Request) {\n\t// Retrieve render\n\tren := context.Get(r, CtxRender).(*render.Render)\n\n\t// Advertise that clients may send Range requests\n\tw.Header().Set(\"Accept-Ranges\", \"bytes\")\n\n\t// Check API version\n\tif version, ok := mux.Vars(r)[\"version\"]; ok {\n\t\t// Check if this API call is supported in the advertised version\n\t\tif !apiVersionSet.Has(version) {\n\t\t\tren.JSON(w, 400, errRes(400, \"unsupported API version: \"+version))\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Check for an ID parameter\n\tpID, ok := mux.Vars(r)[\"id\"]\n\tif !ok {\n\t\tren.JSON(w, 400, errRes(400, \"no integer stream ID provided\"))\n\t\treturn\n\t}\n\n\t// Verify valid integer ID\n\tid, err := strconv.Atoi(pID)\n\tif err != nil {\n\t\tren.JSON(w, 400, errRes(400, \"invalid integer stream ID\"))\n\t\treturn\n\t}\n\n\t// Attempt to load the song with matching ID\n\tsong := &data.Song{ID: id}\n\tif err := song.Load(); err != nil {\n\t\t// Check for invalid ID\n\t\tif err == sql.ErrNoRows {\n\t\t\tren.JSON(w, 404, errRes(404, \"song ID not found\"))\n\t\t\treturn\n\t\t}\n\n\t\t// All other errors\n\t\tlog.Println(err)\n\t\tren.JSON(w, 500, serverErr)\n\t\treturn\n\t}\n\n\t// Attempt to access data stream\n\tstream, err := song.Stream()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\tren.JSON(w, 500, serverErr)\n\t\treturn\n\t}\n\n\t// Generate a string used for logging this operation\n\topStr := fmt.Sprintf(\"[#%05d] %s - %s [%s %dkbps]\", song.ID, song.Artist, song.Title,\n\t\tdata.CodecMap[song.FileTypeID], song.Bitrate)\n\n\t// Attempt to send file stream over HTTP\n\tlog.Println(\"stream: starting:\", opStr)\n\n\t// Pass stream using song's file size, auto-detect MIME type\n\tif err := HTTPStream(song, \"\", song.FileSize, stream, r, w); err != nil {\n\t\t// Check for client reset\n\t\tif strings.Contains(err.Error(), \"connection reset by peer\") || strings.Contains(err.Error(), \"broken pipe\") {\n\t\t\treturn\n\t\t}\n\n\t\t// Check for invalid range, return HTTP 416\n\t\tif err == ErrInvalidRange {\n\t\t\tren.JSON(w, 416, errRes(416, \"invalid HTTP Range header boundaries\"))\n\t\t\treturn\n\t\t}\n\n\t\tlog.Println(\"stream: error:\", err)\n\t\treturn\n\t}\n\n\tlog.Println(\"stream: completed:\", opStr)\n\treturn\n}", "func (s *Transport) Session (user string, options...Option) (*ssh.Session, error) {\n if len(options) == 0 {\n return nil, fmt.Errorf(\"Session must be supplied an authentication option\")\n }\n\n if v, ok := s.session(user); ok {\n return v, nil\n }\n\n opts := &clientOptions{}\n for _, opt := range options {\n opt(opts)\n }\n\n var auth []ssh.AuthMethod\n\n if opts.key == nil {\n switch \"\" {\n case user, opts.pass:\n return nil, fmt.Errorf(\"no valid authentication methods were provided\")\n }\n auth = []ssh.AuthMethod{\n ssh.Password(opts.pass),\n ssh.KeyboardInteractive(func(user, instruction string, questions []string, echos []bool) ([]string, error) {\n // Just send the password back for all questions\n answers := make([]string, len(questions))\n for i, _ := range answers {\n answers[i] = opts.pass\n }\n return answers, nil\n }),\n }\n }else{\n auth = []ssh.AuthMethod{ssh.PublicKeys(opts.key)}\n }\n\n client, err := s.connect(user, &ssh.ClientConfig{User: user, Auth: auth})\n if err != nil {\n return nil, fmt.Errorf(\"problem making initial connection to host %q: %s\", s.device, err)\n }\n\n session, err := client.NewSession()\n if err != nil {\n return nil, fmt.Errorf(\"problem creating session to %q: %s\", s.device, err)\n }\n\n s.mu.Lock()\n defer s.mu.Unlock()\n s.sessions[user] = session\n\n return session, nil\n}", "func (c Client) NewStream(model string, content_type string, options map[string]interface{}) (<-chan Event, io.WriteCloser, error) {\n\ttoken, err := authorization.GetToken(c.watsonClient.Creds)\n\tif err != nil {\n\t\treturn nil, nil, errors.New(\"failed to acquire auth token: \" + err.Error())\n\t}\n\tu, err := url.Parse(c.watsonClient.Creds.Url)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tu.Scheme = \"wss\"\n\tq := url.Values{}\n\tq.Set(\"watson-token\", token)\n\tif len(model) > 0 {\n\t\tq.Set(\"model\", model)\n\t}\n\tu.RawQuery = q.Encode()\n\tu.Path += c.version + \"/recognize\"\n\n\torigin, err := url.Parse(c.watsonClient.Creds.Url)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tconfig := &websocket.Config{\n\t\tLocation: u,\n\t\tOrigin: origin,\n\t\tVersion: websocket.ProtocolVersionHybi13,\n\t}\n\tws, err := websocket.DialConfig(config)\n\tif err != nil {\n\t\treturn nil, nil, errors.New(\"error dialing websocket: \" + err.Error())\n\t}\n\toutput := make(chan Event, 100)\n\ts := stream{\n\t\tinput: output,\n\t\tcontentType: content_type,\n\t\tws: ws,\n\t\toptions: options,\n\t}\n\treturn output, &s, nil\n}", "func (e *MyTest) Stream(ctx context.Context, req *myTest.StreamingRequest, stream myTest.MyTest_StreamStream) error {\n\tlog.Logf(\"Received MyTest.Stream request with count: %d\", req.Count)\n\n\tfor i := 0; i < int(req.Count); i++ {\n\t\tlog.Logf(\"Responding: %d\", i)\n\t\tif err := stream.Send(&myTest.StreamingResponse{\n\t\t\tCount: int64(i),\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (s *Session) OpenReadStream(sType SessionType, fo FilterOptions) (ReadStream, error) {\n\tswitch sType {\n\tcase SessionReadCapture:\n\t\ts.wp.Add()\n\t\tparStream := newSessionStream(s, s.dbo, s.schema, s.wp)\n\t\trs, err := newReadCapStream(parStream, s.cancel, fo)\n\t\tif err != nil {\n\t\t\ts.wp.Done()\n\t\t}\n\t\treturn rs, nil\n\tcase SessionReadPrefix:\n\t\ts.wp.Add()\n\t\tparStream := newSessionStream(s, s.dbo, s.schema, s.wp)\n\t\trs, err := newReadPrefixStream(parStream, s.cancel, fo)\n\t\tif err != nil {\n\t\t\ts.wp.Done()\n\t\t}\n\t\treturn rs, nil\n\tcase SessionReadEntity:\n\t\ts.wp.Add()\n\t\tparStream := newSessionStream(s, s.dbo, s.schema, s.wp)\n\t\tes, err := newReadEntityStream(parStream, s.cancel, fo)\n\t\tif err != nil {\n\t\t\ts.wp.Done()\n\t\t}\n\t\treturn es, nil\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unsupported read stream type\")\n\t}\n}", "func (c *requestWillBeSentClient) GetStream() rpcc.Stream { return c.Stream }", "func (m *Module) Stream() <-chan bar.Output {\n\tch := base.NewChannel()\n\tgo m.worker(ch)\n\treturn ch\n}", "func (s replicatedSession) Open() error {\n\tif err := s.session.Open(); err != nil {\n\t\treturn err\n\t}\n\tfor _, asyncSession := range s.asyncSessions {\n\t\tif err := asyncSession.Open(); err != nil {\n\t\t\ts.log.Error(\"could not open session to async cluster: %v\", zap.Error(err))\n\t\t}\n\t}\n\treturn nil\n}", "func NewStream(r *http.Request, w http.ResponseWriter) (*Stream, error) {\n\tf, ok := w.(http.Flusher)\n\n\tif ok {\n\t\tw.Header().Set(\"Content-Type\", \"text/event-stream\")\n\t\tw.Header().Set(\"Cache-Control\", \"no-cache\")\n\t\tw.Header().Set(\"Connection\", \"keep-alive\")\n\t} else {\n\t\treturn nil, ErrBadStream\n\t}\n\n\treturn &Stream{\n\t\twriter: w,\n\t\tflusher: f,\n\t\trequest: r,\n\t}, nil\n}", "func NewStream(cfg StreamConfig) (Stream, error) {\n\tstream := Stream{}\n\tvar nc *nats.Conn\n\tvar err error\n\n\t// Connect Options.\n\topts := []nats.Option{nats.Name(cfg.Name)}\n\tif cfg.WaitInMinutes != 0 {\n\t\topts = appendWaitOpts(cfg, opts)\n\t}\n\n\t// Provide Authentication information\n\topts = append(opts, nats.UserInfo(cfg.User, cfg.Password))\n\n\t//Connect to NATS\n\tif cfg.URI != \"\" {\n\t\tnc, err = nats.Connect(cfg.URI, opts...)\n\t} else {\n\t\tnc, err = nats.Connect(nats.DefaultURL, opts...)\n\t}\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\n\tstream.Conn = nc\n\treturn stream, err\n}", "func (m *MockEarlyConnection) OpenStreamSync(arg0 context.Context) (quic.Stream, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"OpenStreamSync\", arg0)\n\tret0, _ := ret[0].(quic.Stream)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func NewStream() Stream {\n\treturn []Streamer{}\n}", "func (p *Provider) Stream(streamID int) (*models.Stream, error) {\n\tstreamChan := make(chan *models.Stream, 1)\n\tp.internalRequestChan <- streamRequest{\n\t\trespChan: streamChan,\n\t\tstreamID: streamID,\n\t}\n\tselect {\n\tcase resp := <-streamChan:\n\t\tif resp == nil {\n\t\t\treturn nil, fmt.Errorf(\"stream ID %d not found\", streamID)\n\t\t}\n\t\treturn resp, nil\n\tcase <-time.After(p.queryTimeout):\n\t\tp.logger.Error(\"Stream()\",\n\t\t\tzap.Error(ErrRequestTimedOut),\n\t\t\tzap.Int(\"streamID\", streamID),\n\t\t\tzap.Duration(\"timeout-duration\", p.queryTimeout),\n\t\t)\n\t\treturn nil, ErrRequestTimedOut\n\t}\n}", "func NewStream() *Stream {\n\treturn &Stream{\n\t\trecv: make(chan []byte),\n\t\tdone: make(chan struct{}),\n\t}\n}", "func (_LvRecordableStream *LvRecordableStreamTransactor) StartStream(opts *bind.TransactOpts, _handle string) (*types.Transaction, error) {\n\treturn _LvRecordableStream.contract.Transact(opts, \"startStream\", _handle)\n}", "func NewStreamClient(ctx context.Context, prot api.Protocol, connection types.ClientConnection, host types.Host) Client {\n\tclient := &client{\n\t\tProtocol: prot,\n\t\tConnection: connection,\n\t\tHost: host,\n\t}\n\n\tif factory, ok := streamFactories[prot]; ok {\n\t\tclient.ClientStreamConnection = factory.CreateClientStream(ctx, connection, client, client)\n\t} else {\n\t\treturn nil\n\t}\n\n\tconnection.AddConnectionEventListener(client)\n\tconnection.FilterManager().AddReadFilter(client)\n\tconnection.SetNoDelay(true)\n\n\treturn client\n}", "func (c *subContext) openStream(ctx context.Context, epID epapi.ID, indCh chan<- indication.Indication) error {\n\tresponse, err := c.epClient.Get(ctx, epID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconn, err := c.conns.Connect(fmt.Sprintf(\"%s:%d\", response.IP, response.Port))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclient := termination.NewClient(conn)\n\tresponseCh := make(chan e2tapi.StreamResponse)\n\trequestCh, err := client.Stream(ctx, responseCh)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trequestCh <- e2tapi.StreamRequest{\n\t\tAppID: e2tapi.AppID(c.config.AppID),\n\t\tInstanceID: e2tapi.InstanceID(c.config.InstanceID),\n\t\tSubscriptionID: e2tapi.SubscriptionID(c.sub.ID),\n\t}\n\n\tfor response := range responseCh {\n\t\tindCh <- indication.Indication{\n\t\t\tEncodingType: encoding.Type(response.Header.EncodingType),\n\t\t\tPayload: indication.Payload{\n\t\t\t\tHeader: response.IndicationHeader,\n\t\t\t\tMessage: response.IndicationMessage,\n\t\t\t},\n\t\t}\n\t}\n\treturn nil\n}", "func (c *webSocketHandshakeResponseReceivedClient) GetStream() rpcc.Stream { return c.Stream }", "func NewStream(id string, retriever Retriever, options ...*Option) Stream {\n\tmtrx := createMetrics(id, \"stream\")\n\topt := defaultOptions.merge(options...)\n\n\tedge := newEdge(opt.BufferSize)\n\tinput := newEdge(opt.BufferSize)\n\n\tx := &builder{\n\t\tvertex: vertex{\n\t\t\tid: id,\n\t\t\tvertexType: \"stream\",\n\t\t\tmetrics: mtrx,\n\t\t\tinput: input,\n\t\t\toption: opt,\n\t\t\thandler: func(p []*Packet) {\n\t\t\t\tedge.channel <- p\n\t\t\t},\n\t\t},\n\t\tvertacies: map[string]*vertex{},\n\t\trecorder: func(s1, s2, s3 string, p []*Packet) {},\n\t}\n\n\tx.connector = func(ctx context.Context, b *builder) error {\n\t\ti := retriever(ctx)\n\n\t\tgo func() {\n\t\tLoop:\n\t\t\tfor {\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\tbreak Loop\n\t\t\t\tcase data := <-i:\n\t\t\t\t\tif len(data) < 1 {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\n\t\t\t\t\tpayload := make([]*Packet, len(data))\n\t\t\t\t\tfor i, item := range data {\n\t\t\t\t\t\tpacket := &Packet{\n\t\t\t\t\t\t\tID: uuid.New().String(),\n\t\t\t\t\t\t\tData: item,\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif *x.option.Span {\n\t\t\t\t\t\t\tpacket.newSpan(ctx, mtrx.tracer, \"stream.begin\", id, \"stream\")\n\t\t\t\t\t\t}\n\t\t\t\t\t\tpayload[i] = packet\n\t\t\t\t\t}\n\n\t\t\t\t\tinput.channel <- payload\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t\treturn x.next.cascade(ctx, x, edge)\n\t}\n\n\treturn x\n}", "func NewStream(host core.Host, pid core.PeerID, protoIDs ...core.ProtocolID) (core.Stream, error) {\n\n\tstream, err := host.NewStream(context.Background(), pid, protoIDs...)\n\t// EOF表示底层连接断开, 增加一次重试\n\tif err == io.EOF {\n\t\tlog.Debug(\"NewStream\", \"msg\", \"RetryConnectEOF\")\n\t\tstream, err = host.NewStream(context.Background(), pid, protoIDs...)\n\t}\n\tif err != nil {\n\t\tlog.Error(\"NewStream\", \"pid\", pid.Pretty(), \"msgID\", protoIDs, \" err\", err)\n\t\treturn nil, err\n\t}\n\treturn stream, nil\n}", "func (s *server) Stream(stream pbgame.GameService_StreamServer) error {\n\tdefer utils.PrintPanicStack()\n\n\t// session init\n\tvar sess tp.Session\n\tsess_die := make(chan struct{})\n\tch_agent := s.recv(stream, sess_die)\n\tch_ipc := make(chan *pbgame.Game_Frame, DefaultChIPCSize)\n\n\tdefer func() {\n\t\tregistry.Unregister(sess.UserId)\n\t\tclose(sess_die)\n\t\tlog.Debug(\"stream end:\", sess.UserId)\n\t}()\n\n\t// read metadata from context\n\tmd, ok := metadata.FromContext(stream.Context())\n\tif !ok {\n\t\tlog.Error(\"cannot read metadata from context\")\n\t\treturn ErrorIncorrectFrameType\n\t}\n\t// read key\n\tif len(md[\"userid\"]) == 0 {\n\t\tlog.Error(\"cannot read key:userid from metadata\")\n\t\treturn ErrorIncorrectFrameType\n\t}\n\t// parse userid\n\tuserId, err := strconv.Atoi(md[\"userid\"][0])\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn ErrorIncorrectFrameType\n\t}\n\n\t// register user\n\tsess.UserId = int32(userId)\n\tregistry.Register(sess.UserId, ch_ipc)\n\tlog.Debug(\"userid:\", sess.UserId)\n\n\t// >> main message loop <<\n\tfor {\n\t\tselect {\n\t\tcase frame, ok := <-ch_agent: // frames from agent\n\t\t\tif !ok { // EOF\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tswitch frame.Type {\n\t\t\tcase pbgame.Game_Message: // the passthrough message from client->agent->game\n\t\t\t\t// locate handler by proto number\n\t\t\t\treader := packet.Reader(frame.Message)\n\t\t\t\tc, err := reader.ReadS16()\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(err)\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\th := handler.Handlers[c]\n\t\t\t\tif h == nil {\n\t\t\t\t\tlog.Error(\"service not bind:\", c)\n\t\t\t\t\treturn ErrorServiceNotBind\n\t\t\t\t}\n\n\t\t\t\tlog.Debug(\"msgid:\", c)\n\t\t\t\t// handle request\n\t\t\t\tret := h(&sess, reader)\n\n\t\t\t\t// construct frame & return message from logic\n\t\t\t\tif ret != nil {\n\t\t\t\t\tif err := stream.Send(&pbgame.Game_Frame{Type: pbgame.Game_Message, Message: ret}); err != nil {\n\t\t\t\t\t\tlog.Error(err)\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// session control by logic\n\t\t\t\tif sess.Flag&tp.SessKickOut != 0 { //logic kick out\n\t\t\t\t\tif err := stream.Send(&pbgame.Game_Frame{Type: pbgame.Game_Kick}); err != nil {\n\t\t\t\t\t\tlog.Error(err)\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase pbgame.Game_Ping:\n\t\t\t\tif err := stream.Send(&pbgame.Game_Frame{Type: pbgame.Game_Ping, Message: frame.Message}); err != nil {\n\t\t\t\t\tlog.Error(err)\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tlog.Debug(\"pinged\")\n\t\t\tdefault:\n\t\t\t\tlog.Error(\"incorrect frame type:\", frame.Type)\n\t\t\t\treturn ErrorIncorrectFrameType\n\t\t\t}\n\t\tcase frame := <-ch_ipc: // forward async messages from interprocess(goroutines) communication\n\t\t\tif err := stream.Send(frame); err != nil {\n\t\t\t\tlog.Error(err)\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n}", "func (c *Connection) CreateStream() *Stream {\n\tnextStream := c.maxStream + 1\n\n\t// Client opens odd streams\n\tif c.role == RoleClient {\n\t\tif (nextStream & 1) == 0 {\n\t\t\tnextStream++\n\t\t}\n\t} else {\n\t\tif (nextStream & 1) == 1 {\n\t\t\tnextStream++\n\t\t}\n\t}\n\n\treturn c.ensureStream(nextStream)\n}", "func setUpstream(ctx context.Context, session *lumina.ClientSession) context.Context {\n\treturn context.WithValue(ctx, upstreamContextKey, session)\n}", "func (q *QueryServiceTestHelper) NewStream(gocontext.Context, *grpc.StreamDesc, string, ...grpc.CallOption) (grpc.ClientStream, error) {\n\treturn nil, fmt.Errorf(\"not supported\")\n}", "func (m *Multiplexer) OpenStream(ctx context.Context) (*Stream, error) {\n\t// Create and register the local side of the stream. If we've already\n\t// exhausted local stream identifiers, then we can't open a new stream.\n\tm.streamLock.Lock()\n\tif m.nextOutboundStreamIdentifier == 0 {\n\t\tm.streamLock.Unlock()\n\t\treturn nil, errors.New(\"local stream identifiers exhausted\")\n\t}\n\tstream := newStream(m, m.nextOutboundStreamIdentifier, m.configuration.StreamReceiveWindow)\n\tm.streams[m.nextOutboundStreamIdentifier] = stream\n\tif math.MaxUint64-m.nextOutboundStreamIdentifier < 2 {\n\t\tm.nextOutboundStreamIdentifier = 0\n\t} else {\n\t\tm.nextOutboundStreamIdentifier += 2\n\t}\n\tm.streamLock.Unlock()\n\n\t// If we fail to establish the stream, then defer its closure. We can't use\n\t// the stream's established channel to check this because it could be closed\n\t// by the reader Goroutine after some other error aborts the opening.\n\tvar sentOpenMessage, established bool\n\tdefer func() {\n\t\tif !established {\n\t\t\tstream.close(sentOpenMessage)\n\t\t}\n\t}()\n\n\t// Write the open message and queue it for transmission.\n\tselect {\n\tcase writeBuffer := <-m.writeBufferAvailable:\n\t\twriteBuffer.encodeOpenMessage(stream.identifier, uint64(m.configuration.StreamReceiveWindow))\n\t\tm.writeBufferPending <- writeBuffer\n\t\tsentOpenMessage = true\n\tcase <-ctx.Done():\n\t\treturn nil, context.Canceled\n\tcase <-m.closed:\n\t\treturn nil, ErrMultiplexerClosed\n\t}\n\n\t// Wait for stream acceptance or rejection.\n\tselect {\n\tcase <-stream.established:\n\t\testablished = true\n\t\treturn stream, nil\n\tcase <-stream.remoteClosed:\n\t\treturn nil, ErrStreamRejected\n\tcase <-ctx.Done():\n\t\treturn nil, context.Canceled\n\tcase <-m.closed:\n\t\treturn nil, ErrMultiplexerClosed\n\t}\n}", "func (m *MockPacketHandler) OpenStreamSync() (Stream, error) {\n\tret := m.ctrl.Call(m, \"OpenStreamSync\")\n\tret0, _ := ret[0].(Stream)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *requestInterceptedClient) GetStream() rpcc.Stream { return c.Stream }", "func (c *webTransportClosedClient) GetStream() rpcc.Stream { return c.Stream }", "func (s *server) Stream(stream GameService_StreamServer) error {\n\tdefer PrintPanicStack()\n\t// session init\n\tvar sess Session\n\tsess_die := make(chan struct{})\n\tch_agent := s.recv(stream, sess_die)\n\tch_ipc := make(chan *Game_Frame, DEFAULT_CH_IPC_SIZE)\n\n\tdefer func() {\n\t\tregistry.Unregister(sess.UserId)\n\t\tclose(sess_die)\n\t\tlog.Trace(\"stream end:\", sess.UserId)\n\t}()\n\n\t// read metadata from context\n\tmd, ok := metadata.FromContext(stream.Context())\n\tif !ok {\n\t\tlog.Critical(\"cannot read metadata from context\")\n\t\treturn ERROR_INCORRECT_FRAME_TYPE\n\t}\n\t// read key\n\tif len(md[\"userid\"]) == 0 {\n\t\tlog.Critical(\"cannot read key:userid from metadata\")\n\t\treturn ERROR_INCORRECT_FRAME_TYPE\n\t}\n\t// parse userid\n\tuserid, err := strconv.Atoi(md[\"userid\"][0])\n\tif err != nil {\n\t\tlog.Critical(err)\n\t\treturn ERROR_INCORRECT_FRAME_TYPE\n\t}\n\n\t// register user\n\tsess.UserId = int32(userid)\n\tregistry.Register(sess.UserId, ch_ipc)\n\tlog.Finef(\"userid %v logged in\", sess.UserId)\n\n\t// >> main message loop <<\n\tfor {\n\t\tselect {\n\t\tcase frame, ok := <-ch_agent: // frames from agent\n\t\t\tif !ok { // EOF\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tswitch frame.Type {\n\t\t\tcase Game_Message: // the passthrough message from client->agent->game\n\t\t\t\t// locate handler by proto number\n\t\t\t\treader := packet.Reader(frame.Message)\n\t\t\t\tc, err := reader.ReadS16()\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Critical(err)\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\thandle := client_handler.Handlers[c]\n\t\t\t\tif handle == nil {\n\t\t\t\t\tlog.Criticalf(\"service not bind: %v\", c)\n\t\t\t\t\treturn ERROR_SERVICE_NOT_BIND\n\n\t\t\t\t}\n\n\t\t\t\t// handle request\n\t\t\t\tret := handle(&sess, reader)\n\n\t\t\t\t// construct frame & return message from logic\n\t\t\t\tif ret != nil {\n\t\t\t\t\tif err := stream.Send(&Game_Frame{Type: Game_Message, Message: ret}); err != nil {\n\t\t\t\t\t\tlog.Critical(err)\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// session control by logic\n\t\t\t\tif sess.Flag&SESS_KICKED_OUT != 0 { // logic kick out\n\t\t\t\t\tif err := stream.Send(&Game_Frame{Type: Game_Kick}); err != nil {\n\t\t\t\t\t\tlog.Critical(err)\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase Game_Ping:\n\t\t\t\tif err := stream.Send(&Game_Frame{Type: Game_Ping, Message: frame.Message}); err != nil {\n\t\t\t\t\tlog.Critical(err)\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tlog.Trace(\"pinged\")\n\t\t\tdefault:\n\t\t\t\tlog.Criticalf(\"incorrect frame type: %v\", frame.Type)\n\t\t\t\treturn ERROR_INCORRECT_FRAME_TYPE\n\t\t\t}\n\t\tcase frame := <-ch_ipc: // forward async messages from interprocess(goroutines) communication\n\t\t\tif err := stream.Send(frame); err != nil {\n\t\t\t\tlog.Critical(err)\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n}", "func (e *Helloworlde) Stream(ctx context.Context, req *helloworlde.StreamingRequest, stream helloworlde.Helloworlde_StreamStream) error {\n\tlog.Infof(\"Received Helloworlde.Stream request with count: %d\", req.Count)\n\n\tfor i := 0; i < int(req.Count); i++ {\n\t\tlog.Infof(\"Responding: %d\", i)\n\t\tif err := stream.Send(&helloworlde.StreamingResponse{\n\t\t\tCount: int64(i),\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (s *Session) open(ctx context.Context) error {\n\terr := s.doSession(ctx, func(ctx context.Context, conn *grpc.ClientConn, header *headers.RequestHeader) (*headers.ResponseHeader, interface{}, error) {\n\t\trequest := &api.OpenSessionRequest{\n\t\t\tHeader: header,\n\t\t\tTimeout: &s.Timeout,\n\t\t}\n\t\tclient := api.NewSessionServiceClient(conn)\n\t\tresponse, err := client.OpenSession(ctx, request)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\treturn response.Header, response, nil\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tgo func() {\n\t\tfor range s.ticker.C {\n\t\t\t_ = s.keepAlive(context.TODO())\n\t\t}\n\t}()\n\treturn nil\n}", "func (e *Example) Stream(ctx context.Context, req *example.StreamingRequest, stream example.Example_StreamStream) error {\n\tlog.Logf(\"Received Example.Stream request with count: %d\", req.Count)\n\n\tfor i := 0; i < int(req.Count); i++ {\n\t\tlog.Logf(\"Responding: %d\", i)\n\t\tif err := stream.Send(&example.StreamingResponse{\n\t\t\tCount: int64(i),\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (e *Example) Stream(ctx context.Context, req *example.StreamingRequest, stream example.Example_StreamStream) error {\n\tlog.Logf(\"Received Example.Stream request with count: %d\", req.Count)\n\n\tfor i := 0; i < int(req.Count); i++ {\n\t\tlog.Logf(\"Responding: %d\", i)\n\t\tif err := stream.Send(&example.StreamingResponse{\n\t\t\tCount: int64(i),\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (e *Example) Stream(ctx context.Context, req *example.StreamingRequest, stream example.Example_StreamStream) error {\n\tlog.Logf(\"Received Example.Stream request with count: %d\", req.Count)\n\n\tfor i := 0; i < int(req.Count); i++ {\n\t\tlog.Logf(\"Responding: %d\", i)\n\t\tif err := stream.Send(&example.StreamingResponse{\n\t\t\tCount: int64(i),\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (c *Client) Session(ctx context.Context) (*session.Session, session.Dialer, error) {\n\tm, err := c.getSessionManager()\n\tif err != nil {\n\t\treturn nil, nil, errors.Wrap(err, \"failed to create session manager\")\n\t}\n\tsessionName := \"img\"\n\ts, err := session.NewSession(ctx, sessionName, \"\")\n\tif err != nil {\n\t\treturn nil, nil, errors.Wrap(err, \"failed to create session\")\n\t}\n\tsyncedDirs := make([]filesync.SyncedDir, 0, len(c.localDirs))\n\tfor name, d := range c.localDirs {\n\t\tsyncedDirs = append(syncedDirs, filesync.SyncedDir{Name: name, Dir: d})\n\t}\n\ts.Allow(filesync.NewFSSyncProvider(syncedDirs))\n\ts.Allow(authprovider.NewDockerAuthProvider(os.Stderr))\n\treturn s, sessionDialer(s, m), err\n}", "func WrapStream(s net.Stream) *WrappedStream {\n\treader := bufio.NewReader(s)\n\twriter := bufio.NewWriter(s)\n\t// This is where we pick our specific multicodec. In order to change the\n\t// codec, we only need to change this place.\n\t// See https://godoc.org/github.com/multiformats/go-multicodec/json\n\tdec := json.Multicodec(false).Decoder(reader)\n\tenc := json.Multicodec(false).Encoder(writer)\n\treturn &WrappedStream{\n\t\tstream: s,\n\t\tr: reader,\n\t\tw: writer,\n\t\tenc: enc,\n\t\tdec: dec,\n\t}\n}", "func (c *Connection) Session() (Session, error) {\n\tconnection := c.pump.Connection()\n\tresult := make(chan sessionErr)\n\tc.pump.Inject <- func() {\n\t\ts, err := connection.Session()\n\t\tif err == nil {\n\t\t\ts.Open()\n\t\t}\n\t\tresult <- sessionErr{s, err}\n\t}\n\tse := <-result\n\treturn Session{se.s, c.pump}, se.err\n}", "func (gi *Invoker) StreamRecv(param *common.Params) error {\n\t//gloryPkg := newGloryRequestPackage(\"\", param.MethodName, uint64(common.StreamSendPkg), param.Seq)\n\t//gloryPkg.Params = append(gloryPkg.Params, param.Value)\n\t//gloryPkg.Header.ChanOffset = param.ChanOffset\n\t//gloryPkg.Header.Seq = param.Seq\n\t//if err := gloryPkg.sendToConn(gi.gloryConnClient, gi.handler); err != nil {\n\t//\tlog.Error(\"StreamRecv: gloryPkg.sendToConn(gi.conn, gi.handler) err =\", err)\n\t//\treturn GloryErrorConnErr\n\t//}\n\treturn nil\n}", "func NewStream(ctx context.Context, w http.ResponseWriter) Stream {\n\tresult := &stream{\n\t\tctx: ctx,\n\t\tw: w,\n\t}\n\n\treturn result\n}", "func (e *Payments) Stream(ctx context.Context, req *payments.StreamingRequest, stream payments.Payments_StreamStream) error {\n\tlog.Logf(\"Received Payments.Stream request with count: %d\", req.Count)\n\n\tfor i := 0; i < int(req.Count); i++ {\n\t\tlog.Logf(\"Responding: %d\", i)\n\t\tif err := stream.Send(&payments.StreamingResponse{\n\t\t\tCount: int64(i),\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func NewStream(options ...Option) Stream {\n\ts := Stream{\n\t\ttimeout: defaultTimeout,\n\t\taddr: defaultAddr,\n\t}\n\n\tfor _, opt := range options {\n\t\topt(&s)\n\t}\n\n\treturn s\n}", "func (c *requestServedFromCacheClient) GetStream() rpcc.Stream { return c.Stream }", "func (c *trustTokenOperationDoneClient) GetStream() rpcc.Stream { return c.Stream }", "func (notifee *Notifee) OpenedStream(network.Network, network.Stream) {}", "func (this *ProxyMid) ProxyStream(streamProducer gopcp_stream.StreamProducer, args []interface{}, attachment interface{}, pcpServer *gopcp.PcpServer) (interface{}, error) {\n\t// parse params\n\tvar (\n\t\tserviceType string\n\t\texp gopcp.FunNode\n\t\ttimeout int\n\t)\n\n\terr := utils.ParseArgs(args, []interface{}{&serviceType, &exp, &timeout}, \"wrong signature, expect (proxy, serviceType: string, exp, timeout: int)\")\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar timeoutD = time.Duration(timeout) * time.Second\n\n\tjsonObj := gopcp.ParseAstToJsonObject(exp)\n\n\tswitch arr := jsonObj.(type) {\n\tcase []interface{}:\n\t\t// choose worker\n\t\thandle, err := this.GetWorkerHandler(serviceType, \"\")\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// pipe stream\n\t\tsparams, err := handle.StreamClient.ParamsToStreamParams(append(arr[1:], func(t int, d interface{}) {\n\t\t\t// write response of stream back to client\n\t\t\tswitch t {\n\t\t\tcase gopcp_stream.STREAM_DATA:\n\t\t\t\tstreamProducer.SendData(d, timeoutD)\n\t\t\tcase gopcp_stream.STREAM_END:\n\t\t\t\tstreamProducer.SendEnd(timeoutD)\n\t\t\tdefault:\n\t\t\t\terrMsg, ok := d.(string)\n\t\t\t\tif !ok {\n\t\t\t\t\tstreamProducer.SendError(fmt.Sprintf(\"errored at stream, and responsed error message is not string. d=%v\", d), timeoutD)\n\t\t\t\t} else {\n\t\t\t\t\tstreamProducer.SendError(errMsg, timeoutD)\n\t\t\t\t}\n\t\t\t}\n\t\t}))\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// send a stream request to service\n\t\treturn handle.Call(gopcp.CallResult{append([]interface{}{arr[0]}, sparams...)}, timeoutD)\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"Expect array, but got %v, args=%v\", jsonObj, args)\n\t}\n}", "func (c *webSocketFrameErrorClient) GetStream() rpcc.Stream { return c.Stream }", "func NewStream(executor executor.Executor) executor.Launcher {\n\treturn New(executor, \"stress-ng-stream\", fmt.Sprintf(\"--stream=%d\", StressngStreamProcessNumber.Value()))\n}", "func (r *Runner) Stream(ctx context.Context, study diviner.Study, nparallel int) *Streamer {\n\ts := &Streamer{\n\t\trunner: r,\n\t\tstudy: study,\n\t\tnparallel: nparallel,\n\t\tstopc: make(chan struct{}),\n\t\tdonec: make(chan error),\n\t}\n\tgo func() {\n\t\ts.donec <- s.do(ctx)\n\t}()\n\treturn s\n}", "func (agent *Agent) OpenStream(vbId uint16, flags DcpStreamAddFlag, vbUuid VbUuid, startSeqNo,\n\tendSeqNo, snapStartSeqNo, snapEndSeqNo SeqNo, evtHandler StreamObserver, filter *StreamFilter, cb OpenStreamCallback) (PendingOp, error) {\n\tvar req *memdQRequest\n\thandler := func(resp *memdQResponse, _ *memdQRequest, err error) {\n\t\tif resp != nil && resp.Magic == resMagic {\n\t\t\t// This is the response to the open stream request.\n\t\t\tif err != nil {\n\t\t\t\treq.Cancel()\n\n\t\t\t\t// All client errors are handled by the StreamObserver\n\t\t\t\tcb(nil, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnumEntries := len(resp.Value) / 16\n\t\t\tentries := make([]FailoverEntry, numEntries)\n\t\t\tfor i := 0; i < numEntries; i++ {\n\t\t\t\tentries[i] = FailoverEntry{\n\t\t\t\t\tVbUuid: VbUuid(binary.BigEndian.Uint64(resp.Value[i*16+0:])),\n\t\t\t\t\tSeqNo: SeqNo(binary.BigEndian.Uint64(resp.Value[i*16+8:])),\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tcb(entries, nil)\n\t\t\treturn\n\t\t}\n\n\t\tif err != nil {\n\t\t\treq.Cancel()\n\t\t\tstreamId := noStreamId\n\t\t\tif filter != nil {\n\t\t\t\tstreamId = filter.StreamId\n\t\t\t}\n\t\t\tevtHandler.End(vbId, streamId, err)\n\t\t\treturn\n\t\t}\n\n\t\t// This is one of the stream events\n\t\tswitch resp.Opcode {\n\t\tcase cmdDcpSnapshotMarker:\n\t\t\tvbId := uint16(resp.Vbucket)\n\t\t\tnewStartSeqNo := binary.BigEndian.Uint64(resp.Extras[0:])\n\t\t\tnewEndSeqNo := binary.BigEndian.Uint64(resp.Extras[8:])\n\t\t\tsnapshotType := binary.BigEndian.Uint32(resp.Extras[16:])\n\t\t\tvar streamId uint16\n\t\t\tif resp.FrameExtras != nil && resp.FrameExtras.HasStreamId {\n\t\t\t\tstreamId = resp.FrameExtras.StreamId\n\t\t\t}\n\t\t\tevtHandler.SnapshotMarker(newStartSeqNo, newEndSeqNo, vbId, streamId, SnapshotState(snapshotType))\n\t\tcase cmdDcpMutation:\n\t\t\tvbId := uint16(resp.Vbucket)\n\t\t\tseqNo := binary.BigEndian.Uint64(resp.Extras[0:])\n\t\t\trevNo := binary.BigEndian.Uint64(resp.Extras[8:])\n\t\t\tflags := binary.BigEndian.Uint32(resp.Extras[16:])\n\t\t\texpiry := binary.BigEndian.Uint32(resp.Extras[20:])\n\t\t\tlockTime := binary.BigEndian.Uint32(resp.Extras[24:])\n\t\t\tvar streamId uint16\n\t\t\tif resp.FrameExtras != nil && resp.FrameExtras.HasStreamId {\n\t\t\t\tstreamId = resp.FrameExtras.StreamId\n\t\t\t}\n\t\t\tevtHandler.Mutation(seqNo, revNo, flags, expiry, lockTime, resp.Cas, resp.Datatype, vbId, resp.CollectionID, streamId, resp.Key, resp.Value)\n\t\tcase cmdDcpDeletion:\n\t\t\tvbId := uint16(resp.Vbucket)\n\t\t\tseqNo := binary.BigEndian.Uint64(resp.Extras[0:])\n\t\t\trevNo := binary.BigEndian.Uint64(resp.Extras[8:])\n\t\t\tvar streamId uint16\n\t\t\tif resp.FrameExtras != nil && resp.FrameExtras.HasStreamId {\n\t\t\t\tstreamId = resp.FrameExtras.StreamId\n\t\t\t}\n\t\t\tevtHandler.Deletion(seqNo, revNo, resp.Cas, resp.Datatype, vbId, resp.CollectionID, streamId, resp.Key, resp.Value)\n\t\tcase cmdDcpExpiration:\n\t\t\tvbId := uint16(resp.Vbucket)\n\t\t\tseqNo := binary.BigEndian.Uint64(resp.Extras[0:])\n\t\t\trevNo := binary.BigEndian.Uint64(resp.Extras[8:])\n\t\t\tvar streamId uint16\n\t\t\tif resp.FrameExtras != nil && resp.FrameExtras.HasStreamId {\n\t\t\t\tstreamId = resp.FrameExtras.StreamId\n\t\t\t}\n\t\t\tevtHandler.Expiration(seqNo, revNo, resp.Cas, vbId, resp.CollectionID, streamId, resp.Key)\n\t\tcase cmdDcpEvent:\n\t\t\tvbId := uint16(resp.Vbucket)\n\t\t\tseqNo := binary.BigEndian.Uint64(resp.Extras[0:])\n\t\t\teventCode := StreamEventCode(binary.BigEndian.Uint32(resp.Extras[8:]))\n\t\t\tversion := resp.Extras[12]\n\t\t\tvar streamId uint16\n\t\t\tif resp.FrameExtras != nil && resp.FrameExtras.HasStreamId {\n\t\t\t\tstreamId = resp.FrameExtras.StreamId\n\t\t\t}\n\n\t\t\tswitch eventCode {\n\t\t\tcase StreamEventCollectionCreate:\n\t\t\t\tmanifestUid := binary.BigEndian.Uint64(resp.Value[0:])\n\t\t\t\tscopeId := binary.BigEndian.Uint32(resp.Value[8:])\n\t\t\t\tcollectionId := binary.BigEndian.Uint32(resp.Value[12:])\n\t\t\t\tvar ttl uint32\n\t\t\t\tif version == 1 {\n\t\t\t\t\tttl = binary.BigEndian.Uint32(resp.Value[16:])\n\t\t\t\t}\n\t\t\t\tevtHandler.CreateCollection(seqNo, version, vbId, manifestUid, scopeId, collectionId, ttl, streamId, resp.Key)\n\t\t\tcase StreamEventCollectionDelete:\n\t\t\t\tmanifestUid := binary.BigEndian.Uint64(resp.Value[0:])\n\t\t\t\tscopeId := binary.BigEndian.Uint32(resp.Value[8:])\n\t\t\t\tcollectionId := binary.BigEndian.Uint32(resp.Value[12:])\n\t\t\t\tevtHandler.DeleteCollection(seqNo, version, vbId, manifestUid, scopeId, collectionId, streamId)\n\t\t\tcase StreamEventCollectionFlush:\n\t\t\t\tmanifestUid := binary.BigEndian.Uint64(resp.Value[0:])\n\t\t\t\tcollectionId := binary.BigEndian.Uint32(resp.Value[8:])\n\t\t\t\tevtHandler.FlushCollection(seqNo, version, vbId, manifestUid, collectionId)\n\t\t\tcase StreamEventScopeCreate:\n\t\t\t\tmanifestUid := binary.BigEndian.Uint64(resp.Value[0:])\n\t\t\t\tscopeId := binary.BigEndian.Uint32(resp.Value[8:])\n\t\t\t\tevtHandler.CreateScope(seqNo, version, vbId, manifestUid, scopeId, streamId, resp.Key)\n\t\t\tcase StreamEventScopeDelete:\n\t\t\t\tmanifestUid := binary.BigEndian.Uint64(resp.Value[0:])\n\t\t\t\tscopeId := binary.BigEndian.Uint32(resp.Value[8:])\n\t\t\t\tevtHandler.DeleteScope(seqNo, version, vbId, manifestUid, scopeId, streamId)\n\t\t\tcase StreamEventCollectionChanged:\n\t\t\t\tmanifestUid := binary.BigEndian.Uint64(resp.Value[0:])\n\t\t\t\tcollectionId := binary.BigEndian.Uint32(resp.Value[8:])\n\t\t\t\tttl := binary.BigEndian.Uint32(resp.Value[12:])\n\t\t\t\tevtHandler.ModifyCollection(seqNo, version, vbId, manifestUid, collectionId, ttl, streamId)\n\t\t\t}\n\t\tcase cmdDcpStreamEnd:\n\t\t\tvbId := uint16(resp.Vbucket)\n\t\t\tcode := streamEndStatus(binary.BigEndian.Uint32(resp.Extras[0:]))\n\t\t\tvar streamId uint16\n\t\t\tif resp.FrameExtras != nil && resp.FrameExtras.HasStreamId {\n\t\t\t\tstreamId = resp.FrameExtras.StreamId\n\t\t\t}\n\t\t\tevtHandler.End(vbId, streamId, getStreamEndError(code))\n\t\t\treq.Cancel()\n\t\t}\n\t}\n\n\textraBuf := make([]byte, 48)\n\tbinary.BigEndian.PutUint32(extraBuf[0:], uint32(flags))\n\tbinary.BigEndian.PutUint32(extraBuf[4:], 0)\n\tbinary.BigEndian.PutUint64(extraBuf[8:], uint64(startSeqNo))\n\tbinary.BigEndian.PutUint64(extraBuf[16:], uint64(endSeqNo))\n\tbinary.BigEndian.PutUint64(extraBuf[24:], uint64(vbUuid))\n\tbinary.BigEndian.PutUint64(extraBuf[32:], uint64(snapStartSeqNo))\n\tbinary.BigEndian.PutUint64(extraBuf[40:], uint64(snapEndSeqNo))\n\n\tvar val []byte\n\tval = nil\n\tif filter != nil {\n\t\tconvertedFilter := streamFilter{}\n\t\tfor _, cid := range filter.Collections {\n\t\t\tconvertedFilter.Collections = append(convertedFilter.Collections, fmt.Sprintf(\"%x\", cid))\n\t\t}\n\t\tif filter.Scope != noScopeId {\n\t\t\tconvertedFilter.Scope = fmt.Sprintf(\"%x\", filter.Scope)\n\t\t}\n\t\tif filter.ManifestUid != noManifestUid {\n\t\t\tconvertedFilter.ManifestUid = fmt.Sprintf(\"%x\", filter.ManifestUid)\n\t\t}\n\t\tif filter.StreamId != noStreamId {\n\t\t\tconvertedFilter.StreamId = filter.StreamId\n\t\t}\n\t\tvar err error\n\t\tval, err = json.Marshal(convertedFilter)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq = &memdQRequest{\n\t\tmemdPacket: memdPacket{\n\t\t\tMagic: reqMagic,\n\t\t\tOpcode: cmdDcpStreamReq,\n\t\t\tDatatype: 0,\n\t\t\tCas: 0,\n\t\t\tExtras: extraBuf,\n\t\t\tKey: nil,\n\t\t\tValue: val,\n\t\t\tVbucket: vbId,\n\t\t},\n\t\tCallback: handler,\n\t\tReplicaIdx: 0,\n\t\tPersistent: true,\n\t}\n\treturn agent.dispatchOp(req)\n}" ]
[ "0.63929254", "0.63021433", "0.5947112", "0.59255314", "0.5806985", "0.575277", "0.5727995", "0.57169294", "0.57070553", "0.57049143", "0.5697204", "0.5692881", "0.5660469", "0.56468356", "0.5645025", "0.56316733", "0.56232727", "0.5610288", "0.56046814", "0.5600436", "0.5599777", "0.5575647", "0.55743176", "0.55561626", "0.5540728", "0.5527061", "0.5490817", "0.54860467", "0.54712033", "0.5463482", "0.5448829", "0.5446071", "0.5445168", "0.5434743", "0.54304427", "0.54248035", "0.5406995", "0.5397337", "0.53766006", "0.5371886", "0.5360309", "0.5349414", "0.5339818", "0.5321284", "0.5315535", "0.53138524", "0.5259705", "0.523878", "0.5229162", "0.52269644", "0.5226843", "0.522538", "0.52199125", "0.521966", "0.51928633", "0.5192489", "0.5184864", "0.5177149", "0.51674014", "0.5166062", "0.51652986", "0.5161734", "0.5155227", "0.5151605", "0.5144135", "0.5141819", "0.51394355", "0.51367384", "0.5129518", "0.51253474", "0.51219976", "0.511092", "0.5088796", "0.50876707", "0.5084735", "0.5079769", "0.50779575", "0.5074066", "0.50615466", "0.5056887", "0.5056754", "0.50535727", "0.50533473", "0.50485855", "0.50485855", "0.50485855", "0.50481975", "0.5033851", "0.5033445", "0.50277245", "0.5027397", "0.5027369", "0.501531", "0.50113106", "0.4995331", "0.4991419", "0.4987111", "0.49784276", "0.49773332", "0.4974575", "0.4972112" ]
0.0
-1
DSL returns the attached DSL.
func (e *Element) DSL() func() { return e.DSLFunc }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (a *AttributeDefinition) DSL() func() {\n\treturn a.DSLFunc\n}", "func (t *TraitDefinition) DSL() func() {\n\treturn t.DSLFunc\n}", "func RunHTTPDSL(t *testing.T, dsl func()) *RootExpr {\n\tsetupDSLRun()\n\n\t// run DSL (first pass)\n\tif !eval.Execute(dsl, nil) {\n\t\tt.Fatal(eval.Context.Error())\n\t}\n\n\t// run DSL (second pass)\n\tif err := eval.RunDSL(); err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// return generated root\n\treturn Root\n}", "func (s *DbRecorder) Builder() *squirrel.StatementBuilderType {\n\treturn s.builder\n}", "func (matcher *Matcher) MatchDSL(data map[string]interface{}) bool {\n\tlogExpressionEvaluationFailure := func(matcherName string, err error) {\n\t\tgologger.Warning().Msgf(\"Could not evaluate expression: %s, error: %s\", matcherName, err.Error())\n\t}\n\n\t// Iterate over all the expressions accepted as valid\n\tfor i, expression := range matcher.dslCompiled {\n\t\tif varErr := expressions.ContainsUnresolvedVariables(expression.String()); varErr != nil {\n\t\t\tresolvedExpression, err := expressions.Evaluate(expression.String(), data)\n\t\t\tif err != nil {\n\t\t\t\tlogExpressionEvaluationFailure(matcher.Name, err)\n\t\t\t\treturn false\n\t\t\t}\n\t\t\texpression, err = govaluate.NewEvaluableExpressionWithFunctions(resolvedExpression, dsl.HelperFunctions)\n\t\t\tif err != nil {\n\t\t\t\tlogExpressionEvaluationFailure(matcher.Name, err)\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\n\t\tresult, err := expression.Evaluate(data)\n\t\tif err != nil {\n\t\t\tif matcher.condition == ANDCondition {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tif !matcher.ignoreErr(err) {\n\t\t\t\tgologger.Warning().Msgf(\"[%s] %s\", data[\"template-id\"], err.Error())\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tif boolResult, ok := result.(bool); !ok {\n\t\t\tgologger.Error().Label(\"WRN\").Msgf(\"[%s] The return value of a DSL statement must return a boolean value.\", data[\"template-id\"])\n\t\t\tcontinue\n\t\t} else if !boolResult {\n\t\t\t// If we are in an AND request and a match failed,\n\t\t\t// return false as the AND condition fails on any single mismatch.\n\t\t\tswitch matcher.condition {\n\t\t\tcase ANDCondition:\n\t\t\t\treturn false\n\t\t\tcase ORCondition:\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\t// If the condition was an OR, return on the first match.\n\t\tif matcher.condition == ORCondition {\n\t\t\treturn true\n\t\t}\n\n\t\t// If we are at the end of the dsl, return with true\n\t\tif len(matcher.dslCompiled)-1 == i {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (b *GroupPolicyPresentationRequestBuilder) Definition() *GroupPolicyDefinitionRequestBuilder {\n\tbb := &GroupPolicyDefinitionRequestBuilder{BaseRequestBuilder: b.BaseRequestBuilder}\n\tbb.baseURL += \"/definition\"\n\treturn bb\n}", "func (dsl *DeleteDSL) Send() vpp_clientv1.Reply {\n\treturn dsl.parent.Send()\n}", "func (dsl *DeleteDSL) Send() vpp_clientv1.Reply {\n\treturn dsl.parent.Send()\n}", "func (g *Graph) Definition() *Definition {\n\treturn &Definition{g}\n}", "func (dsl *PutDSL) Delete() linux.DeleteDSL {\n\treturn &DeleteDSL{dsl.parent, dsl.vppPut.Delete()}\n}", "func (dsl *PutDSL) Delete() linux.DeleteDSL {\n\treturn &DeleteDSL{dsl.parent, dsl.vppPut.Delete()}\n}", "func (dsl *DeleteDSL) BD(bdName string) linux.DeleteDSL {\n\tdsl.vppDelete.BD(bdName)\n\treturn dsl\n}", "func (dsl *DeleteDSL) BD(bdName string) linux.DeleteDSL {\n\tdsl.vppDelete.BD(bdName)\n\treturn dsl\n}", "func (dsl *DeleteDSL) Put() linux.PutDSL {\n\treturn &PutDSL{dsl.parent, dsl.vppDelete.Put()}\n}", "func (dsl *DeleteDSL) Put() linux.PutDSL {\n\treturn &PutDSL{dsl.parent, dsl.vppDelete.Put()}\n}", "func (s *service) Builder() *sql.Builder {\n\treturn s.builder\n}", "func (v *View) Definition() Node {\n\treturn v.definition\n}", "func (dsl *PutDSL) Send() vpp_clientv1.Reply {\n\treturn dsl.parent.Send()\n}", "func (dsl *PutDSL) Send() vpp_clientv1.Reply {\n\treturn dsl.parent.Send()\n}", "func RunGRPCDSL(t *testing.T, dsl func()) *RootExpr {\n\tsetupDSLRun()\n\n\t// run DSL (first pass)\n\tif !eval.Execute(dsl, nil) {\n\t\tt.Fatal(eval.Context.Error())\n\t}\n\n\t// run DSL (second pass)\n\tif err := eval.RunDSL(); err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// return generated root\n\treturn Root\n}", "func NewBuilder() Builder {\n\treturn &builder{}\n}", "func GetDialect() SQLDialect { return def.GetDialect() }", "func (dfa *DFAccess) DocumentsDSLQuery(dslQueryString string, nResults int) *SearchResults {\n\treturn dfa.dslQuery(\"documents/dslquery\", dslQueryString, nResults)\n}", "func (dsl *DataChangeDSL) Delete() linux.DeleteDSL {\n\treturn &DeleteDSL{dsl, dsl.vppDataChange.Delete()}\n}", "func (dsl *DataChangeDSL) Delete() linux.DeleteDSL {\n\treturn &DeleteDSL{dsl, dsl.vppDataChange.Delete()}\n}", "func NewDataChangeDSL(txn keyval.ProtoTxn) *DataChangeDSL {\n\tvppDbAdapter := vpp_dbadapter.NewDataChangeDSL(txn)\n\treturn &DataChangeDSL{txn: txn, vppDataChange: vppDbAdapter}\n}", "func NewDataChangeDSL(txn keyval.ProtoTxn) *DataChangeDSL {\n\tvppDbAdapter := vpp_dbadapter.NewDataChangeDSL(txn)\n\treturn &DataChangeDSL{txn: txn, vppDataChange: vppDbAdapter}\n}", "func Builder() builder {\n\treturn builder{\n\t\tfields: make(map[string]interface{}),\n\t\tcontextFields: make(map[string]interface{}),\n\t}\n}", "func (Bgp) Definition() sophos.Definition { return *defBgp }", "func (rb *SlackAttachmentBuilder) Build() SlackAttachment {\n\treturn *rb.v\n}", "func (app *builder) Now() (Selector, error) {\n\tif app.decryptionKey != nil {\n\t\treturn nil, errors.New(\"the decryptionKey is mandatory in order to build a Selector instance\")\n\t}\n\n\tdata := [][]byte{\n\t\tapp.pkAdapter.ToBytes(app.decryptionKey),\n\t}\n\n\tvar content Content\n\tif app.graphbase != nil {\n\t\tdata = append(data, app.graphbase.Hash().Bytes())\n\t\tcontent = createContentWithGraphbase(app.graphbase)\n\t}\n\n\tif app.db != nil {\n\t\tdata = append(data, app.db.Hash().Bytes())\n\t\tcontent = createContentWithDatabase(app.db)\n\t}\n\n\tif app.table != nil {\n\t\tdata = append(data, app.table.Hash().Bytes())\n\t\tcontent = createContentWithTable(app.table)\n\t}\n\n\tif app.set != nil {\n\t\tdata = append(data, app.set.Hash().Bytes())\n\t\tcontent = createContentWithSet(app.set)\n\t}\n\n\tif content == nil {\n\t\treturn nil, errors.New(\"the content (graphbase, database, table, set) is mandatory in order to build a Selector instance\")\n\t}\n\n\thsh, err := app.hashAdapter.FromMultiBytes(data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn createSelector(*hsh, app.decryptionKey, content), nil\n}", "func (storage *Storage) Dialect() string {\n\treturn storage.exec.Dialect()\n}", "func NewBuilder() adapters.Builder {\n\treturn &builder{}\n}", "func (app *languageTestSectionBuilder) Now() (LanguageTestSection, error) {\n\tif app.declarations != nil && len(app.declarations) <= 0 {\n\t\tapp.declarations = nil\n\t}\n\n\tif app.declarations == nil {\n\t\treturn nil, errors.New(\"there must be at least 1 LanguageTestDeclaration instance in order to build a LanguageTestSection instance\")\n\t}\n\n\treturn createLanguageTestSection(app.declarations), nil\n}", "func NewBuilder() Builder {\n\treturn createBuilder()\n}", "func NewBuilder() Builder {\n\treturn createBuilder()\n}", "func NewBuilder() Builder {\n\treturn createBuilder()\n}", "func NewBuilder() Builder {\n\treturn createBuilder()\n}", "func NewBuilder() Builder {\n\treturn createBuilder()\n}", "func NewBuilder() *Builder {\n\treturn &Builder{\n\t\tprojectFunc: func(context.Context) (*gardencorev1beta1.Project, error) {\n\t\t\treturn nil, fmt.Errorf(\"project is required but not set\")\n\t\t},\n\t\tinternalDomainFunc: func() (*gardenerutils.Domain, error) {\n\t\t\treturn nil, fmt.Errorf(\"internal domain is required but not set\")\n\t\t},\n\t}\n}", "func DefaultSQLBuilder() SQLBuilder {\n\treturn SQLBuilder{\n\t\tPath: \"\",\n\t\tTable: DefaultTableSQLBuilder,\n\t\tView: DefaultViewSQLBuilder,\n\t\tEnum: DefaultEnumSQLBuilder,\n\t}\n}", "func (b *Builder) Build(ctx context.Context) (*Garden, error) {\n\tgarden := &Garden{}\n\n\tproject, err := b.projectFunc(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tgarden.Project = project\n\n\tinternalDomain, err := b.internalDomainFunc()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tgarden.InternalDomain = internalDomain\n\n\tdefaultDomains, err := b.defaultDomainsFunc()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tgarden.DefaultDomains = defaultDomains\n\n\treturn garden, nil\n}", "func (f *Function) Definition() Definition {\n\treturn f.def\n}", "func (w *WidgetImplement) Layout() Layout {\n\treturn w.layout\n}", "func (dsl *DeleteDSL) L4Features() linux.DeleteDSL {\n\tdsl.vppDelete.L4Features()\n\treturn dsl\n}", "func (database *Database) Dialect() schema.Dialect {\n\treturn database.dialect\n}", "func NewBuilder() *Builder {\n\treturn &Builder{}\n}", "func NewBuilder() *Builder {\n\treturn &Builder{}\n}", "func NewBuilder() *Builder {\n\treturn &Builder{}\n}", "func NewBuilder() *Builder {\n\treturn &Builder{}\n}", "func NewBuilder() *Builder {\n\treturn &Builder{}\n}", "func (r *Resolver) Language() generated.LanguageResolver { return &languageResolver{r} }", "func (x ApmApplicationEntityOutline) GetLanguage() string {\n\treturn x.Language\n}", "func (d *Document) Body() Body { return d.body }", "func (c *NestedTargetsBuilder) And() *Builder {\n\treturn c.parent\n}", "func (ln *LocalNode) Database() *DB {\n\treturn ln.db\n}", "func (f *ring2Factory) NewBuilder() *resource.Builder {\n\treturn f.kubeBuilderFactory.NewBuilder()\n}", "func ExampleBuilder() {\n\t// Create an SMTP Sender which relies on Go's built-in net/smtp package. Advanced users\n\t// may provide their own Sender, or mock it in unit tests.\n\tsmtpHost := \"smtp.relay.host:25\"\n\tsmtpAuth := smtp.PlainAuth(\"\", \"user\", \"pw\", \"host\")\n\tsender := enmime.NewSMTP(smtpHost, smtpAuth)\n\n\t// MailBuilder is (mostly) immutable, each method below returns a new MailBuilder without\n\t// modifying the original.\n\tmaster := enmime.Builder().\n\t\tFrom(\"Do Not Reply\", \"[email protected]\").\n\t\tSubject(\"Inbucket Newsletter\").\n\t\tText([]byte(\"Text body\")).\n\t\tHTML([]byte(\"<p>HTML body</p>\"))\n\n\t// master is immutable, causing each msg below to have a single recipient.\n\tmsg := master.To(\"Esteemed Customer\", \"[email protected]\")\n\terr := msg.Send(sender)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tmsg = master.To(\"Another Customer\", \"[email protected]\")\n\terr = msg.Send(sender)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (cfg *Config) bind() (*Build, error) {\n\tnamedTypes := cfg.buildNamedTypes()\n\n\tprog, err := cfg.loadProgram(namedTypes, true)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"loading failed\")\n\t}\n\n\timports := buildImports(namedTypes, cfg.Exec.Dir())\n\tcfg.bindTypes(imports, namedTypes, cfg.Exec.Dir(), prog)\n\n\tobjects, err := cfg.buildObjects(namedTypes, prog, imports)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinputs, err := cfg.buildInputs(namedTypes, prog, imports)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tb := &Build{\n\t\tPackageName: cfg.Exec.Package,\n\t\tObjects: objects,\n\t\tInterfaces: cfg.buildInterfaces(namedTypes, prog),\n\t\tInputs: inputs,\n\t\tImports: imports.finalize(),\n\t}\n\n\tif qr, ok := cfg.schema.EntryPoints[\"query\"]; ok {\n\t\tb.QueryRoot = b.Objects.ByName(qr.TypeName())\n\t}\n\n\tif mr, ok := cfg.schema.EntryPoints[\"mutation\"]; ok {\n\t\tb.MutationRoot = b.Objects.ByName(mr.TypeName())\n\t}\n\n\tif sr, ok := cfg.schema.EntryPoints[\"subscription\"]; ok {\n\t\tb.SubscriptionRoot = b.Objects.ByName(sr.TypeName())\n\t}\n\n\tif b.QueryRoot == nil {\n\t\treturn b, fmt.Errorf(\"query entry point missing\")\n\t}\n\n\t// Poke a few magic methods into query\n\tq := b.Objects.ByName(b.QueryRoot.GQLType)\n\tq.Fields = append(q.Fields, Field{\n\t\tType: &Type{namedTypes[\"__Schema\"], []string{modPtr}, \"\"},\n\t\tGQLName: \"__schema\",\n\t\tNoErr: true,\n\t\tGoMethodName: \"ec.introspectSchema\",\n\t\tObject: q,\n\t})\n\tq.Fields = append(q.Fields, Field{\n\t\tType: &Type{namedTypes[\"__Type\"], []string{modPtr}, \"\"},\n\t\tGQLName: \"__type\",\n\t\tNoErr: true,\n\t\tGoMethodName: \"ec.introspectType\",\n\t\tArgs: []FieldArgument{\n\t\t\t{GQLName: \"name\", Type: &Type{namedTypes[\"String\"], []string{}, \"\"}, Object: &Object{}},\n\t\t},\n\t\tObject: q,\n\t})\n\n\treturn b, nil\n}", "func (w *windowTransformation2) getBuilder(t *windowSchemaTemplate, bound execute.Bounds) *table.ArrowBuilder {\n\tkey := w.newWindowGroupKey(t.keyCols, t.keyValues, bound)\n\tbuilder, created := table.GetArrowBuilder(key, w.cache)\n\tif created {\n\t\t// Establish the table schema and initialize the builders.\n\t\tbuilder.Columns = t.cols\n\t\tbuilder.Builders = make([]array.Builder, len(t.cols))\n\t\tfor i, col := range builder.Columns {\n\t\t\tbuilder.Builders[i] = arrow.NewBuilder(col.Type, w.mem)\n\t\t}\n\t}\n\treturn builder\n}", "func (m pWithContext) Underlying() *models.Method {\n\treturn m.Method\n}", "func (database *Database) Wrapper() interface{} {\n\treturn database.wrapper\n}", "func (s *BasevhdlListener) EnterPackage_body_declarative_part(ctx *Package_body_declarative_partContext) {\n}", "func (r *Recorder) Selected() *configservice.ConfigurationRecorder {\n\treturn r.recorder\n}", "func (f *factory) Builder() *resource.Builder {\n\treturn resource.NewBuilder(f.kubeConfigFlags)\n}", "func NewBuilder(context BuilderContext) Builder {\n\treturn &builderImpl{context: context}\n}", "func NewBuilder() *Builder {\n\treturn &Builder{\n\t\tpath: \"\",\n\t\tchecks: make(map[*Predicate]string),\n\t}\n}", "func NewDefinition() *Definition {\n\treturn &Definition{\n\t\tActive: true,\n\t\tPlugins: make([]Plugin, 0),\n\t\tProxy: proxy.NewDefinition(),\n\t}\n}", "func (o BuildSpecOutput) Builder() BuildSpecBuilderPtrOutput {\n\treturn o.ApplyT(func(v BuildSpec) *BuildSpecBuilder { return v.Builder }).(BuildSpecBuilderPtrOutput)\n}", "func (serviceLocator *DefaultServiceLocator) GetViewSQLBuilder(dbName, docID, viewName string, setup, scripts []Query) *ViewSQLChangeSet {\n\tfileName := serviceLocator.localDB.GetDatabaseFileName(dbName)\n\tDBPath := filepath.Join(serviceLocator.GetDBDirPath(), fileName+dbExt)\n\tqualifiedViewName := docID + \"$\" + viewName\n\treturn NewViewSQL(dbName, DBPath, qualifiedViewName, setup, scripts)\n}", "func Builder() *PoolBuilder {\n\n\tif poolBuilderInstance == nil {\n\n\t\tpoolBuilderInstance = new(PoolBuilder)\n\t}\n\n\treturn poolBuilderInstance\n}", "func (m *GraphBaseServiceClient) Admin()(*i7c9d1b36ac198368c1d8bed014b43e2a518b170ee45bf02c8bbe64544a50539a.AdminRequestBuilder) {\n return i7c9d1b36ac198368c1d8bed014b43e2a518b170ee45bf02c8bbe64544a50539a.NewAdminRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func (m *GraphBaseServiceClient) Admin()(*i7c9d1b36ac198368c1d8bed014b43e2a518b170ee45bf02c8bbe64544a50539a.AdminRequestBuilder) {\n return i7c9d1b36ac198368c1d8bed014b43e2a518b170ee45bf02c8bbe64544a50539a.NewAdminRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func DB() Database {\n\treturn db\n}", "func (Interface) Definition() sophos.Definition { return *defInterface }", "func New() *Builder {\n\treturn &Builder{\n\t\tID: getulid(),\n\t}\n}", "func (client *WANDSLLinkConfig1) GetDSLLinkInfo() (NewLinkType string, NewLinkStatus string, err error) {\n\treturn client.GetDSLLinkInfoCtx(context.Background())\n}", "func NewSlackAttachmentBuilder() *SlackAttachmentBuilder {\n\tr := SlackAttachmentBuilder{\n\t\t&SlackAttachment{},\n\t}\n\n\treturn &r\n}", "func (app *builder) Now() (Statement, error) {\n\tif app.hash == nil {\n\t\treturn nil, errors.New(\"the hash is mandatory in order to build a Statement instance\")\n\t}\n\n\tif len(app.incoming) <= 0 {\n\t\tapp.incoming = nil\n\t}\n\n\tif len(app.outgoing) <= 0 {\n\t\tapp.outgoing = nil\n\t}\n\n\timmutable, err := app.immutableBuilder.Create().WithHash(*app.hash).CreatedOn(app.createdOn).Now()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif app.incoming != nil && app.outgoing != nil {\n\t\treturn createStatementWithIncomingAndOutgoing(immutable, app.incoming, app.outgoing), nil\n\t}\n\n\tif app.incoming != nil {\n\t\treturn createStatementWithIncoming(immutable, app.incoming), nil\n\t}\n\n\tif app.outgoing != nil {\n\t\treturn createStatementWithOutgoing(immutable, app.outgoing), nil\n\t}\n\n\treturn createStatement(immutable), nil\n}", "func (f *ring2Factory) NewBuilder() *resource.Builder {\n\tclientMapperFunc := resource.ClientMapperFunc(f.objectMappingFactory.ClientForMapping)\n\tmapper, typer := f.objectMappingFactory.Object()\n\n\tunstructuredClientMapperFunc := resource.ClientMapperFunc(f.objectMappingFactory.UnstructuredClientForMapping)\n\n\tcategoryExpander := f.objectMappingFactory.CategoryExpander()\n\n\treturn resource.NewBuilder(\n\t\t&resource.Mapper{\n\t\t\tRESTMapper: mapper,\n\t\t\tObjectTyper: typer,\n\t\t\tClientMapper: clientMapperFunc,\n\t\t\tDecoder: InternalVersionDecoder(),\n\t\t},\n\t\t&resource.Mapper{\n\t\t\tRESTMapper: mapper,\n\t\t\tObjectTyper: typer,\n\t\t\tClientMapper: unstructuredClientMapperFunc,\n\t\t\tDecoder: unstructured.UnstructuredJSONScheme,\n\t\t},\n\t\tcategoryExpander,\n\t)\n}", "func (env *Zlisp) ImportPackageBuilder() {\n\tenv.AddBuilder(\"infixExpand\", InfixBuilder)\n\tenv.AddBuilder(\"infix\", InfixBuilder)\n\tenv.AddBuilder(\":\", ColonAccessBuilder)\n\tenv.AddBuilder(\"sys\", SystemBuilder)\n\tenv.AddBuilder(\"struct\", StructBuilder)\n\tenv.AddBuilder(\"func\", FuncBuilder)\n\tenv.AddBuilder(\"method\", FuncBuilder)\n\tenv.AddBuilder(\"interface\", InterfaceBuilder)\n\t//env.AddBuilder(\"package\", PackageBuilder)\n\t//env.AddBuilder(\"import\", ImportBuilder)\n\tenv.AddBuilder(\"var\", VarBuilder)\n\tenv.AddBuilder(\"expectError\", ExpectErrorBuilder)\n\tenv.AddBuilder(\"comma\", CommaBuilder)\n\t//\tenv.AddBuilder(\"&\", AddressOfBuilder)\n\n\tenv.AddBuilder(\"import\", ImportPackageBuilder)\n\n\tenv.AddFunction(\"sliceOf\", SliceOfFunction)\n\tenv.AddFunction(\"ptr\", PointerToFunction)\n}", "func (dfa *DFAccess) FactsDSLQuery(dslQueryString string, nResults int) *SearchResults {\n\treturn dfa.dslQuery(\"facts/dslquery\", dslQueryString, nResults)\n}", "func Builder() MeshBuilder {\n\tonce.Do(func() {\n\t\tmeshBuilder = NewMeshBuilder()\n\t})\n\n\treturn meshBuilder\n}", "func (d *DefaulterBuilder) ToInterfaceImplementation() *astmodel.InterfaceImplementation {\n\tgrp, ver := d.resourceName.PackageReference().GroupVersion()\n\n\t// e.g. grp = \"microsoft.network.azure.com\"\n\t// e.g. resource = \"backendaddresspools\"\n\t// e.g. ver = \"v1\"\n\n\tresource := d.resourceName.Name()\n\n\tgrp = strings.ToLower(grp + astmodel.GroupSuffix)\n\tnonPluralResource := strings.ToLower(resource)\n\tresource = strings.ToLower(d.resourceName.Plural().Name())\n\n\t// e.g. \"mutate-microsoft-network-azure-com-v1-backendaddresspool\"\n\t// note that this must match _exactly_ how controller-runtime generates the path\n\t// or it will not work!\n\tpath := fmt.Sprintf(\"/mutate-%s-%s-%s\", strings.ReplaceAll(grp, \".\", \"-\"), ver, nonPluralResource)\n\n\t// e.g. \"default.v123.backendaddresspool.azure.com\"\n\tname := fmt.Sprintf(\"default.%s.%s.%s\", ver, resource, grp)\n\n\tannotation := fmt.Sprintf(\n\t\t\"+kubebuilder:webhook:path=%s,mutating=true,sideEffects=None,\"+\n\t\t\t\"matchPolicy=Exact,failurePolicy=fail,groups=%s,resources=%s,\"+\n\t\t\t\"verbs=create;update,versions=%s,name=%s,admissionReviewVersions=v1\",\n\t\tpath,\n\t\tgrp,\n\t\tresource,\n\t\tver,\n\t\tname)\n\n\tfuncs := []astmodel.Function{\n\t\tNewResourceFunction(\n\t\t\t\"Default\",\n\t\t\td.resource,\n\t\t\td.idFactory,\n\t\t\td.defaultFunction,\n\t\t\tastmodel.NewPackageReferenceSet(astmodel.GenRuntimeReference)),\n\t\tNewResourceFunction(\n\t\t\t\"defaultImpl\",\n\t\t\td.resource,\n\t\t\td.idFactory,\n\t\t\td.localDefault,\n\t\t\tastmodel.NewPackageReferenceSet(astmodel.GenRuntimeReference)),\n\t}\n\n\t// Add the actual individual default functions\n\tfor _, def := range d.defaults {\n\t\tfuncs = append(funcs, def)\n\t}\n\n\treturn astmodel.NewInterfaceImplementation(\n\t\tastmodel.DefaulterInterfaceName,\n\t\tfuncs...).WithAnnotation(annotation)\n}", "func (b *SqliteBuilder) QueryBuilder() QueryBuilder {\n\treturn b.qb\n}", "func (def *Definition) From() source.Sources {\n\treturn def.g.from\n}", "func (per *PersistStringer) PersistImplBuilder(service *Service) string {\n\tvar dbType string\n\tvar backend string\n\tif service.IsSpanner() {\n\t\tdbType = \"spanner.Client\"\n\t\tbackend = \"Spanner\"\n\t} else {\n\t\tdbType = \"sql.DB\"\n\t\tbackend = \"Sql\"\n\t}\n\tprinter := &Printer{}\n\tprinter.P(\n\t\t\"type %sImpl struct{\\nPERSIST *persist_lib.%s\\nFORWARDED RestOf%sHandlers\\n}\\n\",\n\t\tservice.GetName(),\n\t\tNewPersistHelperName(service),\n\t\tservice.GetName(),\n\t)\n\tprinter.P(\"type RestOf%sHandlers interface{\\n\", service.GetName())\n\tfor _, m := range *service.Methods {\n\t\tspannerBi := m.Service.IsSpanner() && m.IsBidiStreaming()\n\t\tif m.GetMethodOption() == nil || spannerBi {\n\t\t\tif m.IsUnary() {\n\t\t\t\tprinter.P(\n\t\t\t\t\t\"%s(ctx context.Context, req *%s) (*%s, error)\\n\",\n\t\t\t\t\tm.GetName(),\n\t\t\t\t\tm.GetInputType(),\n\t\t\t\t\tm.GetOutputType(),\n\t\t\t\t)\n\t\t\t} else if m.IsServerStreaming() {\n\t\t\t\tprinter.P(\n\t\t\t\t\t\"%s(req *%s, stream %s) error\\n\",\n\t\t\t\t\tm.GetName(),\n\t\t\t\t\tm.GetInputType(),\n\t\t\t\t\tNewStreamType(m),\n\t\t\t\t)\n\t\t\t} else {\n\t\t\t\tprinter.P(\n\t\t\t\t\t\"%s(stream %s) error\\n\",\n\t\t\t\t\tm.GetName(),\n\t\t\t\t\tNewStreamType(m),\n\t\t\t\t)\n\t\t\t}\n\t\t}\n\t}\n\tprinter.P(\"}\\n\")\n\tprinter.PA([]string{\n\t\t\"type %sImplBuilder struct {\\n\",\n\t\t\"err error\\n \",\n\t\t\"rest RestOf%sHandlers\\n\",\n\t\t\"queryHandlers *persist_lib.%sQueryHandlers\\n\",\n\t\t\"i *%sImpl\\n\",\n\t\t\"db %s\\n}\\n\",\n\t\t\"func New%sBuilder() *%sImplBuilder {\\nreturn &%sImplBuilder{i: &%sImpl{}}\\n}\\n\",\n\t},\n\t\tservice.GetName(),\n\t\tservice.GetName(),\n\t\tservice.GetName(),\n\t\tservice.GetName(), dbType,\n\t\tservice.GetName(), service.GetName(), service.GetName(), service.GetName(),\n\t)\n\tprinter.PA([]string{\n\t\t\"func (b *%sImplBuilder) WithRestOfGrpcHandlers(r RestOf%sHandlers) *%sImplBuilder {\\n\",\n\t\t\"b.rest = r\\n return b\\n}\\n\",\n\t},\n\t\tservice.GetName(),\n\t\tservice.GetName(),\n\t\tservice.GetName(),\n\t)\n\tprinter.PA([]string{\n\t\t\"func (b *%sImplBuilder) WithPersistQueryHandlers(p *persist_lib.%sQueryHandlers)\",\n\t\t\"*%sImplBuilder {\\n\",\n\t\t\"b.queryHandlers = p\\n return b\\n}\\n\",\n\t},\n\t\tservice.GetName(),\n\t\tservice.GetName(),\n\t\tservice.GetName(),\n\t)\n\n\t// setup default query functions\n\tprinter.PA([]string{\n\t\t\"func (b *%sImplBuilder) WithDefaultQueryHandlers() *%sImplBuilder {\\n\",\n\t\t\"accessor := persist_lib.New%sClientGetter(&b.db)\\n\",\n\t\t\"queryHandlers := &persist_lib.%sQueryHandlers{\\n\",\n\t},\n\t\tservice.GetName(), service.GetName(),\n\t\tbackend,\n\t\tservice.GetName(),\n\t)\n\tfor _, m := range *service.Methods {\n\t\tif m.GetMethodOption() == nil || (m.Service.IsSpanner() && m.IsBidiStreaming()) {\n\t\t\tcontinue\n\t\t}\n\t\tprinter.P(\n\t\t\t\"%s: persist_lib.Default%s(accessor),\\n\",\n\t\t\tNewPersistHandlerName(m),\n\t\t\tNewPersistHandlerName(m),\n\t\t)\n\t}\n\tprinter.P(\"}\\n b.queryHandlers = queryHandlers\\n return b\\n}\\n\")\n\t// fill in holes with defaults\n\tprinter.PA([]string{\n\t\t\"// set the custom handlers you want to use in the handlers\\n\",\n\t\t\"// this method will make sure to use a default handler if\\n\",\n\t\t\"// the handler is nil.\\n\",\n\t\t\"func (b *%sImplBuilder) WithNilAsDefaultQueryHandlers(p *persist_lib.%sQueryHandlers)\",\n\t\t\"*%sImplBuilder {\\n\",\n\t\t\"accessor := persist_lib.New%sClientGetter(&b.db)\\n\",\n\t},\n\t\tservice.GetName(), service.GetName(),\n\t\tservice.GetName(),\n\t\tbackend,\n\t)\n\tfor _, m := range *service.Methods {\n\t\tif m.GetMethodOption() == nil || (m.Service.IsSpanner() && m.IsBidiStreaming()) {\n\t\t\tcontinue\n\t\t}\n\t\tphn := NewPersistHandlerName(m)\n\t\tprinter.P(\n\t\t\t\"if p.%s == nil {\\np.%s = persist_lib.Default%s(accessor)\\n}\\n\",\n\t\t\tphn, phn, phn,\n\t\t)\n\t}\n\tprinter.P(\"b.queryHandlers = p\\n return b\\n}\\n\")\n\n\t// provide the builder with a client\n\tprinter.PA([]string{\n\t\t\"func (b *%sImplBuilder) With%sClient(c *%s) *%sImplBuilder {\\n\",\n\t\t\"b.db = *c\\n return b\\n}\\n\",\n\t},\n\t\tservice.GetName(), backend, dbType, service.GetName(),\n\t)\n\n\tif service.IsSpanner() {\n\t\tprinter.PA([]string{\n\t\t\t\"func (b *%sImplBuilder) WithSpannerURI(ctx context.Context, uri string) *%sImplBuilder {\\n\",\n\t\t\t\"cli, err := spanner.NewClient(ctx, uri)\\n b.err = err\\n b.db = *cli\\n return b\\n}\\n\",\n\t\t},\n\t\t\tservice.GetName(), service.GetName(),\n\t\t)\n\t} else {\n\t\tprinter.PA([]string{\n\t\t\t\"func (b *%sImplBuilder) WithNewSqlDb(driverName, dataSourceName string) *%sImplBuilder {\\n\",\n\t\t\t\"db, err := sql.Open(driverName, dataSourceName)\\n\",\n\t\t\t\"b.err = err\\n b.db = *db\\n return b\\n}\\n\",\n\t\t},\n\t\t\tservice.GetName(), service.GetName(),\n\t\t)\n\t}\n\t// Build method, returns impl, err\n\tprinter.PA([]string{\n\t\t\"func (b *%sImplBuilder) Build() (*%sImpl, error) {\\n\",\n\t\t\"if b.err != nil {\\n return nil, b.err\\n}\\n\",\n\t\t\"b.i.PERSIST = &persist_lib.%s{Handlers: *b.queryHandlers}\\n\",\n\t\t\"b.i.FORWARDED = b.rest\\n\",\n\t\t\"return b.i, nil\\n}\\n\",\n\t},\n\t\tservice.GetName(), service.GetName(),\n\t\tNewPersistHelperName(service),\n\t)\n\t// MustBuild method, returns impl. Can panic.\n\tprinter.PA([]string{\n\t\t\"func (b *%sImplBuilder) MustBuild() *%sImpl {\\n\",\n\t\t\"s, err := b.Build()\\n\",\n\t\t\"if err != nil {\\n panic(\\\"error in builder: \\\" + err.Error())\\n}\\n\",\n\t\t\"return s\\n}\\n\",\n\t},\n\t\tservice.GetName(), service.GetName(),\n\t)\n\treturn printer.String()\n}", "func (l Lambda) Body() Expression {\n\treturn l.body\n}", "func NewBuilder() resolver.Builder {\n\treturn &etcdBuilder{freq: defaultFreq}\n}", "func (b *Builder) Build() (err error) {\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\t// This code allows us to propagate errors without adding lots of checks\n\t\t\t// for `if err != nil` throughout the construction code. This is only\n\t\t\t// possible because the code does not update shared state and does not\n\t\t\t// manipulate locks.\n\t\t\tif ok, e := errorutil.ShouldCatch(r); ok {\n\t\t\t\terr = e\n\t\t\t} else {\n\t\t\t\tpanic(r)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// TODO (rohany): We shouldn't be modifying the semaCtx passed to the builder\n\t// but we unfortunately rely on mutation to the semaCtx. We modify the input\n\t// semaCtx during building of opaque statements, and then expect that those\n\t// mutations are visible on the planner's semaCtx.\n\n\t// Hijack the input TypeResolver in the semaCtx to record all of the user\n\t// defined types that we resolve while building this query.\n\texistingResolver := b.semaCtx.TypeResolver\n\t// Ensure that the original TypeResolver is reset after.\n\tdefer func() { b.semaCtx.TypeResolver = existingResolver }()\n\ttypeTracker := &optTrackingTypeResolver{\n\t\tres: b.semaCtx.TypeResolver,\n\t\tmetadata: b.factory.Metadata(),\n\t}\n\tb.semaCtx.TypeResolver = typeTracker\n\n\t// Special case for CannedOptPlan.\n\tif canned, ok := b.stmt.(*tree.CannedOptPlan); ok {\n\t\tb.factory.DisableOptimizations()\n\t\t_, err := exprgen.Build(b.catalog, b.factory, canned.Plan)\n\t\treturn err\n\t}\n\n\t// Build the memo, and call SetRoot on the memo to indicate the root group\n\t// and physical properties.\n\toutScope := b.buildStmtAtRoot(b.stmt, nil /* desiredTypes */)\n\n\tphysical := outScope.makePhysicalProps()\n\tb.factory.Memo().SetRoot(outScope.expr, physical)\n\treturn nil\n}", "func RunInvalidHTTPDSL(t *testing.T, dsl func()) error {\n\tsetupDSLRun()\n\n\t// run DSL (first pass)\n\tif !eval.Execute(dsl, nil) {\n\t\treturn eval.Context.Errors\n\t}\n\n\t// run DSL (second pass)\n\tif err := eval.RunDSL(); err != nil {\n\t\treturn err\n\t}\n\n\t// expected an error - didn't get one\n\tt.Fatal(\"expected a DSL evaluation error - got none\")\n\n\treturn nil\n}", "func (client *WANDSLLinkConfig1) SetDSLLinkType(NewLinkType string) (err error) {\n\treturn client.SetDSLLinkTypeCtx(context.Background(),\n\t\tNewLinkType,\n\t)\n}", "func NewMeshBuilder() MeshBuilder {\n\timpl := &meshBuilderImpl{\n\t\tLogger: slf4go.Get(\"smf4go\"),\n\t\tregisters: make(map[string]string),\n\t\textensions: make(map[string]Extension),\n\t\tinjector: sdi4go.New(),\n\t}\n\n\timpl.started.Store(false)\n\n\treturn impl\n}", "func (w *WidgetImplement) Parent() Widget {\n\treturn w.parent\n}", "func (RemoteSyslog) Definition() sophos.Definition { return *defRemoteSyslog }", "func (Ca) Definition() sophos.Definition { return *defCa }", "func (Dyndns) Definition() sophos.Definition { return *defDyndns }", "func NewBuilder() Builder {\n\timmutableBuilder := entities.NewImmutableBuilder()\n\treturn createBuilder(immutableBuilder)\n}", "func NewBuilder() Builder {\n\timmutableBuilder := entities.NewImmutableBuilder()\n\treturn createBuilder(immutableBuilder)\n}", "func GlobalSelector() Builder {\n\treturn globalSelector\n}" ]
[ "0.72912025", "0.71394557", "0.52221245", "0.47624788", "0.47548372", "0.46640116", "0.46548972", "0.46548972", "0.46406454", "0.45281145", "0.45281145", "0.44300607", "0.44300607", "0.44285825", "0.44285825", "0.44161195", "0.43295914", "0.43234164", "0.43234164", "0.42773739", "0.42614874", "0.42560044", "0.42310846", "0.41677216", "0.41677216", "0.41646025", "0.41646025", "0.41447166", "0.4141767", "0.4115633", "0.41069233", "0.40876788", "0.4086984", "0.4074484", "0.40621704", "0.40621704", "0.40621704", "0.40621704", "0.40621704", "0.40563658", "0.40539923", "0.40418613", "0.40151972", "0.39930704", "0.39926082", "0.39668998", "0.39604667", "0.39604667", "0.39604667", "0.39604667", "0.39604667", "0.39567384", "0.39566776", "0.3951987", "0.39460188", "0.3927269", "0.39268494", "0.39253667", "0.39182508", "0.39166805", "0.39121336", "0.39099815", "0.3908311", "0.3897748", "0.38936755", "0.3892574", "0.3891989", "0.3886797", "0.38853818", "0.38742584", "0.38742208", "0.3870539", "0.3870539", "0.3869237", "0.3868479", "0.38656843", "0.38593292", "0.38489148", "0.38482884", "0.38456777", "0.38438237", "0.38427532", "0.38414174", "0.38327911", "0.38324422", "0.3831025", "0.38304955", "0.38214415", "0.38153803", "0.380969", "0.3804419", "0.3804044", "0.38020724", "0.3801985", "0.37962466", "0.37939662", "0.37878996", "0.3785215", "0.3785215", "0.37848726" ]
0.7892379
0
Finalize finalizes the relationships.
func (e *Element) Finalize() { for _, rel := range e.Relationships { rel.Finalize() } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g *Graph) Finalize() error {\n\tfor _, node := range g.Nodes {\n\t\tif node.Kind == \"Cluster\" || node.Kind == \"Namespace\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tif _, ok := g.Relationships[node.UID]; ok {\n\t\t\tcontinue\n\t\t}\n\n\t\tif len(node.GetNamespace()) == 0 {\n\t\t\tcluster, err := g.CoreV1().Cluster()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tg.Relationship(cluster, node.Kind, node)\n\t\t\tcontinue\n\t\t}\n\n\t\tmetadata := metav1.ObjectMeta{Name: node.GetNamespace()}\n\t\tnamespace, err := g.CoreV1().Namespace(&v1.Namespace{ObjectMeta: metadata})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tg.Relationship(namespace, node.Kind, node)\n\t}\n\n\treturn nil\n}", "func (cd *ConnectionDetails) Finalize() error {\n\tcd.Dialect = normalizeSynonyms(cd.Dialect)\n\n\tif cd.Options == nil { // for safety\n\t\tcd.Options = make(map[string]string)\n\t}\n\n\tif cd.URL != \"\" {\n\t\tif err := cd.withURL(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif fin, ok := finalizer[cd.Dialect]; ok {\n\t\tfin(cd)\n\t}\n\n\tif DialectSupported(cd.Dialect) {\n\t\tif cd.Database != \"\" || cd.URL != \"\" {\n\t\t\treturn nil\n\t\t}\n\t\treturn errors.New(\"no database or URL specified\")\n\t}\n\treturn errors.Errorf(\"unsupported dialect '%v'\", cd.Dialect)\n}", "func (r *relation) Close() {\n\tfor _, v := range r.mp {\n\t\tv.Close()\n\t}\n}", "func (s *schemaNode) finalize() (err error) {\n\tif s.snType != sntNotSchema {\n\t\tif s.id != \"\" {\n\t\t\tif s.idURL.IsAbs() {\n\t\t\t\ts.baseURI = s.id\n\t\t\t\ts.baseURIObj = s.idURL\n\t\t\t} else {\n\t\t\t\tnode := s.parent\n\t\t\t\tfor node != nil {\n\t\t\t\t\tif node.baseURI != \"\" {\n\t\t\t\t\t\ts.baseURIObj = node.baseURIObj.ResolveReference(s.idURL)\n\t\t\t\t\t\ts.baseURI = s.baseURIObj.String()\n\t\t\t\t\t\tbreak\n\t\t\t\t\t} else {\n\t\t\t\t\t\tnode = node.parent\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif s.baseURI == \"\" {\n\t\t\t\t\tpanic(\"baseURI must not be empty\")\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tnode := s.parent\n\t\t\tfor node != nil {\n\t\t\t\tif node.baseURI != \"\" {\n\t\t\t\t\ts.baseURI = node.baseURI\n\t\t\t\t\ts.baseURIObj = node.baseURIObj\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tnode = node.parent\n\t\t\t}\n\t\t}\n\n\t\ts.setCanonicalURI()\n\t\terr = s.schema.schemaJar.Add(s)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// If here, s must have baseURI\n\tfor _, v := range s.kvMap {\n\t\tif err = v.finalize(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfor _, v := range s.nodeArr {\n\t\tif err = v.finalize(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (m *modelLoader) finalize(itemMap map[reflect.Type][]interface{}) error {\n\t//fill all relationships we can on our items\n\tfor _, f := range m.relationships {\n\t\titems, ok := itemMap[baseType(f.Struct.Type)]\n\t\tif !ok {\n\t\t\t//this relationship isn't in our item map\n\t\t\tcontinue\n\t\t}\n\n\t\tlookup := make(map[string][]reflect.Value)\n\n\t\t//construct a map with possibilities of this relationship\n\t\tfor _, n := range items {\n\t\t\titemVal := reflect.ValueOf(n).Elem()\n\n\t\t\t//build a key for the attributes of this relationship\n\t\t\tvar sb strings.Builder\n\t\t\tfor i, name := range f.Relationship.ForeignFieldNames {\n\t\t\t\tval := itemVal.FieldByName(name).Interface()\n\n\t\t\t\tif valuer, ok := val.(driver.Valuer); ok {\n\t\t\t\t\tvar err error\n\t\t\t\t\tval, err = valuer.Value()\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tsb.WriteString(fmt.Sprintf(\"[%d:%v]\", i, val))\n\t\t\t}\n\n\t\t\tkey := sb.String()\n\t\t\tlookup[key] = append(lookup[key], itemVal.Addr())\n\t\t}\n\n\t\t//go through all models were tracking and fill in this relationship\n\t\tfor _, item := range m.items {\n\t\t\titemVal := reflect.ValueOf(item).Elem()\n\t\t\trelVal := itemVal.FieldByName(f.Name)\n\n\t\t\t//build a key for the attributes of this relationship\n\t\t\tvar sb strings.Builder\n\t\t\tfor i, name := range f.Relationship.AssociationForeignFieldNames {\n\t\t\t\tval := itemVal.FieldByName(name)\n\t\t\t\tif val.Kind() == reflect.Ptr && !val.IsNil() {\n\t\t\t\t\tval = val.Elem()\n\t\t\t\t}\n\n\t\t\t\tkeyValue := val.Interface()\n\t\t\t\tif valuer, ok := keyValue.(driver.Valuer); ok {\n\t\t\t\t\tkeyValue, _ = valuer.Value()\n\t\t\t\t}\n\t\t\t\tsb.WriteString(fmt.Sprintf(\"[%d:%v]\", i, keyValue))\n\t\t\t}\n\n\t\t\tkey := sb.String()\n\t\t\t//find items corresponding to this item for this relationship\n\t\t\tfor _, newVal := range lookup[key] {\n\t\t\t\t//we have items to fill this relationship, fill it based on the struct\n\t\t\t\tif relVal.Kind() == reflect.Slice {\n\t\t\t\t\t//add the result to our slice\n\t\t\t\t\tif relVal.Type().Elem().Kind() != reflect.Ptr {\n\t\t\t\t\t\t//we have a slice of structs so add the struct we're pointing to\n\t\t\t\t\t\tnewVal = newVal.Elem()\n\t\t\t\t\t}\n\n\t\t\t\t\trelVal.Set(reflect.Append(relVal, newVal))\n\t\t\t\t} else {\n\t\t\t\t\t//we don't have a slice so set the item to the first one we have and move on\n\t\t\t\t\tif relVal.Type().Kind() != reflect.Ptr {\n\t\t\t\t\t\tnewVal = newVal.Elem()\n\t\t\t\t\t}\n\n\t\t\t\t\trelVal.Set(newVal)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (_RefundableCrowdsale *RefundableCrowdsaleTransactorSession) Finalize() (*types.Transaction, error) {\n\treturn _RefundableCrowdsale.Contract.Finalize(&_RefundableCrowdsale.TransactOpts)\n}", "func (_RefundableCrowdsale *RefundableCrowdsaleSession) Finalize() (*types.Transaction, error) {\n\treturn _RefundableCrowdsale.Contract.Finalize(&_RefundableCrowdsale.TransactOpts)\n}", "func (_RefundableCrowdsale *RefundableCrowdsaleTransactor) Finalize(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _RefundableCrowdsale.contract.Transact(opts, \"finalize\")\n}", "func (c *Connections) Finalize() (errs []error) {\n\tfor _, connection := range *c {\n\t\tif cerrs := connection.Finalize(); cerrs != nil {\n\t\t\tfor i, err := range cerrs {\n\t\t\t\tcerrs[i] = fmt.Errorf(\"Connection %q: %v\", connection.Name(), err)\n\t\t\t}\n\t\t\terrs = append(errs, cerrs...)\n\t\t}\n\t}\n\treturn errs\n}", "func (p *Project) Finalize() error {\n\terrs := errors.AggregatedError{}\n\tp.Targets = make(TargetNameMap)\n\tfor name, t := range p.MasterFile.Targets {\n\t\tt.Initialize(name, p)\n\t\terrs.Add(p.Targets.Add(t))\n\t}\n\terrs.AddMany(\n\t\tp.Targets.BuildDeps(),\n\t\tp.Targets.CheckCyclicDeps(),\n\t)\n\n\treturn errs.Aggregate()\n}", "func (a *models.App) Finalize() {\n\tvar err error\n\n\ta.DB, err = sql.Close(\"postgres\", connectionString)\n\tutils.ErrCheck(err)\n}", "func (kor *KubernetesOAMRouter) Finalize(canary *flaggerv1.Canary) error {\n\treturn fmt.Errorf(\"OAM router doesn't do finalize\")\n}", "func (c *Container) Finalize() {\n\tc.PrefixTags(\"Element\", \"Container\")\n\tc.Element.Finalize()\n}", "func (t *Tags) Finalize() {\n\tif t.noFinalize {\n\t\treturn\n\t}\n\n\tvalues := t.values\n\tt.values = nil\n\n\tfor i := range values {\n\t\tvalues[i].Finalize()\n\t}\n\n\tif t.pool == nil {\n\t\treturn\n\t}\n\n\tt.pool.PutTags(Tags{values: values})\n}", "func (s *Statement) Finalize() (e error) {\n\treturn SQLiteError(C.sqlite3_finalize(s.cptr))\n}", "func (_Finalizable *FinalizableTransactor) Finalize(opts *bind.TransactOpts, fin bool) (*types.Transaction, error) {\n\treturn _Finalizable.contract.Transact(opts, \"finalize\", fin)\n}", "func (h *HostExpr) Finalize() {\n\tif h.Variables == nil {\n\t\th.Variables = &AttributeExpr{Type: &Object{}}\n\t}\n}", "func (t *Tag) Finalize() {\n\tif t.noFinalize {\n\t\treturn\n\t}\n\tif t.Name != nil {\n\t\tt.Name.Finalize()\n\t\tt.Name = nil\n\t}\n\tif t.Value != nil {\n\t\tt.Value.Finalize()\n\t\tt.Value = nil\n\t}\n}", "func (r *Reaper) Finalize() error {\n\tif r == nil || r.released {\n\t\treturn nil\n\t}\n\tif r.finalized {\n\t\treturn kerror.New(kerror.EIllegal, \"reaper has already called destructors\")\n\t}\n\tdefer func() {\n\t\tr.finalized = true\n\t}()\n\treturn reap(r.destructors...)\n}", "func (c *Controller) Finalize() error {\n\tif c.fixedNs {\n\t\treturn nil\n\t}\n\treturn c.client.Namespaces().Delete(c.namespace.Name, nil)\n}", "func (_CRLv0 *CRLv0Transactor) Finalize(opts *bind.TransactOpts, fin bool) (*types.Transaction, error) {\n\treturn _CRLv0.contract.Transact(opts, \"finalize\", fin)\n}", "func (_OwnerProxyRegistry *OwnerProxyRegistryTransactor) Finalize(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _OwnerProxyRegistry.contract.Transact(opts, \"finalize\")\n}", "func (_FinalizableCrowdsaleImpl *FinalizableCrowdsaleImplTransactor) Finalize(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _FinalizableCrowdsaleImpl.contract.Transact(opts, \"finalize\")\n}", "func (s *SparkCoreAdaptor) Finalize() (errs []error) {\n\treturn\n}", "func (b *BluetoothAdapter) Finalize() (errs []error) {\n\treturn\n}", "func (c *Component) Finalize() {\n\tc.PrefixTags(\"Element\", \"Component\")\n\tc.Element.Finalize()\n}", "func Finalize() error {\n\tperr := C.MPI_Finalize()\n\tif perr != 0 {\n\t\treturn errors.New(\"Error initializing MPI\")\n\t}\n\treturn nil\n}", "func (_TransferProxyRegistry *TransferProxyRegistryTransactor) Finalize(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _TransferProxyRegistry.contract.Transact(opts, \"finalize\")\n}", "func (b *ClientAdaptor) Finalize() (err error) {\n\treturn b.Disconnect()\n}", "func Finalize() error {\n\tC.Py_Finalize()\n\treturn nil\n}", "func (f *Finalizer) Finalize(\n\ttop float64, bottom float64, minSegLen float64,\n\tprimitives *graphics.Primitives) error {\n\tfor _, ll := range f.lifelines {\n\t\tif err := f.finalizeOne(ll, top, bottom, minSegLen, primitives); err != nil {\n\t\t\treturn fmt.Errorf(\"finalizeOne: %v\", err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *TelemetryConfig) Finalize() {\n\tif c == nil {\n\t\treturn\n\t}\n\n\td := DefaultTelemetryConfig()\n\n\tif c.MetricsPrefix == nil {\n\t\tc.MetricsPrefix = d.MetricsPrefix\n\t}\n\n\tc.Stdout.Finalize()\n\tc.DogStatsD.Finalize()\n\tc.Prometheus.Finalize()\n}", "func (a *APIManagedUnbinder) Finalize(ctx context.Context, mg Managed) error {\n\t// TODO(negz): We probably want to delete the managed resource here if its\n\t// reclaim policy is delete, rather than relying on garbage collection, per\n\t// https://github.com/crossplaneio/crossplane/issues/550\n\tmg.SetBindingPhase(v1alpha1.BindingPhaseUnbound)\n\tmg.SetClaimReference(nil)\n\treturn errors.Wrap(IgnoreNotFound(a.client.Update(ctx, mg)), errUpdateManaged)\n}", "func (_CRLv0 *CRLv0TransactorSession) Finalize(fin bool) (*types.Transaction, error) {\n\treturn _CRLv0.Contract.Finalize(&_CRLv0.TransactOpts, fin)\n}", "func (p *Permit) Finalize() {\n\t// nothing to do here anymore, preserving for future use.\n}", "func onFinalize() {\n\t//\n}", "func (ct *ConnectionTransportTLS) Finalize() {\n\tct.mutex.Lock()\n\tdefer ct.mutex.Unlock()\n\tif ct.transport != nil {\n\t\tct.transport.Close()\n\t}\n\tct.transport = ct.stagedTransport\n\tct.stagedTransport = nil\n\tct.srvRemote.Reset()\n}", "func (_FinalizableCrowdsaleImpl *FinalizableCrowdsaleImplSession) Finalize() (*types.Transaction, error) {\n\treturn _FinalizableCrowdsaleImpl.Contract.Finalize(&_FinalizableCrowdsaleImpl.TransactOpts)\n}", "func (_FinalizableCrowdsaleImpl *FinalizableCrowdsaleImplTransactorSession) Finalize() (*types.Transaction, error) {\n\treturn _FinalizableCrowdsaleImpl.Contract.Finalize(&_FinalizableCrowdsaleImpl.TransactOpts)\n}", "func (c *common) Finalized() bool {\n\tnumPreferences := c.preferences.Len()\n\tc.ctx.Log.Verbo(\"Conflict graph has %d preferred transactions\",\n\t\tnumPreferences)\n\treturn numPreferences == 0\n}", "func (_CRLv0 *CRLv0Session) Finalize(fin bool) (*types.Transaction, error) {\n\treturn _CRLv0.Contract.Finalize(&_CRLv0.TransactOpts, fin)\n}", "func (a *I2cBusAdaptor) Finalize() error {\n\ta.mutex.Lock()\n\tdefer a.mutex.Unlock()\n\n\tvar err error\n\tfor _, bus := range a.buses {\n\t\tif bus != nil {\n\t\t\tif e := bus.Close(); e != nil {\n\t\t\t\terr = multierror.Append(err, e)\n\t\t\t}\n\t\t}\n\t}\n\ta.buses = nil\n\treturn err\n}", "func (b *Block) Finalize(endorsements []*endorsement.Endorsement, ts time.Time) error {\n\tif len(b.endorsements) != 0 {\n\t\treturn errors.New(\"the block has been finalized\")\n\t}\n\tb.endorsements = endorsements\n\tb.commitTime = ts\n\n\treturn nil\n}", "func (ut *todoPayload) Finalize() {\n\tvar defaultCompleted = false\n\tif ut.Completed == nil {\n\t\tut.Completed = &defaultCompleted\n\t}\n\tvar defaultCreated, _ = time.Parse(time.RFC3339, \"1978-06-30T10:00:00+09:00\")\n\tif ut.Created == nil {\n\t\tut.Created = &defaultCreated\n\t}\n\tvar defaultModified, _ = time.Parse(time.RFC3339, \"1978-06-30T10:00:00+09:00\")\n\tif ut.Modified == nil {\n\t\tut.Modified = &defaultModified\n\t}\n}", "func (_Finalizable *FinalizableTransactorSession) Finalize(fin bool) (*types.Transaction, error) {\n\treturn _Finalizable.Contract.Finalize(&_Finalizable.TransactOpts, fin)\n}", "func connectionFinalizer(c *Connection) {\n\tc.Close()\n}", "func connectionFinalizer(c *Connection) {\n\tc.Close()\n}", "func (connection *Connection) Close() {\n\tif connection.adabasToData != nil {\n\t\t_ = connection.adabasToData.BackoutTransaction()\n\t\tconnection.adabasToData.Close()\n\t}\n\tif connection.adabasToMap != nil {\n\t\t_ = connection.adabasToMap.BackoutTransaction()\n\t\tconnection.adabasToMap.Close()\n\t}\n}", "func (e *engineImpl) Finalize(\n\tchain engine.ChainReader, header *block.Header,\n\tstate *state.DB, txs []*types.Transaction,\n\treceipts []*types.Receipt, outcxs []*types.CXReceipt,\n\tincxs []*types.CXReceiptsProof, stks []*staking.StakingTransaction,\n) (*types.Block, *big.Int, error) {\n\n\t// Accumulate any block and uncle rewards and commit the final state root\n\t// Header seems complete, assemble into a block and return\n\tpayout, err := AccumulateRewards(\n\t\tchain, state, header, e.Rewarder(), e.Slasher(), e.Beaconchain(),\n\t)\n\tif err != nil {\n\t\treturn nil, nil, ctxerror.New(\"cannot pay block reward\").WithCause(err)\n\t}\n\n\t// Withdraw unlocked tokens to the delegators' accounts\n\t// Only do such at the last block of an epoch\n\tif header.ShardID() == shard.BeaconChainShardID && len(header.ShardState()) > 0 {\n\t\tvalidators, err := chain.ReadValidatorList()\n\t\tif err != nil {\n\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed to read active validators\").WithCause(err)\n\t\t}\n\t\t// Payout undelegated/unlocked tokens\n\t\tfor _, validator := range validators {\n\t\t\twrapper := state.GetStakingInfo(validator)\n\t\t\tif wrapper != nil {\n\t\t\t\tfor i := range wrapper.Delegations {\n\t\t\t\t\tdelegation := &wrapper.Delegations[i]\n\t\t\t\t\ttotalWithdraw := delegation.RemoveUnlockedUndelegations(header.Epoch(), wrapper.LastEpochInCommittee)\n\t\t\t\t\tstate.AddBalance(delegation.DelegatorAddress, totalWithdraw)\n\t\t\t\t}\n\t\t\t\tif err := state.UpdateStakingInfo(validator, wrapper); err != nil {\n\t\t\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed update validator info\").WithCause(err)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\terr = errors.New(\"[Finalize] validator came back empty \" + common2.MustAddressToBech32(validator))\n\t\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed getting validator info\").WithCause(err)\n\t\t\t}\n\t\t}\n\n\t\t// Set the LastEpochInCommittee field for all external validators in the upcoming epoch.\n\t\tnewShardState, err := header.GetShardState()\n\t\tif err != nil {\n\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed to read shard state\").WithCause(err)\n\t\t}\n\t\tprocessed := make(map[common.Address]struct{})\n\t\tfor i := range newShardState.Shards {\n\t\t\tshard := newShardState.Shards[i]\n\t\t\tfor j := range shard.Slots {\n\t\t\t\tslot := shard.Slots[j]\n\t\t\t\tif slot.EffectiveStake != nil { // For external validator\n\t\t\t\t\t_, ok := processed[slot.EcdsaAddress]\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tprocessed[slot.EcdsaAddress] = struct{}{}\n\t\t\t\t\t\twrapper := state.GetStakingInfo(slot.EcdsaAddress)\n\t\t\t\t\t\twrapper.LastEpochInCommittee = newShardState.Epoch\n\n\t\t\t\t\t\tif err := state.UpdateStakingInfo(slot.EcdsaAddress, wrapper); err != nil {\n\t\t\t\t\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed update validator info\").WithCause(err)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\theader.SetRoot(state.IntermediateRoot(chain.Config().IsS3(header.Epoch())))\n\treturn types.NewBlock(header, txs, receipts, outcxs, incxs, stks), payout, nil\n}", "func (a *APIManagedFinalizerRemover) Finalize(ctx context.Context, mg Managed) error {\n\tmeta.RemoveFinalizer(mg, managedFinalizerName)\n\treturn errors.Wrap(a.client.Update(ctx, mg), errUpdateManaged)\n}", "func (c *DedupConfig) Finalize() {\n\tif c.Enabled == nil {\n\t\tc.Enabled = Bool(false ||\n\t\t\tTimeDurationPresent(c.MaxStale) ||\n\t\t\tStringPresent(c.Prefix) ||\n\t\t\tTimeDurationPresent(c.TTL) ||\n\t\t\tTimeDurationPresent(c.BlockQueryWaitTime))\n\t}\n\n\tif c.MaxStale == nil {\n\t\tc.MaxStale = TimeDuration(DefaultDedupMaxStale)\n\t}\n\n\tif c.Prefix == nil {\n\t\tc.Prefix = String(DefaultDedupPrefix)\n\t}\n\n\tif c.TTL == nil {\n\t\tc.TTL = TimeDuration(DefaultDedupTTL)\n\t}\n\n\tif c.BlockQueryWaitTime == nil {\n\t\tc.BlockQueryWaitTime = TimeDuration(DefaultDedupBlockQueryWaitTime)\n\t}\n}", "func (tdb *TyposDatasetBuilder) Finalize() interface{} {\n\t// deduplicate\n\ttypos := make([]Typo, 0, len(tdb.typos))\n\tpairs := map[string]bool{}\n\tfor _, t := range tdb.typos {\n\t\tid := t.Wrong + \"|\" + t.Correct\n\t\tif _, exists := pairs[id]; !exists {\n\t\t\tpairs[id] = true\n\t\t\ttypos = append(typos, t)\n\t\t}\n\t}\n\treturn TyposResult{Typos: typos}\n}", "func (s *GCPCKMSSeal) Finalize(_ context.Context) error {\n\treturn nil\n}", "func (a *APIManagedStatusUnbinder) Finalize(ctx context.Context, mg Managed) error {\n\t// TODO(negz): We probably want to delete the managed resource here if its\n\t// reclaim policy is delete, rather than relying on garbage collection, per\n\t// https://github.com/crossplaneio/crossplane/issues/550\n\tmg.SetBindingPhase(v1alpha1.BindingPhaseUnbound)\n\tmg.SetClaimReference(nil)\n\n\tif err := a.client.Update(ctx, mg); err != nil {\n\t\treturn errors.Wrap(IgnoreNotFound(err), errUpdateManaged)\n\t}\n\n\treturn errors.Wrap(IgnoreNotFound(a.client.Status().Update(ctx, mg)), errUpdateManagedStatus)\n}", "func (_Crowdsale *CrowdsaleTransactor) Finalise(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Crowdsale.contract.Transact(opts, \"finalise\")\n}", "func (conn *Conn) Close() {\n\tif conn.belongsToPool == nil {\n\t\tconn.close()\n\t} else {\n\t\tconn.belongsToPool.Release(conn)\n\t}\n}", "func (_Finalizable *FinalizableSession) Finalize(fin bool) (*types.Transaction, error) {\n\treturn _Finalizable.Contract.Finalize(&_Finalizable.TransactOpts, fin)\n}", "func (res *Results) Finalize() {\n res.Replies = len(res.Took)\n res.min()\n res.max()\n res.avg()\n res.med()\n res.pct()\n\n // Code counts\n for _, code := range res.Code {\n if code < 100 { // ignore\n } else if code < 200 {\n res.Code1xx++\n } else if code < 300 {\n res.Code2xx++\n } else if code < 400 {\n res.Code3xx++\n } else if code < 500 {\n res.Code4xx++\n } else if code < 600 {\n res.Code5xx++\n }\n }\n\n // Error counts\n res.ErrorsTotal = len(res.Errors)\n\n for _, err := range res.Errors {\n e := err.(*url.Error).Err.(*net.OpError).Error()\n if strings.Contains(e, \"connection refused\") {\n res.ErrorsConnRefused++\n } else if strings.Contains(e, \"connection reset\") {\n res.ErrorsConnReset++\n } else if strings.Contains(e, \"connection timed out\") {\n res.ErrorsConnTimeout++\n } else if strings.Contains(e, \"no free file descriptors\") {\n res.ErrorsFdUnavail++\n } else if strings.Contains(e, \"no such host\") {\n res.ErrorsAddrUnavail++\n } else {\n res.ErrorsOther++\n }\n }\n}", "func (kt *SharedKeybaseTransport) Finalize() {\n\tkt.mutex.Lock()\n\tdefer kt.mutex.Unlock()\n\tkt.transport = kt.stagedTransport\n\tkt.stagedTransport = nil\n}", "func (it *ContentRunFinalizeIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (m *Message) Finalize() []byte {\n\t// Update length, which is the first 4 bytes of the header.\n\tusermem.ByteOrder.PutUint32(m.buf, uint32(len(m.buf)))\n\n\t// Align the message. Note that the message length in the header (set\n\t// above) is the useful length of the message, not the total aligned\n\t// length. See net/netlink/af_netlink.c:__nlmsg_put.\n\taligned := alignUp(len(m.buf), linux.NLMSG_ALIGNTO)\n\tm.putZeros(aligned - len(m.buf))\n\treturn m.buf\n}", "func EncodeFinalize(request *Message, db uint32, stmt uint32) {\n\trequest.putUint32(db)\n\trequest.putUint32(stmt)\n\n\trequest.putHeader(bindings.RequestFinalize)\n}", "func (t *ObjectType) Close() error {\n\tif t == nil {\n\t\treturn nil\n\t}\n\tt.mu.Lock()\n\tattributes, d := t.Attributes, t.dpiObjectType\n\tt.Attributes, t.dpiObjectType = nil, nil\n\tt.mu.Unlock()\n\n\tif d == nil {\n\t\treturn nil\n\t}\n\tvar released bool\n\tdefer func() {\n\t\tif !released {\n\t\t\tC.dpiObjectType_release(d)\n\t\t}\n\t}()\n\tif t.CollectionOf != nil {\n\t\tif err := t.CollectionOf.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfor _, attr := range attributes {\n\t\tif err := attr.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif Log != nil {\n\t\tLog(\"msg\", \"ObjectType.Close\", \"name\", t.Name)\n\t}\n\tif C.dpiObjectType_release(d) == C.DPI_FAILURE {\n\t\treturn errors.Errorf(\"error on close object type: %w\", t.getError())\n\t}\n\treleased = true\n\n\treturn nil\n}", "func (c *ProjectFinalizerController) finalize(namespace *v1.Namespace) error {\n\t// if namespace is not terminating, ignore it\n\tif namespace.Status.Phase != v1.NamespaceTerminating {\n\t\treturn nil\n\t}\n\n\t// if we already processed this namespace, ignore it\n\tif projectutil.Finalized(namespace) {\n\t\treturn nil\n\t}\n\n\t// we have removed content, so mark it finalized by us\n\t_, err := projectutil.Finalize(c.client, namespace)\n\treturn err\n}", "func (e *ApplyJoinExec) Close() error {\n\treturn nil\n}", "func (e *NestedLoopJoinExec) Close() error {\n\te.resultRows = nil\n\te.innerRows = nil\n\treturn e.BigExec.Close()\n}", "func (conn *Connection) Free() {\n\tif conn.release {\n\t\t// Py_BEGIN_ALLOW_THREADS\n\t\tconn.Rollback()\n\t\tconn.srvMtx.Lock()\n\t\tC.OCISessionRelease(conn.handle, conn.environment.errorHandle, nil,\n\t\t\t0, C.OCI_DEFAULT)\n\t\t// Py_END_ALLOW_THREADS\n\t\tconn.srvMtx.Unlock()\n\t} else if !conn.attached {\n\t\tif conn.sessionHandle != nil {\n\t\t\t// Py_BEGIN_ALLOW_THREADS\n\t\t\tconn.Rollback()\n\t\t\tconn.srvMtx.Lock()\n\t\t\tC.OCISessionEnd(conn.handle, conn.environment.errorHandle,\n\t\t\t\tconn.sessionHandle, C.OCI_DEFAULT)\n\t\t\t// Py_END_ALLOW_THREADS\n\t\t\tconn.srvMtx.Unlock()\n\t\t}\n\t\tif conn.serverHandle != nil {\n\t\t\tC.OCIServerDetach(conn.serverHandle,\n\t\t\t\tconn.environment.errorHandle, C.OCI_DEFAULT)\n\t\t}\n\t}\n}", "func (u *CryptohomeClient) InstallAttributesFinalize(ctx context.Context) error {\n\tout, err := u.binary.installAttributesFinalize(ctx)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to finalize Install Attributes with the following output %q\", out)\n\t}\n\tif !strings.Contains(out, installAttributesFinalizeSuccessOutput) {\n\t\treturn errors.Errorf(\"failed to finalize Install Attributes, incorrect output message %q\", out)\n\t}\n\treturn nil\n}", "func orientableFinalizer(o *Orientable) {\n\truntime.SetFinalizer(o, func(o *Orientable) { gobject.Unref(o) })\n}", "func (i *IntransitiveActivity) Clean() {\n\ti.BCC = nil\n\ti.Bto = nil\n}", "func (gqm *GQLManager) Close() error {\n\n\t// persist underlying schema data\n\n\tlog.Println(\"...saving gql schema\")\n\ts := gqm.Schema\n\tcontextPath := fmt.Sprintf(\"./contexts/%s/%s/gql\", gqm.UserId, gqm.ContextName)\n\terr := os.MkdirAll(contextPath, os.ModePerm)\n\tif err != nil {\n\t\tlog.Println(\"unable to create schema folder: \", err)\n\t\treturn err\n\t}\n\tf, err := os.Create(contextPath + \"/schema.json\")\n\tif err != nil {\n\t\tlog.Println(\"unable to persist schema:\", err)\n\t\treturn err\n\t}\n\n\tenc := json.NewEncoder(f)\n\tif err := enc.Encode(&s); err != nil {\n\t\tlog.Println(\"json encoding error: \", err)\n\t\treturn err\n\t}\n\n\tlog.Println(\"...schema saved.\")\n\n\treturn nil\n\n}", "func (n *node) finalize(currentNum int) int {\n\tn.nodeNum = currentNum\n\tcurrentNum++\n\n\tfor i := 0; i < 2; i++ {\n\t\tswitch n.children[i].recordType {\n\t\tcase recordTypeFixedNode,\n\t\t\trecordTypeNode:\n\t\t\tcurrentNum = n.children[i].node.finalize(currentNum)\n\t\tdefault:\n\t\t}\n\t}\n\n\treturn currentNum\n}", "func finalizer(a *Adapter) {\n\ta.db.Destroy()\n}", "func (o *Objects) Close() error {\n\to.gcExit <- 1\n\to.Flush()\n\treturn nil\n}", "func TestFinalize(t *testing.T) {\n\tnumGRBefore := runtime.NumGoroutine()\n\t// Create a set of 100 agreement components, and finalize them immediately\n\tfor i := 0; i < 100; i++ {\n\t\tc, _ := agreement.WireAgreement(50)\n\t\tc.FinalizeRound()\n\t}\n\n\t// Ensure we have freed up all of the resources associated with these components\n\tnumGRAfter := runtime.NumGoroutine()\n\t// We should have roughly the same amount of goroutines\n\tassert.InDelta(t, numGRBefore, numGRAfter, 10.0)\n}", "func Finalise() error {\n\tif !globalData.initialised {\n\t\treturn fault.NotInitialised\n\t}\n\n\tglobalData.log.Info(\"shutting down…\")\n\tglobalData.log.Flush()\n\n\t// stop background\n\tglobalData.background.Stop()\n\n\t// release message bus\n\tmessagebus.Bus.Announce.Release()\n\n\tglobalData.log.Info(\"start backing up peer data…\")\n\tif err := receptor.Backup(globalData.backupFile, globalData.receptors.Connectable()); err != nil {\n\t\tglobalData.log.Errorf(\"fail to backup peer data: %s\", err.Error())\n\t}\n\n\t// finally...\n\tglobalData.initialised = false\n\n\tglobalData.log.Info(\"finished\")\n\tglobalData.log.Flush()\n\n\treturn nil\n}", "func (dpos *DummyDpos) Finalize(*types.Block) error { return nil }", "func (mu *MuHash) Finalize() Hash {\n\tblake, err := blake2b.New256([]byte(\"MuHashFinalize\"))\n\tif err != nil {\n\t\tpanic(errors.Wrap(err, \"this should never happen. MuHashFinalize is less than 64 bytes\"))\n\t}\n\tvar serialized SerializedMuHash\n\tmu.serializeInner(&serialized)\n\tvar res Hash\n\tblake.Write(serialized[:])\n\tblake.Sum(res[:0])\n\treturn res\n}", "func Finalize(kubeClient clientset.Interface, namespace *kapi.Namespace) (result *kapi.Namespace, err error) {\n\tif Finalized(namespace) {\n\t\treturn namespace, nil\n\t}\n\n\t// there is a potential for a resource conflict with base kubernetes finalizer\n\t// as a result, we handle resource conflicts in case multiple finalizers try\n\t// to finalize at same time\n\tfor {\n\t\tresult, err = finalizeInternal(kubeClient, namespace, false)\n\t\tif err == nil {\n\t\t\treturn result, nil\n\t\t}\n\n\t\tif !kerrors.IsConflict(err) {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tnamespace, err = kubeClient.Core().Namespaces().Get(namespace.Name)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n}", "func (w *WrappedWriter) Finalize() (_id.ID, error) {\n\tid, err := w.writer.Finalize()\n\n\tif w.onFinalize != nil {\n\t\terr2 := w.onFinalize(id, err)\n\t\tif err2 != nil {\n\t\t\treturn _id.ID{}, err2\n\t\t}\n\t}\n\n\treturn id, err\n}", "func (r *Reaper) Finalized() bool {\n\tif r == nil {\n\t\treturn false\n\t}\n\treturn r.finalized\n}", "func Close() {\n\tlog4go.Debug(\"resources destroy, pid:%v\", os.Getpid())\n\tfor name, r := range resources {\n\t\terr := r.Close()\n\t\tif err != nil {\n\t\t\tlog4go.Error(\"resources[%s] destroy failed:%s\", name, err.Error())\n\t\t} else {\n\t\t\tlog4go.Info(\"resources[%s] destroy finish\", name)\n\t\t}\n\t}\n}", "func (a *Activity) Finish() {\n\tC.ANativeActivity_finish(a.cptr())\n}", "func (joint Joint) Free() {\n\n\tif joint.joint != nil {\n\t\tC.skeltrack_joint_free(joint.joint)\n\t}\n}", "func (iter *BatchObjectIter) Close() {\n\tclose(iter.oidCh)\n}", "func (ic *Context) Finalize() {\n\tfor _, f := range ic.cancelFuncs {\n\t\tf()\n\t}\n\tic.cancelFuncs = nil\n}", "func Finalize(s *gtk.TreeSelection) {\n\tdefer cfg.Env.Trace(\"Finalize\")()\n\tswitch Action {\n\tcase u2f.Registered, u2f.Authenticated:\n\tcase u2f.MissingKey, u2f.RegistrationFailed, u2f.AuthenticationFailed:\n\t\tDeselect(s)\n\t}\n}", "func (pp *PermuteProtocol) Finalize(ciphertext *bfv.Ciphertext, permutation []uint64, crs *ring.Poly, share RefreshShare, ciphertextOut *bfv.Ciphertext) {\n\tpp.Decrypt(ciphertext, share.RefreshShareDecrypt, pp.tmp1)\n\tpp.Permute(pp.tmp1, permutation, pp.tmp1)\n\tpp.Recrypt(pp.tmp1, crs, share.RefreshShareRecrypt, ciphertextOut)\n}", "func (in *IndefiniteObserver) End() {\n\tdefer func() {\n\t\tin.finalizers = nil\n\t\tin.subs = nil\n\t}()\n\n\tfor _, fl := range in.finalizers {\n\t\tfl()\n\t}\n\n\tfor _, sub := range in.subs {\n\t\tif sub.observer == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tsub.End()\n\t}\n}", "func (qm MergeQuerier) Close() error {\n\treturn nil\n}", "func (n *Notary) finalize(acc *wallet.Account, tx *transaction.Transaction, h util.Uint256) error {\n\tnotaryWitness := transaction.Witness{\n\t\tInvocationScript: append([]byte{byte(opcode.PUSHDATA1), 64}, acc.PrivateKey().SignHashable(uint32(n.Network), tx)...),\n\t\tVerificationScript: []byte{},\n\t}\n\tfor i, signer := range tx.Signers {\n\t\tif signer.Account == n.Config.Chain.GetNotaryContractScriptHash() {\n\t\t\ttx.Scripts[i] = notaryWitness\n\t\t\tbreak\n\t\t}\n\t}\n\tnewTx, err := updateTxSize(tx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to update completed transaction's size: %w\", err)\n\t}\n\n\tn.pushNewTx(newTx, h)\n\n\treturn nil\n}", "func (m *PooledWrapper) Finalize(context.Context, ...wrapping.Option) error {\n\treturn nil\n}", "func (c *Container) Close() {\n\tfor _, d := range c.dependencies {\n\t\tif dep, ok := d.(Dependency); ok {\n\t\t\tdep.Close()\n\t\t}\n\t}\n}", "func (c *PrometheusConfig) Finalize() {\n\tif c == nil {\n\t\treturn\n\t}\n\n\td := DefaultPrometheusConfig()\n\n\tif c.Port == nil {\n\t\tc.Port = d.Port\n\t}\n\n\tif c.CachePeriod == nil {\n\t\tc.CachePeriod = d.CachePeriod\n\t}\n}", "func (c *AwsClient) Cleanup() error {\n\tlogrus.Info(\"Deleting own route table\")\n\n\tmyRouteTable, err := c.getRouteTable(\n\t\t[]*ec2.Filter{\n\t\t\t{\n\t\t\t\tName: aws.String(\"tag:name\"),\n\t\t\t\tValues: aws.StringSlice([]string{uniquePrefix}),\n\t\t\t},\n\t\t},\n\t)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to read route table: %s\", err)\n\t}\n\n\tlogrus.Debugf(\"Disassociating route tableID: %s\", *myRouteTable.RouteTableId)\n\tfor _, assoc := range myRouteTable.Associations {\n\t\t_, err := c.aws.DisassociateRouteTable(&ec2.DisassociateRouteTableInput{\n\t\t\tAssociationId: assoc.RouteTableAssociationId,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Failed to disassociate route table %s\", err)\n\t\t}\n\t}\n\n\tlogrus.Debugf(\"Deleting route tableID: %s\", myRouteTable.RouteTableId)\n\t_, err = c.aws.DeleteRouteTable(&ec2.DeleteRouteTableInput{\n\t\tRouteTableId: myRouteTable.RouteTableId,\n\t})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to delete route table %s\", err)\n\t}\n\n\treturn nil\n}", "func (it *ConsortiumManagementProposeIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (_Crowdsale *CrowdsaleTransactorSession) Finalise() (*types.Transaction, error) {\n\treturn _Crowdsale.Contract.Finalise(&_Crowdsale.TransactOpts)\n}", "func (sub *Subscription) End() {\n\tsub.observer = nil\n\n\t// Run finalizers for subscription.\n\tfor _, fl := range sub.handlers {\n\t\tfl()\n\t}\n}", "func (tool *CommandLineTool) Finalize() error {\n\treturn nil\n}", "func (process *Process) Finish() {\n\tprocess.Action.Finish()\n\tprocess.updateInDatabase()\n}" ]
[ "0.74310845", "0.6169468", "0.6152649", "0.5992198", "0.59840745", "0.5963255", "0.59058857", "0.589553", "0.5820891", "0.5803074", "0.5767071", "0.56872946", "0.565301", "0.5554321", "0.5549271", "0.5516171", "0.5505405", "0.54880655", "0.54810625", "0.5472923", "0.5462089", "0.54520273", "0.5445166", "0.5393117", "0.53825176", "0.536733", "0.5329668", "0.53131324", "0.531262", "0.5307311", "0.5294734", "0.52930695", "0.5287907", "0.528648", "0.52607965", "0.5256698", "0.5230217", "0.5219017", "0.5218365", "0.5215748", "0.51598495", "0.51385826", "0.5105946", "0.5077117", "0.507677", "0.5058602", "0.5058602", "0.5049915", "0.501618", "0.5007232", "0.49793816", "0.497833", "0.4968132", "0.4958915", "0.49548316", "0.4951204", "0.49461502", "0.49252295", "0.49081892", "0.49037403", "0.489747", "0.4886186", "0.4878344", "0.48776287", "0.48750535", "0.48602545", "0.48529592", "0.48371345", "0.4832192", "0.48312974", "0.4807002", "0.48019266", "0.47988233", "0.4788223", "0.47837973", "0.47687092", "0.4748838", "0.47478583", "0.4745326", "0.47443327", "0.47378042", "0.47266012", "0.4725547", "0.47190872", "0.47184494", "0.47168", "0.47162998", "0.4712112", "0.47099832", "0.4690674", "0.46889398", "0.4682928", "0.46816337", "0.4680131", "0.46791375", "0.46728134", "0.46697748", "0.46691084", "0.46618384", "0.46603355" ]
0.8170965
0
GetElement returns the underlying element.
func (e *Element) GetElement() *Element { return e }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (e emptyBackEnd) GetElement(i *Element) interface{} {\n\treturn \"not nil!\"\n}", "func (x *Path) GetElement() []string {\n\tif x != nil {\n\t\treturn x.Element\n\t}\n\treturn nil\n}", "func (m *TimeoutMap) GetElement(id string) (Element, bool) {\n\tm.mu.RLock()\n\tdefer m.mu.RUnlock()\n\tvar el Element = nil\n\tvalue, ok := m.elements[id] //先查找\n\tif ok { //如果找得到\n\t\tel = value.GetElement().(Element) //则返回结果\n\t}\n\treturn el, ok\n}", "func (obj *identifier) Element() Element {\n\treturn obj.element\n}", "func (m Elements) Get(t ElementID) []byte {\n\tfor _, pair := range m {\n\t\tif pair.ID == t {\n\t\t\treturn pair.Value\n\t\t}\n\t}\n\treturn nil\n}", "func (rht *RHT) Get(key string) Element {\n\tif queue, ok := rht.elementQueueMapByKey[key]; ok {\n\t\titem := queue.Peek()\n\t\tif item.isRemoved {\n\t\t\treturn nil\n\t\t}\n\t\treturn item.value\n\t}\n\n\treturn nil\n}", "func (q *SimpleQueue) Get() (el interface{}, err error) {\n\tq.Lock()\n\tdefer q.Unlock()\n\tif q.length > 0 {\n\t\tel, q.elements = q.elements[0], q.elements[1:]\n\t\tq.length--\n\t\treturn el, nil\n\t} else {\n\t\treturn nil, fmt.Errorf(\"empty queue\")\n\t}\n}", "func (l *ListItem) GetElem(index int) Element {\n\t//if index will exceed bounds of array\n\tif index >= len(l.contents) {\n\t\t//counter starts at bottom, looks for \"Text\" to stop, or stops at end\n\t\tfor index = 0; index < len(l.contents) && l.contents[index].GetType() != \"span\"; index++ {\n\t\t}\n\t}\n\t//returns requested Element, \"Text\" Element, or last Element.. in that preference depending on existence\n\treturn l.contents[index]\n}", "func (ft *FieldType) GetElem(idx int) string {\n\treturn ft.elems[idx]\n}", "func (e *Element) Get() string {\n\ts := <- e.ch\n\te.ch <- s\n\treturn s\n}", "func (l List) Element() Type {\n\treturn l.element\n}", "func (uni *UniformMatrix3f) GetElement(col, row int, v float32) float32 {\n\n\treturn uni.v[col*3+row]\n}", "func (lruCache *LRUCache) GetElem(k string) (*list.Element, error) {\n\tif elem, exists := lruCache.m[k]; !exists {\n\t\treturn nil, errors.New(\"element does not exist in cache\")\n\t} else {\n\t\t// Element is accessed again therefore move to the front (make most recently used)\n\t\tlruCache.l.MoveToFront(elem)\n\t\treturn elem, nil\n\t}\n\n}", "func (p *idElementPool) get() *idElement {\n\treturn p.pool.Get().(*idElement)\n}", "func (c *CacheManager) GetElement(filename string, element string) (string, error) {\n\treturn filepath.Join(c.cachedDir, filename, element), nil\n}", "func (e *Entity) GetElementByID(string) Object { return nil }", "func getElementBytes(stub shim.ChaincodeStubInterface, elementKey string) ([]byte, error) {\n\telementBytes, err := stub.GetState(elementKey)\n\tif err != nil {\n\t\treturn nil, err\n\t} else if elementBytes == nil {\n\t\treturn nil, fmt.Errorf(\"no element with key %s\", elementKey)\n\t}\n\treturn elementBytes, nil\n}", "func (arr *ArrayADT) GetElement(index int) int {\n\treturn arr.data[index]\n}", "func (t *Tree) Get(rr dns.RR) *Elem {\n\tif t.Root == nil {\n\t\treturn nil\n\t}\n\tn := t.Root.search(rr)\n\tif n == nil {\n\t\treturn nil\n\t}\n\treturn n.Elem\n}", "func (e *Common) Element() string {\n\treturn e.name\n}", "func (l *list) ElementAt(index int) interface{} {\n\treturn l.elements[index]\n}", "func (t *Type) Elem() *Type", "func (e *Element) Get(key string) (interface{}, bool) {\n\tif e.items == nil {\n\t\te.items = make(map[string]interface{})\n\t}\n\tv, ok := e.items[key]\n\treturn v, ok\n}", "func getElementStruct(stub shim.ChaincodeStubInterface, elementKey string, element interface{}) error {\n\telementBytes, err := getElementBytes(stub, elementKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn bytesToStruct(elementBytes, element)\n}", "func (a *Array) Elem() Type { return a.elem }", "func (queue *PriorityQueue) Peek() *Element {\n\treturn queue.At(0)\n}", "func (b *BaseElement) GetParent() (e ElementI) {\n\treturn b.Parent\n}", "func (f *Feature) Element() string {\n\treturn string(f.element)\n}", "func (c *Chan) Elem() Type {\n\treturn c.elem\n}", "func (queue *PriorityQueue) Pop() *Element {\n\treturn heap.Pop(&queue.heap).(*Element)\n}", "func (s *Stack) GetTop() Elem {\n\ts.lock.RLock()\n\tdefer s.lock.RUnlock()\n\treturn s.data[s.top]\n}", "func (list *ArrayList) Get(index int) (interface{}, bool) {\n\tif !list.boundCheck(index) {\n\t\treturn nil, false\n\t}\n\n\treturn list.elements[index], true\n}", "func (e *Element) Item() interface{} { return e.item }", "func (b *Board) Get(p Position) Piece {\n\treturn *b.At(p)\n}", "func (e *Element) Value() interface{} {\n\treturn e.value\n}", "func (e *Element) Value() interface{} {\n\treturn e.value\n}", "func (s Series) Elem(i int) Element {\n\treturn s.elements.Elem(i)\n}", "func (b *BaseElement) GetValue() string {\n\treturn b.Value\n}", "func (p *Pointer) Elem() Type { return p.base }", "func (this *LinkedList) Get(index int) interface{} {\n\tif index < 0 || index >= this.Size() {\n\t\tpanic(\"index out of bound\")\n\t}\n\tpe := this.head\n\tfor i := 0; i < index; i++ {\n\t\tpe = pe.next\n\t}\n\treturn pe.elem\n}", "func (c *Compound) Get() string {\n\treturn c.curRenderable\n}", "func (vm *VirtualMachine) getValueForElement(e quads.Element) interface{} {\n\tif strings.Contains(e.ID(), \"ptr_\") {\n\t\tmemblock := vm.getMemBlockForAddr(e.GetAddr())\n\t\tptrAddr, ok := memblock.Get(e.GetAddr()).(float64)\n\t\tif !ok {\n\t\t\tlog.Fatalf(\"Error: (getValueForElement) couldn't cast index to float64\")\n\t\t}\n\n\t\tauxElement := quads.NewElement(int(ptrAddr), e.ID(), e.Type(), \"\")\n\t\tmemblock = vm.getMemBlockForElement(auxElement)\n\t\trealValue := memblock.Get(int(ptrAddr))\n\t\treturn realValue\n\t}\n\tmemblock := vm.getMemBlockForElement(e)\n\treturn memblock.Get(e.GetAddr())\n}", "func (vector *Vector) ElementAt(index int) interface{} {\n\treturn (*vector)[index]\n}", "func (p *IdlePool) Get() io.Closer {\n\tp.Lock()\n\tdefer p.Unlock()\n\tfor i, c := range p.elems {\n\t\tif c == nil {\n\t\t\tcontinue\n\t\t}\n\t\tp.elems[i] = nil\n\t\treturn c\n\t}\n\treturn nil\n}", "func (t *ListType) Elem() DataType { return t.elem }", "func Get(i interface{}, elem ...interface{}) (val interface{}, ok bool) {\n\tv, ok := valueGet(reflect.ValueOf(i), elem...)\n\tif ok {\n\t\treturn v.Interface(), ok\n\t}\n\treturn nil, false\n}", "func (it *subIterator) Get() Item {\n\tif !it.IsValid() {\n\t\treturn nil\n\t}\n\n\treturn it.iterator.node.item\n}", "func (h Htmldoc) GetElementByID(id string) Element {\n\treturn Element{\n\t\tel: h.el.Call(\"getElementById\", id),\n\t}\n}", "func (v Value) ActiveElement() Value {\n return Value{v.Get(\"activeElement\")}\n}", "func (m *matrixComplex) Get(row int, col int) complex128 { return m.elements[row][col] }", "func (list *ArrayList) Get(index int) (interface{}, bool) {\n\n\tif !list.withinRange(index) {\n\t\treturn nil, false\n\t}\n\n\treturn list.elements[index], true\n}", "func (ss *SyncScratchPad) Get(key storage.Key) (storage.Element, error) {\n\tss.mutex.RLock()\n\tdefer ss.mutex.RUnlock()\n\treturn ss.store.Get(key)\n}", "func (doc *Document) GetElementContent(el *Element) ([]byte, error) {\n\tbuf := make([]byte, el.Size)\n\n\t_, err := io.ReadFull(doc.r, buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tel.Bytes = append(el.Bytes, buf...)\n\treturn buf, nil\n}", "func rcGetArrayElem(p *TCompiler, code *TCode) (*value.Value, error) {\n\tvar v *value.Value = nil\n\tb := p.regGet(code.B)\n\tc := p.regGet(code.C)\n\tif b.Type == value.Array {\n\t\tidx := c.ToInt()\n\t\tv = b.ArrayGet(idx)\n\t\tif v == nil { // 値がなければ作る\n\t\t\tv = value.NewNullPtr()\n\t\t\tb.ArraySet(idx, v)\n\t\t}\n\t\tp.regSet(code.A, v)\n\t} else if b.Type == value.Hash {\n\t\tv = b.HashGet(c.ToString())\n\t\tp.regSet(code.A, v)\n\t}\n\tp.moveNext()\n\treturn v, nil\n}", "func (sl *Slice) Elem(idx int) Ki {\n\treturn (*sl)[idx]\n}", "func (v *Posit8x4) Get(i int) Posit8 { return v.impl[i] }", "func (it *Iterator) Get() []byte {\n\treturn (*Item)(it.iter.Get()).Bytes()\n}", "func (q *Queue) Peek() interface{} {\n\tq.mu.Lock()\n\tdefer q.mu.Unlock()\n\tif el := q.elements.Front(); el != nil {\n\t\treturn el.Value\n\t} else {\n\t\treturn nil\n\t}\n}", "func (q *Queue) Peek() interface{} {\n\tq.mu.Lock()\n\tdefer q.mu.Unlock()\n\tif el := q.elements.Front(); el != nil {\n\t\treturn el.Value\n\t} else {\n\t\treturn nil\n\t}\n}", "func (c *GeoBass) Get(p Point) (interface{}, error) {\n\tc.m.RLock()\n\tdefer c.m.RUnlock()\n\thash, err := getHash(p, c.precision)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to hash input point: %v\", err)\n\t}\n\tvalue, ok := c.items[hash]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"unable to find element: %v\", err)\n\t}\n\treturn value, nil\n}", "func (o *InlineObject54) GetGet() string {\n\tif o == nil || o.Get == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Get\n}", "func (arr *Array) Elem(i int) *Term {\n\treturn arr.elems[i]\n}", "func (s *Storage) Get(key string) (mvcc.Element, bool) {\n\treturn s.memtable.Get(key)\n}", "func (i *Item) Get() interface{} {\n\treturn i.payload\n}", "func (m *Map) Elem() Type { return m.elem }", "func (m *matrixComplex) GetElements() [][]complex128 { return m.elements }", "func (vm *VirtualMachine) getMemBlockForElement(elem quads.Element) Memory {\n\tif strings.Contains(elem.ID(), \"self_\") {\n\t\tstrElements := strings.Split(elem.ID(), \"_\")\n\t\tif len(strElements) < 2 {\n\t\t\tlog.Fatalf(\"Error: (getMemBlockForElement) unexpected object attribute id format\")\n\t\t}\n\t\tobjInstanceAddr, err := strconv.Atoi(strElements[1])\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Error: (getMemBlockForElement) couldn't cast objInstanceAddr to int\")\n\t\t}\n\t\tif objInstanceAddr == -1 {\n\t\t\treturn vm.currentSelf\n\t\t}\n\t\tmemblock := vm.getMemBlockForAddr(objInstanceAddr)\n\t\tobjInstance, ok := memblock.Get(objInstanceAddr).(Memory)\n\t\tif !ok {\n\t\t\tlog.Fatalf(\"Error: (getMemBlockForElement) couldn't cast %v to Memory\", memblock.Get(objInstanceAddr))\n\t\t}\n\t\treturn objInstance\n\t}\n\treturn vm.getMemBlockForAddr(elem.GetAddr())\n}", "func (entry *ElementEntry) Pop() Element {\n\tif entry.IsEmpty() {\n\t\treturn nil\n\t}\n\n\tfirstElement := entry.elementSlice[0]\n\tentry.elementSlice = entry.elementSlice[1:len(entry.elementSlice)]\n\treturn firstElement\n}", "func Elem(v reflect.Value) reflect.Value {\n\tswitch v.Kind() {\n\tcase reflect.Interface, reflect.Ptr:\n\t\treturn Elem(v.Elem())\n\tdefault:\n\t\treturn v\n\t}\n}", "func (list *ArrayList[T]) Get(index int) (T, bool) {\n\tif index < 0 || index >= list.Size() {\n\t\tvar zero T\n\t\treturn zero, false\n\t}\n\treturn list.elems[index], true\n}", "func (list *SkipList) Get(key decimal.Decimal) *Element {\n\tlist.mutex.Lock()\n\tdefer list.mutex.Unlock()\n\n\tvar prev *elementNode = &list.elementNode\n\tvar next *Element\n\n\tfor i := list.maxLevel - 1; i >= 0; i-- {\n\t\tnext = prev.next[i]\n\n\t\tfor next != nil && !list.cmp(key, next.key) {\n\t\t\tprev = &next.elementNode\n\t\t\tnext = next.next[i]\n\t\t}\n\t}\n\n\tif next != nil && list.cmp(next.key, key) {\n\t\treturn next\n\t}\n\n\treturn nil\n}", "func (n *NetworkInterface) Get() (string, error) {\n\tn.mu.Lock()\n\tdefer n.mu.Unlock()\n\t//fmt.Println(\"qu len: \", len(n.Queue))\n\tif len(n.Queue) > 0 {\n\t\ttoReturn := n.Queue[0]\n\t\tn.Queue = n.Queue[1:]\n\t\treturn toReturn, nil\n\t}\n\treturn \"\", errors.New(\"Empty\")\n}", "func (p *Pool) Get() (x interface{}) {\n\tp.init()\n\treturn p.p.Get()\n}", "func (v *atomicReference) Get() interface{} {\n\treturn v.value.Load().(pointer).referent\n}", "func rcGetArrayElemI(p *TCompiler, code *TCode) (*value.Value, error) {\n\tvar v *value.Value = nil\n\trb := p.regGet(code.B)\n\tif rb.Type == value.Array {\n\t\tv := rb.ArrayGet(code.C)\n\t\tp.regSet(code.A, v)\n\t}\n\tp.moveNext()\n\treturn v, nil\n}", "func (q *Queue) Pop() interface{} {\n\tq.mu.Lock()\n\tdefer q.mu.Unlock()\n\tel := q.elements.Front()\n\tif el == nil {\n\t\treturn nil\n\t}\n\treturn q.elements.Remove(el)\n}", "func (q *Queue) Pop() interface{} {\n\tq.mu.Lock()\n\tdefer q.mu.Unlock()\n\tel := q.elements.Front()\n\tif el == nil {\n\t\treturn nil\n\t}\n\treturn q.elements.Remove(el)\n}", "func (l *LRU) Get(ID T,w http.ResponseWriter, r *http.Request) (T, bool) {\r\n//\tdefer l.dump(w,r)\r\n\tval, ok := l.cache[ID]\r\n\tif !ok {\r\n\t\treturn nil, ok\r\n\t}\r\n\tl.link.MoveToFront(val)\r\n\tel := val.Value.(Element)\r\n\treturn el.val, true\r\n}", "func (e *Element) Node() *js.Object {\n\treturn e.node\n}", "func (list *SkipList) Get(key uint64) *Element {\n\tlist.mutex.Lock()\n\tdefer list.mutex.Unlock()\n\n\tvar prev *elementNode = &list.elementNode\n\tvar next *Element\n\n\tfor i := list.maxLevel - 1; i >= 0; i-- {\n\t\tnext = prev.next[i]\n\t\t// search in level i\n\t\tfor next != nil && next.key < key {\n\t\t\tprev = &next.elementNode\n\t\t\tnext = next.next[i]\n\t\t}\n\t}\n\n\tif next != nil && next.key == key {\n\t\treturn next\n\t}\n\treturn nil\n}", "func (q *Queue) Peek() *Element {\n\tif q.Count <= 0 {\n\t\treturn nil\n\t}\n\treturn q.Tail\n}", "func (l *List) Get(v interface{} /* val */) *El {\n\tcur := l.search(v, true, false)\n\n\tif cur == nil || l.less(v, cur.val) {\n\t\treturn nil\n\t}\n\n\treturn cur\n}", "func (f *Fifo) ReadElem() (*Tensor, error) {\n\topts, err := f.GetOptionWithByteSize(ROFifoElemDataSize, C.sizeof_int)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\telemSize, err := ROFifoElemDataSize.Decode(opts, 1)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar metaData unsafe.Pointer\n\tsize := C.uint(elemSize.(uint))\n\tdata := C.malloc(C.sizeof_char * C.ulong(elemSize.(uint)))\n\n\ts := C.ncs_FifoReadElem(f.handle, data, &size, &metaData)\n\n\tif Status(s) != StatusOK {\n\t\treturn nil, fmt.Errorf(\"Failed to read FIFO element: %s\", Status(s))\n\t}\n\n\treturn &Tensor{\n\t\tData: C.GoBytes(data, C.int(size)),\n\t}, nil\n}", "func (f * LinkedList) getElmt(idx int) (*Element) {\n\tif (idx >= f.length) {\n\t\treturn nil\n\t} else {\n\t\ti := 0\n\t\tcurrentElmt := f.start\n\t\tfor i != idx {\n\t\t\tcurrentElmt = currentElmt.next\n\t\t\ti++\n\t\t}\n\t\treturn currentElmt\n\t}\n}", "func (p *BufferPool) Get() []byte {\n\treturn p.p.Get().([]byte)\n}", "func (e *ExprGetElementPtr) Type() types.Type {\n\t// Cache element type if not present.\n\tif e.ElemType == nil {\n\t\tswitch typ := e.Src.Type().(type) {\n\t\tcase *types.PointerType:\n\t\t\te.ElemType = typ.ElemType\n\t\tcase *types.VectorType:\n\t\t\tt, ok := typ.ElemType.(*types.PointerType)\n\t\t\tif !ok {\n\t\t\t\tpanic(fmt.Errorf(\"invalid vector element type; expected *types.Pointer, got %T\", typ.ElemType))\n\t\t\t}\n\t\t\te.ElemType = t.ElemType\n\t\tdefault:\n\t\t\tpanic(fmt.Errorf(\"support for source type %T not yet implemented\", typ))\n\t\t}\n\t}\n\t// Cache type if not present.\n\tif e.Typ == nil {\n\t\te.Typ = gepType(e.ElemType, e.Indices)\n\t}\n\treturn e.Typ\n}", "func (*XMLDocument) GetElementByID(elementId string) (w window.Element) {\n\tmacro.Rewrite(\"$_.getElementById($1)\", elementId)\n\treturn w\n}", "func (c *Capabilities) Get(capability string) *Capability {\n\treturn c.m[capability]\n}", "func (f *Int64Slice) Get() interface{} {\n\treturn *f\n}", "func (e *PollerRegistry) Get(key string) (*poller.Poller, error) {\n\tif val, ok := e.Registry[key]; ok {\n\t\treturn val, nil\n\t}\n\treturn nil, ErrElementNotFound\n}", "func (b *BaseElement) GetX() int32 {\n\treturn b.x\n}", "func (p *Pool) Get(h util.Uint256) *payload.Extensible {\n\tp.lock.RLock()\n\tdefer p.lock.RUnlock()\n\n\telem, ok := p.verified[h]\n\tif !ok {\n\t\treturn nil\n\t}\n\treturn elem.Value.(*payload.Extensible)\n}", "func (hat *HashedArrayTree) Get(index int) (interface{}, error) {\n\tif !hat.validIndex(index) {\n\t\treturn nil, ErrIndexOutOfRange\n\t}\n\tti, li := hat.topIndex(index), hat.leafIndex(index)\n\treturn hat.top[ti][li], nil\n}", "func (r *Wrapper) Get(name string) any {\n\tval, ok := r.Lookup(name)\n\tif !ok {\n\t\treturn nil\n\t}\n\treturn val\n}", "func(list *List) Get(index int) (interface{}, bool) {\n\tif !list.withinRange(index){\n\t\treturn nil, false\n\t}\n\n\treturn list.elements[index], true\n}", "func (g *NormalGrid) Get(row, col int) *Cell {\n\tif row < 0 || row > g.rows-1 {\n\t\treturn nil\n\t}\n\tif col < 0 || col > g.cols-1 {\n\t\treturn nil\n\t}\n\treturn g.grid[row][col]\n}", "func (s *SequenceItemValue) GetValue() interface{} { return s.elements }", "func (c *Compound) GetRGBA() *image.RGBA {\n\tc.lock.RLock()\n\trgba := c.subRenderables[c.curRenderable].GetRGBA()\n\tc.lock.RUnlock()\n\treturn rgba\n}", "func (*XMLDocument) ActiveElement() (activeElement window.Element) {\n\tmacro.Rewrite(\"$_.activeElement\")\n\treturn activeElement\n}", "func (sll *SingleLinkedList) Get(index int) interface{} {\n\treturn sll.getNode(index).value\n}" ]
[ "0.68505967", "0.63871956", "0.6307817", "0.62885356", "0.62882924", "0.62737614", "0.6238111", "0.6231037", "0.61693996", "0.61534953", "0.6141453", "0.6129889", "0.6086939", "0.59937763", "0.592392", "0.59199315", "0.58983856", "0.58807564", "0.58547294", "0.5854002", "0.57835233", "0.5779439", "0.5764602", "0.5761751", "0.5761483", "0.5690564", "0.5638468", "0.56341857", "0.5608866", "0.5601809", "0.56000555", "0.5596377", "0.55933714", "0.55770105", "0.55769306", "0.55769306", "0.5576084", "0.5550253", "0.5516187", "0.55062413", "0.54813564", "0.54802364", "0.5476854", "0.5449202", "0.54450125", "0.5434471", "0.5412789", "0.5412291", "0.53890985", "0.5361564", "0.5351014", "0.5350566", "0.53149414", "0.5314056", "0.53050494", "0.5280886", "0.5279377", "0.526932", "0.526932", "0.52552235", "0.5242224", "0.52397776", "0.52235687", "0.52137136", "0.5208033", "0.52044415", "0.5194561", "0.5191816", "0.51843584", "0.51765436", "0.5176104", "0.5175081", "0.5173545", "0.51567906", "0.51514333", "0.514792", "0.514792", "0.51307285", "0.5127321", "0.5125111", "0.512297", "0.51126575", "0.5111892", "0.51098585", "0.510901", "0.50925326", "0.5091782", "0.5081169", "0.50766104", "0.5074547", "0.506918", "0.5066848", "0.5055808", "0.5054971", "0.504676", "0.50344354", "0.5034102", "0.50310934", "0.50281644", "0.5026628" ]
0.82764304
0
MergeTags adds the given tags. It skips tags already present in e.Tags.
func (e *Element) MergeTags(tags ...string) { e.Tags = mergeTags(e.Tags, tags) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func MergeTags(generalTags []*Tag, infraTags []*Tag) []*Tag {\n\tvar dupMap = make(map[string]bool)\n\tfor _, tag := range generalTags {\n\t\tdupMap[tag.Key] = true\n\t}\n\tfor _, tag := range infraTags {\n\t\tif _, exists := dupMap[tag.Key]; !exists {\n\t\t\tgeneralTags = append(generalTags, tag)\n\t\t}\n\t}\n\treturn generalTags\n}", "func mergeTags(generalTags []*tag, infraTags []*tag) []*tag {\n\tvar dupMap = make(map[string]bool)\n\tfor _, tag := range generalTags {\n\t\tdupMap[tag.key] = true\n\t}\n\tfor _, tag := range infraTags {\n\t\tif _, exists := dupMap[tag.key]; !exists {\n\t\t\tgeneralTags = append(generalTags, tag)\n\t\t}\n\t}\n\treturn generalTags\n}", "func mergeTags(t1, t2 []Tag) []Tag {\n\tn := len(t1) + len(t2)\n\tif n == 0 {\n\t\treturn nil\n\t}\n\n\tout := make([]Tag, 0, n)\n\tout = append(out, t1...)\n\tout = append(out, t2...)\n\n\treturn SortTags(out)\n}", "func mergeTags(existing string, tags []string) string {\n\tif existing == \"\" {\n\t\treturn strings.Join(tags, \",\")\n\t}\n\told := strings.Split(existing, \",\")\n\tvar merged []string\n\tfor _, o := range old {\n\t\tfound := false\n\t\tfor _, tag := range tags {\n\t\t\tif tag == o {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tmerged = append(merged, o)\n\t\t}\n\t}\n\tfor _, tag := range tags {\n\t\tfound := false\n\t\tfor _, o := range merged {\n\t\t\tif tag == o {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tmerged = append(merged, tag)\n\t\t}\n\t}\n\treturn strings.Join(merged, \",\")\n}", "func (auo *ArticleUpdateOne) AddTags(t ...*Tag) *ArticleUpdateOne {\n\tids := make([]int, len(t))\n\tfor i := range t {\n\t\tids[i] = t[i].ID\n\t}\n\treturn auo.AddTagIDs(ids...)\n}", "func (p *plugin) concatTags(tags1 *structtag.Tags, tags2 *structtag.Tags) (*structtag.Tags, error) {\n\tif tags1.Len() == 0 {\n\t\treturn tags2, nil\n\t}\n\tif tags2.Len() == 0 {\n\t\treturn tags1, nil\n\t}\n\n\tfor _, t2 := range tags2.Tags() {\n\t\tvar found bool\n\t\tfor _, t1 := range tags1.Tags() {\n\t\t\tif t1.Key == t2.Key {\n\t\t\t\tif len(t1.Name) == 0 {\n\t\t\t\t\tt1.Name = t2.Name\n\t\t\t\t}\n\t\t\t\tif t1.Options == nil || len(t1.Options) == 0 {\n\t\t\t\t\tt1.Options = t2.Options\n\t\t\t\t}\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tvar err error\n\t\t\ts := tags1.String() + \" \" + t2.String()\n\t\t\ttags1, err = structtag.Parse(s)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"failed to parse tags '%s': %s\", s, err.Error())\n\t\t\t}\n\t\t}\n\t}\n\n\treturn tags1, nil\n}", "func (au *ArticleUpdate) AddTags(t ...*Tag) *ArticleUpdate {\n\tids := make([]int, len(t))\n\tfor i := range t {\n\t\tids[i] = t[i].ID\n\t}\n\treturn au.AddTagIDs(ids...)\n}", "func (cli *OpsGenieAlertV2Client) AddTags(req alertsv2.AddTagsToAlertRequest) (*AsyncRequestResponse, error) {\n\treturn cli.sendAsyncPostRequest(&req)\n}", "func MergeTagSlices(Original []TagInformation, ToAdd []TagInformation) []TagInformation {\n\t//Rules:\n\t//We do not care about meta-tags\n\t//Tags in ToAdd win\n\t//Exlusionary tags win after tags in ToAdd\n\n\t//First, remove duplicates from original that exist in ToAdd\n\tfor Index := 0; Index < len(ToAdd); Index++ {\n\t\tif ToAdd[Index].IsMeta {\n\t\t\tcontinue //Skip Metatags\n\t\t}\n\t\t//Standard tag confirmed, scan for duplicates\n\t\tfor ScanIndex := 0; ScanIndex < len(Original); ScanIndex++ {\n\t\t\tif Original[ScanIndex].IsMeta {\n\t\t\t\tcontinue //Skip comparing metas\n\t\t\t}\n\t\t\tif Original[ScanIndex].ID == ToAdd[Index].ID {\n\t\t\t\t//Remove and resize\n\t\t\t\tOriginal = append(Original[:ScanIndex], Original[ScanIndex+1:]...)\n\t\t\t\t//we just need to continue current scan from the same ScanIndex\n\t\t\t\tScanIndex--\n\t\t\t}\n\t\t}\n\t}\n\n\t//Now we can fall back to RemoveDuplicateTags to cleanup any other issues\n\treturn RemoveDuplicateTags(append(Original, ToAdd...))\n}", "func JoinTags(tags ...Tag) TagSet {\n\tvar result TagSet\n\tresult = append(result, tags...)\n\treturn result\n}", "func (app *Configurable) AddTags(parameters map[string]string) interfaces.AppFunction {\n\ttags, failed := app.processTagsParameter(parameters)\n\tif failed {\n\t\treturn nil\n\t}\n\n\ttransform := transforms.NewTags(tags)\n\treturn transform.AddTags\n}", "func addTags(s selection, args []string) {\n\tif len(args) < 1 {\n\t\tlog.Fatal(`Usage: A addtags <tags> [options]\n<tags>:\tcomma-separated tags to add, e.g. json,xml\n[options]:\toptions to add, e.g. 'json=omitempty'`)\n\t}\n\targuments := []string{\n\t\t\"-file\", s.filename(), \"-modified\", \"-format\", \"json\", \"-line\", s.lineSel(), \"-add-tags\", args[0],\n\t}\n\tif len(args) > 1 {\n\t\targuments = append(arguments, \"-add-options\", args[1])\n\t}\n\tbuf := runWithStdin(s.archive(), \"gomodifytags\", arguments...)\n\tvar out gomodifytagsOutput\n\tif err := json.Unmarshal([]byte(buf), &out); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif err := s.win.Addr(\"%d,%d\", out.Start, out.End); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif _, err := s.win.Write(\"data\", []byte(strings.Join(out.Lines, \"\\n\")+\"\\n\")); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tshowAddr(s.win, s.start)\n\tif len(out.Errs) != 0 {\n\t\tfmt.Fprintln(os.Stderr, strings.Join(out.Errs, \"\\n\"))\n\t}\n}", "func (t Tags) Merge(key string, value ...string) {\n\tfor _, v := range value {\n\t\tcurrent := t.GetAll(key)\n\t\tfound := false\n\t\tfor _, cv := range current {\n\t\t\tif v == cv {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tt.Add(key, v)\n\t\t}\n\t}\n}", "func (s *Sample) AddTags(tags ...string) {\n\tif s.Tags == nil {\n\t\ts.Tags = make([]string, 0)\n\t}\n\ts.Tags = append(s.Tags, tags...)\n}", "func (m Measurement) AddTags(tags map[string]string) Measurement {\n\tfor name, value := range tags {\n\t\tm.AddTag(name, value)\n\t}\n\treturn m\n}", "func TagsDiff(sqsTags map[string]string, newTags map[string]string) (removed, added map[string]string) {\n\tremoved = map[string]string{}\n\tfor k, v := range sqsTags {\n\t\tif _, ok := newTags[k]; !ok {\n\t\t\tremoved[k] = v\n\t\t}\n\t}\n\n\tadded = map[string]string{}\n\tfor k, newV := range newTags {\n\t\tif oldV, ok := sqsTags[k]; !ok || oldV != newV {\n\t\t\tadded[k] = newV\n\t\t}\n\t}\n\treturn\n}", "func (m Tags) Append(tags ...model.Tag) Tags {\n\tresult := Tags{}\n\tfor n, t := range m {\n\t\tresult[n] = t\n\t}\n\tfor _, t := range tags {\n\t\tresult[*t.Scope] = t\n\t}\n\treturn result\n}", "func (ac *ArticleCreate) AddTags(t ...*Tag) *ArticleCreate {\n\tids := make([]int64, len(t))\n\tfor i := range t {\n\t\tids[i] = t[i].ID\n\t}\n\treturn ac.AddTagIDs(ids...)\n}", "func (t TagSet) Merge(more TagSet) TagSet {\n\tmerged := t[:]\n\treturn append(merged, more...)\n}", "func (p *Poly) AddTags(tags map[string]string) {\n\tif tags == nil {\n\t\treturn\n\t}\n\n\tfor key, value := range tags {\n\t\tp.Tags[key] = value\n\t}\n}", "func DiffTags(spec []v1alpha1.Tag, current []ecr.Tag) (addTags []ecr.Tag, remove []string) {\n\taddMap := make(map[string]string, len(spec))\n\tfor _, t := range spec {\n\t\taddMap[t.Key] = t.Value\n\t}\n\tremoveMap := map[string]struct{}{}\n\tfor _, t := range current {\n\t\tif addMap[aws.StringValue(t.Key)] == aws.StringValue(t.Value) {\n\t\t\tdelete(addMap, aws.StringValue(t.Key))\n\t\t\tcontinue\n\t\t}\n\t\tremoveMap[aws.StringValue(t.Key)] = struct{}{}\n\t}\n\tfor k, v := range addMap {\n\t\taddTags = append(addTags, ecr.Tag{Key: aws.String(k), Value: aws.String(v)})\n\t}\n\tfor k := range removeMap {\n\t\tremove = append(remove, k)\n\t}\n\treturn\n}", "func ExampleELB_AddTags_shared00() {\n\tsvc := elb.New(session.New())\n\tinput := &elb.AddTagsInput{\n\t\tLoadBalancerNames: []*string{\n\t\t\taws.String(\"my-load-balancer\"),\n\t\t},\n\t\tTags: []*elb.Tag{\n\t\t\t{\n\t\t\t\tKey: aws.String(\"project\"),\n\t\t\t\tValue: aws.String(\"lima\"),\n\t\t\t},\n\t\t\t{\n\t\t\t\tKey: aws.String(\"department\"),\n\t\t\t\tValue: aws.String(\"digital-media\"),\n\t\t\t},\n\t\t},\n\t}\n\n\tresult, err := svc.AddTags(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase elb.ErrCodeAccessPointNotFoundException:\n\t\t\t\tfmt.Println(elb.ErrCodeAccessPointNotFoundException, aerr.Error())\n\t\t\tcase elb.ErrCodeTooManyTagsException:\n\t\t\t\tfmt.Println(elb.ErrCodeTooManyTagsException, aerr.Error())\n\t\t\tcase elb.ErrCodeDuplicateTagKeysException:\n\t\t\t\tfmt.Println(elb.ErrCodeDuplicateTagKeysException, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (c *Context) AddTags(tags ...string) {\n\tnewTags := make([]string, len(c.tags), len(c.tags)+len(tags))\n\tcopy(newTags, c.tags)\n\tc.tags = append(newTags, tags...)\n}", "func (q *Query) AddTags(raw []string) {\n\thashed := HashTags(raw)\n\tq.Raw = append(q.Raw, raw...)\n\tq.Hashed = append(q.Hashed, hashed...)\n}", "func (a *Client) AddTags(params *AddTagsParams, authInfo runtime.ClientAuthInfoWriter) (*AddTagsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewAddTagsParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"addTags\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/app/rest/builds/{buildLocator}/tags\",\n\t\tProducesMediaTypes: []string{\"application/json\", \"application/xml\", \"text/plain\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/xml\", \"text/plain\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &AddTagsReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*AddTagsOK), nil\n\n}", "func (client *Client) AddTags(args *AddTagsArgs) error {\n\tresponse := AddTagsResponse{}\n\terr := client.Invoke(\"AddTags\", args, &response)\n\treturn err\n}", "func (th *TagHandler) uniqueFilterAndAddTags(mName string, mHostname *gostatsd.Source, mTags *gostatsd.Tags) bool {\n\tif len(th.filters) == 0 {\n\t\t*mTags = uniqueTags(*mTags, th.tags)\n\t\treturn true\n\t}\n\n\tdropTags := map[string]struct{}{}\n\n\tfor _, filter := range th.filters {\n\t\tif len(filter.MatchMetrics) > 0 && !filter.MatchMetrics.MatchAny(mName) { // returns false if nothing present\n\t\t\t// name doesn't match an include, stop\n\t\t\tcontinue\n\t\t}\n\n\t\t// this list may be empty, and therefore return false\n\t\tif filter.ExcludeMetrics.MatchAny(mName) { // returns false if nothing present\n\t\t\t// name matches an exclude, stop\n\t\t\tcontinue\n\t\t}\n\n\t\tif len(filter.MatchTags) > 0 && !filter.MatchTags.MatchAnyMultiple(*mTags) { // returns false if either list is empty\n\t\t\t// no tags match\n\t\t\tcontinue\n\t\t}\n\n\t\tif filter.DropMetric {\n\t\t\treturn false\n\t\t}\n\n\t\tfor _, dropFilter := range filter.DropTags {\n\t\t\tfor _, tag := range *mTags {\n\t\t\t\tif dropFilter.Match(tag) {\n\t\t\t\t\tdropTags[tag] = present\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tif filter.DropHost {\n\t\t\t*mHostname = \"\"\n\t\t}\n\t}\n\n\t*mTags = uniqueTagsWithSeen(dropTags, *mTags, th.tags)\n\treturn true\n}", "func Tags(err error) []Tag {\n\treturn deepAppendTags(nil, err)\n}", "func SortTags(tags []v1alpha1.Tag, ecrTags []ecr.Tag) {\n\tsort.Slice(tags, func(i, j int) bool {\n\t\treturn tags[i].Key < tags[j].Key\n\t})\n\n\tsort.Slice(ecrTags, func(i, j int) bool {\n\t\treturn *ecrTags[i].Key < *ecrTags[j].Key\n\t})\n}", "func (sp *Space) AddTags(tags ...string) {\n\tfor _, shape := range *sp {\n\t\tshape.AddTags(tags...)\n\t}\n}", "func (d *Default) AddTags(ctx context.Context, sp *spec.Spec, opts *DefaultOptions) error {\n\t// Tags are applied to the schemas, if needed.\n\t// With refNames we can also apply the same tags\n\t// to the schemas that are referenced, but are already extracted,\n\t// in which case setting tags in place is not enough, and has no effect.\n\taddTagsFunc := func(tags map[string][]string, refNames *[]string) spec.SchemaWalker {\n\t\treturn func(path spec.SchemaPath) error {\n\t\t\tif len(tags) == 0 {\n\t\t\t\treturn errors.New(\"should stop\")\n\t\t\t}\n\t\t\tsm := path.Last()\n\n\t\t\tif sm.Tags == nil {\n\t\t\t\tsm.Tags = make(map[string][]string, len(opts.Tags))\n\t\t\t}\n\n\t\t\tactualTgs := make(map[string][]string)\n\n\t\t\tfor k, tg := range tags {\n\t\t\t\tnewTags := make([]string, len(tg))\n\t\t\t\tcopy(newTags, tg)\n\n\t\t\t\tactualTgs[k] = newTags\n\t\t\t}\n\n\t\t\tfor k, tg := range sm.Tags {\n\t\t\t\tnewTags := make([]string, len(tg))\n\t\t\t\tcopy(newTags, tg)\n\n\t\t\t\tactualTgs[k] = newTags\n\t\t\t}\n\n\t\t\tfor _, tag := range actualTgs {\n\t\t\t\tname := sm.FieldName\n\n\t\t\t\tif name == \"\" {\n\t\t\t\t\tname = sm.OriginalName\n\t\t\t\t}\n\n\t\t\t\tfor i, tagPart := range tag {\n\t\t\t\t\ttempl, err := template.New(\"tag\").Funcs(sprig.TxtFuncMap()).Parse(tagPart)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn fmt.Errorf(\"unexpected error when parsing tags template: %w\", err)\n\t\t\t\t\t}\n\n\t\t\t\t\ttagBuf := &bytes.Buffer{}\n\t\t\t\t\terr = templ.Execute(tagBuf, &TagTemplateValues{\n\t\t\t\t\t\tFieldName: name,\n\t\t\t\t\t\tType: sm.Name,\n\t\t\t\t\t\tDescription: sm.Description,\n\t\t\t\t\t})\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn fmt.Errorf(\"unexpected error when parsing tags template: %w\", err)\n\t\t\t\t\t}\n\n\t\t\t\t\ttag[i] = tagBuf.String()\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tsm.Tags = actualTgs\n\n\t\t\tif sm.Name != \"\" {\n\t\t\t\tif refNames != nil {\n\t\t\t\t\t*refNames = append(*refNames, sm.Name)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tfor _, s := range sp.Schemas {\n\t\terr := s.Walk(addTagsFunc(opts.Tags, nil), true)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfor _, p := range sp.Paths {\n\t\tfor _, o := range p.Operations {\n\t\t\terr := d.addTagsToOperation(ctx, sp, o, addTagsFunc, opts)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tfor _, cb := range o.Callbacks {\n\t\t\t\tfor _, cbPath := range cb {\n\t\t\t\t\tfor _, cbOp := range cbPath.Operations {\n\t\t\t\t\t\terr := d.addTagsToOperation(ctx, sp, cbOp, addTagsFunc, opts)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\treturn err\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\t}\n\n\treturn nil\n}", "func updateTags(nTags ...string) {\n\td := map[string]bool{}\n\tnewTags := append(_tags, nTags...)\n\tfor _, t := range newTags {\n\t\td[strings.ToLower(t)] = true\n\t}\n\n\t_newTags := make([]string, 0)\n\tfor k := range d {\n\t\t_newTags = append(_newTags, k)\n\t}\n\t_tags = _newTags\n\t// Update tags to disk\n\tutil.StoreJSON(ml.FileOthers, \"tags\", _tags)\n}", "func (o *TransactionSplit) SetTags(v []string) {\n\to.Tags = v\n}", "func (v *GetVero) AddTags(id interface{}, tags ...interface{}) error {\n\n\tuID, err := checkID(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tt, err := checkTags(tags...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ta := getveroTag{\n\t\tAuthToken: v.AuthToken,\n\t\tID: uID,\n\t\tAdd: t,\n\t\tRemove: nil,\n\t}\n\tb, err := json.Marshal(a)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = sendToVeroPut(b, tagsURL)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n\n}", "func (z *Zone) AddTags(tags ...string) int {\n\ts := NewSet(z.Tags...)\n\tn := len(s)\n\ts.Add(tags...)\n\td := len(s) - n\n\tif d != 0 {\n\t\tz.Tags = s.Values()\n\t}\n\treturn d\n}", "func (m *Manager) AddTag(tags tag.Tags) error {\n\tm.mutex.Lock()\n\tdefer m.mutex.Unlock()\n\tadded := false\n\tfor _, tag := range tags {\n\t\tadded = added || m.tags.Add(tag)\n\t}\n\tif !added {\n\t\treturn nil\n\t}\n\treturn m.self.WriteTag(m.tags)\n}", "func (m NodeMetadata) Extend(name, label, description string, tags ...string) NodeMetadata {\n\tif len(m.Description) == 0 {\n\t\tm.Description = description\n\t}\n\tif len(m.Label) == 0 {\n\t\tm.Label = label\n\t}\n\tif len(m.Name) == 0 {\n\t\tm.Name = name\n\t}\n\tfor _, t := range tags {\n\t\tvar found bool\n\t\tfor _, ft := range m.Tags {\n\t\t\tif ft == t {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tm.Tags = append(m.Tags, t)\n\t\t}\n\t}\n\treturn m\n}", "func (o *GroupReplaceRequest) SetTags(v []string) {\n\to.Tags = &v\n}", "func (s *server) SetExtraTags(tags []string) {\n\ts.extraTags = tags\n}", "func (o TagsList) Append(objects ...elemental.Identifiable) elemental.Identifiables {\n\n\tout := append(TagsList{}, o...)\n\tfor _, obj := range objects {\n\t\tout = append(out, obj.(*Tag))\n\t}\n\n\treturn out\n}", "func (auo *ArticleUpdateOne) RemoveTags(t ...*Tag) *ArticleUpdateOne {\n\tids := make([]int, len(t))\n\tfor i := range t {\n\t\tids[i] = t[i].ID\n\t}\n\treturn auo.RemoveTagIDs(ids...)\n}", "func (o *EventAttributes) SetTags(v []string) {\n\to.Tags = v\n}", "func (e *Event) AddTag(tag string) {\n\tdata := e.Data()\n\ttags := data[\"tags\"].(Tags)\n\tlength := len(tags)\n\tidx := sort.SearchStrings(tags, tag)\n\tif idx >= length {\n\t\tdata[\"tags\"] = append(tags, tag)\n\t} else if tags[idx] != tag && length < 1024 {\n\t\t// Max 1024 tags - this also prevents CWE-190 from CodeQL (Size computation for allocation may overflow) due to length+1 without bound check\n\t\tif length+1 > cap(tags) {\n\t\t\toldTags := tags\n\t\t\ttags = make(Tags, length+1)\n\t\t\tcopy(tags, oldTags[:idx])\n\t\t\tcopy(tags[idx+1:], oldTags[idx:])\n\t\t} else {\n\t\t\ttags = tags[:length+1]\n\t\t\tcopy(tags[idx+1:], tags[idx:])\n\t\t}\n\t\ttags[idx] = tag\n\t\tdata[\"tags\"] = tags\n\t}\n}", "func elbTags(tags map[string]string) []*elb.Tag {\n\tvar e []*elb.Tag\n\n\tfor k, v := range tags {\n\t\te = append(e, elbTag(k, v))\n\t}\n\n\treturn e\n}", "func ExcludeTags(rii RegInvImage, excludedTags map[Tag]bool) RegInvImage {\n\tfiltered := make(RegInvImage)\n\tfor imageName, digestTags := range rii {\n\t\tfor digest, tags := range digestTags {\n\t\t\tfor _, tag := range tags {\n\t\t\t\tif _, excludeMe := excludedTags[tag]; excludeMe {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif filtered[imageName] == nil {\n\t\t\t\t\tfiltered[imageName] = make(DigestTags)\n\t\t\t\t}\n\t\t\t\tfiltered[imageName][digest] = append(\n\t\t\t\t\tfiltered[imageName][digest],\n\t\t\t\t\ttag)\n\t\t\t}\n\t\t}\n\t}\n\treturn filtered\n}", "func (o EntityRecognizerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *EntityRecognizer) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (me *NamedExecHelper) SetIgnoreTags(tags ...string) *NamedExecHelper {\r\n\tme.ignoreTags = append(me.ignoreTags, tags...)\r\n\treturn me\r\n}", "func CompareTags(tags []v1alpha1.Tag, ecrTags []ecr.Tag) bool {\n\tif len(tags) != len(ecrTags) {\n\t\treturn false\n\t}\n\n\tSortTags(tags, ecrTags)\n\n\tfor i, t := range tags {\n\t\tif t.Key != aws.StringValue(ecrTags[i].Key) || t.Value != aws.StringValue(ecrTags[i].Value) {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func (s *Wave) SetTags(v map[string]*string) *Wave {\n\ts.Tags = v\n\treturn s\n}", "func (o *ShortenBitlinkBodyAllOf) SetTags(v []string) {\n\to.Tags = &v\n}", "func Tags(tags map[string]string) ChangeOption {\n\treturn changeOption{\n\t\tapplier: applierFunc(\n\t\t\tfunc(caller caller, co interface{}) {\n\t\t\t\tco.(*secret.UpdateSetRequest).Tags = tags\n\t\t\t},\n\t\t),\n\t}\n}", "func (o *FiltersNet) SetTags(v []string) {\n\to.Tags = &v\n}", "func (t Tags) Import(tags map[string][]string) {\n\tfor k, v := range tags {\n\t\tt.Merge(k, v...)\n\t}\n}", "func (el *ZapEventLogger) SetTags(ctx context.Context, tags map[string]interface{}) {\n\tspan := opentrace.SpanFromContext(ctx)\n\tif span == nil {\n\t\t_, file, line, _ := runtime.Caller(1)\n\t\tlog.Errorf(\"SetTags with no Span in context called on %s:%d\", path.Base(file), line)\n\t\treturn\n\t}\n\tfor k, v := range tags {\n\t\tspan.SetTag(k, v)\n\t}\n}", "func (o *FiltersVirtualGateway) SetTags(v []string) {\n\to.Tags = &v\n}", "func mergeRightTags(tagsLeft, tagsRight map[string]string) map[string]string {\n\tif tagsLeft == nil && tagsRight == nil {\n\t\treturn nil\n\t}\n\tif len(tagsRight) == 0 {\n\t\treturn tagsLeft\n\t}\n\tif len(tagsLeft) == 0 {\n\t\treturn tagsRight\n\t}\n\n\tresult := make(map[string]string, len(tagsLeft)+len(tagsRight))\n\tfor k, v := range tagsLeft {\n\t\tresult[k] = v\n\t}\n\tfor k, v := range tagsRight {\n\t\tresult[k] = v\n\t}\n\treturn result\n}", "func (o *NewData) SetTags(v []string) {\n\to.Tags = &v\n}", "func processTags(tags []string) []string {\n\tif tags == nil {\n\t\treturn nil\n\t}\n\tprocessedTags := []string{}\n\tusedTags := make(map[string]bool)\n\tfor _, tag := range tags {\n\t\ttag = strings.TrimSpace(tag)\n\t\tif len(tag) > 0 && !usedTags[tag] {\n\t\t\tprocessedTags = append(processedTags, tag)\n\t\t\tusedTags[tag] = true\n\t\t}\n\t}\n\treturn processedTags\n}", "func (o EventIntegrationOutput) Tags() EventIntegrationTagArrayOutput {\n\treturn o.ApplyT(func(v *EventIntegration) EventIntegrationTagArrayOutput { return v.Tags }).(EventIntegrationTagArrayOutput)\n}", "func (o *FiltersVmGroup) SetTags(v []string) {\n\to.Tags = &v\n}", "func (o *GetIPAMsubnetsParams) SetTags(tags *string) {\n\to.Tags = tags\n}", "func (au *ArticleUpdate) RemoveTags(t ...*Tag) *ArticleUpdate {\n\tids := make([]int, len(t))\n\tfor i := range t {\n\t\tids[i] = t[i].ID\n\t}\n\treturn au.RemoveTagIDs(ids...)\n}", "func (o ExtensionAssociationOutput) Tags() ExtensionAssociationTagArrayOutput {\n\treturn o.ApplyT(func(v *ExtensionAssociation) ExtensionAssociationTagArrayOutput { return v.Tags }).(ExtensionAssociationTagArrayOutput)\n}", "func (s *Application) SetTags(v map[string]*string) *Application {\n\ts.Tags = v\n\treturn s\n}", "func (tuo *TagUpdateOne) AddGroupTags(g ...*GroupTag) *TagUpdateOne {\n\tids := make([]int, len(g))\n\tfor i := range g {\n\t\tids[i] = g[i].ID\n\t}\n\treturn tuo.AddGroupTagIDs(ids...)\n}", "func (options *ReplaceJobOptions) SetTags(tags []string) *ReplaceJobOptions {\n\toptions.Tags = tags\n\treturn options\n}", "func (o *PostMultiNodeDeviceParams) SetTags(tags *string) {\n\to.Tags = tags\n}", "func (o AnycastEipAddressOutput) Tags() pulumi.MapOutput {\n\treturn o.ApplyT(func(v *AnycastEipAddress) pulumi.MapOutput { return v.Tags }).(pulumi.MapOutput)\n}", "func (options *ReplaceSharedDatasetOptions) SetTags(tags []string) *ReplaceSharedDatasetOptions {\n\toptions.Tags = tags\n\treturn options\n}", "func (e *Element) PrefixTags(tags ...string) {\n\tprefix := strings.Join(tags, \",\")\n\tif e.Tags == \"\" {\n\t\te.Tags = prefix\n\t\treturn\n\t}\n\te.Tags = mergeTags(prefix, strings.Split(e.Tags, \",\"))\n}", "func (o *Job) SetTags(v map[string]string) {\n\to.Tags = &v\n}", "func (o *Member) SetTags(v map[string]string) {\n\to.Tags = &v\n}", "func (o EipAddressOutput) Tags() pulumi.MapOutput {\n\treturn o.ApplyT(func(v *EipAddress) pulumi.MapOutput { return v.Tags }).(pulumi.MapOutput)\n}", "func (s *ArchiveWaveOutput) SetTags(v map[string]*string) *ArchiveWaveOutput {\n\ts.Tags = v\n\treturn s\n}", "func (o *SyntheticsBrowserTest) SetTags(v []string) {\n\to.Tags = v\n}", "func (o *DataExportQuery) SetTags(v []string) {\n\to.Tags = &v\n}", "func (o ListenerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v Listener) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (m *MockCloudManager) AttachTags(tagsId []string, resourceId string, resourceType string) (err error) {\n\treturn nil\n}", "func (s *EvaluationForm) SetTags(v map[string]*string) *EvaluationForm {\n\ts.Tags = v\n\treturn s\n}", "func AddTags(ctx context.Context, tags ...Tag) context.Context {\n\tif tagsContainer, ok := ctx.Value(tagsKey).(*tagsContainer); ok && tagsContainer != nil {\n\t\ttagsContainer.Tags = append(tagsContainer.Tags, tags...)\n\t\treturn ctx\n\t}\n\treturn context.WithValue(ctx, tagsKey, &tagsContainer{\n\t\tTags: tags,\n\t})\n}", "func (t tagger) deleteTags(modFullTags []string) error {\n\tfor _, modFullTag := range modFullTags {\n\t\tfmt.Printf(\"Deleting tag %v\\n\", modFullTag)\n\t\tcmd := exec.Command(\"git\", \"tag\", \"-d\", modFullTag)\n\t\tif output, err := cmd.CombinedOutput(); err != nil {\n\t\t\treturn fmt.Errorf(\"could not delete tag %v:\\n%v (%v)\", modFullTag, string(output), err)\n\t\t}\n\t}\n\treturn nil\n}", "func Tags(list ...namedTagEventList) map[string]imageapi.TagEventList {\n\tm := make(map[string]imageapi.TagEventList, len(list))\n\tfor _, tag := range list {\n\t\tm[tag.name] = tag.events\n\t}\n\treturn m\n}", "func (bh *Header) Tags(fn func(t Tag, value string)) {\n\tif fn == nil {\n\t\treturn\n\t}\n\tfn(versionTag, bh.Version)\n\tif bh.SortOrder != UnknownOrder {\n\t\tfn(sortOrderTag, bh.SortOrder.String())\n\t}\n\tif bh.GroupOrder != GroupNone {\n\t\tfn(groupOrderTag, bh.GroupOrder.String())\n\t}\n\tfor _, tp := range bh.otherTags {\n\t\tfn(tp.tag, tp.value)\n\t}\n}", "func (o AppOutput) Tags() AppTagArrayOutput {\n\treturn o.ApplyT(func(v *App) AppTagArrayOutput { return v.Tags }).(AppTagArrayOutput)\n}", "func (e *ErrorEntry) SetTags(tags []string) *ErrorEntry {\n\te.Tags = tags\n\treturn e\n}", "func (tu *TransactionUpdate) AddTag(t ...*Tag) *TransactionUpdate {\n\tids := make([]uuid.UUID, len(t))\n\tfor i := range t {\n\t\tids[i] = t[i].ID\n\t}\n\treturn tu.AddTagIDs(ids...)\n}", "func MissingTags(srcTags []string, dstTags []string) []string {\n\tmissingTags := []string{}\n\n\tfor _, srcTag := range srcTags {\n\t\tif !stringInSlice(srcTag, dstTags) {\n\t\t\tmissingTags = append(missingTags, srcTag)\n\t\t}\n\t}\n\n\treturn missingTags\n}", "func (tuo *TransactionUpdateOne) AddTag(t ...*Tag) *TransactionUpdateOne {\n\tids := make([]uuid.UUID, len(t))\n\tfor i := range t {\n\t\tids[i] = t[i].ID\n\t}\n\treturn tuo.AddTagIDs(ids...)\n}", "func (options *UpdateActionOptions) SetTags(tags []string) *UpdateActionOptions {\n\toptions.Tags = tags\n\treturn options\n}", "func (cl *ConfigLocks) Merge(c2 *ConfigLocks) {\n\tcl.Bits |= c2.Bits\n\tfor t := range c2.Tags {\n\t\tcl.Tags[t] = struct{}{}\n\t}\n}", "func (s *UpdateWaveOutput) SetTags(v map[string]*string) *UpdateWaveOutput {\n\ts.Tags = v\n\treturn s\n}", "func (b *Bag) Merge(from *Bag) {\n\tb.Add(from.Items()...)\n}", "func Tags(tags map[string]string) Option {\n\treturn optionFunc(func(r *runtime) {\n\t\tr.tags = tags\n\t})\n}", "func (f *flags) Tags() []string {\n\tif f.opt.Tags == modeFlagUnUsed {\n\t\treturn nil\n\t}\n\treturn f.exclude(f.getvalue(f.opt.Tags, \"tags\", \"t\"),\n\t\tf.getvalue(f.opt.ExcludeTags, \"nt\", \"\"))\n\n}", "func (s *CreateWaveOutput) SetTags(v map[string]*string) *CreateWaveOutput {\n\ts.Tags = v\n\treturn s\n}", "func (m *Client) AddOrganizationTags(arg0 context.Context, arg1 int64, arg2 []zendesk.Tag) ([]zendesk.Tag, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"AddOrganizationTags\", arg0, arg1, arg2)\n\tret0, _ := ret[0].([]zendesk.Tag)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (s *CreateControlPanelInput) SetTags(v map[string]*string) *CreateControlPanelInput {\n\ts.Tags = v\n\treturn s\n}", "func (options *CreateSharedDatasetOptions) SetTags(tags []string) *CreateSharedDatasetOptions {\n\toptions.Tags = tags\n\treturn options\n}", "func (s *CreateWaveInput) SetTags(v map[string]*string) *CreateWaveInput {\n\ts.Tags = v\n\treturn s\n}", "func (o *ImageImportManifest) SetTags(v []string) {\n\to.Tags = v\n}" ]
[ "0.69267917", "0.66822976", "0.6641655", "0.6398739", "0.60571116", "0.60490227", "0.59776914", "0.5947697", "0.5940782", "0.59327805", "0.58563393", "0.583582", "0.5833119", "0.58270645", "0.5826986", "0.5785081", "0.57704574", "0.5765203", "0.57195115", "0.57146055", "0.57086104", "0.5691742", "0.5623111", "0.5600554", "0.55874634", "0.5557007", "0.548824", "0.54658127", "0.5448353", "0.5400391", "0.53794", "0.537355", "0.5338179", "0.53205436", "0.5303035", "0.52967846", "0.52917403", "0.52900934", "0.52589965", "0.5251154", "0.52222276", "0.52206266", "0.5219615", "0.5218611", "0.52174765", "0.5187616", "0.5183938", "0.51728386", "0.5168859", "0.516702", "0.51647025", "0.51627976", "0.5156918", "0.5150673", "0.51466215", "0.5140712", "0.512206", "0.51140815", "0.5112594", "0.5109212", "0.510734", "0.51066124", "0.5101115", "0.5096564", "0.50922996", "0.5091608", "0.50900215", "0.5086761", "0.50813794", "0.5076521", "0.50733787", "0.50652456", "0.5057862", "0.5057789", "0.50519085", "0.50509876", "0.50434893", "0.50416356", "0.50369227", "0.50365746", "0.5036544", "0.5026759", "0.5016576", "0.5006819", "0.50034475", "0.49984688", "0.49943554", "0.49926937", "0.49851456", "0.4985076", "0.49828827", "0.4981592", "0.49755734", "0.49740237", "0.49637645", "0.4961673", "0.49604794", "0.49555427", "0.4955329", "0.49533355" ]
0.72073734
0
PrefixTags adds the given tags to the beginning of the comma separated list.
func (e *Element) PrefixTags(tags ...string) { prefix := strings.Join(tags, ",") if e.Tags == "" { e.Tags = prefix return } e.Tags = mergeTags(prefix, strings.Split(e.Tags, ",")) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TagsHasPrefix(v string) predicate.Project {\n\treturn predicate.Project(func(s *sql.Selector) {\n\t\ts.Where(sql.HasPrefix(s.C(FieldTags), v))\n\t})\n}", "func (e *Extractor) NamesFromTagWithPrefix(tag string, prefix string) (out []string, err error) {\n\n\tif err := e.isValidStruct(); err != nil {\n\t\treturn nil, err\n\t}\n\n\ts := reflect.ValueOf(e.StructAddr).Elem()\n\tfields := e.fields(s)\n\n\tfor _, field := range fields {\n\t\tval, ok := field.tags.Lookup(tag)\n\t\tif !ok {\n\t\t\tcontinue\n\t\t}\n\t\tkey, omit := e.parseOmitempty(val, field.value)\n\t\tif omit {\n\t\t\tcontinue\n\t\t}\n\t\tout = append(out, strings.TrimSpace(prefix+key))\n\t}\n\n\treturn\n}", "func prefixSlice(pre string, slice []string) (ret []string) {\n\tfor _, i := range slice {\n\t\tret = append(ret, pre+i)\n\t}\n\n\treturn\n}", "func (o LookupManagedPrefixListResultOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v LookupManagedPrefixListResult) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func prependAll(prefix string, in []string) (out []string) {\n\tfor _, s := range in {\n\t\tout = append(out, prefix+s)\n\t}\n\treturn\n}", "func SetPrefix(prefix string) string {\n\tdefer logger.SetPrefix(prefix)\n\told := tags[0]\n\ttags[0] = prefix\n\treturn old\n}", "func TagHasPrefix(t kappnavv1.Tag, prefix string) bool {\n\treturn strings.HasPrefix(string(t), prefix)\n}", "func prefixJoin(prefix string, array []string, separator string) (result string) {\n\tif len(array) == 0 {\n\t\treturn\n\t}\n\tfor index, val := range array {\n\t\tif index == 0 {\n\t\t\tresult = val\n\t\t} else {\n\t\t\tresult = join(result, concat(prefix, val), separator)\n\t\t}\n\t}\n\treturn\n}", "func IncomingDTagTransferRequestsPrefix(recipient string) []byte {\n\treturn append(DTagTransferRequestPrefix, []byte(recipient)...)\n}", "func TrimPrefix(objects []string, prefix string) []string {\n\tvar results []string\n\tfor _, object := range objects {\n\t\tresults = append(results, strings.TrimPrefix(object, prefix))\n\t}\n\treturn results\n}", "func (s *IPSet) AddPrefix(p IPPrefix) { s.AddRange(p.Range()) }", "func (s *IPSet) AddPrefix(p IPPrefix) { s.AddRange(p.Range()) }", "func getListPrefix(opt *Options, s *goquery.Selection) string {\n\tif isWrapperListItem(s) {\n\t\treturn \"\"\n\t}\n\n\tparent := s.Parent()\n\tif parent.Is(\"ul\") {\n\t\treturn opt.BulletListMarker + \" \"\n\t} else if parent.Is(\"ol\") {\n\t\tcurrentIndex := s.Index() + 1\n\n\t\tlastIndex := parent.Children().Last().Index() + 1\n\t\tmaxLength := len(strconv.Itoa(lastIndex))\n\n\t\t// pad the numbers so that all prefix numbers in the list take up the same space\n\t\t// `%02d.` -> \"01. \"\n\t\tformat := `%0` + strconv.Itoa(maxLength) + `d. `\n\t\treturn fmt.Sprintf(format, currentIndex)\n\t}\n\t// If the HTML is malformed and the list element isn't in a ul or ol, return no prefix\n\treturn \"\"\n}", "func VertexListPrefix(graph string) []byte {\n\treturn bytes.Join([][]byte{vertexPrefix, []byte(graph), {}}, []byte{0})\n}", "func (o GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyOutput) Prefixes() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKey) []string { return v.Prefixes }).(pulumi.StringArrayOutput)\n}", "func (c *HelpCountryCode) SetPrefixes(value []string) {\n\tc.Flags.Set(0)\n\tc.Prefixes = value\n}", "func (b UpdateBuilder) Prefix(sql string, args ...interface{}) UpdateCondition {\n\treturn builder.Append(b, \"Prefixes\", Expr(sql, args...)).(UpdateBuilder)\n}", "func (r IPRange) Prefixes() []IPPrefix {\n\tif !r.Valid() {\n\t\treturn nil\n\t}\n\tvar makePrefix prefixMaker\n\tif r.From.Is4() {\n\t\tmakePrefix = func(ip16 ip16, bits uint8) IPPrefix {\n\t\t\treturn IPPrefix{IPFrom16([16]byte(ip16)), bits - 12*8}\n\t\t}\n\t} else {\n\t\tmakePrefix = func(ip16 ip16, bits uint8) IPPrefix {\n\t\t\treturn IPPrefix{IPv6Raw([16]byte(ip16)), bits}\n\t\t}\n\t}\n\ta16, b16 := ip16(r.From.As16()), ip16(r.To.As16())\n\treturn appendRangePrefixes(nil, makePrefix, a16, b16)\n}", "func JoinTags(tags ...Tag) TagSet {\n\tvar result TagSet\n\tresult = append(result, tags...)\n\treturn result\n}", "func prependToLines(para, prefix string) string {\n\tlines := strings.Split(para, \"\\n\")\n\tfor i, line := range lines {\n\t\tlines[i] = prefix + line\n\t}\n\treturn strings.Join(lines, \"\\n\")\n}", "func PrefixLines(lines []string, prefix string) string {\n\tvar outputLines []string\n\n\tfor _, line := range lines {\n\t\tif line == \"\" {\n\t\t\toutputLines = append(outputLines, line)\n\t\t} else {\n\t\t\toutputLines = append(outputLines, prefix+line)\n\t\t}\n\t}\n\n\treturn strings.Join(outputLines, \"\\n\")\n}", "func prefixLines(prefix, lines string) string {\n\treturn prefix + strings.Replace(lines, \"\\n\", \"\\n\"+prefix, -1)\n}", "func (o GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyPtrOutput) Prefixes() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKey) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Prefixes\n\t}).(pulumi.StringArrayOutput)\n}", "func addTagsToLabels(tags map[string]string, labelNamePrefix string, labels map[string]string) {\n\tfor k, v := range tags {\n\t\tlabels[labelNamePrefix+\"_\"+sanitizeLabelName(k)] = v\n\t}\n}", "func CreatePrefixList(pkg string) []string {\n\tif pkg == \"\" {\n\t\treturn []string{\"\"}\n\t}\n\n\tnumDots := 0\n\t// one pass to pre-allocate the returned slice\n\tfor i := 0; i < len(pkg); i++ {\n\t\tif pkg[i] == '.' {\n\t\t\tnumDots++\n\t\t}\n\t}\n\tif numDots == 0 {\n\t\treturn []string{pkg, \"\"}\n\t}\n\n\tprefixes := make([]string, numDots+2)\n\t// second pass to fill in returned slice\n\tfor i := 0; i < len(pkg); i++ {\n\t\tif pkg[i] == '.' {\n\t\t\tprefixes[numDots] = pkg[:i]\n\t\t\tnumDots--\n\t\t}\n\t}\n\tprefixes[0] = pkg\n\n\treturn prefixes\n}", "func (ub *UpdateBuilder) Prefix(\n\tsql string,\n\targs ...interface{},\n) *UpdateBuilder {\n\tub.sql = ub.sql.Prefix(sql, args...)\n\treturn ub\n}", "func (m *Model) FieldsPrefix(prefix string, fieldNamesOrMapStruct ...interface{}) *Model {\n\tfields := m.getFieldsFrom(fieldNamesOrMapStruct...)\n\tif len(fields) == 0 {\n\t\treturn m\n\t}\n\tgstr.PrefixArray(fields, prefix+\".\")\n\treturn m.appendFieldsByStr(gstr.Join(fields, \",\"))\n}", "func PrefixSlice(a []string, w string) []string {\n\tif a == nil {\n\t\treturn nil\n\t}\n\tif len(a) == 0 {\n\t\treturn []string{}\n\t}\n\tret := make([]string, len(a))\n\tsb := strings.Builder{}\n\tfor i := 0; i < len(a); i++ {\n\t\tsb.Grow(len(a[i]) + len(w))\n\t\tsb.WriteString(w)\n\t\tsb.WriteString(a[i])\n\t\tret[i] = sb.String()\n\t\tsb.Reset()\n\t}\n\n\treturn ret\n}", "func (b DeleteBuilder) Prefix(sql string, args ...interface{}) DeleteCondition {\n\treturn builder.Append(b, \"Prefixes\", Expr(sql, args...)).(DeleteBuilder)\n}", "func addPrefix(a []string, prefix string) []string {\n\tvar b []string\n\tfor i := 0; i < len(a); i++ {\n\t\tb = append(b, strings.Join([]string{prefix, a[i]}, \":\"))\n\t}\n\treturn b\n}", "func FilterPrefix(stringSet sets.String, prefix string, ignoreCase bool) sets.String {\n\tif prefix == \"\" {\n\t\treturn stringSet\n\t}\n\treturn filterSet(stringSet, prefix, ignoreCase, strings.HasPrefix)\n}", "func (o *CloudSnapshotAccount) SetNormalizedTags(normalizedTags []string) {\n\n\to.NormalizedTags = normalizedTags\n}", "func PackTags(in map[string]string) string {\n\ttags := []string{}\n\n\tfor k, v := range in {\n\t\ttags = append(tags, fmt.Sprintf(\"%s:%s\", k, v))\n\t}\n\n\tsort.Strings(tags)\n\n\treturn strings.Join(tags, \",\")\n}", "func (f *filter) Prefix(prefix string) Filter {\n\tif f.prefixed || prefix == \"\" {\n\t\treturn f\n\t}\n\n\tif len(f.allowed) > 0 {\n\t\tfor _, r := range f.allowed {\n\t\t\tr.prefix(prefix)\n\t\t}\n\n\t\tf.prefixed = true\n\t}\n\n\treturn f\n}", "func (o *SparseCloudSnapshotAccount) SetNormalizedTags(normalizedTags []string) {\n\n\to.NormalizedTags = &normalizedTags\n}", "func Prefix(pre []byte) []byte {\n\treturn append(pre, KeySeparator...)\n}", "func (q *queryImpl) Prefix(sql string, args ...interface{}) Query {\n\tq.prefixes = append(q.suffixes, ExprString{sql, args})\n\treturn q\n}", "func (q *Query) Prefix(sql string, args ...interface{}) *Query {\n\tq.prefixes = append(q.suffixes, Expr(sql, args...))\n\treturn q\n}", "func NamePrefixes() []string {\n\ts := UserID.String()\n\treturn []string{\"<@\" + s + \">\", \"<@!\" + s + \">\", \"1dot\"}\n}", "func (l *DockerLib) SetPrefix(p string) {\n\tl.prefix = p\n}", "func Tags(val []string) zap.Field {\n\treturn zap.Field{Key: FieldTags, Type: zapcore.SkipType, Interface: val}\n}", "func Prefixed(prefix string) Option {\n\treturn func(opts *logOptions) {\n\t\tif opts.prefix != `` {\n\t\t\topts.prefix = fmt.Sprintf(`%s.%s`, opts.prefix, prefix)\n\t\t\treturn\n\t\t}\n\t\topts.prefix = prefix\n\t}\n}", "func (o PublishingOutput) DocTagPrefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Publishing) *string { return v.DocTagPrefix }).(pulumi.StringPtrOutput)\n}", "func rangeValuePrefix(ss ...[]byte) []byte {\n\treturn buildRangeValue(0, ss...)\n}", "func (f *msgFilter) SetPrefix(p string) {\n\tf.prefix = p\n}", "func (mlog *MultiLogger) SetPrefix(prefix string) {\n\tmlog.prefix = []byte(prefix)\n}", "func addPrefixToLines(input, prefix string) string {\n\toutput := \"\"\n\tscanner := bufio.NewScanner(strings.NewReader(input))\n\tfor scanner.Scan() {\n\t\toutput += prefix + scanner.Text() + \"\\n\"\n\t}\n\toutput = strings.TrimRight(output, \"\\n\")\n\treturn output\n}", "func (o TransferJobTransferSpecObjectConditionsOutput) IncludePrefixes() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v TransferJobTransferSpecObjectConditions) []string { return v.IncludePrefixes }).(pulumi.StringArrayOutput)\n}", "func (o InventoryFilterOutput) Prefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InventoryFilter) *string { return v.Prefix }).(pulumi.StringPtrOutput)\n}", "func (o *SignalPersonName) SetPrefix(v string) {\n\to.Prefix.Set(&v)\n}", "func (o SecurityGroupRuleOutput) PrefixListIds() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *SecurityGroupRule) pulumi.StringArrayOutput { return v.PrefixListIds }).(pulumi.StringArrayOutput)\n}", "func prependIfMissing(str string, prefix string, ignoreCase bool, prefixes ...string) string {\n if IsEmpty(prefix) || internalStartsWith(str, prefix, ignoreCase) {\n\t\treturn str\n\t}\n\tfor _, pref := range prefixes {\n\t\tif pref == \"\" || internalStartsWith(str, pref, ignoreCase) {\n\t\t\treturn str\n\t\t}\n\t}\n\treturn prefix + str\n}", "func (in *ActionIpAddressIndexInput) SetPrefix(value int64) *ActionIpAddressIndexInput {\n\tin.Prefix = value\n\n\tif in._selectedParameters == nil {\n\t\tin._selectedParameters = make(map[string]interface{})\n\t}\n\n\tin._selectedParameters[\"Prefix\"] = nil\n\treturn in\n}", "func mergeTags(existing string, tags []string) string {\n\tif existing == \"\" {\n\t\treturn strings.Join(tags, \",\")\n\t}\n\told := strings.Split(existing, \",\")\n\tvar merged []string\n\tfor _, o := range old {\n\t\tfound := false\n\t\tfor _, tag := range tags {\n\t\t\tif tag == o {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tmerged = append(merged, o)\n\t\t}\n\t}\n\tfor _, tag := range tags {\n\t\tfound := false\n\t\tfor _, o := range merged {\n\t\t\tif tag == o {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tmerged = append(merged, tag)\n\t\t}\n\t}\n\treturn strings.Join(merged, \",\")\n}", "func (r RawValues) FilterPrefix(prefix string) RawValues {\n\tfiltered := make(RawValues, 0)\n\tfor _, r := range r {\n\t\tif match.HasPrefix(r.Value, prefix) {\n\t\t\tfiltered = append(filtered, r)\n\t\t}\n\t}\n\treturn filtered\n}", "func (o GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyResponseOutput) Prefixes() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyResponse) []string { return v.Prefixes }).(pulumi.StringArrayOutput)\n}", "func (o PublishingResponseOutput) DocTagPrefix() pulumi.StringOutput {\n\treturn o.ApplyT(func(v PublishingResponse) string { return v.DocTagPrefix }).(pulumi.StringOutput)\n}", "func EdgeListPrefix(graph string) []byte {\n\treturn bytes.Join([][]byte{edgePrefix, []byte(graph), {}}, []byte{0})\n}", "func TagsLT(v string) predicate.Project {\n\treturn predicate.Project(func(s *sql.Selector) {\n\t\ts.Where(sql.LT(s.C(FieldTags), v))\n\t})\n}", "func (q *Query) AddTags(raw []string) {\n\thashed := HashTags(raw)\n\tq.Raw = append(q.Raw, raw...)\n\tq.Hashed = append(q.Hashed, hashed...)\n}", "func (s *SlackNotify) SetPrefix(v string) {\n\ts.prefix = v\n}", "func (o TransferJobTransferSpecObjectConditionsPtrOutput) IncludePrefixes() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *TransferJobTransferSpecObjectConditions) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.IncludePrefixes\n\t}).(pulumi.StringArrayOutput)\n}", "func (o BucketNotificationTopicOutput) FilterPrefix() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketNotificationTopic) *string { return v.FilterPrefix }).(pulumi.StringPtrOutput)\n}", "func promGroupTags(args []parse.Node) (parse.Tags, error) {\n\ttags := make(parse.Tags)\n\tcsvTags := strings.Split(args[1].(*parse.StringNode).Text, \",\")\n\tfor _, k := range csvTags {\n\t\ttags[k] = struct{}{}\n\t}\n\treturn tags, nil\n}", "func prefixDirectory(directory string, names []string) []string {\n\tif directory == \".\" {\n\t\treturn names\n\t}\n\tret := make([]string, len(names))\n\tfor i, name := range names {\n\t\tret[i] = filepath.Join(directory, name)\n\t}\n\treturn ret\n}", "func prefixDirectory(directory string, names []string) []string {\n\tif directory == \".\" {\n\t\treturn names\n\t}\n\tret := make([]string, len(names))\n\tfor i, name := range names {\n\t\tret[i] = filepath.Join(directory, name)\n\t}\n\treturn ret\n}", "func prefixDirectory(directory string, names []string) []string {\n\tif directory == \".\" {\n\t\treturn names\n\t}\n\tret := make([]string, len(names))\n\tfor i, name := range names {\n\t\tret[i] = filepath.Join(directory, name)\n\t}\n\treturn ret\n}", "func prefixDirectory(directory string, names []string) []string {\n\tif directory == \".\" {\n\t\treturn names\n\t}\n\tret := make([]string, len(names))\n\tfor i, name := range names {\n\t\tret[i] = filepath.Join(directory, name)\n\t}\n\treturn ret\n}", "func ExampleBucket_MapPrefix() {\n\tbx, _ := buckets.Open(tempfile())\n\tdefer os.Remove(bx.Path())\n\tdefer bx.Close()\n\n\t// Create a new things bucket.\n\tthings, _ := bx.New([]byte(\"things\"))\n\n\t// Setup items to insert.\n\titems := []struct {\n\t\tKey, Value []byte\n\t}{\n\t\t{[]byte(\"A\"), []byte(\"1\")}, // `A` prefix match\n\t\t{[]byte(\"AA\"), []byte(\"2\")}, // match\n\t\t{[]byte(\"AAA\"), []byte(\"3\")}, // match\n\t\t{[]byte(\"AAB\"), []byte(\"2\")}, // match\n\t\t{[]byte(\"B\"), []byte(\"O\")},\n\t\t{[]byte(\"BA\"), []byte(\"0\")},\n\t\t{[]byte(\"BAA\"), []byte(\"0\")},\n\t}\n\n\t// Insert 'em.\n\tif err := things.Insert(items); err != nil {\n\t\tfmt.Printf(\"could not insert items in `things` bucket: %v\\n\", err)\n\t}\n\n\t// Now collect each item whose key starts with \"A\".\n\tprefix := []byte(\"A\")\n\n\t// Setup slice of items.\n\ttype item struct {\n\t\tKey, Value []byte\n\t}\n\tresults := []item{}\n\n\t// Anon func to map over matched keys.\n\tdo := func(k, v []byte) error {\n\t\tresults = append(results, item{k, v})\n\t\treturn nil\n\t}\n\n\tif err := things.MapPrefix(do, prefix); err != nil {\n\t\tfmt.Printf(\"could not map items with prefix %s: %v\\n\", prefix, err)\n\t}\n\n\tfor _, item := range results {\n\t\tfmt.Printf(\"%s -> %s\\n\", item.Key, item.Value)\n\t}\n\t// Output:\n\t// A -> 1\n\t// AA -> 2\n\t// AAA -> 3\n\t// AAB -> 2\n}", "func WithPrefix(prefix string) OptFunc {\n\treturn func(l *Logger) {\n\t\tl.SetPrefix(prefix)\n\t}\n}", "func (c *config) PrefixKeys(prefix string) []string {\n c.m.Lock()\n defer c.m.Unlock()\n\n keys := []string{}\n for k, _ := range c.conf {\n if strings.HasPrefix(k, prefix) {\n keys = append(keys, k)\n }\n }\n return keys\n}", "func SrcEdgeListPrefix(graph string) []byte {\n\treturn bytes.Join([][]byte{srcEdgePrefix, []byte(graph), {}}, []byte{0})\n}", "func ConcatTagValues(tagValues []string) string {\n\tif len(tagValues) == 0 {\n\t\treturn \"\"\n\t}\n\treturn strings.Join(tagValues, \",\")\n}", "func DeleteTags(elemTags map[string]string, blacklist map[string]bool) {\n\tfor key, isPrefix := range blacklist {\n\t\tif isPrefix {\n\t\t\tfor tag := range elemTags {\n\t\t\t\tif strings.HasPrefix(tag, key) {\n\t\t\t\t\tdelete(elemTags, tag)\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tdelete(elemTags, key)\n\t\t}\n\t}\n}", "func (s *IPSet) Prefixes() []IPPrefix {\n\tvar out []IPPrefix\n\tfor _, r := range s.Ranges() {\n\t\tout = append(out, r.Prefixes()...)\n\t}\n\treturn out\n}", "func (s *IPSet) Prefixes() []IPPrefix {\n\tvar out []IPPrefix\n\tfor _, r := range s.Ranges() {\n\t\tout = append(out, r.Prefixes()...)\n\t}\n\treturn out\n}", "func prefixDirectory(directory string, names []string) []string {\r\n\tif directory == \".\" {\r\n\t\treturn names\r\n\t}\r\n\tret := make([]string, len(names))\r\n\tfor i, name := range names {\r\n\t\tret[i] = filepath.Join(directory, name)\r\n\t}\r\n\treturn ret\r\n}", "func (kps *KubernetesPrefixSource) Prefixes() []string {\n\treturn kps.prefixes.Load()\n}", "func (_TokensNetwork *TokensNetworkCaller) SignaturePrefix(opts *bind.CallOpts) (string, error) {\n\tvar (\n\t\tret0 = new(string)\n\t)\n\tout := ret0\n\terr := _TokensNetwork.contract.Call(opts, out, \"signature_prefix\")\n\treturn *ret0, err\n}", "func (o BucketLifecycleRuleItemConditionOutput) MatchesPrefix() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v BucketLifecycleRuleItemCondition) []string { return v.MatchesPrefix }).(pulumi.StringArrayOutput)\n}", "func PrefixWithSep(prefix, sep, v string) string {\n\tif len(prefix) > 0 {\n\t\treturn prefix + sep + v\n\t}\n\treturn v\n}", "func cleanTags(tagCSV string) []string {\n\tclean := []string{}\n\ttags := strings.Split(tagCSV, \",\")\n\tfor _, v := range tags {\n\t\ttag := strings.TrimSpace(strings.ToLower(v))\n\t\tif len(tag) > 0 {\n\t\t\tclean = append(clean, tag)\n\t\t}\n\t}\n\treturn clean\n}", "func (f *InteractiveFiller) makePrefix(field *desc.FieldDescriptor) string {\n\treturn makePrefix(f.prefixFormat, field, f.state.ancestor, f.state.hasAncestorAndHasRepeatedField)\n}", "func GetIgnoreTagPrefixes(ctx *pulumi.Context) string {\n\treturn config.Get(ctx, \"aws:ignoreTagPrefixes\")\n}", "func (s AwsEc2SecurityGroupPrefixListId) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.PrefixListId != nil {\n\t\tv := *s.PrefixListId\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"PrefixListId\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\treturn nil\n}", "func SetPrefix(prefix string) { std.SetPrefix(prefix) }", "func (c *SearchCall) Prefixed(prefixed bool) *SearchCall {\n\tc.urlParams_.Set(\"prefixed\", fmt.Sprint(prefixed))\n\treturn c\n}", "func VariablesPrefix(vars []Variable, prefix string) []Variable {\n\tres := make([]Variable, len(vars))\n\tfor i := range vars {\n\t\ts := vars[i]\n\t\ts.Name = prefix + s.Name\n\t\tres[i] = s\n\t}\n\treturn res\n}", "func SetTags(n [10]string) [10]string {\n\told := tags\n\ttags = n\n\treturn old\n}", "func (a *AbbrFieldNamer) SetPrefix(s string) {\n\ta.Prefix = s\n}", "func (ctx *PlikContext) UpdateLoggerPrefix(prefix string) {\n\tstr := \"\"\n\tif ip, ok := ctx.Get(\"RemoteIp\"); ok {\n\t\tstr += fmt.Sprintf(\"[%s]\", ip)\n\t}\n\tif uploadID, ok := ctx.Get(\"UploadId\"); ok {\n\t\tstr += fmt.Sprintf(\"[%s]\", uploadID)\n\t}\n\tif fileName, ok := ctx.Get(\"FileName\"); ok {\n\t\tstr += fmt.Sprintf(\"[%s]\", fileName)\n\t}\n\tctx.SetPrefix(str + prefix)\n}", "func ListPrefix(ctx context.Context, prefix string) (KeyValuePairs, error) {\n\tv, err := Client().ListPrefix(ctx, prefix)\n\tTrace(\"ListPrefix\", err, logrus.Fields{fieldPrefix: prefix, fieldNumEntries: len(v)})\n\treturn v, err\n}", "func ProcessTags(t, sep string, repl ...string) []string {\n\tr := regexp.MustCompile(\"(?i)\" + strings.Join(repl, \"|\"))\n\tout := r.ReplaceAllString(strings.ToLower(t), \"\")\n\tout = strings.Replace(out, \"and\", sep, -1)\n\treturn TrimSpacesList(strings.Split(out, sep))\n}", "func FilterEntriesByPrefix(prefix string, entries []string) []string {\n\tvar result []string\n\tfor _, entry := range entries {\n\t\tif strings.HasPrefix(entry, prefix) {\n\t\t\tresult = append(result, entry)\n\t\t}\n\t}\n\treturn result\n}", "func (z *Zone) AddTags(tags ...string) int {\n\ts := NewSet(z.Tags...)\n\tn := len(s)\n\ts.Add(tags...)\n\td := len(s) - n\n\tif d != 0 {\n\t\tz.Tags = s.Values()\n\t}\n\treturn d\n}", "func PrefixPaths(paths []string, prefix string) []string {\n\tresult := make([]string, len(paths))\n\tfor i, path := range paths {\n\t\tresult[i] = filepath.Join(prefix, path)\n\t}\n\treturn result\n}", "func (s Sticker) TagList() []string {\n\ttags := strings.Split(s.Tags, \",\")\n\tfor i := range tags {\n\t\ttags[i] = strings.TrimSpace(tags[i])\n\t}\n\treturn tags\n}", "func SetPrefix(p string) {\n\tprefix = p\n}", "func (moves Moves) ApplyPrefix(prefix path.Path) Moves {\n\tnewMoves := make(Moves, len(moves))\n\tfor i, mv := range moves {\n\t\tnewMoves[i] = mv.ApplyPrefix(prefix)\n\t}\n\n\treturn newMoves\n}", "func (l *ContextLogger) AddPrefix(prefix interface{}) {\n\tl.prefixes = append(l.prefixes, prefix)\n}" ]
[ "0.6147257", "0.59784806", "0.5961276", "0.5960882", "0.5945469", "0.55871814", "0.54802436", "0.54000705", "0.53906906", "0.5346082", "0.5337324", "0.5337324", "0.53358513", "0.52466637", "0.52123576", "0.51916856", "0.518039", "0.5169964", "0.5168189", "0.51646787", "0.51174843", "0.510715", "0.5090357", "0.5077158", "0.50734323", "0.5058536", "0.504805", "0.5045239", "0.50246584", "0.4989945", "0.49681944", "0.49650082", "0.4958204", "0.49582002", "0.4951566", "0.4946797", "0.49312866", "0.49266568", "0.49147326", "0.4912101", "0.48970532", "0.48933807", "0.48821747", "0.4878316", "0.48697653", "0.48673308", "0.48660854", "0.48572344", "0.4856988", "0.48527578", "0.48509634", "0.48509502", "0.48391023", "0.48237452", "0.48201698", "0.48170543", "0.48050824", "0.4798326", "0.47952744", "0.479346", "0.4790071", "0.47808796", "0.47738054", "0.47683266", "0.47646114", "0.47646114", "0.47646114", "0.47646114", "0.4763486", "0.47497922", "0.4747208", "0.47470787", "0.47396562", "0.47344533", "0.4733677", "0.4733677", "0.47322842", "0.47293714", "0.47246733", "0.47206557", "0.47201648", "0.4716585", "0.47086158", "0.46957794", "0.4695723", "0.4679344", "0.46695665", "0.46585524", "0.46570984", "0.46544597", "0.46530345", "0.46480808", "0.46444535", "0.46410513", "0.4637709", "0.4629087", "0.46260893", "0.462377", "0.46235022", "0.4620024" ]
0.8500971
0
mergeTags merges the comma separated tags in old with the ones in tags and returns a comma separated string with the results.
func mergeTags(existing string, tags []string) string { if existing == "" { return strings.Join(tags, ",") } old := strings.Split(existing, ",") var merged []string for _, o := range old { found := false for _, tag := range tags { if tag == o { found = true break } } if !found { merged = append(merged, o) } } for _, tag := range tags { found := false for _, o := range merged { if tag == o { found = true break } } if !found { merged = append(merged, tag) } } return strings.Join(merged, ",") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func mergeTags(t1, t2 []Tag) []Tag {\n\tn := len(t1) + len(t2)\n\tif n == 0 {\n\t\treturn nil\n\t}\n\n\tout := make([]Tag, 0, n)\n\tout = append(out, t1...)\n\tout = append(out, t2...)\n\n\treturn SortTags(out)\n}", "func (p *plugin) concatTags(tags1 *structtag.Tags, tags2 *structtag.Tags) (*structtag.Tags, error) {\n\tif tags1.Len() == 0 {\n\t\treturn tags2, nil\n\t}\n\tif tags2.Len() == 0 {\n\t\treturn tags1, nil\n\t}\n\n\tfor _, t2 := range tags2.Tags() {\n\t\tvar found bool\n\t\tfor _, t1 := range tags1.Tags() {\n\t\t\tif t1.Key == t2.Key {\n\t\t\t\tif len(t1.Name) == 0 {\n\t\t\t\t\tt1.Name = t2.Name\n\t\t\t\t}\n\t\t\t\tif t1.Options == nil || len(t1.Options) == 0 {\n\t\t\t\t\tt1.Options = t2.Options\n\t\t\t\t}\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tvar err error\n\t\t\ts := tags1.String() + \" \" + t2.String()\n\t\t\ttags1, err = structtag.Parse(s)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"failed to parse tags '%s': %s\", s, err.Error())\n\t\t\t}\n\t\t}\n\t}\n\n\treturn tags1, nil\n}", "func MergeTags(generalTags []*Tag, infraTags []*Tag) []*Tag {\n\tvar dupMap = make(map[string]bool)\n\tfor _, tag := range generalTags {\n\t\tdupMap[tag.Key] = true\n\t}\n\tfor _, tag := range infraTags {\n\t\tif _, exists := dupMap[tag.Key]; !exists {\n\t\t\tgeneralTags = append(generalTags, tag)\n\t\t}\n\t}\n\treturn generalTags\n}", "func mergeTags(generalTags []*tag, infraTags []*tag) []*tag {\n\tvar dupMap = make(map[string]bool)\n\tfor _, tag := range generalTags {\n\t\tdupMap[tag.key] = true\n\t}\n\tfor _, tag := range infraTags {\n\t\tif _, exists := dupMap[tag.key]; !exists {\n\t\t\tgeneralTags = append(generalTags, tag)\n\t\t}\n\t}\n\treturn generalTags\n}", "func JoinTags(tags ...Tag) TagSet {\n\tvar result TagSet\n\tresult = append(result, tags...)\n\treturn result\n}", "func (e *Element) MergeTags(tags ...string) {\n\te.Tags = mergeTags(e.Tags, tags)\n}", "func TagsDiff(sqsTags map[string]string, newTags map[string]string) (removed, added map[string]string) {\n\tremoved = map[string]string{}\n\tfor k, v := range sqsTags {\n\t\tif _, ok := newTags[k]; !ok {\n\t\t\tremoved[k] = v\n\t\t}\n\t}\n\n\tadded = map[string]string{}\n\tfor k, newV := range newTags {\n\t\tif oldV, ok := sqsTags[k]; !ok || oldV != newV {\n\t\t\tadded[k] = newV\n\t\t}\n\t}\n\treturn\n}", "func (m *Entry) CommaTags() string {\n\ttags := []string{}\n\tif m.Tags != nil {\n\t\tfor _, tag := range m.Tags {\n\t\t\ttags = append(tags, tag.Name)\n\t\t}\n\t}\n\treturn strings.Join(tags, \",\")\n}", "func DiffTags(spec []v1alpha1.Tag, current []ecr.Tag) (addTags []ecr.Tag, remove []string) {\n\taddMap := make(map[string]string, len(spec))\n\tfor _, t := range spec {\n\t\taddMap[t.Key] = t.Value\n\t}\n\tremoveMap := map[string]struct{}{}\n\tfor _, t := range current {\n\t\tif addMap[aws.StringValue(t.Key)] == aws.StringValue(t.Value) {\n\t\t\tdelete(addMap, aws.StringValue(t.Key))\n\t\t\tcontinue\n\t\t}\n\t\tremoveMap[aws.StringValue(t.Key)] = struct{}{}\n\t}\n\tfor k, v := range addMap {\n\t\taddTags = append(addTags, ecr.Tag{Key: aws.String(k), Value: aws.String(v)})\n\t}\n\tfor k := range removeMap {\n\t\tremove = append(remove, k)\n\t}\n\treturn\n}", "func ConvertTagsToString(tags []string) string {\n\tfinalString := make([]string, 0)\n\ttagChar := \"#\"\n\tfor _, value := range tags {\n\t\ttrimmed := strings.TrimSpace(value)\n\t\tfinalString = append(finalString, fmt.Sprintf(\"%s%s\", tagChar, trimmed))\n\t}\n\treturn strings.Join(finalString, \" \")\n}", "func ConcatTagValues(tagValues []string) string {\n\tif len(tagValues) == 0 {\n\t\treturn \"\"\n\t}\n\treturn strings.Join(tagValues, \",\")\n}", "func (t Tags) Merge(key string, value ...string) {\n\tfor _, v := range value {\n\t\tcurrent := t.GetAll(key)\n\t\tfound := false\n\t\tfor _, cv := range current {\n\t\t\tif v == cv {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tt.Add(key, v)\n\t\t}\n\t}\n}", "func MergeTagSlices(Original []TagInformation, ToAdd []TagInformation) []TagInformation {\n\t//Rules:\n\t//We do not care about meta-tags\n\t//Tags in ToAdd win\n\t//Exlusionary tags win after tags in ToAdd\n\n\t//First, remove duplicates from original that exist in ToAdd\n\tfor Index := 0; Index < len(ToAdd); Index++ {\n\t\tif ToAdd[Index].IsMeta {\n\t\t\tcontinue //Skip Metatags\n\t\t}\n\t\t//Standard tag confirmed, scan for duplicates\n\t\tfor ScanIndex := 0; ScanIndex < len(Original); ScanIndex++ {\n\t\t\tif Original[ScanIndex].IsMeta {\n\t\t\t\tcontinue //Skip comparing metas\n\t\t\t}\n\t\t\tif Original[ScanIndex].ID == ToAdd[Index].ID {\n\t\t\t\t//Remove and resize\n\t\t\t\tOriginal = append(Original[:ScanIndex], Original[ScanIndex+1:]...)\n\t\t\t\t//we just need to continue current scan from the same ScanIndex\n\t\t\t\tScanIndex--\n\t\t\t}\n\t\t}\n\t}\n\n\t//Now we can fall back to RemoveDuplicateTags to cleanup any other issues\n\treturn RemoveDuplicateTags(append(Original, ToAdd...))\n}", "func updateTags(nTags ...string) {\n\td := map[string]bool{}\n\tnewTags := append(_tags, nTags...)\n\tfor _, t := range newTags {\n\t\td[strings.ToLower(t)] = true\n\t}\n\n\t_newTags := make([]string, 0)\n\tfor k := range d {\n\t\t_newTags = append(_newTags, k)\n\t}\n\t_tags = _newTags\n\t// Update tags to disk\n\tutil.StoreJSON(ml.FileOthers, \"tags\", _tags)\n}", "func (o TriggerBuildOutput) Tags() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v TriggerBuild) []string { return v.Tags }).(pulumi.StringArrayOutput)\n}", "func (a *Client) ReplaceTags(params *ReplaceTagsParams, authInfo runtime.ClientAuthInfoWriter) (*ReplaceTagsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewReplaceTagsParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"replaceTags\",\n\t\tMethod: \"PUT\",\n\t\tPathPattern: \"/app/rest/builds/{buildLocator}/tags\",\n\t\tProducesMediaTypes: []string{\"application/json\", \"application/xml\", \"text/plain\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/xml\", \"text/plain\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ReplaceTagsReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*ReplaceTagsOK), nil\n\n}", "func getTags(operation *openapi3.Operation) string {\n\treturn strings.Replace(operation.Tags[0], \" \", \"-\", -1)\n}", "func buildTags() string {\n\treturn *tags\n}", "func (t TagSet) Merge(more TagSet) TagSet {\n\tmerged := t[:]\n\treturn append(merged, more...)\n}", "func (o EntityRecognizerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *EntityRecognizer) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (t tagger) deleteTags(modFullTags []string) error {\n\tfor _, modFullTag := range modFullTags {\n\t\tfmt.Printf(\"Deleting tag %v\\n\", modFullTag)\n\t\tcmd := exec.Command(\"git\", \"tag\", \"-d\", modFullTag)\n\t\tif output, err := cmd.CombinedOutput(); err != nil {\n\t\t\treturn fmt.Errorf(\"could not delete tag %v:\\n%v (%v)\", modFullTag, string(output), err)\n\t\t}\n\t}\n\treturn nil\n}", "func (o TrackerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Tracker) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o LookupListenerResultOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v LookupListenerResult) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func mergeRightTags(tagsLeft, tagsRight map[string]string) map[string]string {\n\tif tagsLeft == nil && tagsRight == nil {\n\t\treturn nil\n\t}\n\tif len(tagsRight) == 0 {\n\t\treturn tagsLeft\n\t}\n\tif len(tagsLeft) == 0 {\n\t\treturn tagsRight\n\t}\n\n\tresult := make(map[string]string, len(tagsLeft)+len(tagsRight))\n\tfor k, v := range tagsLeft {\n\t\tresult[k] = v\n\t}\n\tfor k, v := range tagsRight {\n\t\tresult[k] = v\n\t}\n\treturn result\n}", "func sanitizeTags(tags string) string {\n\ttagsSlice := strings.Split(tags, \",\")\n\tfor i, el := range tagsSlice {\n\t\ttagsSlice[i] = strings.TrimSpace(el)\n\t}\n\tif len(tagsSlice[len(tagsSlice)-1]) > 1 {\n\t\treturn (strings.Join(tagsSlice, \"|\"))\n\t} else {\n\t\treturn (strings.Join(tagsSlice[:len(tagsSlice)-1], \"|\"))\n\t}\n}", "func (o ListenerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v Listener) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func mergeResourceTags(one map[string]string, two map[string]*string) (map[string]*string, bool) {\n\tmergedResourceTags := map[string]*string{}\n\t// \"Copy\" two into a new map.\n\tfor key, value := range two {\n\t\tmergedResourceTags[key] = to.Ptr(*value)\n\t}\n\tchanged := false\n\t// Merge \"one\" into the new map, setting changed if we had to make a modification,\n\t// this is for determining whether we needed to make an update to the existing tags.\n\tfor key, value := range one {\n\t\tval, ok := mergedResourceTags[key]\n\t\tif !ok || *val != value {\n\t\t\tmergedResourceTags[key] = to.Ptr(value)\n\t\t\tchanged = true\n\t\t}\n\t}\n\treturn mergedResourceTags, changed\n}", "func (o NetworkOutput) Tags() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *Network) pulumi.StringArrayOutput { return v.Tags }).(pulumi.StringArrayOutput)\n}", "func ProcessTags(t, sep string, repl ...string) []string {\n\tr := regexp.MustCompile(\"(?i)\" + strings.Join(repl, \"|\"))\n\tout := r.ReplaceAllString(strings.ToLower(t), \"\")\n\tout = strings.Replace(out, \"and\", sep, -1)\n\treturn TrimSpacesList(strings.Split(out, sep))\n}", "func (o IotHubDeviceUpdateAccountOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *IotHubDeviceUpdateAccount) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func uniqueTagsWithSeen(seen map[string]struct{}, t1 gostatsd.Tags, t2 gostatsd.Tags) gostatsd.Tags {\n\tlast := len(t1)\n\tfor idx := 0; idx < last; {\n\t\ttag := t1[idx]\n\t\tif _, ok := seen[tag]; ok {\n\t\t\tlast--\n\t\t\tt1[idx] = t1[last]\n\t\t\tt1 = t1[:last]\n\t\t} else {\n\t\t\tseen[tag] = present\n\t\t\tidx++\n\t\t}\n\t}\n\n\tfor _, tag := range t2 {\n\t\tif _, ok := seen[tag]; !ok {\n\t\t\tt1 = append(t1, tag)\n\t\t}\n\t}\n\n\treturn t1\n}", "func (o AnalyzerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Analyzer) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o AnalyzerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Analyzer) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func splitTags(tags string) (string, tagOptions) {\n\tres := strings.Split(tags, \",\")\n\treturn res[0], res[1:]\n}", "func PackTags(in map[string]string) string {\n\ttags := []string{}\n\n\tfor k, v := range in {\n\t\ttags = append(tags, fmt.Sprintf(\"%s:%s\", k, v))\n\t}\n\n\tsort.Strings(tags)\n\n\treturn strings.Join(tags, \",\")\n}", "func (o LookupTestCaseResultOutput) Tags() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v LookupTestCaseResult) []string { return v.Tags }).(pulumi.StringArrayOutput)\n}", "func TestMergeWithPreviousBuilds(t *testing.T) {\n\tbuilds := MergeWithPreviousBuilds([]graph.Artifact{artifact(\"img1\", \"tag1_1\"), artifact(\"img2\", \"tag2_1\")}, nil)\n\ttestutil.CheckDeepEqual(t, \"img1:tag1_1,img2:tag2_1\", tags(builds))\n\n\tbuilds = MergeWithPreviousBuilds([]graph.Artifact{artifact(\"img1\", \"tag1_2\")}, builds)\n\ttestutil.CheckDeepEqual(t, \"img1:tag1_2,img2:tag2_1\", tags(builds))\n\n\tbuilds = MergeWithPreviousBuilds([]graph.Artifact{artifact(\"img2\", \"tag2_2\")}, builds)\n\ttestutil.CheckDeepEqual(t, \"img1:tag1_2,img2:tag2_2\", tags(builds))\n\n\tbuilds = MergeWithPreviousBuilds([]graph.Artifact{artifact(\"img1\", \"tag1_3\"), artifact(\"img2\", \"tag2_3\")}, builds)\n\ttestutil.CheckDeepEqual(t, \"img1:tag1_3,img2:tag2_3\", tags(builds))\n}", "func (p *Provider) formatTags(tags map[string]string) []string {\n\tformattedTags := []string{}\n\tfor name, value := range tags {\n\t\tformattedTags = append(formattedTags, fmt.Sprintf(\"%s:%s\", name, value))\n\t}\n\n\treturn formattedTags\n}", "func mergeTagSets(s1, s2, scratch tagSet) tagSet {\n\ta := scratch\n\ti, j, k := 0, 0, 0\n\tfor ; i < len(s1) && j < len(s2) && k < len(a); k++ {\n\t\tif s1[i].key == s2[j].key {\n\t\t\ta[k] = s2[j]\n\t\t\ti++\n\t\t\tj++\n\t\t} else if s1[i].key < s2[j].key {\n\t\t\ta[k] = s1[i]\n\t\t\ti++\n\t\t} else {\n\t\t\ta[k] = s2[j]\n\t\t\tj++\n\t\t}\n\t}\n\tif i < len(s1) {\n\t\tk += copy(a[k:], s1[i:])\n\t}\n\tif j < len(s2) {\n\t\tk += copy(a[k:], s2[j:])\n\t}\n\treturn a[:k]\n}", "func mergeAlternately(word1 string, word2 string) string {\n\tvar buf bytes.Buffer\n\tfor i := range word1 {\n\t\tbuf.WriteByte(word1[i])\n\t\tif i < len(word2) {\n\t\t\tbuf.WriteByte(word2[i])\n\t\t}\n\t}\n\n\tif len(word1) < len(word2) {\n\t\tbuf.WriteString(word2[len(word1):])\n\t}\n\treturn buf.String()\n}", "func (o ListenerRuleOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *ListenerRule) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func promGroupTags(args []parse.Node) (parse.Tags, error) {\n\ttags := make(parse.Tags)\n\tcsvTags := strings.Split(args[1].(*parse.StringNode).Text, \",\")\n\tfor _, k := range csvTags {\n\t\ttags[k] = struct{}{}\n\t}\n\treturn tags, nil\n}", "func (o PhpAppLayerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *PhpAppLayer) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o DatastoreFileshareOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *DatastoreFileshare) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o CrawlerOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Crawler) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o LinuxWebAppOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *LinuxWebApp) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o TargetGroupOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v TargetGroup) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o LookupGroupResultOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v LookupGroupResult) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o LookupSharedImageResultOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v LookupSharedImageResult) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func saveNewTags(tx *sql.Tx, tags []string) error {\n\tif len(tags) != 0 {\n\t\tquery := \"insert into tags (\" + tagsInsertFields + \") values \"\n\t\targs := make([]interface{}, len(tags))\n\n\t\tfor index, tag := range tags {\n\t\t\tquery += fmt.Sprintf(\"($%d),\", index+1)\n\t\t\targs[index] = tag\n\t\t}\n\t\tquery = strings.TrimSuffix(query, \",\")\n\n\t\tquery += \" on conflict do nothing\"\n\t\tif stmt, err := tx.Prepare(query); err != nil {\n\t\t\treturn err\n\t\t} else {\n\t\t\t_, err := stmt.Exec(args...)\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (o FaqOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Faq) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func uniqueTags(t1 gostatsd.Tags, t2 gostatsd.Tags) gostatsd.Tags {\n\treturn uniqueTagsWithSeen(map[string]struct{}{}, t1, t2)\n}", "func (o AppMonitorOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *AppMonitor) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o TrackerOutput) TagsAll() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Tracker) pulumi.StringMapOutput { return v.TagsAll }).(pulumi.StringMapOutput)\n}", "func (o BucketReplicationConfigRuleFilterAndOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigRuleFilterAnd) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o LookupServerResultOutput) Tags() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v LookupServerResult) []string { return v.Tags }).(pulumi.StringArrayOutput)\n}", "func (o FleetOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Fleet) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o GraphOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Graph) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o ApplicationOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Application) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o AnomalySubscriptionOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *AnomalySubscription) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func processTags(tags []string) []string {\n\tif tags == nil {\n\t\treturn nil\n\t}\n\tprocessedTags := []string{}\n\tusedTags := make(map[string]bool)\n\tfor _, tag := range tags {\n\t\ttag = strings.TrimSpace(tag)\n\t\tif len(tag) > 0 && !usedTags[tag] {\n\t\t\tprocessedTags = append(processedTags, tag)\n\t\t\tusedTags[tag] = true\n\t\t}\n\t}\n\treturn processedTags\n}", "func (o EntityRecognizerOutput) TagsAll() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *EntityRecognizer) pulumi.StringMapOutput { return v.TagsAll }).(pulumi.StringMapOutput)\n}", "func (o NetworkOutput) AllTags() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *Network) pulumi.StringArrayOutput { return v.AllTags }).(pulumi.StringArrayOutput)\n}", "func (o DatabaseReplicaOutput) Tags() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *DatabaseReplica) pulumi.StringArrayOutput { return v.Tags }).(pulumi.StringArrayOutput)\n}", "func (o TaskOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Task) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (DBConnection *MariaDBPlugin) BulkAddTag(TagID uint64, OldTagID uint64, LinkerID uint64) error {\n\t//Prevent adding alias\n\ttagInfo, err := DBConnection.GetTag(TagID, false)\n\toldTagInfo, err2 := DBConnection.GetTag(OldTagID, false)\n\tif err != nil || err2 != nil {\n\t\treturn errors.New(\"Failed to validate tags\")\n\t}\n\n\t//If this is an alias, then add aliasedid instead\n\tif tagInfo.IsAlias {\n\t\tTagID = tagInfo.AliasedID\n\t}\n\n\t//Similiarly convert oldTag if it is an alias\n\tif oldTagInfo.IsAlias {\n\t\tOldTagID = oldTagInfo.AliasedID\n\t}\n\n\tif _, err := DBConnection.DBHandle.Exec(\"INSERT INTO ImageTags (TagID, ImageID, LinkerID) SELECT ?, ImageID, ? FROM ImageTags WHERE TagID=? AND ImageID NOT IN (SELECT ImageID FROM ImageTags WHERE TagID=?);\", TagID, LinkerID, OldTagID, TagID); err != nil {\n\t\tlogging.WriteLog(logging.LogLevelError, \"MariaDBPlugin/BulkAddTag\", strconv.FormatUint(LinkerID, 10), logging.ResultFailure, []string{\"Tag not added to image\", strconv.FormatUint(OldTagID, 10), strconv.FormatUint(TagID, 10), err.Error()})\n\t\treturn err\n\t}\n\tlogging.WriteLog(logging.LogLevelError, \"MariaDBPlugin/BulkAddTag\", strconv.FormatUint(LinkerID, 10), logging.ResultSuccess, []string{\"Tags added\", strconv.FormatUint(OldTagID, 10), strconv.FormatUint(TagID, 10)})\n\treturn nil\n}", "func (b *builder) collapsedTags(ts []*Tag, count int, flatTags bool) []*Tag {\n\tts = SortTags(ts, flatTags)\n\tif len(ts) <= count {\n\t\treturn ts\n\t}\n\n\ttagGroups := make([][]*Tag, count)\n\tfor i, t := range (ts)[:count] {\n\t\ttagGroups[i] = []*Tag{t}\n\t}\n\tfor _, t := range (ts)[count:] {\n\t\tg, d := 0, tagDistance(t, tagGroups[0][0])\n\t\tfor i := 1; i < count; i++ {\n\t\t\tif nd := tagDistance(t, tagGroups[i][0]); nd < d {\n\t\t\t\tg, d = i, nd\n\t\t\t}\n\t\t}\n\t\ttagGroups[g] = append(tagGroups[g], t)\n\t}\n\n\tvar nts []*Tag\n\tfor _, g := range tagGroups {\n\t\tl, w, c := b.tagGroupLabel(g)\n\t\tnts = append(nts, &Tag{\n\t\t\tName: l,\n\t\t\tFlat: w,\n\t\t\tCum: c,\n\t\t})\n\t}\n\treturn SortTags(nts, flatTags)\n}", "func Tags(err error) []Tag {\n\treturn deepAppendTags(nil, err)\n}", "func (o SubnetOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Subnet) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func ReplaceHTMLTags(text string) string {\n\tvar (\n\t\tparser = html.NewTokenizer(strings.NewReader(text))\n\t\ttagStack = stack.New()\n\t\ttextToTag = map[int]string{}\n\t)\n\n\tfor {\n\t\tnode := parser.Next()\n\t\tswitch node {\n\t\tcase html.ErrorToken:\n\t\t\tresult := strings.Replace(textToTag[0], \"&nbsp;\", \" \", -1)\n\t\t\treturn result\n\t\tcase html.TextToken:\n\t\t\tt := string(parser.Text())\n\t\t\ttextToTag[tagStack.Len()] = strings.Join([]string{textToTag[tagStack.Len()], t}, \"\")\n\t\tcase html.StartTagToken:\n\t\t\ttagName, hasAttr := parser.TagName()\n\t\t\tif string(tagName) == scriptTag {\n\t\t\t\t// We can skip script tags, as they are invisible for the user, but we can indicate that there are\n\t\t\t\t// scripts in the task. To skip tag, it is necessary to call Next() two times:\n\t\t\t\t// 1) returns TextToken with the script body\n\t\t\t\t// 2) returns EndTagToken for the closed script tag\n\t\t\t\t// Usually script tag doesn't have any neste tags, so this aproach should work\n\t\t\t\tlog.Printf(\"[INFO] Skipping script tag\")\n\t\t\t\tparser.Next()\n\t\t\t\tparser.Next()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ttag := Tag{Tag: string(tagName), Attrs: map[string]string{}}\n\t\t\tif hasAttr {\n\t\t\t\tfor {\n\t\t\t\t\tattr, val, moreAttr := parser.TagAttr()\n\t\t\t\t\tif DEBUG {\n\t\t\t\t\t\tlog.Printf(\"[DEBUG] Found attr %s\", attr)\n\t\t\t\t\t}\n\t\t\t\t\ttag.Attrs[string(attr)] = string(val)\n\t\t\t\t\tif !moreAttr {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tif DEBUG {\n\t\t\t\tlog.Printf(\"[DEBUG] Found tag %q\", tag)\n\t\t\t}\n\t\t\ttagStack.Push(tag)\n\t\tcase html.EndTagToken:\n\t\t\tvar (\n\t\t\t\taddText string\n\t\t\t\ttagNo = tagStack.Len()\n\t\t\t\ttag = tagStack.Pop()\n\t\t\t\tclosedTag, _ = parser.TagName()\n\t\t\t)\n\t\t\tif tag.(Tag).Tag != string(closedTag) {\n\t\t\t\tlog.Printf(\"[WARNING] Found closed tag %q but expected %q\", closedTag, tag)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif DEBUG {\n\t\t\t\tlog.Printf(\"[DEBUG] Found end of tag %q\", closedTag)\n\t\t\t}\n\t\t\tswitch tag.(Tag).Tag {\n\t\t\tcase iTag:\n\t\t\t\taddText = fmt.Sprintf(\"_%s_\", textToTag[tagNo])\n\t\t\tcase bTag, strongTag:\n\t\t\t\taddText = fmt.Sprintf(\"*%s*\", textToTag[tagNo])\n\t\t\tcase aTag:\n\t\t\t\t// if strings.Compare(string(attr), \"href\") == 0 {\n\t\t\t\taddText = fmt.Sprintf(\"[%s](%s)\", textToTag[tagNo], tag.(Tag).Attrs[\"href\"])\n\t\t\t\t// }\n\t\t\tdefault:\n\t\t\t\taddText = textToTag[tagNo]\n\t\t\t}\n\t\t\ttextToTag[tagStack.Len()] = strings.Join([]string{textToTag[tagStack.Len()], addText}, \"\")\n\t\t\tdelete(textToTag, tagNo)\n\t\t}\n\t}\n}", "func (o LinkOutput) Tags() LinkTagArrayOutput {\n\treturn o.ApplyT(func(v *Link) LinkTagArrayOutput { return v.Tags }).(LinkTagArrayOutput)\n}", "func SortTags(tags []string, sortTag SortTag) []string {\n\tswitch sortTag {\n\tcase SortTagReverse:\n\t\tfor i := len(tags)/2 - 1; i >= 0; i-- {\n\t\t\topp := len(tags) - 1 - i\n\t\t\ttags[i], tags[opp] = tags[opp], tags[i]\n\t\t}\n\t\treturn tags\n\tcase SortTagLexicographical:\n\t\tsort.Strings(tags)\n\t\treturn tags\n\tcase SortTagSemver:\n\t\tsemverIsh := func(s string) string {\n\t\t\ts = strings.TrimLeftFunc(s, func(r rune) bool {\n\t\t\t\treturn !unicode.IsNumber(r)\n\t\t\t})\n\t\t\tif vt := fmt.Sprintf(\"v%s\", s); semver.IsValid(vt) {\n\t\t\t\treturn vt\n\t\t\t}\n\t\t\treturn \"\"\n\t\t}\n\t\tsort.Slice(tags, func(i, j int) bool {\n\t\t\tif c := semver.Compare(semverIsh(tags[i]), semverIsh(tags[j])); c > 0 {\n\t\t\t\treturn true\n\t\t\t} else if c < 0 {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tif c := strings.Count(tags[i], \".\") - strings.Count(tags[j], \".\"); c > 0 {\n\t\t\t\treturn true\n\t\t\t} else if c < 0 {\n\t\t\t\treturn false\n\t\t\t}\n\n\t\t\treturn strings.Compare(tags[i], tags[j]) < 0\n\t\t})\n\t\treturn tags\n\tdefault:\n\t\treturn tags\n\t}\n}", "func (o LookupMulticastDomainResultOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v LookupMulticastDomainResult) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o CloudConfigurationRuleOutput) Tags() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *CloudConfigurationRule) pulumi.StringArrayOutput { return v.Tags }).(pulumi.StringArrayOutput)\n}", "func (o AnalyticsConfigurationFilterOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v AnalyticsConfigurationFilter) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o LookupContactResultOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v LookupContactResult) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o ReportGroupOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *ReportGroup) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (g *GitLocal) Tags(dir string) ([]string, error) {\n\treturn g.GitCLI.Tags(dir)\n}", "func (m Tags) Append(tags ...model.Tag) Tags {\n\tresult := Tags{}\n\tfor n, t := range m {\n\t\tresult[n] = t\n\t}\n\tfor _, t := range tags {\n\t\tresult[*t.Scope] = t\n\t}\n\treturn result\n}", "func (o RunBookOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *RunBook) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o ThingGroupOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *ThingGroup) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o EventIntegrationOutput) Tags() EventIntegrationTagArrayOutput {\n\treturn o.ApplyT(func(v *EventIntegration) EventIntegrationTagArrayOutput { return v.Tags }).(EventIntegrationTagArrayOutput)\n}", "func (o LanguageModelOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *LanguageModel) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func CucumberTagsListToString(tags []string) string {\n\tvar tagList []string\n\tfor _, tag := range tags {\n\t\ttagList = append(tagList, \"@\"+tag)\n\t}\n\treturn strings.Join(tagList, \",\")\n}", "func (src *prometheusMetricsSource) buildTags(m *dto.Metric) map[string]string {\n\tresult := map[string]string{}\n\tfor k, v := range src.tags {\n\t\tif len(v) > 0 {\n\t\t\tresult[k] = v\n\t\t}\n\t}\n\tfor _, lp := range m.Label {\n\t\tif len(lp.GetValue()) > 0 {\n\t\t\tresult[lp.GetName()] = lp.GetValue()\n\t\t}\n\t}\n\treturn result\n}", "func (o RegistryGeoreplicationOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v RegistryGeoreplication) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (o *Options) tagMetrics(rowTags []tag.Tag, addlTags []string) []string {\n\tfinalTags := make([]string, len(o.Tags), len(o.Tags)+len(rowTags)+len(addlTags))\n\tcopy(finalTags, o.Tags)\n\tfor key := range rowTags {\n\t\tfinalTags = append(finalTags,\n\t\t\trowTags[key].Key.Name()+\":\"+rowTags[key].Value)\n\t}\n\tfinalTags = append(finalTags, addlTags...)\n\treturn finalTags\n}", "func (o NetworkInsightsAnalysisOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *NetworkInsightsAnalysis) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (r *LogGroup) Tags() pulumi.MapOutput {\n\treturn (pulumi.MapOutput)(r.s.State[\"tags\"])\n}", "func (o AppOutput) Tags() AppTagArrayOutput {\n\treturn o.ApplyT(func(v *App) AppTagArrayOutput { return v.Tags }).(AppTagArrayOutput)\n}", "func (t *table) appendTags(cr *colReader) {\n\tfor j := range t.cols {\n\t\tv := t.tags[j]\n\t\tif v != nil {\n\t\t\tb := arrow.NewStringBuilder(t.alloc)\n\t\t\tb.Reserve(cr.l)\n\t\t\tb.ReserveData(cr.l * len(v))\n\t\t\tfor i := 0; i < cr.l; i++ {\n\t\t\t\tb.Append(v)\n\t\t\t}\n\t\t\tcr.cols[j] = b.NewArray()\n\t\t\tb.Release()\n\t\t}\n\t}\n}", "func (o AnalyzerOutput) TagsAll() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *Analyzer) pulumi.StringMapOutput { return v.TagsAll }).(pulumi.StringMapOutput)\n}", "func parseTags(tag string) (ret block.Dict) {\n\tif len(tag) > 0 {\n\t\ttags := make(block.Dict)\n\t\tvisitTags(tag, func(k, v string) {\n\t\t\ttags[k] = parseCommas(v)\n\t\t})\n\t\tret = tags\n\t}\n\treturn\n}", "func addTags(s selection, args []string) {\n\tif len(args) < 1 {\n\t\tlog.Fatal(`Usage: A addtags <tags> [options]\n<tags>:\tcomma-separated tags to add, e.g. json,xml\n[options]:\toptions to add, e.g. 'json=omitempty'`)\n\t}\n\targuments := []string{\n\t\t\"-file\", s.filename(), \"-modified\", \"-format\", \"json\", \"-line\", s.lineSel(), \"-add-tags\", args[0],\n\t}\n\tif len(args) > 1 {\n\t\targuments = append(arguments, \"-add-options\", args[1])\n\t}\n\tbuf := runWithStdin(s.archive(), \"gomodifytags\", arguments...)\n\tvar out gomodifytagsOutput\n\tif err := json.Unmarshal([]byte(buf), &out); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif err := s.win.Addr(\"%d,%d\", out.Start, out.End); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif _, err := s.win.Write(\"data\", []byte(strings.Join(out.Lines, \"\\n\")+\"\\n\")); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tshowAddr(s.win, s.start)\n\tif len(out.Errs) != 0 {\n\t\tfmt.Fprintln(os.Stderr, strings.Join(out.Errs, \"\\n\"))\n\t}\n}", "func (o LookupManagedPrefixListResultOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v LookupManagedPrefixListResult) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func combine(original, override map[string]string) map[string]string {\n\t// We know the size must be at least the length of the existing tag map, but\n\t// since values can be overridden we cannot assume the length is the sum of\n\t// both inputs.\n\tcombined := make(map[string]string, len(original))\n\n\tfor k, v := range original {\n\t\tcombined[k] = v\n\t}\n\tfor k, v := range override {\n\t\tcombined[k] = v\n\t}\n\n\treturn combined\n}", "func (o BucketLifecycleConfigurationV2RuleFilterAndOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v BucketLifecycleConfigurationV2RuleFilterAnd) map[string]string { return v.Tags }).(pulumi.StringMapOutput)\n}", "func CompareTags(tags []v1alpha1.Tag, ecrTags []ecr.Tag) bool {\n\tif len(tags) != len(ecrTags) {\n\t\treturn false\n\t}\n\n\tSortTags(tags, ecrTags)\n\n\tfor i, t := range tags {\n\t\tif t.Key != aws.StringValue(ecrTags[i].Key) || t.Value != aws.StringValue(ecrTags[i].Value) {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func (o PlacementGroupOutput) Tags() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *PlacementGroup) pulumi.StringMapOutput { return v.Tags }).(pulumi.StringMapOutput)\n}", "func (e *Element) PrefixTags(tags ...string) {\n\tprefix := strings.Join(tags, \",\")\n\tif e.Tags == \"\" {\n\t\te.Tags = prefix\n\t\treturn\n\t}\n\te.Tags = mergeTags(prefix, strings.Split(e.Tags, \",\"))\n}" ]
[ "0.65004945", "0.64422244", "0.6354086", "0.62970436", "0.6101939", "0.59882367", "0.59166086", "0.5755861", "0.5537153", "0.5528983", "0.5473174", "0.5382972", "0.53700674", "0.5263985", "0.52288485", "0.5228463", "0.5219212", "0.52181876", "0.5166654", "0.5160324", "0.5131595", "0.5130929", "0.5126762", "0.5125288", "0.50954765", "0.5093467", "0.5087405", "0.5071751", "0.50020957", "0.49972856", "0.4989088", "0.4986165", "0.4986165", "0.49657562", "0.4952281", "0.49476495", "0.49473384", "0.494595", "0.4932123", "0.4919427", "0.49149218", "0.4911493", "0.49054968", "0.49043748", "0.48894024", "0.48867774", "0.4882855", "0.48826644", "0.48792806", "0.48792598", "0.48300597", "0.48232442", "0.4823161", "0.48130885", "0.4811566", "0.4800262", "0.48001212", "0.478677", "0.47807005", "0.4780679", "0.47802985", "0.4777557", "0.47757107", "0.47645092", "0.47523415", "0.4752283", "0.4744913", "0.47415334", "0.47387806", "0.4736645", "0.47357377", "0.47346368", "0.4730713", "0.4729671", "0.47242656", "0.47220632", "0.47218436", "0.47206128", "0.47178826", "0.47070816", "0.4700596", "0.46939602", "0.46900198", "0.46897084", "0.46815524", "0.46703568", "0.46571836", "0.4647138", "0.46383485", "0.4634992", "0.46349734", "0.46315885", "0.46287146", "0.46272957", "0.46214592", "0.46166286", "0.4613823", "0.4608713", "0.4608091", "0.4605325" ]
0.7932081
0
do nothing, it's a posix specific feature at the moment
func suspend() {}
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Server) isUnixSocket() bool {\n\treturn false\n}", "func fakeSyscall(duration time.Duration) {\n\truntime.Entersyscall()\n\tfor start := runtime.Nanotime(); runtime.Nanotime()-start < int64(duration); {\n\t}\n\truntime.Exitsyscall()\n}", "func testUnixAddr() string {\n\tf, err := os.CreateTemp(\"\", \"uwsgi-exporter-test\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\taddr := f.Name()\n\tf.Close()\n\tos.Remove(addr)\n\treturn addr\n}", "func Posixfy(path string) string {\n\treturn strings.Replace(path, \"\\\\\", \"/\", -1)\n}", "func fixSyscallName(prefix string, name KProbeName) string {\n\t// see get_syscall_fname in bcc\n\n\tparts := strings.Split(string(name), \"/\")\n\tprobeType := parts[0]\n\trawName := strings.TrimPrefix(parts[1], \"sys_\")\n\n\tout := probeType + \"/\" + prefix + rawName\n\n\treturn out\n}", "func nonblockHack(file *os.File) error {\n\treturn unix.SetNonblock(int(file.Fd()), true)\n}", "func clearErrno()", "func main() {\n\tfile, err := os.Open(\"./temp\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer file.Close()\n\tfdnum := file.Fd()\n\tfmt.Printf(\"fd: %b %b %b %b\\n\", byte(fdnum), byte(fdnum>>8), byte(fdnum>>16), byte(fdnum>>24))\n\tfmt.Printf(\"ready to send fd: %d\\n\", fdnum)\n\t// 编码fd编译传送给其他进程\n\tdata := syscall.UnixRights(int(fdnum))\n\traddr, err := net.ResolveUnixAddr(\"unix\", socksPath)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\t// 连接UnixSock\n\tconn, err := net.DialUnix(\"unix\", nil, raddr)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfmt.Printf(\"has dial, time:%v\\n\", time.Now())\n\ttime.Sleep(10 * time.Second)\n\t// 发送msg\n\tn, oobn, err := conn.WriteMsgUnix(nil, data, nil)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfmt.Printf(\"has write, time:%v\\n\", time.Now())\n\ttime.Sleep(10 * time.Second)\n\tfmt.Printf(\"WriteMsgUnix = n:%d, oobn:%d; want 1, %d\\n\", n, oobn, len(data))\n\tfmt.Printf(\"write %d data success\\n\", n)\n}", "func osfix(workingdir string) (exported map[string]reflect.Value) {\n\texported = stdlib.Symbols[\"os\"]\n\texported[\"Getwd\"] = reflect.ValueOf(func() (string, error) {\n\t\treturn workingdir, nil\n\t})\n\texported[\"Chdir\"] = reflect.ValueOf(func(path string) error {\n\t\treturn &os.PathError{\n\t\t\tPath: path,\n\t\t\tErr: errors.New(\"bearded-wookie disallows changing the current directory at this time\"),\n\t\t}\n\t})\n\texported[\"Exit\"] = reflect.ValueOf(func(code int) {\n\t\tpanic(1)\n\t})\n\n\treturn exported\n}", "func syscall_rawSyscall(fn, a1, a2, a3 uintptr) (r1, r2, err uintptr) {\n\tlibcCall(unsafe.Pointer(abi.FuncPCABI0(syscall)), unsafe.Pointer(&fn))\n\treturn\n}", "func MustCompilePOSIX(str string) *NRegexp {\n\treturn &NRegexp{Regexp: regexp.MustCompilePOSIX(str)}\n}", "func parseStatPOSIX(line string) (plugin.EntryAttributes, string, error) {\n\tvar attr plugin.EntryAttributes\n\tsegments := strings.SplitN(line, \" \", 6)\n\tif len(segments) != 6 {\n\t\treturn attr, \"\", fmt.Errorf(\"Stat did not return 6 components: %v\", line)\n\t}\n\n\tsize, err := strconv.ParseUint(segments[0], 10, 64)\n\tif err != nil {\n\t\treturn attr, \"\", err\n\t}\n\tattr.SetSize(size)\n\n\tatime, err := parseTime(segments[1])\n\tif err != nil {\n\t\treturn attr, \"\", err\n\t}\n\tattr.SetAtime(atime)\n\n\tmtime, err := parseTime(segments[2])\n\tif err != nil {\n\t\treturn attr, \"\", err\n\t}\n\tattr.SetMtime(mtime)\n\n\tctime, err := parseTime(segments[3])\n\tif err != nil {\n\t\treturn attr, \"\", err\n\t}\n\tattr.SetCtime(ctime)\n\n\tmode, err := munge.ToFileMode(\"0x\" + segments[4])\n\tif err != nil {\n\t\treturn attr, \"\", err\n\t}\n\tattr.SetMode(mode)\n\n\treturn attr, segments[5], nil\n}", "func SYSCALL() { ctx.SYSCALL() }", "func RunningInUserNS() bool {\r\n\treturn false\r\n}", "func CompilePOSIX(expr string) (*NRegexp, error) {\n\tre, err := regexp.CompilePOSIX(expr)\n\treturn &NRegexp{Regexp: re}, err\n}", "func IsWindows() bool { return false }", "func ensureNoOsFs(name string, fs afero.Fs) error {\n\tif _, ok := fs.(*afero.OsFs); ok {\n\t\treturn fmt.Errorf(\"gofakes3: invalid OsFs passed to %s,. s3afero backends assume they have control over the filesystem's root. use afero.NewBasePathFs() to avoid misery\", name)\n\t}\n\treturn nil\n}", "func sysMount(device, target, mType string, flag uintptr, data string) error {\n\tif err := syscall.Mount(device, target, mType, flag, data); err != nil {\n\t\treturn err\n\t}\n\n\t// If we have a bind mount or remount, remount...\n\tif flag&syscall.MS_BIND == syscall.MS_BIND &&\n\t\tflag&syscall.MS_RDONLY == syscall.MS_RDONLY {\n\t\treturn syscall.Mount(\n\t\t\tdevice, target, mType, flag|syscall.MS_REMOUNT, data)\n\t}\n\treturn nil\n}", "func OSIsUnix(os string) bool {\n\tfor _, goos := range OSUnix {\n\t\tif os == goos {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func fixStdioPermissions(u *user.ExecUser) error {\n\tvar null syscall.Stat_t\n\tif err := syscall.Stat(\"/dev/null\", &null); err != nil {\n\t\treturn err\n\t}\n\tfor _, fd := range []uintptr{\n\t\tos.Stdin.Fd(),\n\t\tos.Stderr.Fd(),\n\t\tos.Stdout.Fd(),\n\t} {\n\t\tvar s syscall.Stat_t\n\t\tif err := syscall.Fstat(int(fd), &s); err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// skip chown of /dev/null if it was used as one of the STDIO fds.\n\t\tif s.Rdev == null.Rdev {\n\t\t\tcontinue\n\t\t}\n\t\tif err := syscall.Fchown(int(fd), u.Uid, u.Gid); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func Xfopen(tls TLS, path, mode uintptr) uintptr {\n\tp := GoString(path)\n\tvar u uintptr\n\tswitch p {\n\tcase os.Stderr.Name():\n\t\tu = stderr\n\tcase os.Stdin.Name():\n\t\tu = stdin\n\tcase os.Stdout.Name():\n\t\tu = stdout\n\tdefault:\n\t\tvar f *os.File\n\t\tvar err error\n\t\tswitch mode := GoString(mode); mode {\n\t\tcase \"a\":\n\t\t\tif f, err = os.OpenFile(p, os.O_RDWR|os.O_CREATE|os.O_APPEND, 0666); err != nil {\n\t\t\t\tswitch {\n\t\t\t\tcase os.IsPermission(err):\n\t\t\t\t\ttls.setErrno(errno.XEPERM)\n\t\t\t\tdefault:\n\t\t\t\t\ttls.setErrno(errno.XEACCES)\n\t\t\t\t}\n\t\t\t}\n\t\tcase \"r\", \"rb\":\n\t\t\tif f, err = os.OpenFile(p, os.O_RDONLY, 0666); err != nil {\n\t\t\t\tswitch {\n\t\t\t\tcase os.IsNotExist(err):\n\t\t\t\t\ttls.setErrno(errno.XENOENT)\n\t\t\t\tcase os.IsPermission(err):\n\t\t\t\t\ttls.setErrno(errno.XEPERM)\n\t\t\t\tdefault:\n\t\t\t\t\ttls.setErrno(errno.XEACCES)\n\t\t\t\t}\n\t\t\t}\n\t\tcase \"r+b\":\n\t\t\tif f, err = os.OpenFile(p, os.O_RDWR, 0666); err != nil {\n\t\t\t\tswitch {\n\t\t\t\tcase os.IsNotExist(err):\n\t\t\t\t\ttls.setErrno(errno.XENOENT)\n\t\t\t\tcase os.IsPermission(err):\n\t\t\t\t\ttls.setErrno(errno.XEPERM)\n\t\t\t\tdefault:\n\t\t\t\t\ttls.setErrno(errno.XEACCES)\n\t\t\t\t}\n\t\t\t}\n\t\tcase \"w\", \"wb\":\n\t\t\tif f, err = os.OpenFile(p, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0666); err != nil {\n\t\t\t\tswitch {\n\t\t\t\tcase os.IsPermission(err):\n\t\t\t\t\ttls.setErrno(errno.XEPERM)\n\t\t\t\tdefault:\n\t\t\t\t\ttls.setErrno(errno.XEACCES)\n\t\t\t\t}\n\t\t\t}\n\t\tdefault:\n\t\t\tpanic(mode)\n\t\t}\n\t\tif f != nil {\n\t\t\tu = Xmalloc(tls, ptrSize)\n\t\t\tfiles.add(f, u)\n\t\t}\n\t}\n\treturn u\n}", "func TestEndpoints_LocalCreateUnixSocket(t *testing.T) {\n\tendpoints, config, cleanup := newEndpoints(t)\n\tdefer cleanup()\n\n\trequire.NoError(t, endpoints.Up(config))\n\n\tpath := endpoints.LocalSocketPath()\n\tassert.NoError(t, httpGetOverUnixSocket(path))\n\n\t// The unix socket file gets removed after shutdown.\n\tcleanup()\n\tassert.Equal(t, false, shared.PathExists(path))\n}", "func _check_oom(tls *crt.TLS, _x unsafe.Pointer) {\n\tif _x == nil {\n\t\tcrt.Xfprintf(tls, (*crt.XFILE)(Xstderr), str(1151))\n\t\tcrt.Xexit(tls, int32(1))\n\t}\n}", "func LockingPOSIX() MountOption {\n\treturn func(conf *mountConfig) error {\n\t\tconf.initFlags |= InitPOSIXLocks\n\t\treturn nil\n\t}\n}", "func mknodChar0Overlay(cleansedOriginalPath string) error {\n\tdir := filepath.Dir(cleansedOriginalPath)\n\ttmp, err := ioutil.TempDir(dir, \"mc0o\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to create a tmp directory under %s with error %v\", dir, err)\n\t}\n\tdefer os.RemoveAll(tmp)\n\tlower := filepath.Join(tmp, \"l\")\n\tupper := filepath.Join(tmp, \"u\")\n\twork := filepath.Join(tmp, \"w\")\n\tmerged := filepath.Join(tmp, \"m\")\n\tfor _, s := range []string{lower, upper, work, merged} {\n\t\tif err := os.MkdirAll(s, 0700); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to mkdir %s with error %v\", s, err)\n\t\t}\n\t}\n\tdummyBase := \"d\"\n\tlowerDummy := filepath.Join(lower, dummyBase)\n\tif err := ioutil.WriteFile(lowerDummy, []byte{}, 0600); err != nil {\n\t\treturn fmt.Errorf(\"failed to create a dummy lower file %s with error %v\", lowerDummy, err)\n\t}\n\tmOpts := fmt.Sprintf(\"lowerdir=%s,upperdir=%s,workdir=%s\", lower, upper, work)\n\t// docker/pkg/mount.Mount() requires procfs to be mounted. So we use syscall.Mount() directly instead.\n\tif err := syscall.Mount(\"overlay\", merged, \"overlay\", uintptr(0), mOpts); err != nil {\n\t\treturn fmt.Errorf(\"failed to mount overlay (%s) on %s with error %v\", mOpts, merged, err)\n\t}\n\tmergedDummy := filepath.Join(merged, dummyBase)\n\tif err := os.Remove(mergedDummy); err != nil {\n\t\tsyscall.Unmount(merged, 0)\n\t\treturn fmt.Errorf(\"failed to unlink %s with error %v\", mergedDummy, err)\n\t}\n\tif err := syscall.Unmount(merged, 0); err != nil {\n\t\treturn fmt.Errorf(\"failed to unmount %s with error %v\", merged, err)\n\t}\n\tupperDummy := filepath.Join(upper, dummyBase)\n\tif err := isChar0(upperDummy); err != nil {\n\t\treturn err\n\t}\n\tif err := os.Rename(upperDummy, cleansedOriginalPath); err != nil {\n\t\treturn fmt.Errorf(\"failed to rename %s to %s with error %v\", upperDummy, cleansedOriginalPath, err)\n\t}\n\treturn nil\n}", "func TestStraceRun(t *testing.T) {\n\tStraceRun(\"echo asdf > /dev/null\", nil, \"\")\n}", "func syscall_syscall(fn, a1, a2, a3 uintptr) (r1, r2, err uintptr) {\n\tentersyscall()\n\tlibcCall(unsafe.Pointer(abi.FuncPCABI0(syscall)), unsafe.Pointer(&fn))\n\texitsyscall()\n\treturn\n}", "func syscallMode(p PermissionBits) (o uint32) {\n\to |= uint32(p)\n\n\tif p.Setuid() {\n\t\to |= syscall.S_ISUID\n\t}\n\tif p.Setgid() {\n\t\to |= syscall.S_ISGID\n\t}\n\tif p.Sticky() {\n\t\to |= syscall.S_ISVTX\n\t}\n\treturn\n}", "func syscall_rawSyscall10X(fn, a1, a2, a3, a4, a5, a6, a7, a8, a9, a10 uintptr) (r1, r2, err uintptr) {\n\tlibcCall(unsafe.Pointer(abi.FuncPCABI0(syscall10X)), unsafe.Pointer(&fn))\n\treturn\n}", "func checkErrno(err error) error {\n\te, ok := err.(syscall.Errno)\n\tif !ok {\n\t\treturn err\n\t}\n\n\tif e == 0 {\n\t\treturn nil\n\t}\n\n\treturn err\n}", "func ensureUnixLineEndings(ctx *gcp.Context, file ...string) error {\n\tisWriteable, err := ctx.IsWritable(file...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !isWriteable {\n\t\treturn nil\n\t}\n\n\tpath := filepath.Join(file...)\n\tdata, err := ctx.ReadFile(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata = bytes.ReplaceAll(data, []byte{'\\r', '\\n'}, []byte{'\\n'})\n\n\tif err := ctx.WriteFile(path, data, os.FileMode(0755)); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func setPlatform(ea *ExtractionArgs) (success bool) {\n\tswitch platform := runtime.GOOS; platform {\n\tcase osFREEBSD, osLINUX:\n\t\tea.Extractor = extractSectionUnix\n\t\tif ea.Verbose {\n\t\t\tea.ArArgs = append(ea.ArArgs, \"xv\")\n\t\t} else {\n\t\t\tea.ArArgs = append(ea.ArArgs, \"x\")\n\t\t}\n\t\tea.ObjectTypeInArchive = fileTypeELFOBJECT\n\t\tsuccess = true\n\tcase osDARWIN:\n\t\tea.Extractor = extractSectionDarwin\n\t\tea.ArArgs = append(ea.ArArgs, \"-x\")\n\t\tif ea.Verbose {\n\t\t\tea.ArArgs = append(ea.ArArgs, \"-v\")\n\t\t}\n\t\tea.ObjectTypeInArchive = fileTypeMACHOBJECT\n\t\tsuccess = true\n\tdefault:\n\t\tLogError(\"Unsupported platform: %s.\", platform)\n\t}\n\treturn\n}", "func unixPath(s string) string {\n\ts = filepath.Clean(s)\n\ts = strings.Replace(s, \"\\\\\", \"/\", -1)\n\n\t// Duplicate clean for trailing slashes that were previously windows ones.\n\treturn filepath.Clean(s)\n}", "func PrintSyscall(sc int32) string {\n\tvar syscalls = map[int32]string{\n\t\t0: \"read\",\n\t\t1: \"write\",\n\t\t2: \"open\",\n\t\t3: \"close\",\n\t\t4: \"newstat\",\n\t\t5: \"fstat\",\n\t\t6: \"newlstat\",\n\t\t7: \"poll\",\n\t\t8: \"lseek\",\n\t\t9: \"mmap\",\n\t\t10: \"mprotect\",\n\t\t11: \"munmap\",\n\t\t12: \"brk\",\n\t\t13: \"rt_sigaction\",\n\t\t14: \"rt_sigprocmask\",\n\t\t15: \"rt_sigreturn\",\n\t\t16: \"ioctl\",\n\t\t17: \"pread64\",\n\t\t18: \"pwrite64\",\n\t\t19: \"readv\",\n\t\t20: \"writev\",\n\t\t21: \"access\",\n\t\t22: \"pipe\",\n\t\t23: \"select\",\n\t\t24: \"sched_yield\",\n\t\t25: \"mremap\",\n\t\t26: \"msync\",\n\t\t27: \"mincore\",\n\t\t28: \"madvise\",\n\t\t29: \"shmget\",\n\t\t30: \"shmat\",\n\t\t31: \"shmctl\",\n\t\t32: \"dup\",\n\t\t33: \"dup2\",\n\t\t34: \"pause\",\n\t\t35: \"nanosleep\",\n\t\t36: \"getitimer\",\n\t\t37: \"alarm\",\n\t\t38: \"setitimer\",\n\t\t39: \"getpid\",\n\t\t40: \"sendfile\",\n\t\t41: \"socket\",\n\t\t42: \"connect\",\n\t\t43: \"accept\",\n\t\t44: \"sendto\",\n\t\t45: \"recvfrom\",\n\t\t46: \"sendmsg\",\n\t\t47: \"recvmsg\",\n\t\t48: \"shutdown\",\n\t\t49: \"bind\",\n\t\t50: \"listen\",\n\t\t51: \"getsockname\",\n\t\t52: \"getpeername\",\n\t\t53: \"socketpair\",\n\t\t54: \"setsockopt\",\n\t\t55: \"getsockopt\",\n\t\t56: \"clone\",\n\t\t57: \"fork\",\n\t\t58: \"vfork\",\n\t\t59: \"execve\",\n\t\t60: \"exit\",\n\t\t61: \"wait4\",\n\t\t62: \"kill\",\n\t\t63: \"uname\",\n\t\t64: \"semget\",\n\t\t65: \"semop\",\n\t\t66: \"semctl\",\n\t\t67: \"shmdt\",\n\t\t68: \"msgget\",\n\t\t69: \"msgsnd\",\n\t\t70: \"msgrcv\",\n\t\t71: \"msgctl\",\n\t\t72: \"fcntl\",\n\t\t73: \"flock\",\n\t\t74: \"fsync\",\n\t\t75: \"fdatasync\",\n\t\t76: \"truncate\",\n\t\t77: \"ftruncate\",\n\t\t78: \"getdents\",\n\t\t79: \"getcwd\",\n\t\t80: \"chdir\",\n\t\t81: \"fchdir\",\n\t\t82: \"rename\",\n\t\t83: \"mkdir\",\n\t\t84: \"rmdir\",\n\t\t85: \"creat\",\n\t\t86: \"link\",\n\t\t87: \"unlink\",\n\t\t88: \"symlink\",\n\t\t89: \"readlink\",\n\t\t90: \"chmod\",\n\t\t91: \"fchmod\",\n\t\t92: \"chown\",\n\t\t93: \"fchown\",\n\t\t94: \"lchown\",\n\t\t95: \"umask\",\n\t\t96: \"gettimeofday\",\n\t\t97: \"getrlimit\",\n\t\t98: \"getrusage\",\n\t\t99: \"sysinfo\",\n\t\t100: \"times\",\n\t\t101: \"ptrace\",\n\t\t102: \"getuid\",\n\t\t103: \"syslog\",\n\t\t104: \"getgid\",\n\t\t105: \"setuid\",\n\t\t106: \"setgid\",\n\t\t107: \"geteuid\",\n\t\t108: \"getegid\",\n\t\t109: \"setpgid\",\n\t\t110: \"getppid\",\n\t\t111: \"getpgrp\",\n\t\t112: \"setsid\",\n\t\t113: \"setreuid\",\n\t\t114: \"setregid\",\n\t\t115: \"getgroups\",\n\t\t116: \"setgroups\",\n\t\t117: \"setresuid\",\n\t\t118: \"getresuid\",\n\t\t119: \"setresgid\",\n\t\t120: \"getresgid\",\n\t\t121: \"getpgid\",\n\t\t122: \"setfsuid\",\n\t\t123: \"setfsgid\",\n\t\t124: \"getsid\",\n\t\t125: \"capget\",\n\t\t126: \"capset\",\n\t\t127: \"rt_sigpending\",\n\t\t128: \"rt_sigtimedwait\",\n\t\t129: \"rt_sigqueueinfo\",\n\t\t130: \"rt_sigsuspend\",\n\t\t131: \"sigaltstack\",\n\t\t132: \"utime\",\n\t\t133: \"mknod\",\n\t\t134: \"uselib\",\n\t\t135: \"personality\",\n\t\t136: \"ustat\",\n\t\t137: \"statfs\",\n\t\t138: \"fstatfs\",\n\t\t139: \"sysfs\",\n\t\t140: \"getpriority\",\n\t\t141: \"setpriority\",\n\t\t142: \"sched_setparam\",\n\t\t143: \"sched_getparam\",\n\t\t144: \"sched_setscheduler\",\n\t\t145: \"sched_getscheduler\",\n\t\t146: \"sched_get_priority_max\",\n\t\t147: \"sched_get_priority_min\",\n\t\t148: \"sched_rr_get_interval\",\n\t\t149: \"mlock\",\n\t\t150: \"munlock\",\n\t\t151: \"mlockall\",\n\t\t152: \"munlockall\",\n\t\t153: \"vhangup\",\n\t\t154: \"modify_ldt\",\n\t\t155: \"pivot_root\",\n\t\t156: \"sysctl\",\n\t\t157: \"prctl\",\n\t\t158: \"arch_prctl\",\n\t\t159: \"adjtimex\",\n\t\t160: \"setrlimit\",\n\t\t161: \"chroot\",\n\t\t162: \"sync\",\n\t\t163: \"acct\",\n\t\t164: \"settimeofday\",\n\t\t165: \"mount\",\n\t\t166: \"umount\",\n\t\t167: \"swapon\",\n\t\t168: \"swapoff\",\n\t\t169: \"reboot\",\n\t\t170: \"sethostname\",\n\t\t171: \"setdomainname\",\n\t\t172: \"iopl\",\n\t\t173: \"ioperm\",\n\t\t174: \"create_module\",\n\t\t175: \"init_module\",\n\t\t176: \"delete_module\",\n\t\t177: \"get_kernel_syms\",\n\t\t178: \"query_module\",\n\t\t179: \"quotactl\",\n\t\t180: \"nfsservctl\",\n\t\t181: \"getpmsg\",\n\t\t182: \"putpmsg\",\n\t\t183: \"afs\",\n\t\t184: \"tuxcall\",\n\t\t185: \"security\",\n\t\t186: \"gettid\",\n\t\t187: \"readahead\",\n\t\t188: \"setxattr\",\n\t\t189: \"lsetxattr\",\n\t\t190: \"fsetxattr\",\n\t\t191: \"getxattr\",\n\t\t192: \"lgetxattr\",\n\t\t193: \"fgetxattr\",\n\t\t194: \"listxattr\",\n\t\t195: \"llistxattr\",\n\t\t196: \"flistxattr\",\n\t\t197: \"removexattr\",\n\t\t198: \"lremovexattr\",\n\t\t199: \"fremovexattr\",\n\t\t200: \"tkill\",\n\t\t201: \"time\",\n\t\t202: \"futex\",\n\t\t203: \"sched_setaffinity\",\n\t\t204: \"sched_getaffinity\",\n\t\t205: \"set_thread_area\",\n\t\t206: \"io_setup\",\n\t\t207: \"io_destroy\",\n\t\t208: \"io_getevents\",\n\t\t209: \"io_submit\",\n\t\t210: \"io_cancel\",\n\t\t211: \"get_thread_area\",\n\t\t212: \"lookup_dcookie\",\n\t\t213: \"epoll_create\",\n\t\t214: \"epoll_ctl_old\",\n\t\t215: \"epoll_wait_old\",\n\t\t216: \"remap_file_pages\",\n\t\t217: \"getdents64\",\n\t\t218: \"set_tid_address\",\n\t\t219: \"restart_syscall\",\n\t\t220: \"semtimedop\",\n\t\t221: \"fadvise64\",\n\t\t222: \"timer_create\",\n\t\t223: \"timer_settime\",\n\t\t224: \"timer_gettime\",\n\t\t225: \"timer_getoverrun\",\n\t\t226: \"timer_delete\",\n\t\t227: \"clock_settime\",\n\t\t228: \"clock_gettime\",\n\t\t229: \"clock_getres\",\n\t\t230: \"clock_nanosleep\",\n\t\t231: \"exit_group\",\n\t\t232: \"epoll_wait\",\n\t\t233: \"epoll_ctl\",\n\t\t234: \"tgkill\",\n\t\t235: \"utimes\",\n\t\t236: \"vserver\",\n\t\t237: \"mbind\",\n\t\t238: \"set_mempolicy\",\n\t\t239: \"get_mempolicy\",\n\t\t240: \"mq_open\",\n\t\t241: \"mq_unlink\",\n\t\t242: \"mq_timedsend\",\n\t\t243: \"mq_timedreceive\",\n\t\t244: \"mq_notify\",\n\t\t245: \"mq_getsetattr\",\n\t\t246: \"kexec_load\",\n\t\t247: \"waitid\",\n\t\t248: \"add_key\",\n\t\t249: \"request_key\",\n\t\t250: \"keyctl\",\n\t\t251: \"ioprio_set\",\n\t\t252: \"ioprio_get\",\n\t\t253: \"inotify_init\",\n\t\t254: \"inotify_add_watch\",\n\t\t255: \"inotify_rm_watch\",\n\t\t256: \"migrate_pages\",\n\t\t257: \"openat\",\n\t\t258: \"mkdirat\",\n\t\t259: \"mknodat\",\n\t\t260: \"fchownat\",\n\t\t261: \"futimesat\",\n\t\t262: \"newfstatat\",\n\t\t263: \"unlinkat\",\n\t\t264: \"renameat\",\n\t\t265: \"linkat\",\n\t\t266: \"symlinkat\",\n\t\t267: \"readlinkat\",\n\t\t268: \"fchmodat\",\n\t\t269: \"faccessat\",\n\t\t270: \"pselect6\",\n\t\t271: \"ppoll\",\n\t\t272: \"unshare\",\n\t\t273: \"set_robust_list\",\n\t\t274: \"get_robust_list\",\n\t\t275: \"splice\",\n\t\t276: \"tee\",\n\t\t277: \"sync_file_range\",\n\t\t278: \"vmsplice\",\n\t\t279: \"move_pages\",\n\t\t280: \"utimensat\",\n\t\t281: \"epoll_pwait\",\n\t\t282: \"signalfd\",\n\t\t283: \"timerfd_create\",\n\t\t284: \"eventfd\",\n\t\t285: \"fallocate\",\n\t\t286: \"timerfd_settime\",\n\t\t287: \"timerfd_gettime\",\n\t\t288: \"accept4\",\n\t\t289: \"signalfd4\",\n\t\t290: \"eventfd2\",\n\t\t291: \"epoll_create1\",\n\t\t292: \"dup3\",\n\t\t293: \"pipe2\",\n\t\t294: \"ionotify_init1\",\n\t\t295: \"preadv\",\n\t\t296: \"pwritev\",\n\t\t297: \"rt_tgsigqueueinfo\",\n\t\t298: \"perf_event_open\",\n\t\t299: \"recvmmsg\",\n\t\t300: \"fanotify_init\",\n\t\t301: \"fanotify_mark\",\n\t\t302: \"prlimit64\",\n\t\t303: \"name_tohandle_at\",\n\t\t304: \"open_by_handle_at\",\n\t\t305: \"clock_adjtime\",\n\t\t306: \"sycnfs\",\n\t\t307: \"sendmmsg\",\n\t\t308: \"setns\",\n\t\t309: \"getcpu\",\n\t\t310: \"process_vm_readv\",\n\t\t311: \"process_vm_writev\",\n\t\t312: \"kcmp\",\n\t\t313: \"finit_module\",\n\t\t314: \"sched_setattr\",\n\t\t315: \"sched_getattr\",\n\t\t316: \"renameat2\",\n\t\t317: \"seccomp\",\n\t\t318: \"getrandom\",\n\t\t319: \"memfd_create\",\n\t\t320: \"kexec_file_load\",\n\t\t321: \"bpf\",\n\t\t322: \"execveat\",\n\t\t323: \"userfaultfd\",\n\t\t324: \"membarrier\",\n\t\t325: \"mlock2\",\n\t\t326: \"copy_file_range\",\n\t\t327: \"preadv2\",\n\t\t328: \"pwritev2\",\n\t\t329: \"pkey_mprotect\",\n\t\t330: \"pkey_alloc\",\n\t\t331: \"pkey_free\",\n\t\t332: \"statx\",\n\t\t333: \"io_pgetevents\",\n\t\t334: \"rseq\",\n\t}\n\tvar res string\n\tif scName, ok := syscalls[sc]; ok {\n\t\tres = scName\n\t} else {\n\t\tres = strconv.Itoa(int(sc))\n\t}\n\treturn res\n}", "func isWellKnownFS(fn string) bool {\n\tvar fs syscall.Statfs_t\n\terr := syscall.Statfs(fn, &fs)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif fs.Type == EXT4_SUPER_MAGIC || fs.Type == TMPFS_MAGIC {\n\t\treturn true\n\t}\n\treturn false\n}", "func testFullCycleFilesystem(t *testing.T) {\n\tt.Log(\"TODO\")\n}", "func PosixShellEscape(input string) string {\n\tif input == \"\" {\n\t\treturn \"''\"\n\t}\n\n\tvar sb strings.Builder\n\tsb.Grow(len(input) * 2)\n\n\tescape := false\n\tfor _, c := range []byte(input) {\n\t\tmode := posixModeTable[c]\n\t\tswitch mode {\n\t\tcase quo:\n\t\t\tsb.WriteByte(c)\n\t\t\tescape = true\n\t\tcase \"\":\n\t\t\tsb.WriteByte(c)\n\t\tdefault:\n\t\t\tsb.WriteString(string(mode))\n\t\t\tescape = true\n\t\t}\n\t}\n\n\tif escape {\n\t\treturn `\"` + sb.String() + `\"`\n\t}\n\n\treturn sb.String()\n}", "func syscallMode(i os.FileMode) (o uint32) {\n\to |= uint32(i.Perm())\n\tif i&os.ModeSetuid != 0 {\n\t\to |= unix.S_ISUID\n\t}\n\tif i&os.ModeSetgid != 0 {\n\t\to |= unix.S_ISGID\n\t}\n\tif i&os.ModeSticky != 0 {\n\t\to |= unix.S_ISVTX\n\t}\n\tif i&os.ModeNamedPipe != 0 {\n\t\to |= unix.S_IFIFO\n\t}\n\tif i&os.ModeDevice != 0 {\n\t\tswitch i & os.ModeCharDevice {\n\t\tcase 0:\n\t\t\to |= unix.S_IFBLK\n\t\tdefault:\n\t\t\to |= unix.S_IFCHR\n\t\t}\n\t}\n\treturn\n}", "func syscall_syscallX(fn, a1, a2, a3 uintptr) (r1, r2, err uintptr) {\n\tentersyscall()\n\tlibcCall(unsafe.Pointer(abi.FuncPCABI0(syscallX)), unsafe.Pointer(&fn))\n\texitsyscall()\n\treturn\n}", "func testLddBadSo(t *testing.T) {\n\ttempDir, err := ioutil.TempDir(\"\", \"ldd\")\n\tif err != nil {\n\t\tt.Fatalf(\"TempDir: %v\", err)\n\t}\n\tdefer os.RemoveAll(tempDir)\n\tif err := os.Setenv(\"LD_LIBRARY_PATH\", tempDir); err != nil {\n\t\tt.Fatalf(\"Setting LDD_LIBRARY_PATH to %v: want nil, got %v\", tempDir, err)\n\t}\n\tif _, err := Ldd([]string{\"/bin/date\"}); err == nil {\n\t\tt.Fatalf(\"Ldd on /bin/date: want err, got nil\")\n\t}\n\tt.Logf(\"Err on bad dir is %v\", err)\n\n}", "func unset_pid(opt *Opts) bool {\n\tif opt.mnt != \"\" {\n\t\tif opt.pid != false {\n\t\t\tif err := syscall.Unmount(\"/proc\", 0); err != nil {\n\t\t\t\tfmt.Println(\"Error unsetting PID namespace.\")\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t}\n\t} else {\n\t\treturn false\n\t}\n\treturn true\n}", "func Xrename(tls TLS, oldpath, newpath uintptr) int32 {\n\tpanic(\"TOD\")\n}", "func TestStraceParse2Basic(t *testing.T) {\n\tnopen := 0\n\tnexec := 0\n\tfor _, l := range straceout {\n\t\tif strings.Contains(l, \" open(\") {\n\t\t\tnopen++\n\t\t}\n\t\tif strings.Contains(l, \" execve(\") {\n\t\t\tnexec++\n\t\t}\n\t}\n\tsyscalls := map[string]int{}\n\tfor info := range StraceParse2(StraceParse1(ChanFromList(straceout))) {\n\t\tsyscalls[info.syscall]++\n\t}\n\tif nopen != syscalls[\"open\"] {\n\t\tt.Errorf(\"\\\"open\\\" count mismatch: %d != %d\", nopen, syscalls[\"open\"])\n\t}\n\tif nexec != syscalls[\"execve\"] {\n\t\tt.Errorf(\"\\\"execve\\\" count mismatch: %d != %d\", nexec, syscalls[\"execve\"])\n\t}\n}", "func Xlstat(tls TLS, file, buf uintptr) int32 {\n\tpanic(\"TODO\")\n}", "func dos2unix(b []byte) []byte {\n\tconst cr, lf = '\\r', '\\n'\n\n\treturn bytes.ReplaceAll(bytes.ReplaceAll(bytes.ReplaceAll(b, []byte{cr, lf}, []byte{lf}), []byte{lf, cr}, []byte{lf}), []byte{cr}, []byte{lf})\n}", "func osDetector() {\n\tfmt.Print(\"Go runs on \")\n\tswitch os := runtime.GOOS; os {\n\tcase \"darwin\":\n\t\tfmt.Println(\"OS X.\")\n\tcase \"linux\":\n\t\tfmt.Println(\"Linux.\")\n\tdefault:\n\t\t// freebsd, openbsd,\n\t\t// plan9, windows...\n\t\tfmt.Printf(\"%s.\\n\", os)\n\t}\n}", "func checkSystem() derrors.Error {\n\treturn nil\n}", "func Xpopen(tls TLS, command, typ uintptr) uintptr {\n\tpanic(\"TODO popen\")\n}", "func mstartm0() {\n\t// Create an extra M for callbacks on threads not created by Go.\n\t// An extra M is also needed on Windows for callbacks created by\n\t// syscall.NewCallback. See issue #6751 for details.\n\tif (iscgo || GOOS == \"windows\") && !cgoHasExtraM {\n\t\tcgoHasExtraM = true\n\t\tnewextram()\n\t}\n\tinitsig(false)\n}", "func UnixSocket(socket string) func(*Attacker) {\n\treturn func(a *Attacker) {\n\t\tif tr, ok := a.client.Transport.(*http.Transport); socket != \"\" && ok {\n\t\t\ttr.DialContext = func(_ context.Context, _, _ string) (net.Conn, error) {\n\t\t\t\treturn net.Dial(\"unix\", socket)\n\t\t\t}\n\t\t}\n\t}\n}", "func TestNofail(t *testing.T) {\n\tdir := test_helpers.InitFS(t)\n\tmnt := dir + \".mnt\"\n\ttest_helpers.MountOrFatal(t, dir, mnt, \"-nofail\", \"-extpass=echo test\")\n\tdefer test_helpers.UnmountPanic(mnt)\n}", "func Xfchown(tls *TLS, fd int32, owner, group uint32) int32 {\n\tpanic(\"TODO fchown\")\n}", "func cleanRecordName(name string) string {\n\tstr := name\n\tif strings.HasPrefix(name, \"\\\\052\") {\n\t\tstr = strings.Replace(name, \"\\\\052\", \"*\", 1)\n\t\tlog.Printf(\"[DEBUG] Replacing octal \\\\052 for * in: %s\", name)\n\t}\n\treturn str\n}", "func fullFilePath2(fSys *afero.Afero, fn string) string {\n\txType := reflect.TypeOf(fSys)\n\txValue := reflect.ValueOf(fSys)\n\tfmt.Println(xType, xValue) // Os: *afero.Afero &{0x85d228} // MemMap: *afero.Afero &{0xc00007bb60}\n\tif fn == \"none\" || fn == \"off\" {\n\t\treturn fn\n\t}\n\tvar e error\n\tvar s string\n\t//if xValue < &{0xc000000000} {\n\ts, e = filepath.Abs(fn)\n\tmsg.InfoOnErr(e, fmt.Sprintf(\"failed to parse %s\\n\", fn))\n\t//}\n\treturn s\n}", "func (e *FileEvent) ignoreLinux() bool {\n\t// Ignore anything the inotify API says to ignore\n\tif e.mask&sys_IN_IGNORED == sys_IN_IGNORED {\n\t\treturn true\n\t}\n\n\t// If the event is not a DELETE or RENAME, the file must exist.\n\t// Otherwise the event is ignored.\n\t// *Note*: this was put in place because it was seen that a MODIFY\n\t// event was sent after the DELETE. This ignores that MODIFY and\n\t// assumes a DELETE will come or has come if the file doesn't exist.\n\tif !(e.IsDelete() || e.IsRename()) {\n\t\t_, statErr := os.Lstat(e.Name)\n\t\treturn os.IsNotExist(statErr)\n\t}\n\treturn false\n}", "func createMockUnixUname(kernelVersion string) func(*unix.Utsname) error {\n\treturn func(uname *unix.Utsname) error {\n\t\tvar kernelVersionAsByteArr [65]byte\n\t\tcopy(kernelVersionAsByteArr[:], kernelVersion)\n\t\tuname.Release = kernelVersionAsByteArr\n\t\treturn nil\n\t}\n}", "func forceUnmount(target string) (err error) {\n\t// Simple retry logic for unmount\n\tfor i := 0; i < 10; i++ {\n\t\tif err = sysUnmount(target, 0); err == nil {\n\t\t\treturn nil\n\t\t}\n\t\ttime.Sleep(100 * time.Millisecond)\n\t}\n\treturn\n}", "func isUnixAddrResolvable(fl FieldLevel) bool {\n\t_, err := net.ResolveUnixAddr(\"unix\", fl.Field().String())\n\n\treturn err == nil\n}", "func osName() string {\r\n\treturn runtime.GOOS\r\n}", "func testCmdMountFilesystem(t *testing.T) {\n\tt.Log(\"TODO\")\n}", "func testCmdUnmountFilesystem(t *testing.T) {\n\tt.Log(\"TODO\")\n}", "func noescape(p unsafe.Pointer) unsafe.Pointer {\n\tx := uintptr(p)\n\treturn unsafe.Pointer(x ^ 0) //nolint:staticcheck\n}", "func noescape(p unsafe.Pointer) unsafe.Pointer {\n\tx := uintptr(p)\n\treturn unsafe.Pointer(x ^ 0) //nolint:staticcheck\n}", "func WindowsServerChecks() {\n\n}", "func sharedUnixSocketVolumeMount() apiv1.VolumeMount {\n\treturn apiv1.VolumeMount{\n\t\tName: sidecar.UnixDomainSocketVolume,\n\t\tMountPath: pluggableComponentSocketMountPath,\n\t}\n}", "func Renameatx_np(fromfd int, from string, tofd int, to string, flags uint) (err error) {\n\tvar _p0 *byte\n\t_p0, err = unix.BytePtrFromString(from)\n\tif err != nil {\n\t\treturn\n\t}\n\tvar _p1 *byte\n\t_p1, err = unix.BytePtrFromString(to)\n\tif err != nil {\n\t\treturn\n\t}\n\t_, _, e1 := syscall_syscall6(libc_renameatx_np_trampoline_addr, uintptr(fromfd), uintptr(unsafe.Pointer(_p0)), uintptr(tofd), uintptr(unsafe.Pointer(_p1)), uintptr(flags), 0)\n\tif e1 != 0 {\n\t\terr = e1\n\t}\n\treturn\n}", "func Xchdir(tls *TLS, path uintptr) int32 {\n\tpanic(\"TODO chdir\")\n}", "func syscall_syscall10(fn, a1, a2, a3, a4, a5, a6, a7, a8, a9, a10 uintptr) (r1, r2, err uintptr) {\n\tentersyscall()\n\tlibcCall(unsafe.Pointer(abi.FuncPCABI0(syscall10)), unsafe.Pointer(&fn))\n\texitsyscall()\n\treturn\n}", "func (info *unixFileInfo) fill() {\n\tinfo.mode = os.FileMode(info.sys.Mode & 0777)\n\tswitch info.sys.Mode & unix.S_IFMT {\n\tcase unix.S_IFBLK:\n\t\tinfo.mode |= os.ModeDevice\n\tcase unix.S_IFCHR:\n\t\tinfo.mode |= os.ModeDevice | os.ModeCharDevice\n\tcase unix.S_IFDIR:\n\t\tinfo.mode |= os.ModeDir\n\tcase unix.S_IFIFO:\n\t\tinfo.mode |= os.ModeNamedPipe\n\tcase unix.S_IFLNK:\n\t\tinfo.mode |= os.ModeSymlink\n\tcase unix.S_IFREG:\n\t\t// nothing to do\n\tcase unix.S_IFSOCK:\n\t\tinfo.mode |= os.ModeSocket\n\t}\n\tif info.sys.Mode&unix.S_ISGID != 0 {\n\t\tinfo.mode |= os.ModeSetgid\n\t}\n\tif info.sys.Mode&unix.S_ISUID != 0 {\n\t\tinfo.mode |= os.ModeSetuid\n\t}\n\tif info.sys.Mode&unix.S_ISVTX != 0 {\n\t\tinfo.mode |= os.ModeSticky\n\t}\n}", "func getSyscallPrefix() (string, error) {\n\n\tsyscallPrefixes := []string{\n\t\t\"__sys_\",\n\t\t\"sys_\",\n\t\t\"__x64_sys_\",\n\t\t\"__x32_compat_sys_\",\n\t\t\"__ia32_compat_sys_\",\n\t\t\"__arm64_sys_\",\n\t\t\"__s390x_sys_\",\n\t\t\"__s390_sys_\",\n\t}\n\n\tkallsyms := path.Join(util.GetProcRoot(), \"kallsyms\")\n\tfile, err := os.Open(kallsyms)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tscanner := bufio.NewScanner(file)\n\n\tfor scanner.Scan() {\n\t\tline := scanner.Text()\n\t\tfor _, prefix := range syscallPrefixes {\n\t\t\tif strings.HasSuffix(line, \" \"+prefix+\"socket\") {\n\t\t\t\treturn prefix, nil\n\t\t\t}\n\t\t}\n\t}\n\n\treturn \"\", fmt.Errorf(\"could not get syscall prefix\")\n}", "func getSysctrlEnv(env []string) []string {\n\tfoundLC := false\n\tfor i, line := range env {\n\t\tif strings.HasPrefix(line, \"LC_ALL\") {\n\t\t\tenv[i] = \"LC_ALL=C\"\n\t\t\tfoundLC = true\n\t\t}\n\t}\n\tif !foundLC {\n\t\tenv = append(env, \"LC_ALL=C\")\n\t}\n\treturn env\n}", "func libc_unsetenv(name *byte) int32", "func normalizeOS(name string) string {\n\tsp := strings.SplitN(name, \" \", 3)\n\tif len(sp) != 3 || sp[1] != \"NT\" {\n\t\treturn name\n\t}\n\n\tswitch sp[2] {\n\tcase \"5.0\":\n\t\treturn \"Windows 2000\"\n\tcase \"5.01\":\n\t\treturn \"Windows 2000, Service Pack 1 (SP1)\"\n\tcase \"5.1\":\n\t\treturn \"Windows XP\"\n\tcase \"5.2\":\n\t\treturn \"Windows XP x64 Edition\"\n\tcase \"6.0\":\n\t\treturn \"Windows Vista\"\n\tcase \"6.1\":\n\t\treturn \"Windows 7\"\n\tcase \"6.2\":\n\t\treturn \"Windows 8\"\n\tcase \"6.3\":\n\t\treturn \"Windows 8.1\"\n\tcase \"10.0\":\n\t\treturn \"Windows 10\"\n\t}\n\treturn name\n}", "func isSystemdBasedOS() bool {\n\treturn exec.Command(\"/usr/bin/pidof\", \"systemd\").Run() == nil || exec.Command(\"/bin/pidof\", \"systemd\").Run() == nil\n}", "func daemonise() {\n\t// Drop privileges by switching to nobody user and group\n\tif _, _, err := syscall.Syscall(syscall.SYS_SETGID, 65534, 0, 0); err != 0 {\n\t\tos.Exit(1)\n\t}\n\tif _, _, err := syscall.Syscall(syscall.SYS_SETUID, 65534, 0, 0); err != 0 {\n\t\tos.Exit(1)\n\t}\n\n\t// Do first fork\n\tpid, _, _ := syscall.Syscall(syscall.SYS_FORK, 0, 0, 0)\n\n\t// Exit in parent process\n\tswitch pid {\n\tcase 0:\n\t\t// Child process, carry on\n\t\tbreak\n\tdefault:\n\t\t// Parent process, exit cleanly\n\t\tos.Exit(0)\n\t}\n\n\t// Call setsid\n\t_, err := syscall.Setsid()\n\tif err != nil {\n\t\tos.Exit(1)\n\t}\n\n\t// Fork again\n\tpid, _, _ = syscall.Syscall(syscall.SYS_FORK, 0, 0, 0)\n\n\t// Exit in parent again\n\tswitch pid {\n\tcase 0:\n\t\t// Child process, carry on\n\t\tbreak\n\tdefault:\n\t\t// Parent process, exit cleanly\n\t\tos.Exit(0)\n\t}\n\n\t// Clear umask\n\tsyscall.Umask(0)\n\n\t// Change working directory\n\terr = syscall.Chdir(\"/\")\n\tif err != nil {\n\t\tos.Exit(1)\n\t}\n\n\t// Duplicate /dev/null to stdin, stdout and stderr\n\tnullFile, err := os.OpenFile(\"/dev/null\", os.O_RDWR, 0)\n\tif err != nil {\n\t\tos.Exit(1)\n\t}\n\tnullFd := nullFile.Fd()\n\tsyscall.Dup2(int(nullFd), int(os.Stdin.Fd()))\n\tsyscall.Dup2(int(nullFd), int(os.Stdout.Fd()))\n\tsyscall.Dup2(int(nullFd), int(os.Stderr.Fd()))\n\n}", "func syscall_syscall10X(fn, a1, a2, a3, a4, a5, a6, a7, a8, a9, a10 uintptr) (r1, r2, err uintptr) {\n\tentersyscall()\n\tlibcCall(unsafe.Pointer(abi.FuncPCABI0(syscall10X)), unsafe.Pointer(&fn))\n\texitsyscall()\n\treturn\n}", "func set_uid(opt *Opts) bool {\n\tif opt.uid == false {\n\t\treturn false\n\t}\n\treturn true\n}", "func unixFdatasync(fd int) (err error) {\n\t_, _, e1 := unix.Syscall(unix.SYS_FDATASYNC, uintptr(fd), 0, 0)\n\tif e1 != 0 {\n\t\terr = errnoErr(e1)\n\t}\n\treturn\n}", "func defaultSolarisIfNameCmd() []string {\n\treturn []string{\"/usr/sbin/route\", \"-n\", \"get\", \"default\"}\n}", "func sharedUnixSocketVolume() apiv1.Volume {\n\treturn apiv1.Volume{\n\t\tName: sidecar.UnixDomainSocketVolume,\n\t\tVolumeSource: apiv1.VolumeSource{\n\t\t\tEmptyDir: &apiv1.EmptyDirVolumeSource{},\n\t\t},\n\t}\n}", "func (e *ProgramEnv) Clearenv() {\n\tos.Clearenv()\n}", "func masterSock() string {\r\n\ts := \"/var/tmp/824-mr-\"\r\n\ts += strconv.Itoa(os.Getuid())\r\n\treturn s\r\n}", "func runLinux(path, parameters string, seconds int) (string, bool) {\n\n\tctx, cancel := context.WithTimeout(context.Background(), time.Duration(seconds)*time.Second)\n\tdefer cancel()\n\n\t// Runs as nobody\n\tcommand := exec.CommandContext(ctx, `setpriv`, `--no-new-privs`, `--reuid=nobody`, path, `-PARAMETERS=`+parameters)\n\n\toutputBytes, err := command.CombinedOutput()\n\toutput := strings.TrimSuffix(string(outputBytes), \"\\n\")\n\n\tif ctx.Err() == context.DeadlineExceeded {\n\t\treturn errors.Render(errors.ExecutionTimeout, path), false\n\t}\n\n\tif err != nil {\n\t\treturn err.Error() + \"\\n\" + output, false\n\t}\n\n\treturn output, true\n}", "func (mfs *MountedFS) DisableFsIDCheck() { mfs.checkFsID = false }", "func ExtensionPOSIXRename() *sshfx.ExtensionPair {\n\treturn &sshfx.ExtensionPair{\n\t\tName: extensionPOSIXRename,\n\t\tData: \"1\",\n\t}\n}", "func setRootless() error {\n\tinitRootless = true\n\tfile, err := os.Open(uidMapPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\tbuf := bufio.NewReader(file)\n\tfor {\n\t\tline, _, err := buf.ReadLine()\n\t\tif err != nil {\n\t\t\tif err == io.EOF {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t\tif line == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tvar parseError = errors.Errorf(\"Failed to parse uid map file %s\", uidMapPath)\n\t\t// if the container id (id[0]) is 0 (root inside the container)\n\t\t// has a mapping to the host id (id[1]) that is not root, then\n\t\t// it can be determined that the host user is running rootless\n\t\tids := strings.Fields(string(line))\n\t\t// do some sanity checks\n\t\tif len(ids) != 3 {\n\t\t\treturn parseError\n\t\t}\n\t\tuserNSUid, err := strconv.ParseUint(ids[0], 10, 0)\n\t\tif err != nil {\n\t\t\treturn parseError\n\t\t}\n\t\thostUID, err := strconv.ParseUint(ids[1], 10, 0)\n\t\tif err != nil {\n\t\t\treturn parseError\n\t\t}\n\t\trangeUID, err := strconv.ParseUint(ids[2], 10, 0)\n\t\tif err != nil || rangeUID == 0 {\n\t\t\treturn parseError\n\t\t}\n\n\t\tif userNSUid == 0 && hostUID != 0 {\n\t\t\trootlessLog.Info(\"Running as rootless\")\n\t\t\tisRootless = true\n\t\t\treturn nil\n\t\t}\n\t}\n}", "func fcntl(fd int, cmd int, arg int) (val int, errno int) {\n r0, _, e1 := syscall.Syscall(syscall.SYS_FCNTL, uintptr(fd), uintptr(cmd), uintptr(arg))\n val = int(r0)\n errno = int(e1)\n return\n}", "func IgnoreIoError(_ int, _ error) {}", "func LUtimesNano(path string, ts []syscall.Timespec) error {\n\t// These are not currently available in syscall\n\tatFdCwd := -100\n\tatSymLinkNoFollow := 0x100\n\n\tvar _path *byte\n\t_path, err := syscall.BytePtrFromString(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, _, err := syscall.Syscall6(syscall.SYS_UTIMENSAT, uintptr(atFdCwd), uintptr(unsafe.Pointer(_path)), uintptr(unsafe.Pointer(&ts[0])), uintptr(atSymLinkNoFollow), 0, 0); err != 0 && err != syscall.ENOSYS {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func breakMountpath(t *testing.T, mpath, suffix string) {\n\tos.Rename(mpath, mpath+suffix)\n\tf, err := os.OpenFile(mpath, os.O_CREATE|os.O_WRONLY, cos.PermRWR)\n\tif err != nil {\n\t\tt.Errorf(\"Failed to create file: %v\", err)\n\t}\n\tf.Close()\n}", "func set_pid(opt *Opts) bool {\n\t// Check if option mount was set, if not, return false\n\tif opt.mnt != \"\" {\n\t\tif opt.pid != false {\n\t\t\tif err := syscall.Mount(\"proc\", \"proc\", \"proc\", 0, \"\"); err != nil {\n\t\t\t\tfmt.Println(\"Error setting PID namespace.\")\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t\t//\t\t\tif err := syscall.Mount(\"dev\", \"/dev\", \"devtmpfs\", 0, \"\"); err != nil {\n\t\t\t//\t\t\t\tfmt.Println(\"Error setting PID namespace.\")\n\t\t\t//\t\t\t\tos.Exit(1)\n\t\t\t//\t\t\t}\n\t\t}\n\t} else {\n\t\tif opt.pid != false {\n\t\t\tfmt.Println(\"Error: option -pid require -mnt.\")\n\t\t}\n\t\treturn false\n\t}\n\treturn true\n}", "func Xfdopen(tls TLS, fd int32, mode uintptr) uintptr {\n\tpanic(\"TODO\")\n}", "func TestDialUnixNonex(t *testing.T) {\n\tt.Logf(\"Running DialUnix test to nonexistant database %s\", TEST_DBNAMEBAD)\n\tdb, err = DialUnix(TEST_SOCK, TEST_USER, TEST_PASSWD, TEST_DBNAMEBAD)\n\tif err != nil {\n\t\tt.Logf(\"Error %s\", err)\n\t}\n\tif cErr, ok := err.(*ClientError); ok {\n\t\tif cErr.Errno != 1044 {\n\t\t\tt.Logf(\"Error #%d received, expected #1044\", cErr.Errno)\n\t\t\tt.Fail()\n\t\t}\n\t}\n}", "func parseOptions(options string) (int, string) {\n\tvar (\n\t\tflag int\n\t\tdata []string\n\t)\n\n\tflags := map[string]struct {\n\t\tclear bool\n\t\tflag int\n\t}{\n\t\t\"defaults\": {false, 0},\n\t\t\"ro\": {false, RDONLY},\n\t\t\"rw\": {true, RDONLY},\n\t\t\"suid\": {true, NOSUID},\n\t\t\"nosuid\": {false, NOSUID},\n\t\t\"dev\": {true, NODEV},\n\t\t\"nodev\": {false, NODEV},\n\t\t\"exec\": {true, NOEXEC},\n\t\t\"noexec\": {false, NOEXEC},\n\t\t\"sync\": {false, SYNCHRONOUS},\n\t\t\"async\": {true, SYNCHRONOUS},\n\t\t\"dirsync\": {false, DIRSYNC},\n\t\t\"remount\": {false, REMOUNT},\n\t\t\"mand\": {false, MANDLOCK},\n\t\t\"nomand\": {true, MANDLOCK},\n\t\t\"atime\": {true, NOATIME},\n\t\t\"noatime\": {false, NOATIME},\n\t\t\"diratime\": {true, NODIRATIME},\n\t\t\"nodiratime\": {false, NODIRATIME},\n\t\t\"bind\": {false, BIND},\n\t\t\"rbind\": {false, RBIND},\n\t\t\"unbindable\": {false, UNBINDABLE},\n\t\t\"runbindable\": {false, RUNBINDABLE},\n\t\t\"private\": {false, PRIVATE},\n\t\t\"rprivate\": {false, RPRIVATE},\n\t\t\"shared\": {false, SHARED},\n\t\t\"rshared\": {false, RSHARED},\n\t\t\"slave\": {false, SLAVE},\n\t\t\"rslave\": {false, RSLAVE},\n\t\t\"relatime\": {false, RELATIME},\n\t\t\"norelatime\": {true, RELATIME},\n\t\t\"strictatime\": {false, STRICTATIME},\n\t\t\"nostrictatime\": {true, STRICTATIME},\n\t}\n\n\tfor _, o := range strings.Split(options, \",\") {\n\t\t// If the option does not exist in the flags table or the flag\n\t\t// is not supported on the platform,\n\t\t// then it is a data value for a specific fs type\n\t\tif f, exists := flags[o]; exists && f.flag != 0 {\n\t\t\tif f.clear {\n\t\t\t\tflag &= ^f.flag\n\t\t\t} else {\n\t\t\t\tflag |= f.flag\n\t\t\t}\n\t\t} else {\n\t\t\tdata = append(data, o)\n\t\t}\n\t}\n\treturn flag, strings.Join(data, \",\")\n}", "func Nanosec() int64 {\n\treturn syscall.Nanosec()\n}", "func unixServer(addr string) (io.ReadCloser, error) {\n\treturn listenServer(\"unix\", addr)\n}", "func IsSysErrNoSys(err error) bool {\n\tif err == syscall.ENOSYS {\n\t\treturn true\n\t}\n\tpathErr, ok := err.(*os.PathError)\n\treturn ok && pathErr.Err == syscall.ENOSYS\n}", "func _sockaddr_un(budpath ustr.Ustr) []uint8 {\n\tret := make([]uint8, 2, 16)\n\t// len\n\twriten(ret, 1, 0, len(budpath))\n\t// family\n\twriten(ret, 1, 1, defs.AF_UNIX)\n\t// path\n\tret = append(ret, budpath...)\n\tret = append(ret, 0)\n\treturn ret\n}", "func WindowsWorkstationChecks() {\n\n}", "func syscall_rawSyscall6X(fn, a1, a2, a3, a4, a5, a6 uintptr) (r1, r2, err uintptr) {\n\tlibcCall(unsafe.Pointer(abi.FuncPCABI0(syscall6X)), unsafe.Pointer(&fn))\n\treturn\n}", "func getFileSystem(frame *rtda.Frame) {\n\tthread := frame.Thread\n\tunixFsClass := frame.GetClassLoader().LoadClass(\"java/io/UnixFileSystem\")\n\tif unixFsClass.InitializationNotStarted() {\n\t\tframe.NextPC = thread.PC // undo getFileSystem\n\t\tthread.InitClass(unixFsClass)\n\t\treturn\n\t}\n\n\tunixFsObj := unixFsClass.NewObj()\n\tframe.PushRef(unixFsObj)\n\n\t// call <init>\n\tframe.PushRef(unixFsObj) // this\n\tconstructor := unixFsClass.GetDefaultConstructor()\n\tthread.InvokeMethod(constructor)\n}" ]
[ "0.5275571", "0.5275274", "0.526875", "0.52434105", "0.5234334", "0.51867586", "0.51269925", "0.50681794", "0.5060849", "0.5052121", "0.49770987", "0.49730974", "0.49512863", "0.48924834", "0.4852543", "0.48440093", "0.48321727", "0.48252714", "0.481353", "0.48109722", "0.47940248", "0.47898918", "0.47732627", "0.47522303", "0.47320879", "0.4731716", "0.47315636", "0.4726914", "0.47121915", "0.47089526", "0.46995264", "0.46958518", "0.46944642", "0.46912155", "0.4691188", "0.4686707", "0.4683847", "0.46758214", "0.4674659", "0.4674279", "0.46731225", "0.4673055", "0.4670205", "0.4666794", "0.46653596", "0.4663634", "0.46584496", "0.46517816", "0.4651606", "0.4649758", "0.4646562", "0.46353891", "0.46224216", "0.4601296", "0.46012676", "0.4597151", "0.45963025", "0.4591086", "0.45875275", "0.45707196", "0.45695767", "0.45674598", "0.45674598", "0.45625022", "0.45582724", "0.45575005", "0.4556062", "0.4543661", "0.45318937", "0.45247212", "0.45169795", "0.45119387", "0.45110223", "0.45103535", "0.45055565", "0.4504967", "0.44975543", "0.44966033", "0.44929805", "0.4489453", "0.44802144", "0.4470263", "0.44681555", "0.44648322", "0.44641605", "0.44634375", "0.44628146", "0.44606668", "0.44552097", "0.44532758", "0.44465062", "0.44464037", "0.4446236", "0.44442555", "0.44431743", "0.4442797", "0.44409937", "0.44349486", "0.44344532", "0.44334504", "0.44307822" ]
0.0
-1
Git Call git command with arguments. All print out displayed. It returns git Return code.
func Do(opts ...string) int { colorCyan, colorReset := utils.DefColor(36) log.Printf("%sgit %s%s\n", colorCyan, strings.Join(opts, " "), colorReset) return utils.RunCmd("git", opts...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Git(cmd string, args ...string) (res *exec.Cmd, stdout, stderr *bytes.Buffer) {\n\tcmdArgs := make([]string, 1)\n\tcmdArgs[0] = cmd\n\tcmdArgs = append(cmdArgs, args...)\n\tres = exec.Command(gitCmd, cmdArgs...)\n\tstdout, stderr = new(bytes.Buffer), new(bytes.Buffer)\n\tres.Stdout, res.Stderr = stdout, stderr\n\treturn\n}", "func Git(argv []string, cmdr cmd.Commander) error {\n\tusage := executable.Render(`\nValid commands for git:\n\ngit:remote Adds git remote of application to repository\ngit:remove Removes git remote of application from repository\n\nUse '{{.Name}} help [command]' to learn more.\n`)\n\n\tswitch argv[0] {\n\tcase \"git:remote\":\n\t\treturn gitRemote(argv, cmdr)\n\tcase \"git:remove\":\n\t\treturn gitRemove(argv, cmdr)\n\tcase \"git\":\n\t\tfmt.Print(usage)\n\t\treturn nil\n\tdefault:\n\t\tPrintUsage(cmdr)\n\t\treturn nil\n\t}\n}", "func execGitCommand(ctx context.Context, store store.Store, repositoryID int, args ...string) (string, error) {\n\trepo, err := repositoryIDToRepo(ctx, store, repositoryID)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tcmd := gitserver.DefaultClient.Command(\"git\", args...)\n\tcmd.Repo = repo\n\tout, err := cmd.CombinedOutput(ctx)\n\treturn string(bytes.TrimSpace(out)), errors.Wrap(err, \"gitserver.Command\")\n}", "func Run(args ...string) (string, error) {\n\t// TODO: use exex.CommandContext here and refactor.\n\t/* #nosec */\n\tcmd := exec.Command(\"git\", args...)\n\tlog.WithField(\"args\", args).Debug(\"running git\")\n\tbts, err := cmd.CombinedOutput()\n\tlog.WithField(\"output\", string(bts)).\n\t\tDebug(\"git result\")\n\tif err != nil {\n\t\treturn \"\", errors.New(string(bts))\n\t}\n\treturn string(bts), nil\n}", "func Run(args ...string) (output string, err error) {\n\tvar cmd = exec.Command(\"git\", args...)\n\tbts, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn \"\", errors.New(string(bts))\n\t}\n\treturn string(bts), err\n}", "func (r *Repo) Git(cmd string, args ...string) (res *exec.Cmd, out, err *bytes.Buffer) {\n\tvar path string\n\tif r.WorkDir == \"\" {\n\t\tpath = r.GitDir\n\t} else {\n\t\tpath = r.WorkDir\n\t}\n\tres, out, err = Git(cmd, args...)\n\tres.Dir = path\n\treturn\n}", "func withGit(url, commit, folder string, verbose bool) error {\n\tvar out bytes.Buffer\n\tdata := map[string]string{\n\t\t\"workDir\": folder,\n\t\t\"repoDir\": RepoDir(url),\n\t\t\"remote\": url,\n\t\t\"ref\": commit,\n\t}\n\tif err := gitTemplate.Execute(&out, data); err != nil {\n\t\treturn errors.Wrap(err, \"unable to create git script\")\n\t}\n\tscript := out.String()\n\tif verbose {\n\t\tfmt.Println(script)\n\t}\n\n\tcmd := exec.Command(\"bash\", \"-c\", script)\n\tif verbose {\n\t\tcmdStdout, err := cmd.StdoutPipe()\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"unable to create StdOut pipe for bash\")\n\t\t}\n\t\tstdoutScanner := bufio.NewScanner(cmdStdout)\n\t\tgo func() {\n\t\t\tfor stdoutScanner.Scan() {\n\t\t\t\tfmt.Println(\"bash: \" + stdoutScanner.Text())\n\t\t\t}\n\t\t}()\n\n\t\tcmdStderr, err := cmd.StderrPipe()\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"unable to create StdErr pipe for bash\")\n\t\t}\n\t\tstderrScanner := bufio.NewScanner(cmdStderr)\n\t\tgo func() {\n\t\t\tfor stderrScanner.Scan() {\n\t\t\t\tfmt.Println(\"bash: \" + stderrScanner.Text())\n\t\t\t}\n\t\t}()\n\t}\n\terr := cmd.Start()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"unable to start cloning with git\")\n\t}\n\terr = cmd.Wait()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"unable to clone with git\")\n\t}\n\treturn nil\n}", "func Get(opts ...string) (string, error) {\n\tgotrace.Trace(\"RUNNING: git %s\", strings.Join(opts, \" \"))\n\tout, err := exec.Command(\"git\", opts...).Output()\n\treturn string(out), err\n}", "func (s *GitTestHelper) RunGit(args ...string) bool {\n\treturn s.Exec(s.Git(), args...)\n}", "func main() {\n\n\tfmt.Println(\"Testing Git Access\")\n\n\tfmt.Println(\"Hello World!!\")\n\n}", "func (c *config) command(args ...string) (out string, err error) {\n\n\tcwd, _ := os.Getwd()\n\tpathArgs := []string{\"-C\", fmt.Sprintf(\"%v/%v\", cwd, TempDir)}\n\targs = append(pathArgs, args...)\n\n\tcmd := exec.Command(\"git\", args...)\n\n\t// verbose mode\n\tif c.useVerbose {\n\t\tlog.Println(cmd.Args)\n\t}\n\n\t// makes sure the TempDir exist\n\tif _, err := os.Stat(fmt.Sprintf(\"%v/%v\", cwd, TempDir)); os.IsNotExist(err) {\n\t\terr := os.Mkdir(TempDir, 0777)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\n\t// starts command\n\tvar output byteSlice\n\toutput, err = cmd.Output()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tout = output.ToString()\n\treturn out, err\n}", "func gitPush(c *cli.Context) error {\n\t// inputs\n\tremote, branch := c.Args().Get(0), c.Args().Get(1)\n\texpected(remote, branch)\n\n\t// execute\n\terr := exec.Command(\"git\", \"push\", remote, branch).Run()\n\tcheckErr(err)\n\n\t// outputs\n\tfmt.Println(\"remote:\" + remote)\n\tfmt.Println(\"branch:\" + branch)\n\treturn nil\n}", "func RunGitCommand(command string, formating ...interface{}) error {\n\tcmd := prepareGitCommand(command, formating...)\n\treturn cmd.Run()\n}", "func (e *osExecutor) RunGitOperation(commands []string, path string) error {\n\tvar args []string\n\targs = append(args, \"git\")\n\targs = append(args, \"-C\")\n\targs = append(args, path)\n\targs = append(args, commands...)\n\n\tcmd := exec.Command(args[0], args[1:]...)\n\n\treturn runLoudly(cmd, e.verboseMode)\n}", "func prepareGitCommand(command string, formating ...interface{}) *exec.Cmd {\n\tif formating != nil {\n\t\tcommand = fmt.Sprintf(command, formating...)\n\t}\n\targs := strings.Split(command, \" \")\n\tcmd := exec.Command(\"git\", args...)\n\tcmd.Stderr = os.Stderr\n\treturn cmd\n}", "func commit(msg string) {\n\tgit := exec.Command(\"git\", \"commit\", \"-m\", msg)\n\tgitOut, err := git.Output()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfmt.Println(string(gitOut))\n}", "func (gitCommand *GitCommand) Run(wait bool) (io.ReadCloser, error) {\n\tcmd := exec.Command(\"git\", gitCommand.Args...)\n\tstdout, err := cmd.StdoutPipe()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif gitCommand.ProcInput != nil {\n\t\tcmd.Stdin = gitCommand.ProcInput\n\t}\n\n\tif err := cmd.Start(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif wait {\n\t\terr = cmd.Wait()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn stdout, nil\n}", "func (repo *TestRepo) GitCommand(t *testing.T, args ...string) *exec.Cmd {\n\tt.Helper()\n\n\tgitArgs := []string{\"-C\", repo.Path}\n\tgitArgs = append(gitArgs, args...)\n\n\t//nolint:gosec // The args all come from the test code.\n\tcmd := exec.Command(\"git\", gitArgs...)\n\tcmd.Env = CleanGitEnv()\n\treturn cmd\n}", "func main() {\n\n/*\n\tCheckArgs(\"<repository-path>\")\n\tpath := os.Args[1]\n\tfmt.Println(path)\n\tr, err := git.PlainOpen(path)\n\tCheckIfError(err)\n\n\tInfo(\"git push\")\n\t// push using default options\n\terr = r.Push(&git.PushOptions{})\n\tCheckIfError(err)\n*/\n\topenFile()\n}", "func main() {\n\tCheckArgs(\"<path>\")\n\tpath := os.Args[1]\n\n\t// We instantiate a new repository targeting the given path (the .git folder)\n\tr, err := git.PlainOpen(path)\n\tCheckIfError(err)\n\n\t// Length of the HEAD history\n\tInfo(\"git rev-list HEAD --count\")\n\n\t// ... retrieving the HEAD reference\n\tref, err := r.Head()\n\tCheckIfError(err)\n\n\t// ... retrieves the commit history\n\tcIter, err := r.Log(&git.LogOptions{From: ref.Hash()})\n\tCheckIfError(err)\n\n\t// ... just iterates over the commits\n\tvar cCount int\n\terr = cIter.ForEach(func(c *object.Commit) error {\n\t\tcCount++\n\n\t\treturn nil\n\t})\n\tCheckIfError(err)\n\n\tfmt.Println(cCount)\n}", "func (cu *CmdUtils) execCmd(cmd string, args []string, workingDir string) (string, error) {\n\ttoExecute := exec.Command(\"git\", args...)\n\ttoExecute.Dir = workingDir\n\tstdoutStderr, err := toExecute.CombinedOutput()\n\tif err != nil {\n\t\treturn string(stdoutStderr), fmt.Errorf(\"unable to execute command %s due to %w, %s\", cmd, err, string(stdoutStderr))\n\t}\n\tlog.Debug().Str(\"output\", string(stdoutStderr)).Msg(\"execution finished\")\n\treturn string(stdoutStderr), nil\n}", "func MocksForFindGit(ctx context.Context, cmd *exec.Command) error {\n\tif strings.Contains(cmd.Name, \"git\") && len(cmd.Args) == 1 && cmd.Args[0] == \"--version\" {\n\t\t_, err := cmd.CombinedOutput.Write([]byte(\"git version 99.99.1\"))\n\t\treturn err\n\t}\n\treturn nil\n}", "func (git *Git) Run() error {\n\tcmd := exec.Command(\"git\", \"clone\", \"--no-checkout\", git.URL, git.SourcePath)\n\n\t_, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"git clone --no-checkout %s %s. %s\", git.URL, git.SourcePath, err)\n\t}\n\n\tcmd = exec.Command(\"git\", \"remote\", \"add\", \"composer\", git.URL)\n\tcmd.Dir = git.SourcePath\n\n\t_, err = cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"git remote add composer %s. %s\", git.URL, err)\n\t}\n\n\tcmd = exec.Command(\"git\", \"fecth\", \"composer\")\n\tcmd.Dir = git.SourcePath\n\tcmd.CombinedOutput()\n\n\tcmd = exec.Command(\"git\", \"checkout\", \"-b\", git.Version, \"composer/\"+git.Version)\n\tcmd.Dir = git.SourcePath\n\tcmd.CombinedOutput()\n\n\tcmd = exec.Command(\"git\", \"reset\", \"--hard\", git.Source[\"reference\"])\n\tcmd.Dir = git.SourcePath\n\tcmd.CombinedOutput()\n\n\t_, err = os.Stat(git.SourcePath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tgit.PathExist = true\n\treturn nil\n}", "func CompsCommand(runtime *program.SubRuntime) (retval int, err string) {\n\targv := runtime.Argv\n\n\t// parse the repo uri\n\turi, e := repos.NewRepoURI(argv[0])\n\tif e != nil {\n\t\terr = constants.StringUnparsedRepoName\n\t\tretval = constants.ErrorInvalidRepo\n\t\treturn\n\t}\n\n\t// print each component on one line\n\tfor _, comp := range uri.Components() {\n\t\tfmt.Println(comp)\n\t}\n\n\t// and finish\n\treturn\n}", "func GitCommit() string { return gitCommit }", "func (g *GitUtil) ExecCommand(args ...string) (string, error) {\n\tcmd := exec.Command(\"git\", args...)\n\tcmd.Dir = g.TopLevelPath\n\n\t// Capture output\n\tvar out bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = &stderr\n\n\t// Run\n\terr := cmd.Run()\n\n\tif err != nil {\n\t\tlog.Debugln(stderr.String())\n\t}\n\n\treturn string(bytes.TrimSpace(out.Bytes())), err\n}", "func (m *MockIExec) DoGit(dir string, args ...string) (string, error) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{dir}\n\tfor _, a := range args {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"DoGit\", varargs...)\n\tret0, _ := ret[0].(string)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c Command) Run() error {\n\tif c.fs != nil {\n\t\tif err := c.fs.Parse(os.Args[1:]); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif *c.version {\n\t\tfmt.Println(c.b.Version(\"changelog\"))\n\t\treturn nil\n\t}\n\n\t// history is always written to stdout\n\tif *c.history {\n\t\t*c.toStdOut = true\n\t}\n\n\tl := flash.New(flash.WithDebug(*c.debug))\n\n\tgitCmd, err := git.New(l)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttoplevelDir, err := gitCmd.TopLevelDir()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// resolv changelog file path\n\tabsChangelogPath := filepath.Join(toplevelDir, *c.file)\n\tc.file = &absChangelogPath\n\n\tgitCmd.Noop = c.noop\n\n\tif err := c.validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif *c.initConfig {\n\t\treturn c.runWriteConfig(l)\n\t}\n\n\tif !gitCmd.IsRepo() {\n\t\treturn errors.New(\"current folder is not a git repository\")\n\t}\n\n\tcfg, err := config.Load(\".\")\n\tif err != nil {\n\t\tif err != config.ErrEmpty && err != config.ErrNotFound {\n\t\t\treturn err\n\t\t}\n\n\t\tl.Debugw(\"no changelog config file found - using default config\", \"path\", filepath.Join(\".\", config.FileName))\n\n\t\tcfg = &config.Default\n\t}\n\n\tvar dst io.Writer = os.Stdout\n\n\tif !*c.toStdOut {\n\t\tf, err := os.OpenFile(*c.file, os.O_CREATE|os.O_RDWR, 0600)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer f.Close() // nolint: gosec\n\t\tdst = f\n\t}\n\n\tif *c.history {\n\t\treturn c.runHistory(dst, l, *cfg, gitCmd)\n\t}\n\n\thasTags, err := gitCmd.HasTags()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !hasTags {\n\t\treturn c.runInit(dst, l, *cfg, gitCmd)\n\t}\n\n\treturn c.runRelease(dst, l, *cfg, gitCmd)\n}", "func (ss *Sources) gitFetch(spec v1.SourceSpec) (string, error) {\n\tp := ss.repoPath(spec)\n\t_, err := os.Stat(p)\n\tif os.IsNotExist(err) {\n\t\t// Clone new repo.\n\t\td, _ := filepath.Split(p)\n\t\terr = os.MkdirAll(d, 0750)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\t_, _, err = exe.Run(ss.Log, &exe.Opt{Dir: d}, \"\", \"git\", \"clone\", urlWithToken(spec.URL, spec.Token))\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\t_, _, err = exe.Run(ss.Log, &exe.Opt{Dir: p}, \"\", \"git\", \"checkout\", spec.Ref)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tss.Log.Info(\"GIT-clone\", \"url\", spec.URL, \"ref\", spec.Ref)\n\t} else {\n\t\t// Pull existing repo content.\n\t\t_, _, err = exe.Run(ss.Log, &exe.Opt{Dir: p}, \"\", \"git\", \"pull\", \"origin\", spec.Ref)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tss.Log.V(2).Info(\"GIT-pull\", \"url\", spec.URL, \"ref\", spec.Ref)\n\t}\n\n\t// Get hash.\n\th, _, err := exe.Run(ss.Log, &exe.Opt{Dir: p}, \"\", \"git\", \"rev-parse\", spec.Ref)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\th = strings.TrimRight(h, \"\\n\\r\")\n\tif len(h) == 0 {\n\t\treturn \"\", fmt.Errorf(\"expected git hash\")\n\t}\n\n\treturn h, nil\n}", "func (gc *GitCommand) Run() error {\n\targs := append(gc.reqArgs(), gc.addlArgs...)\n\treturn exec.Command(gc.binName, args...).Run()\n}", "func gitVersion() string {\n\tcmd := exec.Command(\"git\", \"rev-list\", \"--max-count=1\", \"--pretty=format:'%ad-%h'\", \"--date=short\", \"HEAD\")\n\tcmd.Dir = camRoot\n\tout, err := cmd.Output()\n\tif err != nil {\n\t\tlog.Fatalf(\"Error running git rev-list in %s: %v\", camRoot, err)\n\t}\n\tv := strings.TrimSpace(string(out))\n\tif m := gitVersionRx.FindStringSubmatch(v); m != nil {\n\t\tv = m[0]\n\t} else {\n\t\tpanic(\"Failed to find git version in \" + v)\n\t}\n\tcmd = exec.Command(\"git\", \"diff\", \"--exit-code\")\n\tcmd.Dir = camRoot\n\tif err := cmd.Run(); err != nil {\n\t\tv += \"+\"\n\t}\n\treturn v\n}", "func gi(args string) string {\n\n\tif len(args) == 0 {\n\t\targs = defaultGitignoreItems\n\t}\n\n\tcommand := \"curl -fLw '\\n' https://www.gitignore.io/api/\\\"${(j:,:)@}\\\" \"\n\tcommand += strings.Join(args, \" \")\n\n\treturn Shell(command)\n}", "func (o *StepGitForkAndCloneOptions) Run() error {\n\tif o.Dir == \"\" {\n\t\tdir, err := os.Getwd()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\to.Dir = dir\n\t}\n\tgitURL := \"\"\n\tif len(o.Args) > 1 {\n\t\treturn errors.Errorf(\"Must specify exactly one git url but was %v\", o.Args)\n\t} else if len(o.Args) == 0 {\n\t\tif os.Getenv(\"REPO_URL\") != \"\" {\n\t\t\tgitURL = os.Getenv(\"REPO_URL\")\n\t\t}\n\t} else {\n\t\tgitURL = o.Args[0]\n\t}\n\n\tif gitURL == \"\" {\n\t\treturn errors.Errorf(\"Must specify a git url on the CLI or using the environment variable REPO_URL\")\n\t}\n\tprovider, err := o.GitProviderForURL(gitURL, \"git username\")\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"getting git provider for %s\", gitURL)\n\t}\n\tdir, baseRef, upstreamInfo, forkInfo, err := gits.ForkAndPullRepo(gitURL, o.Dir, o.BaseRef, \"master\", provider, o.Git(), \"\")\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"forking and pulling %s\", gitURL)\n\t}\n\to.OutDir = dir\n\tif o.PrintOutDir {\n\t\t// Output the directory so it can be used in a script\n\t\t// Must use fmt.Print() as we need to write to stdout\n\t\tfmt.Print(dir)\n\t}\n\tif forkInfo != nil {\n\t\tlog.Logger().Infof(\"Forked %s to %s, pulled it into %s and checked out %s\", util.ColorInfo(upstreamInfo.HTMLURL), util.ColorInfo(forkInfo.HTMLURL), util.ColorInfo(dir), util.ColorInfo(baseRef))\n\t} else {\n\t\tlog.Logger().Infof(\"Pulled %s (%s) into %s\", upstreamInfo.URL, baseRef, dir)\n\t}\n\n\treturn nil\n}", "func Status() string {\n\treturn run.Capture(\"git status\")\n}", "func Run(conf *Config, fs sys.FS, env sys.Env, storageDriver storagedriver.StorageDriver) error {\n\tlog.Debug(\"Running git hook\")\n\n\tbuilderKey, err := builderconf.GetBuilderKey()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tkubeClient, err := client.NewInCluster()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"couldn't reach the api server (%s)\", err)\n\t}\n\n\tscanner := bufio.NewScanner(os.Stdin)\n\tfor scanner.Scan() {\n\t\tline := scanner.Text()\n\t\toldRev, newRev, refName, err := readLine(line)\n\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"reading STDIN (%s)\", err)\n\t\t}\n\n\t\tlog.Debug(\"read [%s,%s,%s]\", oldRev, newRev, refName)\n\n\t\t// if we're processing a receive-pack on an existing repo, run a build\n\t\tif strings.HasPrefix(conf.SSHOriginalCommand, \"git-receive-pack\") {\n\t\t\tif err := build(conf, storageDriver, kubeClient, fs, env, builderKey, newRev); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn scanner.Err()\n}", "func GitCommit(tb testing.TB, msg string) {\n\ttb.Helper()\n\tout, err := fakeGit(\"commit\", \"--allow-empty\", \"-m\", msg)\n\trequire.NoError(tb, err)\n\trequire.Contains(tb, out, \"main\", msg)\n}", "func gitInfo(dir string) (string, time.Time) {\n\tfname := dir + \"/.git/HEAD\"\n\tbuf, err := os.ReadFile(fname)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tbuf = bytes.TrimSpace(buf)\n\tvar githash string\n\tif len(buf) == 40 {\n\t\tgithash = string(buf[:40])\n\t} else if bytes.HasPrefix(buf, []byte(\"ref: \")) {\n\t\tfname = dir + \"/.git/\" + string(buf[5:])\n\t\tbuf, err = os.ReadFile(fname)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tgithash = string(buf[:40])\n\t} else {\n\t\tlog.Fatalf(\"githash cannot be recovered from %s\", fname)\n\t}\n\tloadTime := time.Now()\n\treturn githash, loadTime\n}", "func (git *Git) Exec(subcmd string, args ...string) (string, error) {\n\tb, err := git.Command(subcmd, args...).CombinedOutput()\n\n\t// Chop last newline\n\tl := len(b)\n\tif l > 0 && b[l-1] == '\\n' {\n\t\tb = b[:l-1]\n\t}\n\n\t// Make output in oneline in error cases\n\tif err != nil {\n\t\tfor i := range b {\n\t\t\tif b[i] == '\\n' {\n\t\t\t\tb[i] = ' '\n\t\t\t}\n\t\t}\n\t}\n\n\treturn string(b), err\n}", "func GitDiff(options ...string) []string {\n\tcommand := &GitCommand{Args: append([]string{\"diff\"}, options...)}\n\n\toutput := command.RunAndGetOutput()\n\treturn strings.Split(string(output), \"\\x00\")\n}", "func main() {\n\targs := os.Args\n\tcwd := args[1]\n\n\tgitFolders := allGitFolders(cwd)\n\tfmt.Println(cwd)\n\tfmt.Println(gitFolders)\n\n\tb, err := ioutil.ReadFile(cwd + \"/.git/gitname.txt\") // just pass the file name\n\tcheckError(err)\n\n\tfmt.Println(\"Current git project: \", string(b))\n\tfmt.Println(\"All git projects in this repo: \")\n\tif len(gitFolders) == 0 {\n\t\tfmt.Println(\"There are no other git projects in this directory.\")\n\t\treturn\n\t}\n\tfor index, folder := range gitFolders {\n\t\tname, err := ioutil.ReadFile(folder + \"/gitname.txt\")\n\t\tcheckError(err)\n\t\tfmt.Println(strconv.Itoa(index) + \". \" + string(name))\n\t}\n\n\tfmt.Println(\"Type the number of git project you'd like to switch to. Type anything else to exit.\")\n\treader := bufio.NewReader(os.Stdin)\n\ttext, _ := reader.ReadString('\\n')\n\ttext = strings.TrimSpace(text)\n\tinput, err := strconv.Atoi(text)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif input >= 0 && input < len(gitFolders) {\n\t\toldName := gitFolders[input]\n\t\tos.Rename(oldName, \"temp\")\n\t\tos.Rename(\".git\", oldName)\n\t\tos.Rename(\"temp\", \".git\")\n\t\tcopyFileContents(cwd+\"/.git/ReadMe.txt\", cwd+\"/ReadMe.txt\")\n\t\tcopyFileContents(cwd+\"/.git/.gitignore\", cwd+\"/.gitignore\")\n\t\tcopyFileContents(cwd+\"/.git/.gitattributes\", cwd+\"/.gitattributes\")\n\n\t}\n}", "func executeCommand(config *GitXargsConfig, repositoryDir string, repo *github.Repository, worktree *git.Worktree) error {\n\n\tlogger := logging.GetLogger(\"git-xargs\")\n\n\tif len(config.Args) < 1 {\n\t\treturn errors.WithStackTrace(NoCommandSuppliedErr{})\n\t}\n\n\tcmdArgs := config.Args\n\n\tcmd := exec.Command(cmdArgs[0], cmdArgs[1:]...)\n\tcmd.Dir = repositoryDir\n\n\tlogger.WithFields(logrus.Fields{\n\t\t\"Repo\": repo.GetName(),\n\t\t\"Directory\": repositoryDir,\n\t\t\"Command\": config.Args,\n\t}).Debug(\"Executing command against local clone of repo...\")\n\n\tstdoutStdErr, err := cmd.CombinedOutput()\n\n\tif err != nil {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Error\": err,\n\t\t}).Debug(\"Error getting output of command execution\")\n\t\t// Track the command error against the repo\n\t\tconfig.Stats.TrackSingle(CommandErrorOccurredDuringExecution, repo)\n\t\treturn errors.WithStackTrace(err)\n\t}\n\n\tlogger.WithFields(logrus.Fields{\n\t\t\"CombinedOutput\": string(stdoutStdErr),\n\t}).Debug(\"Received output of command run\")\n\n\tstatus, statusErr := worktree.Status()\n\n\tif statusErr != nil {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Error\": statusErr,\n\t\t\t\"Repo\": repo.GetName(),\n\t\t\t\"Dir\": repositoryDir,\n\t\t}).Debug(\"Error looking up worktree status\")\n\n\t\t// Track the status check failure\n\t\tconfig.Stats.TrackSingle(WorktreeStatusCheckFailedCommand, repo)\n\t\treturn errors.WithStackTrace(statusErr)\n\t}\n\n\t// If the supplied command resulted in any changes, we need to stage, add and commit them\n\tif !status.IsClean() {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Repo\": repo.GetName(),\n\t\t}).Debug(\"Local repository worktree no longer clean, will stage and add new files and commit changes\")\n\n\t\t// Track the fact that worktree changes were made following execution\n\t\tconfig.Stats.TrackSingle(WorktreeStatusDirty, repo)\n\n\t\tfor filepath := range status {\n\t\t\tif status.IsUntracked(filepath) {\n\t\t\t\tfmt.Printf(\"Found untracked file %s. Adding to stage\", filepath)\n\t\t\t\t_, addErr := worktree.Add(filepath)\n\t\t\t\tif addErr != nil {\n\t\t\t\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\t\t\t\"Error\": addErr,\n\t\t\t\t\t\t\"Filepath\": filepath,\n\t\t\t\t\t}).Debug(\"Error adding file to git stage\")\n\t\t\t\t\t// Track the file staging failure\n\t\t\t\t\tconfig.Stats.TrackSingle(WorktreeAddFileFailed, repo)\n\t\t\t\t\treturn errors.WithStackTrace(addErr)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t} else {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Repo\": repo.GetName(),\n\t\t}).Debug(\"Local repository status is clean - nothing to stage or commit\")\n\n\t\t// Track the fact that repo had no file changes post command execution\n\t\tconfig.Stats.TrackSingle(WorktreeStatusClean, repo)\n\t}\n\n\treturn nil\n}", "func FindGit(ctx context.Context) (string, int, int, error) {\n\tmtx.Lock()\n\tdefer mtx.Unlock()\n\tif git == \"\" {\n\t\tgitPath, err := osexec.LookPath(\"git\")\n\t\tif err != nil {\n\t\t\treturn \"\", 0, 0, skerr.Wrapf(err, \"Failed to find git\")\n\t\t}\n\t\tmaj, min, err := Version(ctx, gitPath)\n\t\tif err != nil {\n\t\t\treturn \"\", 0, 0, skerr.Wrapf(err, \"Failed to obtain git version\")\n\t\t}\n\t\tsklog.Infof(\"Git is %s; version %d.%d\", gitPath, maj, min)\n\t\tisFromCIPD := IsFromCIPD(gitPath)\n\t\tisFromCIPDVal := 0\n\t\tif isFromCIPD {\n\t\t\tisFromCIPDVal = 1\n\t\t}\n\t\tmetrics2.GetInt64Metric(\"git_from_cipd\").Update(int64(isFromCIPDVal))\n\t\tgit = gitPath\n\t\tgitVersionMajor = maj\n\t\tgitVersionMinor = min\n\t}\n\treturn git, gitVersionMajor, gitVersionMinor, nil\n}", "func commit(srvChan chan string, channel, nick, hostname string, args []string) {\n\ttype repoJSON struct {\n\t\tId int\n\t\tOwner map[string]interface{}\n\t\tName string\n\t\tFull_name string\n\t\tDescription string\n\t\tPrivate bool\n\t\tFork bool\n\t\tUrl string\n\t\tHtml_url string\n\t}\n\ttype commitJSON struct {\n\t\tSha string\n\t\tCommit map[string]interface{}\n\t\tUrl string\n\t\tHtml_url string\n\t\tComments_url string\n\t\tAuthor map[string]interface{}\n\t\tCommitter map[string]interface{}\n\t\tParents map[string]interface{}\n\t}\n\ttype urlJSON struct {\n\t\tKind string\n\t\tId string\n\t\tLongUrl string\n\t}\n\tmessage := \"NOTICE \" + channel + \" :\"\n\tsince := rand.Intn(1000000)\n\tres, err := http.Get(\"https://api.github.com/repositories?since=\" + fmt.Sprintf(\"%d\", since))\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tbody, err := ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tvar repos []repoJSON\n\tjson.Unmarshal(body, &repos)\n\tfullName := repos[rand.Intn(len(repos))].Full_name\n\tres, err = http.Get(\"https://api.github.com/repos/\" + fullName + \"/commits\")\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tbody, err = ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tvar commits []commitJSON\n\tjson.Unmarshal(body, &commits)\n\tif len(commits) < 1 {\n\t\tcommit(srvChan, channel, nick, hostname, args) //try again\n\t\treturn\n\t} else {\n\t\tcommitNum := rand.Intn(len(commits))\n\t\tcommitMsg := commits[commitNum].Commit[\"message\"].(string)\n\n\t\tAPIkey, err := ioutil.ReadFile(\"APIkey\")\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\turlReader := strings.NewReader(`{\"longUrl\": \"` + commits[commitNum].Html_url + `\"}`)\n\t\tc := http.Client{}\n\t\tres, err := c.Post(\"https://www.googleapis.com/urlshortener/v1/url?key=\" + string(APIkey), \"application/json\", urlReader)\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tbody, err := ioutil.ReadAll(res.Body)\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tvar googUrl urlJSON\n\t\tjson.Unmarshal(body, &googUrl)\n\t\tmessage += strings.Split(commitMsg, \"\\n\")[0] + \" | \" + googUrl.Id\n\t}\n\tsrvChan <- message\n\tlog.Println(message)\n}", "func main() {\n\trepoPath := filepath.Join(os.Getenv(\"GOPATH\"), \"src/github.com/libgit2/git2go\")\n\tgitRepo, err := git.OpenRepository(repoPath)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tcommitOid, err := gitRepo.Head()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tblob, _ := gitRepo.LookupBlob(commitOid.Target())\n\tlog.Println(blob)\n\t// commit, err := gitRepo.LookupCommit(commitOid)\n\tcommit, err := gitRepo.LookupCommit(commitOid.Target())\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tcommitTree, err := commit.Tree()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\toptions, err := git.DefaultDiffOptions()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\toptions.IdAbbrev = 40\n\toptions.InterhunkLines = 0\n\toptions.Flags = git.DiffIncludeUntracked\n\tvar parentTree *git.Tree\n\tif commit.ParentCount() > 0 {\n\t\tparentTree, err = commit.Parent(0).Tree()\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}\n\tgitDiff, err := gitRepo.DiffTreeToTree(parentTree, commitTree, &options)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tfindOpts, err := git.DefaultDiffFindOptions()\n\tfindOpts.Flags = git.DiffFindBreakRewrites\n\terr = gitDiff.FindSimilar(&findOpts)\n\n\t// Show all file patch diffs in a commit.\n\tfiles := make([]string, 0)\n\thunks := make([]git.DiffHunk, 0)\n\tlines := make([]git.DiffLine, 0)\n\tpatches := make([]string, 0)\n\terr = gitDiff.ForEach(func(file git.DiffDelta, progress float64) (git.DiffForEachHunkCallback, error) {\n\t\tpatch, err := gitDiff.Patch(len(patches))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tdefer patch.Free()\n\t\tpatchStr, err := patch.String()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpatches = append(patches, patchStr)\n\n\t\tfiles = append(files, file.OldFile.Path)\n\t\treturn func(hunk git.DiffHunk) (git.DiffForEachLineCallback, error) {\n\t\t\thunks = append(hunks, hunk)\n\t\t\treturn func(line git.DiffLine) error {\n\t\t\t\tlines = append(lines, line)\n\t\t\t\treturn nil\n\t\t\t}, nil\n\t\t}, nil\n\t}, git.DiffDetailLines)\n\n\tlog.Println(\"files: \", files, \"\\n\")\n\tlog.Println(\"hunks: \", hunks, \"\\n\")\n\tlog.Println(\"lines: \", lines, \"\\n\")\n\tlog.Println(\"patches: \", patches, \"\\n\")\n}", "func gitClone(repo, dir string) error {\n\tlog.Printf(\"cloning %s\\n\", repo)\n\n\t_, err := git.PlainClone(dir, false, &git.CloneOptions{\n\t\tURL: repo,\n\t\tProgress: os.Stdout,\n\t})\n\treturn err\n}", "func (j *DSGit) GetGitOps(ctx *Ctx, thrN int) (ch chan error, err error) {\n\tworker := func(c chan error, url string) (e error) {\n\t\tdefer func() {\n\t\t\tif c != nil {\n\t\t\t\tc <- e\n\t\t\t}\n\t\t}()\n\t\tvar (\n\t\t\tsout string\n\t\t\tserr string\n\t\t)\n\t\tcmdLine := []string{GitOpsCommand, url}\n\t\tvar env map[string]string\n\t\tif GitOpsNoCleanup {\n\t\t\tenv = map[string]string{\"SKIP_CLEANUP\": \"1\"}\n\t\t}\n\t\tsout, serr, e = ExecCommand(ctx, cmdLine, \"\", env)\n\t\tif e != nil {\n\t\t\tif GitOpsFailureFatal {\n\t\t\t\tPrintf(\"error executing %v: %v\\n%s\\n%s\\n\", cmdLine, e, sout, serr)\n\t\t\t} else {\n\t\t\t\tPrintf(\"WARNING: error executing %v: %v\\n%s\\n%s\\n\", cmdLine, e, sout, serr)\n\t\t\t\te = nil\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\ttype resultType struct {\n\t\t\tLoc int `json:\"loc\"`\n\t\t\tPls []RawPLS `json:\"pls\"`\n\t\t}\n\t\tvar data resultType\n\t\te = jsoniter.Unmarshal([]byte(sout), &data)\n\t\tif e != nil {\n\t\t\tif GitOpsFailureFatal {\n\t\t\t\tPrintf(\"error unmarshaling from %v\\n\", sout)\n\t\t\t} else {\n\t\t\t\tPrintf(\"WARNING: error unmarshaling from %v\\n\", sout)\n\t\t\t\te = nil\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tj.Loc = data.Loc\n\t\tfor _, f := range data.Pls {\n\t\t\tfiles, _ := strconv.Atoi(f.Files)\n\t\t\tblank, _ := strconv.Atoi(f.Blank)\n\t\t\tcomment, _ := strconv.Atoi(f.Comment)\n\t\t\tcode, _ := strconv.Atoi(f.Code)\n\t\t\tj.Pls = append(\n\t\t\t\tj.Pls,\n\t\t\t\tPLS{\n\t\t\t\t\tLanguage: f.Language,\n\t\t\t\t\tFiles: files,\n\t\t\t\t\tBlank: blank,\n\t\t\t\t\tComment: comment,\n\t\t\t\t\tCode: code,\n\t\t\t\t},\n\t\t\t)\n\t\t}\n\t\treturn\n\t}\n\tif thrN <= 1 {\n\t\treturn nil, worker(nil, j.URL)\n\t}\n\tch = make(chan error)\n\tgo func() { _ = worker(ch, j.URL) }()\n\treturn ch, nil\n}", "func branchGit() {\n\n}", "func gitState(repoExisted bool, out string) repoState {\n\tif !repoExisted {\n\t\treturn stateNew\n\t}\n\tif lines := strings.Split(out, \"\\n\"); len(lines) > 2 {\n\t\treturn stateChanged\n\t}\n\treturn stateUnchanged\n}", "func GitDiff(path string) string {\n\targs := []string{\"rev-parse\", \"--abbrev-ref\", \"HEAD\"}\n\tbranch := strings.TrimSuffix(string(common.GitRun(path, args, true)), \"\\n\")\n\n\targs = []string{\"rev-parse\", \"--abbrev-ref\", \"@{upstream}\"}\n\tupstream := strings.TrimSuffix(string(common.GitRun(path, args, false)), \"\\n\")\n\tif upstream == \"\" {\n\t\tupstream = fmt.Sprintf(\"origin/%v\", branch)\n\t}\n\n\targs = []string{\"rev-list\", \"--left-right\", branch + \"...\" + upstream}\n\toutput := common.GitRun(path, args, false)\n\n\tresult := \"\"\n\tif len(output) == 0 {\n\t\tresult = \"✓\"\n\t} else {\n\t\tbehind := 0\n\t\tahead := 0\n\t\tfor _, rev := range strings.Split(string(output), \"\\n\") {\n\t\t\tif rev == \"\" {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\trev = string(rev[0])\n\t\t\tif rev == \">\" {\n\t\t\t\tbehind++\n\t\t\t}\n\t\t\tif rev == \"<\" {\n\t\t\t\tahead++\n\t\t\t}\n\t\t}\n\n\t\tif ahead > 0 {\n\t\t\tresult = fmt.Sprintf(\"▲%d\", ahead)\n\t\t}\n\t\tif behind > 0 {\n\t\t\tresult = fmt.Sprintf(\"%v▼%d\", result, behind)\n\t\t}\n\t}\n\n\treturn result\n}", "func (h *stiGit) Checkout(repo, ref string) error {\n\topts := cmd.CommandOpts{\n\t\tStdout: os.Stdout,\n\t\tStderr: os.Stderr,\n\t\tDir: repo,\n\t}\n\tif log.Is(1) {\n\t\treturn h.RunWithOptions(opts, \"git\", \"checkout\", \"--quiet\", ref)\n\t}\n\treturn h.RunWithOptions(opts, \"git\", \"checkout\", ref)\n}", "func Version(ctx context.Context, git string) (int, int, error) {\n\tout, err := exec.RunCwd(ctx, \".\", git, \"--version\")\n\tif err != nil {\n\t\treturn -1, -1, err\n\t}\n\tm := gitVersionRegex.FindStringSubmatch(out)\n\tif m == nil {\n\t\treturn -1, -1, fmt.Errorf(\"Failed to parse the git version from output: %q\", out)\n\t}\n\tif len(m) != 3 {\n\t\treturn -1, -1, fmt.Errorf(\"Failed to parse the git version from output: %q\", out)\n\t}\n\tmajor, err := strconv.Atoi(m[1])\n\tif err != nil {\n\t\treturn -1, -1, fmt.Errorf(\"Failed to parse the git version from output: %q\", out)\n\t}\n\tminor, err := strconv.Atoi(m[2])\n\tif err != nil {\n\t\treturn -1, -1, fmt.Errorf(\"Failed to parse the git version from output: %q\", out)\n\t}\n\treturn major, minor, nil\n}", "func (gitCommand *GitCommand) RunAndGetOutput() []byte {\n\tstdout, err := gitCommand.Run(false)\n\tif err != nil {\n\t\treturn []byte{}\n\t}\n\n\tdata, err := ioutil.ReadAll(stdout)\n\tif err != nil {\n\t\treturn []byte{}\n\t}\n\n\treturn data\n}", "func (c *CmdGitMdget) Run() error {\n\tcli, err := GetGitClient(c.G())\n\tctx := context.Background()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar res []keybase1.GitRepoResult\n\tif len(c.folder) > 0 {\n\t\tfolder, err := ParseTLF(c.folder)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tres, err = cli.GetGitMetadata(ctx, folder)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tres, err = cli.GetAllGitMetadata(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tjsonStr, err := json.MarshalIndent(res, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Println(string(jsonStr))\n\treturn nil\n}", "func gitInit(_ *cli.Context) error {\n\n\terr := exec.Command(\"git\", \"init\").Run()\n\tcheckErr(err)\n\treturn nil\n}", "func (git *Git) Command(subcmd string, args ...string) *exec.Cmd {\n\t// e.g. 'git diff --cached' -> 'git -C /path/to/repo diff --cached'\n\ta := append([]string{\"-C\", git.root, subcmd}, args...)\n\tcmd := exec.Command(git.bin, a...)\n\treturn cmd\n}", "func TestGitCommandCheckout(t *testing.T) {\n\ttype scenario struct {\n\t\ttestName string\n\t\tcommand func(string, ...string) *exec.Cmd\n\t\ttest func(error)\n\t\tforce bool\n\t}\n\n\tscenarios := []scenario{\n\t\t{\n\t\t\t\"Checkout\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"checkout\", \"test\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t\tfalse,\n\t\t},\n\t\t{\n\t\t\t\"Checkout forced\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"checkout\", \"--force\", \"test\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t\ttrue,\n\t\t},\n\t}\n\n\tfor _, s := range scenarios {\n\t\tt.Run(s.testName, func(t *testing.T) {\n\t\t\tgitCmd := NewDummyGitCommand()\n\t\t\tgitCmd.OSCommand.Command = s.command\n\t\t\ts.test(gitCmd.Checkout(\"test\", CheckoutOptions{Force: s.force}))\n\t\t})\n\t}\n}", "func main() {\n\tcmd.Execute(version, gitCommit, buildDate)\n}", "func (o *RunOptions) Git() gits.Gitter {\n\tif o.Gitter == nil {\n\t\to.Gitter = gits.NewGitCLI()\n\t}\n\treturn o.Gitter\n}", "func Checkout(c *git.Client, args []string) error {\n\tflags := flag.NewFlagSet(\"checkout\", flag.ExitOnError)\n\tflags.SetOutput(flag.CommandLine.Output())\n\tflags.Usage = func() {\n\t\tflag.Usage()\n\t\tfmt.Fprintf(flag.CommandLine.Output(), \"\\n\\nOptions:\\n\")\n\t\tflags.PrintDefaults()\n\t}\n\toptions := git.CheckoutOptions{}\n\n\tflags.BoolVar(&options.Quiet, \"quiet\", false, \"Quiet. Suppress feedback messages.\")\n\tflags.BoolVar(&options.Quiet, \"q\", false, \"Alias of --quiet.\")\n\n\tprogress := flags.Bool(\"progress\", true, \"Report progress to standard error stream\")\n\tnoprogress := flags.Bool(\"no-progress\", false, \"Override --progress and suppress progress reporting\")\n\n\tflags.BoolVar(&options.Force, \"force\", false, \"When switching branches, proceed even if the index differs from HEAD\")\n\tflags.BoolVar(&options.Force, \"f\", false, \"Alias of --force.\")\n\n\tours := flags.Bool(\"ours\", false, \"Use stage2 for checking out unmerged paths from the index\")\n\ttheirs := flags.Bool(\"theirs\", false, \"Use stage3 for checking out unmerged paths from the index\")\n\n\tb := flags.String(\"b\", \"\", \"Create a new branch\")\n\tB := flags.String(\"B\", \"\", \"Create a new branch, overwriting if it exists\")\n\n\ttrack := flags.String(\"track\", \"\", \"When creating a new branch, set upstream to branch\")\n\tt := flags.String(\"t\", \"\", \"Alias of --track\")\n\tnotrack := flags.Bool(\"no-track\", false, \"Override --track and do not set upstream\")\n\n\tflags.BoolVar(&options.CreateReflog, \"l\", true, \"Create the new branch's reflog\")\n\tflags.BoolVar(&options.Detach, \"detach\", false, \"Checkout in detached head state.\")\n\torphan := flags.String(\"orphan\", \"\", \"Create a new branch with no parents\")\n\n\tflags.BoolVar(&options.IgnoreSkipWorktreeBits, \"ignore-skip-worktree-bits\", false, \"Check out regardless of skip worktree bit\")\n\n\tflags.BoolVar(&options.Merge, \"merge\", false, \"Perform three-way merge with local modifications if switching branches\")\n\tflags.BoolVar(&options.Merge, \"m\", false, \"Alias of --merge\")\n\n\tflags.StringVar(&options.ConflictStyle, \"conflict\", \"merge\", \"Use style to display conflicts (valid values are merge or diff3) (Not implemented)\")\n\n\tflags.BoolVar(&options.Patch, \"patch\", false, \"Interactively select hunks to discard (not implemented\")\n\tflags.BoolVar(&options.Patch, \"p\", false, \"Alias of --patch\")\n\n\tflags.BoolVar(&options.IgnoreOtherWorktrees, \"ignore-other-worktrees\", false, \"Unused, for compatibility with git only.\")\n\n\tflags.Parse(args)\n\tfiles := flags.Args()\n\n\toptions.Progress = *progress && !*noprogress\n\tif *ours && *theirs {\n\t\treturn fmt.Errorf(\"--ours and --theirs are mutually exclusive.\")\n\t} else if *ours {\n\t\toptions.Stage = git.Stage2\n\t} else if *theirs {\n\t\toptions.Stage = git.Stage3\n\t}\n\n\tif *b != \"\" && *B != \"\" {\n\t\tfmt.Fprintf(flag.CommandLine.Output(), \"-b and -B are mutually exclusive.\\n\")\n\t\tflags.Usage()\n\t\tos.Exit(2)\n\t} else if *b != \"\" {\n\t\toptions.Branch = *b\n\t} else if *B != \"\" {\n\t\toptions.Branch = *B\n\t\toptions.ForceBranch = true\n\t}\n\n\tif *notrack && (*track != \"\" || *t != \"\") {\n\t\tfmt.Fprintf(flag.CommandLine.Output(), \"--track and --no-track are mutually exclusive.\\n\")\n\t\tflags.Usage()\n\t\tos.Exit(2)\n\t} else if !*notrack {\n\t\tif *track != \"\" && *t != \"\" {\n\t\t\tfmt.Fprintf(flag.CommandLine.Output(), \"--track and -t are mutually exclusive.\\n\")\n\t\t\tflags.Usage()\n\t\t\tos.Exit(2)\n\t\t} else if *track != \"\" {\n\t\t\toptions.Track = *track\n\t\t} else if *t != \"\" {\n\t\t\toptions.Track = *t\n\t\t}\n\t}\n\n\tif *orphan != \"\" {\n\t\tif options.Branch != \"\" {\n\t\t\tfmt.Fprintf(flag.CommandLine.Output(), \"--orphan is incompatible with -b/-B\\n\")\n\t\t\tflags.Usage()\n\t\t\tos.Exit(2)\n\t\t}\n\t\toptions.Branch = *orphan\n\t\toptions.OrphanBranch = true\n\t}\n\n\tvar thing string = \"HEAD\"\n\tif len(files) > 0 {\n\t\tf := git.File(files[0])\n\t\tif !f.Exists() {\n\t\t\tthing = files[0]\n\t\t\tfiles = files[1:]\n\t\t}\n\t}\n\n\t// Convert from string to git.File\n\tgfiles := make([]git.File, len(files))\n\tfor i, f := range files {\n\t\tgfiles[i] = git.File(f)\n\t}\n\n\treturn git.Checkout(c, options, thing, gfiles)\n}", "func parseGitCommand(sshOriginalCommand string) (command string, repopath string, err error) {\n\tr, err := regexp.Compile(`git-([a-z-]+) '/?([\\w-+@][\\w-+.@]*/)?([\\w-]+)\\.git'`)\n\tif err != nil {\n\t\treturn \"\", \"\", errors.New(\"parseGitCommand(): could not compile regex\")\n\t}\n\n\tm := r.FindStringSubmatch(sshOriginalCommand)\n\tif len(m) != 4 {\n\t\treturn \"\", \"\", errors.New(\"parseGitCommand(): Invalid GIT command\")\n\t}\n\n\treturn m[1], m[2] + m[3], nil\n}", "func (o *StepVerifyGitOptions) Run() error {\n\tlog.Logger().Infof(\"Verifying the git config\\n\")\n\n\tgha, err := o.IsGitHubAppMode()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"while checking if github app mode is enabled\")\n\t}\n\n\tif gha && o.GitHubAppOwner == \"\" {\n\t\tlog.Logger().Infof(\"this command does nothing if using github app mode and no %s option specified\", optionGitHubAppOwner)\n\t\treturn nil\n\t}\n\n\tvar authSvc auth.ConfigService\n\tif gha {\n\t\tauthSvc, err = o.GitAuthConfigServiceGitHubAppMode(\"github\")\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"when creating auth config service using GitAuthConfigServiceGitHubAppMode\")\n\t\t}\n\t} else {\n\t\tauthSvc, err = o.GitAuthConfigService()\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"when creating auth config service using GitAuthConfigService\")\n\t\t}\n\t}\n\n\tconfig := authSvc.Config()\n\tif config == nil {\n\t\treturn fmt.Errorf(\"git auth config is empty\")\n\t}\n\n\tservers := config.Servers\n\tif len(servers) == 0 {\n\t\treturn fmt.Errorf(\"no git servers found in the auth configuration\")\n\t}\n\tinfo := util.ColorInfo\n\tpipeUserValid := false\n\tfor _, server := range servers {\n\t\tfor _, userAuth := range server.Users {\n\t\t\tlog.Logger().Infof(\"Verifying username %s at git server %s at %s\\n\",\n\t\t\t\tinfo(userAuth.Username), info(server.Name), info(server.URL))\n\n\t\t\tprovider, err := gits.CreateProvider(server, userAuth, o.Git())\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"creating git provider for %s at git server %s\",\n\t\t\t\t\tuserAuth.Username, server.URL)\n\t\t\t}\n\n\t\t\tif strings.HasSuffix(provider.CurrentUsername(), \"[bot]\") {\n\t\t\t\tpipeUserValid = true\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\torgs, err := provider.ListOrganisations()\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"listing the organisations for %s at git server %s\",\n\t\t\t\t\tuserAuth.Username, server.URL)\n\t\t\t}\n\t\t\torgNames := []string{}\n\t\t\tfor _, org := range orgs {\n\t\t\t\torgNames = append(orgNames, org.Login)\n\t\t\t}\n\t\t\tsort.Strings(orgNames)\n\t\t\tlog.Logger().Infof(\"Found %d organisations in git server %s: %s\\n\",\n\t\t\t\tlen(orgs), info(server.URL), info(strings.Join(orgNames, \", \")))\n\t\t\tif config.PipeLineServer == server.URL && config.PipeLineUsername == userAuth.Username {\n\t\t\t\tpipeUserValid = true\n\t\t\t}\n\t\t}\n\t}\n\n\tif pipeUserValid {\n\t\tlog.Logger().Infof(\"Validated pipeline user %s on git server %s\", util.ColorInfo(config.PipeLineUsername), util.ColorInfo(config.PipeLineServer))\n\t} else {\n\t\treturn errors.Errorf(\"pipeline user %s on git server %s not valid\", util.ColorError(config.PipeLineUsername), util.ColorError(config.PipeLineServer))\n\t}\n\n\tlog.Logger().Infof(\"Git tokens seem to be setup correctly\\n\")\n\treturn nil\n}", "func (gc *GitCommand) Output() (string, error) {\n\targs := append(gc.reqArgs(), gc.addlArgs...)\n\tout, err := exec.Command(gc.binName, args...).Output()\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(out), nil\n}", "func printDiffs(c *git.Client, options git.DiffCommonOptions, diffs []git.HashDiff) error {\n\tif err := git.GeneratePatch(c, options, diffs, nil); err != nil {\n\t\treturn err\n\t}\n\tif options.ExitCode {\n\t\tif len(diffs) > 0 {\n\t\t\tos.Exit(1)\n\t\t}\n\t}\n\treturn nil\n}", "func (c CommitCmd) Run(args []string) int {\n\n\tvar yes bool\n\tcmdFlags := flag.NewFlagSet(\"commit\", flag.ContinueOnError)\n\tcmdFlags.BoolVar(&yes, \"yes\", false, \"\")\n\tcmdFlags.Usage = func() { c.UI.Output(c.Help()) }\n\tif err := cmdFlags.Parse(args); err != nil {\n\t\treturn 1\n\t}\n\n\tconfirm := yes\n\tif !confirm {\n\t\tresponse, err := c.UI.Ask(\"Save time for last commit (y/n)?\")\n\t\tif err != nil {\n\t\t\treturn 0\n\t\t}\n\t\tconfirm = strings.TrimSpace(strings.ToLower(response)) == \"y\"\n\t}\n\n\tif confirm {\n\t\tif _, err := metric.Process(false); err != nil {\n\t\t\tc.UI.Error(err.Error())\n\t\t\treturn 1\n\t\t}\n\t}\n\treturn 0\n}", "func gitPath() (string, error) {\n\tgitOnce.Do(func() {\n\t\tpath, err := exec.LookPath(\"git\")\n\t\tif err != nil {\n\t\t\tgitOnce.err = err\n\t\t\treturn\n\t\t}\n\t\tif runtime.GOOS == \"plan9\" {\n\t\t\tgitOnce.err = errors.New(\"plan9 git does not support the full git command line\")\n\t\t}\n\t\tgitOnce.path = path\n\t})\n\n\treturn gitOnce.path, gitOnce.err\n}", "func TestGitCommandGetBranchGraph(t *testing.T) {\n\tgitCmd := NewDummyGitCommand()\n\tgitCmd.OSCommand.Command = func(cmd string, args ...string) *exec.Cmd {\n\t\tassert.EqualValues(t, \"git\", cmd)\n\t\tassert.EqualValues(t, []string{\"log\", \"--graph\", \"--color=always\", \"--abbrev-commit\", \"--decorate\", \"--date=relative\", \"--pretty=medium\", \"test\", \"--\"}, args)\n\t\treturn secureexec.Command(\"echo\")\n\t}\n\t_, err := gitCmd.GetBranchGraph(\"test\")\n\tassert.NoError(t, err)\n}", "func GitMerge(pr *github.PullRequest, message string) error {\n\n\tfilepath := git.GetRepositoryFilePath(pr.Head.Repository.FullName)\n\tremoteRepositoryURL := git.GenerateCloneURL(pr.Head.Repository.FullName)\n\n\tif !git.Exists(filepath) {\n\t\tif _, err := git.Clone(remoteRepositoryURL); err != nil {\n\t\t\tpr.PostComment(\"I could not pull \" + pr.Head.Repository.FullName + \" from GitHub.\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err := git.Fetch(filepath); err != nil {\n\t\tgit.Prune(filepath)\n\t\tpr.PostComment(\"I could not fetch the latest changes from GitHub. Please try again in a few minutes.\")\n\t\treturn err\n\t}\n\n\tif err := git.Checkout(filepath, pr.Head.Ref); err != nil {\n\t\tpr.PostComment(\"I could not checkout \" + pr.Head.Ref + \" locally.\")\n\t\treturn err\n\t}\n\n\tif err := git.Reset(filepath, path.Join(\"origin\", pr.Head.Ref)); err != nil {\n\t\tpr.PostComment(\"I could not checkout \" + pr.Head.Ref + \" locally.\")\n\t\treturn err\n\t}\n\n\tif err := git.Config(filepath, \"user.name\", git.GetName()); err != nil {\n\t\tpr.PostComment(\"I could run git config for user.name on the server.\")\n\t\treturn err\n\t}\n\n\tif err := git.Config(filepath, \"user.email\", git.GetEmail()); err != nil {\n\t\tpr.PostComment(\"I could run git config for user.email on the server.\")\n\t\treturn err\n\t}\n\n\tif err := git.Rebase(filepath, path.Join(\"origin\", pr.Base.Ref)); err != nil {\n\t\tpr.PostComment(\"I could not rebase \" + pr.Head.Ref + \" with \" + pr.Base.Ref + \". There are conflicts.\")\n\t\treturn err\n\t}\n\n\tif err := git.Push(filepath, pr.Head.Ref); err != nil {\n\t\tpr.PostComment(\"I could not push the changes to \" + pr.Base.Ref + \".\")\n\t\treturn err\n\t}\n\n\tif err := git.Checkout(filepath, pr.Base.Ref); err != nil {\n\t\tpr.PostComment(\"I could not checkout \" + pr.Base.Ref + \" locally.\")\n\t\treturn err\n\t}\n\n\tif err := git.Fetch(filepath); err != nil {\n\t\tgit.Prune(filepath)\n\t\tpr.PostComment(\"I could not fetch the latest changes from GitHub. Please try again in a few minutes.\")\n\t\treturn err\n\t}\n\n\tif err := git.Reset(filepath, path.Join(\"origin\", pr.Base.Ref)); err != nil {\n\t\tpr.PostComment(\"I could not checkout \" + pr.Base.Ref + \" locally.\")\n\t\treturn err\n\t}\n\n\tif err := git.Merge(filepath, pr.Head.Ref, message); err != nil {\n\n\t\tpr.PostComment(\"I could not merge \" + pr.Head.Ref + \" into \" + pr.Base.Ref + \".\" + \"\\nNext time use your fingers for more than just picking your nose.\")\n\t\treturn err\n\t}\n\n\tif err := git.Push(filepath, pr.Base.Ref); err != nil {\n\t\tpr.PostComment(\"I could not push the changes to \" + pr.Base.Ref + \".\")\n\t\treturn err\n\t}\n\n\tpr.PostComment(\"I just merged \" + pr.Head.Ref + \" into \" + pr.Base.Ref+ \"\\nProbably you heard this before:\\nI did it, but I did not enjoy it...\")\n\treturn nil\n}", "func GitCommitPush(diretorio string, mensagemCommit string, username string, password string) error {\n\n\t// Abrir um repositório já existente\n\tr, err := git.PlainOpen(diretorio)\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"git.PlainOpen\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\tw, err := r.Worktree()\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.Worktree\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// Adds the new file to the staging area.\n\tlogger.Info.Println(\"git add .\")\n\t_, err = w.Add(\".\")\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Add\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// We can verify the current status of the worktree using the method Status.\n\tlogger.Info.Println(\"git status --porcelain\")\n\tstatus, err := w.Status()\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Status\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\tlogger.Info.Println(status)\n\n\t// Commit dos arquivos\n\tlogger.Info.Printf(\"git commit -m %s\\n\\r\", mensagemCommit)\n\tcommit, err := w.Commit(mensagemCommit, &git.CommitOptions{\n\t\tAuthor: &object.Signature{\n\t\t\tName: \"Marcelo Melo\",\n\t\t\tEmail: \"[email protected]\",\n\t\t\tWhen: time.Now(),\n\t\t},\n\t})\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Commit\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// Prints the current HEAD to verify that all worked well.\n\tlogger.Info.Println(\"git show -s\")\n\tobj, err := r.CommitObject(commit)\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.CommitObject\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\tlogger.Info.Println(obj)\n\n\tlogger.Info.Println(\"git push\")\n\terr = r.Push(&git.PushOptions{\n\t\tRemoteName: \"origin\",\n\t\tProgress: os.Stdout,\n\t\tAuth: &githttp.BasicAuth{\n\t\t\tUsername: username,\n\t\t\tPassword: password,\n\t\t}})\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.Push\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Commander) gitInit() error {\n\tc.cmd = exec.Command(\"git\", \"init\")\n\tc.cmd.Stdin, c.cmd.Stdout, c.cmd.Stderr = os.Stdin, os.Stdout, os.Stderr\n\tif err := c.cmd.Run(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.addRemote(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func gitInit(t testing.TB, dir string) {\n\tt.Helper()\n\tmustHaveGit(t)\n\n\tif _, err := run(t, dir, \"git\", \"init\"); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif _, err := run(t, dir, \"git\", \"config\", \"user.name\", \"Go Gopher\"); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif _, err := run(t, dir, \"git\", \"config\", \"user.email\", \"[email protected]\"); err != nil {\n\t\tt.Fatal(err)\n\t}\n}", "func main() {\n\tflag.Parse()\n\tappName := flag.Arg(1)\n\tif appName == \"\" {\n\t\tcommon.LogFail(\"Please specify an app to run the command on\")\n\t}\n\terr := common.VerifyAppName(appName)\n\tif err != nil {\n\t\tcommon.LogFail(err.Error())\n\t}\n\n\tappRoot := strings.Join([]string{common.MustGetEnv(\"DOKKU_ROOT\"), appName}, \"/\")\n\tcmdEnv := map[string]string{\n\t\t\"GIT_DIR\": appRoot,\n\t}\n\tgitGcCmd := common.NewShellCmd(\"git gc --aggressive\")\n\tgitGcCmd.Env = cmdEnv\n\tgitGcCmd.Execute()\n}", "func TestGitCommandGetCommitDifferences(t *testing.T) {\n\ttype scenario struct {\n\t\ttestName string\n\t\tcommand func(string, ...string) *exec.Cmd\n\t\ttest func(string, string)\n\t}\n\n\tscenarios := []scenario{\n\t\t{\n\t\t\t\"Can't retrieve pushable count\",\n\t\t\tfunc(string, ...string) *exec.Cmd {\n\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t},\n\t\t\tfunc(pushableCount string, pullableCount string) {\n\t\t\t\tassert.EqualValues(t, \"?\", pushableCount)\n\t\t\t\tassert.EqualValues(t, \"?\", pullableCount)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Can't retrieve pullable count\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tif args[1] == \"HEAD..@{u}\" {\n\t\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t\t}\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tfunc(pushableCount string, pullableCount string) {\n\t\t\t\tassert.EqualValues(t, \"?\", pushableCount)\n\t\t\t\tassert.EqualValues(t, \"?\", pullableCount)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Retrieve pullable and pushable count\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tif args[1] == \"HEAD..@{u}\" {\n\t\t\t\t\treturn secureexec.Command(\"echo\", \"10\")\n\t\t\t\t}\n\n\t\t\t\treturn secureexec.Command(\"echo\", \"11\")\n\t\t\t},\n\t\t\tfunc(pushableCount string, pullableCount string) {\n\t\t\t\tassert.EqualValues(t, \"11\", pushableCount)\n\t\t\t\tassert.EqualValues(t, \"10\", pullableCount)\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, s := range scenarios {\n\t\tt.Run(s.testName, func(t *testing.T) {\n\t\t\tgitCmd := NewDummyGitCommand()\n\t\t\tgitCmd.OSCommand.Command = s.command\n\t\t\ts.test(gitCmd.GetCommitDifferences(\"HEAD\", \"@{u}\"))\n\t\t})\n\t}\n}", "func GitModified(path string) int {\n\targs := []string{\"diff\", \"--shortstat\"}\n\toutput := common.GitRun(path, args, true)\n\tpat := regexp.MustCompile(`^\\s*(\\d+)`)\n\tmatches := pat.FindAllStringSubmatch(string(output), -1)\n\tif len(matches) > 0 {\n\t\tmodified, err := strconv.Atoi(matches[0][1])\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\treturn modified\n\t}\n\treturn 0\n}", "func (p project) gitClone() error {\n\tif p.SkipClone {\n\t\treturn nil\n\t}\n\tcmd := fmt.Sprintf(\"git clone -b %s %s %s\", p.Branch, p.Repo, localRepos+p.Name)\n\treturn doExec(cmd, \"\")\n}", "func (c cBuild) getGitCommit() string {\n\tif gproc.SearchBinary(\"git\") == \"\" {\n\t\treturn \"\"\n\t}\n\tvar (\n\t\tcmd = `git log -1 --format=\"%cd %H\" --date=format:\"%Y-%m-%d %H:%M:%S\"`\n\t\ts, _ = gproc.ShellExec(cmd)\n\t)\n\tmlog.Debug(cmd)\n\tif s != \"\" {\n\t\tif !gstr.Contains(s, \"fatal\") {\n\t\t\treturn gstr.Trim(s)\n\t\t}\n\t}\n\treturn \"\"\n}", "func (r Repository) Commit(msg string, opt *CommitOptions) (Hash, error) {\n\tif opt == nil {\n\t\topt = &CommitOptions{}\n\t}\n\n\targs := []string{\"commit\"}\n\tif opt.Amend {\n\t\targs = append(args, \"--amend\")\n\t} else {\n\t\targs = append(args, \"-m\", msg)\n\t}\n\n\tvar env []string\n\tif opt.AuthorName != \"\" || opt.AuthorEmail != \"\" {\n\t\tenv = []string{\n\t\t\tfmt.Sprintf(\"GIT_AUTHOR_NAME=%v\", opt.AuthorName),\n\t\t\tfmt.Sprintf(\"GIT_AUTHOR_EMAIL=%v\", opt.AuthorEmail),\n\t\t\tfmt.Sprintf(\"GIT_COMMITTER_NAME=%v\", opt.AuthorName),\n\t\t\tfmt.Sprintf(\"GIT_COMMITTER_EMAIL=%v\", opt.AuthorEmail),\n\t\t}\n\t}\n\tif _, err := r.run(env, opt.Timeout, \"commit\", \"-m\", msg); err != nil {\n\t\treturn Hash{}, err\n\t}\n\tout, err := r.run(nil, 0, \"rev-parse\", \"HEAD\")\n\tif err != nil {\n\t\treturn Hash{}, err\n\t}\n\treturn ParseHash(out)\n}", "func (Stage) Run(ctx *ctx.Context) error {\n\tlog.Debug(\"checking if git exists on PATH\")\n\tif !utils.BinExists(\"git\") {\n\t\treturn ErrGitNotFound\n\t}\n\n\tlog.Debug(\"checking if directory is a git repo\")\n\tif !utils.InRepo() {\n\t\tif err := ctx.CheckDryRun(ErrNotInRepo); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Debug(\"checking if git is in a clean state\")\n\tif isDirty, out := utils.IsDirty(); isDirty {\n\t\tif ctx.AllowDirty {\n\t\t\tlog.Info(\"allowing git to be in a dirty state\")\n\t\t} else {\n\t\t\tif err := ctx.CheckDryRun(ErrDirtyGit); err != nil {\n\t\t\t\tlog.Errorf(\"dirty git state detected\\n\" + out)\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func commit() string {\n\thash, _ := sh.Output(\"git\", \"rev-parse\", \"--short\", \"HEAD\")\n\treturn hash\n}", "func GitVersion() string { return gitVersion }", "func GitStatus() ([]string, error) {\n\tvar (\n\t\terr error\n\t\tmodfiles []string\n\t)\n\n\t// Git status command\n\t// gsArgs := []string{\"diff\", \"--staged\", \"--name-status\"}\n\t// gitStatus := exec.Command(\"git\", gsArgs...)\n\tgsArgs := []string{\"status\", \"--short\"}\n\tgitStatus := exec.Command(\"git\", gsArgs...)\n\n\t// Get stdout and trim the empty last index\n\tfileStatus, err := gitStatus.CombinedOutput()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfsSplit := strings.Split(string(fileStatus), \"\\n\")\n\n\tfor _, status := range fsSplit {\n\n\t\ts := strings.Fields(status)\n\t\tif len(s) == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// With the \"git status --short\" command the staged files are at 0 index.\n\t\t// Unstaged: \" M filename\" has a space before the M.\n\t\t// Staged: \"M filename\" has no space before the M.\n\t\t//\n\t\t// If there is a space we do not want to include it in the commit.\n\t\tif unicode.IsSpace(rune(status[0])) {\n\t\t\tcontinue\n\t\t}\n\n\t\t// With the \"git status --short\" command the staged files are at 0 index.\n\t\t// Unstaged: \"?? directory/\"\n\t\t//\n\t\t// If there is a question mark that means that it is untracked and unstaged.\n\t\tif unicode.IsPunct(rune(status[0])) {\n\t\t\tcontinue\n\t\t}\n\n\t\t// If the file was deleted there is no reason to read that file.\n\t\tif strings.Contains(s[0], \"D\") {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Only get the file path.\n\t\tmodfiles = append(modfiles, s[len(s)-1])\n\t}\n\n\treturn modfiles, err\n}", "func main() {\n\tapp := kingpin.New(\"gg\", \"a git commit & push\")\n\tautoCommit(app) //\tgit add\n\tgetMessage(app) //\tgit commit -m\n\tpushRemote(app) //\tgit push -u\n\tgetCommit(app) //\tgit log --date=short --no-merges --pretty=format:\"%cd (@%cn) %h %s\"\n\tfullCommand(app)\n\tkingpin.MustParse(app.Parse(os.Args[1:]))\n}", "func iRunGitcleanInThatRepo() error {\n\terr := RunGitClean(gitWorkingDir)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func handleRepo(config *Config, sshURL string) {\n\tlog.Println(\"Repo clone\", sshURL)\n\n\ttempCloneName := getTempRepoName(sshURL)\n\n\tsyscall.Chdir(config.TempDir)\n\tos.RemoveAll(\"./\" + tempCloneName)\n\tdefer func() {\n\t\tsyscall.Chdir(config.TempDir)\n\t\tos.RemoveAll(\"./\" + tempCloneName)\n\t}()\n\n\t_, err_clone := exec.Command(config.GitCMD, \"clone\", \"--branch\", config.Branch, sshURL, tempCloneName).Output()\n\tif err_clone != nil {\n\t\tlog.Println(\"Repo cannot be cloned\", sshURL, err_clone)\n\t\treturn\n\t}\n\n\tout_grep, err_grep := execCmdWithOutput(config.GrepCMD, \"-rl\", \"--exclude-dir\", \".git\", config.ReplaceFrom, tempCloneName)\n\tif err_grep != nil {\n\t\tlog.Panic(err_grep)\n\t\treturn\n\t}\n\n\tout_grep_trimmed := strings.Trim(out_grep, \"\\n\\r\\t \")\n\tif out_grep_trimmed == \"\" {\n\t\tlog.Println(\"No match\")\n\t\treturn\n\t}\n\n\tfiles := strings.Split(out_grep_trimmed, \"\\n\")\n\tfor _, fileName := range files {\n\t\thandleFile(config, fileName)\n\t}\n\n\t// Make git operations safe - they have to be called from the directory\n\tmutexInRepoOp.Lock()\n\tsyscall.Chdir(\"./\" + tempCloneName)\n\tdiff, err_diff := execCmdWithOutput(config.GitCMD, \"diff\")\n\tif err_diff != nil {\n\t\tlog.Panic(err_diff)\n\t\treturn\n\t}\n\tlog.Println(diff)\n\n\tif flagCommit {\n\t\tlog.Println(\"Committing changes\")\n\t\t_, err_commit := execCmdWithOutput(config.GitCMD, \"commit\", \"-a\", \"-m\", config.CommitMessage)\n\t\tif err_commit != nil {\n\t\t\tlog.Panic(err_commit)\n\t\t\treturn\n\t\t}\n\n\t\tlog.Println(\"Push to remote\")\n\t\t_, err_push := execCmdWithOutput(config.GitCMD, \"push\", \"origin\", config.Branch)\n\t\tif err_push != nil {\n\t\t\tlog.Panic(err_push)\n\t\t\treturn\n\t\t}\n\t\tlog.Println(\"Commit and push succeed\")\n\t}\n\n\tsyscall.Chdir(config.TempDir)\n\tmutexInRepoOp.Unlock()\n}", "func GitAdd(tb testing.TB) {\n\ttb.Helper()\n\tout, err := fakeGit(\"add\", \"-A\")\n\trequire.NoError(tb, err)\n\trequire.Empty(tb, out)\n}", "func (r *Runner) Run(ctx context.Context, remoteName string, remoteUrl string) error {\n\tlog.Infof(\"running git-remote-dgit on remote %s with url %s\", remoteName, remoteUrl)\n\n\t// get the named remote as reported by git, but then\n\t// create a new remote with only the url specified\n\t// this is for cases when a remote has multiple urls\n\t// specified for push / fetch\n\tnamedRemote, err := r.local.Remote(remoteName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = namedRemote.Config().Validate()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Invalid remote config: %v\", err)\n\t}\n\n\tremote := git.NewRemote(r.local.Storer, &config.RemoteConfig{\n\t\tName: namedRemote.Config().Name,\n\t\tFetch: namedRemote.Config().Fetch,\n\t\tURLs: []string{remoteUrl},\n\t})\n\n\tstdinReader := bufio.NewReader(r.stdin)\n\n\tfor {\n\t\tvar err error\n\n\t\tcommand, err := stdinReader.ReadString('\\n')\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcommand = strings.TrimSpace(command)\n\t\tcommandParts := strings.Split(command, \" \")\n\n\t\tlog.Infof(\"received command on stdin %s\", command)\n\n\t\targs := strings.TrimSpace(strings.TrimPrefix(command, commandParts[0]))\n\t\tcommand = commandParts[0]\n\n\t\tswitch command {\n\t\tcase \"capabilities\":\n\t\t\tr.respond(strings.Join([]string{\n\t\t\t\t\"*push\",\n\t\t\t\t\"*fetch\",\n\t\t\t}, \"\\n\") + \"\\n\")\n\t\t\tr.respond(\"\\n\")\n\t\tcase \"list\":\n\t\t\trefs, err := remote.List(&git.ListOptions{})\n\n\t\t\tif err == transport.ErrRepositoryNotFound && args == \"for-push\" {\n\t\t\t\tr.respond(\"\\n\")\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif err == transport.ErrRepositoryNotFound {\n\t\t\t\treturn fmt.Errorf(msg.RepoNotFound)\n\t\t\t}\n\n\t\t\tif err == transport.ErrEmptyRemoteRepository || len(refs) == 0 {\n\t\t\t\tr.respond(\"\\n\")\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar head string\n\n\t\t\tlistResponse := make([]string, len(refs))\n\t\t\tfor i, ref := range refs {\n\t\t\t\tlistResponse[i] = fmt.Sprintf(\"%s %s\", ref.Hash(), ref.Name())\n\n\t\t\t\t// TODO: set default branch in repo chaintree which\n\t\t\t\t// would become head here\n\t\t\t\t//\n\t\t\t\t// if master head exists, use that\n\t\t\t\tif ref.Name() == \"refs/heads/master\" {\n\t\t\t\t\thead = ref.Name().String()\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tsort.Slice(listResponse, func(i, j int) bool {\n\t\t\t\treturn strings.Split(listResponse[i], \" \")[1] < strings.Split(listResponse[j], \" \")[1]\n\t\t\t})\n\n\t\t\t// if head is empty, use last as default\n\t\t\tif head == \"\" {\n\t\t\t\thead = strings.Split(listResponse[len(listResponse)-1], \" \")[1]\n\t\t\t}\n\n\t\t\tr.respond(\"@%s HEAD\\n\", head)\n\t\t\tr.respond(\"%s\\n\", strings.Join(listResponse, \"\\n\"))\n\t\t\tr.respond(\"\\n\")\n\t\tcase \"push\":\n\t\t\trefSpec := config.RefSpec(args)\n\n\t\t\tendpoint, err := transport.NewEndpoint(remote.Config().URLs[0])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tauth, err := r.auth()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tlog.Debugf(\"auth for push: %s %s\", auth.Name(), auth.String())\n\n\t\t\terr = remote.PushContext(ctx, &git.PushOptions{\n\t\t\t\tRemoteName: remote.Config().Name,\n\t\t\t\tRefSpecs: []config.RefSpec{refSpec},\n\t\t\t\tAuth: auth,\n\t\t\t})\n\n\t\t\tif err == transport.ErrRepositoryNotFound {\n\t\t\t\terr = nil // reset err back to nil\n\t\t\t\tclient, err := dgit.Default()\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\t_, err = client.CreateRepoTree(ctx, endpoint, auth, os.Getenv(\"DGIT_OBJ_STORAGE\"))\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\t// Retry push now that repo exists\n\t\t\t\terr = remote.PushContext(ctx, &git.PushOptions{\n\t\t\t\t\tRemoteName: remote.Config().Name,\n\t\t\t\t\tRefSpecs: []config.RefSpec{refSpec},\n\t\t\t\t\tAuth: auth,\n\t\t\t\t})\n\t\t\t}\n\n\t\t\tdst := refSpec.Dst(plumbing.ReferenceName(\"*\"))\n\t\t\tif err != nil && err != git.NoErrAlreadyUpToDate {\n\t\t\t\tr.respond(\"error %s %s\\n\", dst, err.Error())\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tr.respond(\"ok %s\\n\", dst)\n\t\t\tr.respond(\"\\n\")\n\t\tcase \"fetch\":\n\t\t\tsplitArgs := strings.Split(args, \" \")\n\t\t\tif len(splitArgs) != 2 {\n\t\t\t\treturn fmt.Errorf(\"incorrect arguments for fetch, received %s, expected 'hash refname'\", args)\n\t\t\t}\n\n\t\t\trefName := plumbing.ReferenceName(splitArgs[1])\n\n\t\t\trefSpecs := []config.RefSpec{}\n\n\t\t\tlog.Debugf(\"remote fetch config %v\", remote.Config().Name)\n\n\t\t\tfor _, fetchRefSpec := range remote.Config().Fetch {\n\t\t\t\tif !fetchRefSpec.Match(refName) {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tnewRefStr := \"\"\n\t\t\t\tif fetchRefSpec.IsForceUpdate() {\n\t\t\t\t\tnewRefStr += \"+\"\n\t\t\t\t}\n\t\t\t\tnewRefStr += refName.String() + \":\" + fetchRefSpec.Dst(refName).String()\n\n\t\t\t\tnewRef := config.RefSpec(newRefStr)\n\n\t\t\t\tif err := newRef.Validate(); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tlog.Debugf(\"attempting to fetch on %s\", newRef.String())\n\t\t\t\trefSpecs = append(refSpecs, newRef)\n\t\t\t}\n\n\t\t\terr := remote.FetchContext(ctx, &git.FetchOptions{\n\t\t\t\tRemoteName: remote.Config().Name,\n\t\t\t\tRefSpecs: refSpecs,\n\t\t\t})\n\t\t\tif err != nil && err != git.NoErrAlreadyUpToDate {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tlog.Debugf(\"fetch complete\")\n\t\t\tr.respond(\"\\n\")\n\t\t// Connect can be used for upload / receive pack\n\t\t// case \"connect\":\n\t\t// \tr.respond(\"fallback\\n\")\n\t\tcase \"\": // command stream terminated, return out\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"Command '%s' not handled\", command)\n\t\t}\n\t}\n\n\treturn nil\n}", "func TestGitCommandCurrentBranchName(t *testing.T) {\n\ttype scenario struct {\n\t\ttestName string\n\t\tcommand func(string, ...string) *exec.Cmd\n\t\ttest func(string, string, error)\n\t}\n\n\tscenarios := []scenario{\n\t\t{\n\t\t\t\"says we are on the master branch if we are\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.Equal(t, \"git\", cmd)\n\t\t\t\treturn secureexec.Command(\"echo\", \"master\")\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"master\", name)\n\t\t\t\tassert.EqualValues(t, \"master\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"falls back to git `git branch --contains` if symbolic-ref fails\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\n\t\t\t\tswitch args[0] {\n\t\t\t\tcase \"symbolic-ref\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"symbolic-ref\", \"--short\", \"HEAD\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t\tcase \"branch\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"branch\", \"--contains\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"echo\", \"* master\")\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"master\", name)\n\t\t\t\tassert.EqualValues(t, \"master\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"handles a detached head\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\n\t\t\t\tswitch args[0] {\n\t\t\t\tcase \"symbolic-ref\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"symbolic-ref\", \"--short\", \"HEAD\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t\tcase \"branch\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"branch\", \"--contains\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"echo\", \"* (HEAD detached at 123abcd)\")\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"123abcd\", name)\n\t\t\t\tassert.EqualValues(t, \"(HEAD detached at 123abcd)\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"bubbles up error if there is one\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.Equal(t, \"git\", cmd)\n\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.Error(t, err)\n\t\t\t\tassert.EqualValues(t, \"\", name)\n\t\t\t\tassert.EqualValues(t, \"\", displayname)\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, s := range scenarios {\n\t\tt.Run(s.testName, func(t *testing.T) {\n\t\t\tgitCmd := NewDummyGitCommand()\n\t\t\tgitCmd.OSCommand.Command = s.command\n\t\t\ts.test(gitCmd.CurrentBranchName())\n\t\t})\n\t}\n}", "func parseGitCommand(sshcmd string) (command, name string, err error) {\n\t// The following regex validates the git command, which is in the form:\n\t// <git-command> [<namespace>/]<name>\n\t// with namespace being optional. If a namespace is used, we validate it\n\t// according to the following:\n\t// - a namespace is optional\n\t// - a namespace contains only alphanumerics, underlines, @´s, -´s, +´s\n\t// and periods but it does not start with a period (.)\n\t// - one and exactly one slash (/) separates namespace and the actual name\n\tr, err := regexp.Compile(`(git-[a-z-]+) '/?([\\w-+@][\\w-+.@]*/)?([\\w-]+)\\.git'`)\n\t// r, err := regexp.Compile(`git-(upload|receive)-pack '/?([\\w-+@][\\w-+.@]*/)?([\\w-]+)\\.git'`)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tm := r.FindStringSubmatch(sshcmd)\n\tif len(m) != 4 {\n\t\treturn \"\", \"\", errors.New(\"You've tried to execute some weird command, I'm deliberately denying you to do that, get over it.\")\n\t}\n\treturn m[1], m[2] + m[3], nil\n}", "func RunInPath(path string, args ...string) ([]byte, error) {\n\t// is git installed?\n\t_git, _ := Git()\n\tif _git == \"\" {\n\t\treturn nil, MissingGitError\n\t}\n\n\t// if we have a path, attempt to change into it before executing\n\t// the git command\n\tif path != \"\" {\n\t\tvar _dir string\n\n\t\t// do we have a file or a directory?\n\t\t_info, _err := os.Stat(path)\n\t\tif _err != nil {\n\t\t\treturn nil, _err\n\t\t} else if _info.IsDir() {\n\t\t\t_dir = path\n\t\t} else {\n\t\t\t_dir, _ = filepath.Split(path)\n\t\t}\n\n\t\t_cwd, _err := os.Getwd()\n\t\tif _err != nil {\n\t\t\treturn nil, _err\n\t\t}\n\n\t\t// attempt to change into the given path\n\t\t_err = os.Chdir(_dir)\n\t\tif _err != nil {\n\t\t\treturn nil, _err\n\t\t}\n\t\tdefer os.Chdir(_cwd)\n\t}\n\n\t// execute the git command\n\treturn exec.Command(_git, args...).Output()\n}", "func FindGit() (string, error) {\n\tif !bazel.InBazelTest() {\n\t\treturn exec.LookPath(\"git\")\n\t}\n\tif runtime.GOOS == \"windows\" {\n\t\treturn filepath.Join(bazel.RunfilesDir(), \"external\", \"git_amd64_windows\", \"bin\", \"git.exe\"), nil\n\t} else if runtime.GOOS == \"linux\" {\n\t\treturn filepath.Join(bazel.RunfilesDir(), \"external\", \"git_amd64_linux\", \"bin\", \"git\"), nil\n\t}\n\treturn \"\", skerr.Fmt(\"unsupported runtime.GOOS: %q\", runtime.GOOS)\n}", "func version() {\n fmt.Printf(\"v%s\\ncommit=%s\\n\", versionNumber, commitId)\n}", "func CheckoutIndexCmd(c *git.Client, args []string) error {\n\tflags := flag.NewFlagSet(\"checkout-index\", flag.ExitOnError)\n\tflags.SetOutput(flag.CommandLine.Output())\n\tflags.Usage = func() {\n\t\tflag.Usage()\n\t\tfmt.Fprintf(flag.CommandLine.Output(), \"\\n\\nOptions:\\n\")\n\t\tflags.PrintDefaults()\n\t\t// Some git tests test for a 129 exit code if the commandline\n\t\t// parsing fails for checkout-index.\n\t\tos.Exit(129)\n\t}\n\toptions := git.CheckoutIndexOptions{}\n\n\tflags.BoolVar(&options.UpdateStat, \"index\", false, \"Update stat information for checkout out entries in the index\")\n\tflags.BoolVar(&options.UpdateStat, \"u\", false, \"Alias for --index\")\n\n\tflags.BoolVar(&options.Quiet, \"quiet\", false, \"Be quiet if files exist or are not in index\")\n\tflags.BoolVar(&options.Quiet, \"q\", false, \"Alias for --quiet\")\n\n\tflags.BoolVar(&options.Force, \"force\", false, \"Force overwrite of existing files\")\n\tflags.BoolVar(&options.Force, \"f\", false, \"Alias for --force\")\n\n\tflags.BoolVar(&options.All, \"all\", false, \"Checkout all files in the index.\")\n\tflags.BoolVar(&options.All, \"a\", false, \"Alias for --all\")\n\n\tflags.BoolVar(&options.NoCreate, \"no-create\", false, \"Don't checkout new files, only refresh existing ones\")\n\tflags.BoolVar(&options.NoCreate, \"n\", false, \"Alias for --no-create\")\n\n\tflags.StringVar(&options.Prefix, \"prefix\", \"\", \"When creating files, prepend string\")\n\tflags.StringVar(&options.Stage, \"stage\", \"\", \"Copy files from named stage (unimplemented)\")\n\n\tflags.BoolVar(&options.Temp, \"temp\", false, \"Instead of copying files to a working directory, write them to a temp dir\")\n\n\tstdin := flags.Bool(\"stdin\", false, \"Instead of taking paths from command line, read from stdin\")\n\tflags.BoolVar(&options.NullTerminate, \"z\", false, \"Use nil instead of newline to terminate paths read from stdin\")\n\n\tflags.Parse(args)\n\tfiles := flags.Args()\n\tif *stdin {\n\t\toptions.Stdin = os.Stdin\n\t}\n\n\t// Convert from string to git.File\n\tgfiles := make([]git.File, len(files))\n\tfor i, f := range files {\n\t\tgfiles[i] = git.File(f)\n\t}\n\n\treturn git.CheckoutIndex(c, options, gfiles)\n\n}", "func SignCommitCmd(cfg *config.AppConfig, repo pl.LocalRepo, args *types3.SignCommitArgs) error {\n\tpopulateSignCommitArgsFromRepoConfig(repo, args)\n\n\t// Set merge ID from env if unset\n\tif args.MergeID == \"\" {\n\t\targs.MergeID = strings.ToUpper(os.Getenv(fmt.Sprintf(\"%s_MR_ID\", cfg.GetAppName())))\n\t}\n\n\t// Signing key is required\n\tif args.SigningKey == \"\" {\n\t\treturn ErrMissingPushKeyID\n\t}\n\n\t// Get and unlock the signing key\n\tkey, err := args.KeyUnlocker(cfg, &common.UnlockKeyArgs{\n\t\tKeyStoreID: args.SigningKey,\n\t\tPassphrase: args.PushKeyPass,\n\t\tNoPrompt: args.NoPrompt,\n\t\tTargetRepo: repo,\n\t\tStdout: args.Stdout,\n\t\tPrompt: \"Enter passphrase to unlock the signing key\\n\",\n\t})\n\tif err != nil {\n\t\treturn errors2.Wrap(err, \"failed to unlock the signing key\")\n\t}\n\n\t// Get push key from key (args.SigningKey may not be push key address)\n\tpushKeyID := key.GetPushKeyAddress()\n\n\t// If MergeID is set, validate it.\n\tif args.MergeID != \"\" {\n\t\terr = validation.CheckMergeProposalID(args.MergeID, -1)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(err.(*errors.BadFieldError).Msg)\n\t\t}\n\t}\n\n\t// Get the next nonce, if not set\n\tif args.Nonce == 0 {\n\t\tnonce, err := args.GetNextNonce(pushKeyID, args.RPCClient)\n\t\tif err != nil {\n\t\t\treturn errors2.Wrapf(err, \"failed to get next nonce\")\n\t\t}\n\t\targs.Nonce = cast.ToUint64(nonce)\n\t}\n\n\t// Get the current HEAD reference.\n\thead, err := repo.Head()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get HEAD\")\n\t}\n\tif args.Head != \"\" {\n\t\thead = args.Head\n\t}\n\tif !pl.IsReference(head) {\n\t\thead = plumbing.NewBranchReferenceName(args.Head).String()\n\t}\n\n\t// Get the HEAD reference object\n\theadRef, err := repo.Reference(plumbing.ReferenceName(head), false)\n\tif err != nil {\n\t\treturn errors2.Wrapf(err, \"failed to find reference: %s\", head)\n\t}\n\n\tif err = args.CreateApplyPushTokenToRemote(repo, &server.MakeAndApplyPushTokenToRemoteArgs{\n\t\tTargetRemote: args.Remote,\n\t\tPushKey: key,\n\t\tStderr: args.Stderr,\n\t\tResetTokens: args.ResetTokens,\n\t\tTxDetail: &types.TxDetail{\n\t\t\tFee: util.String(args.Fee),\n\t\t\tValue: util.String(args.Value),\n\t\t\tNonce: args.Nonce,\n\t\t\tPushKeyID: pushKeyID,\n\t\t\tMergeProposalID: args.MergeID,\n\t\t\tReference: head,\n\t\t\tHead: headRef.Hash().String(),\n\t\t},\n\t}); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func PrintUsage() {\n\tfmt.Fprintln(os.Stdout, \"Usage: gitio [-code=] url\\nIf you will be use any code, set code flag\")\n}", "func main() {\n\tvar gitPath = \"/home/pi/project/test\"\n\trepo, err := gogit.PlainOpen(gitPath)\n\tb, e := repo.Branch(\"home\")\n\tfmt.Println(b, e)\n\tpanicError(err)\n\tbranches, err := repo.Branches()\n\tpanicError(err)\n\t//fmt.Println(branches)\n\tbranches.ForEach(func(ref *plumbing.Reference) error {\n\t\tfmt.Println(ref)\n\t\treturn nil\n\t})\n}", "func GitUntracked(path string) int {\n\targs := []string{\"ls-files\", \"--others\", \"--exclude-standard\"}\n\toutput := common.GitRun(path, args, true)\n\treturn len(strings.Split(string(output), \"\\n\")) - 1\n}", "func versionCmd(c *cli.Context) error {\n\tif len(c.Args()) > 0 {\n\t\treturn errors.New(\"'buildah version' does not accept arguments\")\n\t}\n\n\t//converting unix time from string to int64\n\tbuildTime, err := strconv.ParseInt(buildInfo, 10, 64)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Println(\"Version: \", buildah.Version)\n\tfmt.Println(\"Go Version: \", runtime.Version())\n\tfmt.Println(\"Image Spec: \", ispecs.Version)\n\tfmt.Println(\"Runtime Spec: \", rspecs.Version)\n\tfmt.Println(\"CNI Spec: \", cniversion.Current())\n\tfmt.Println(\"libcni Version: \", cniVersion)\n\tfmt.Println(\"Git Commit: \", gitCommit)\n\n\t//Prints out the build time in readable format\n\tfmt.Println(\"Built: \", time.Unix(buildTime, 0).Format(time.ANSIC))\n\tfmt.Println(\"OS/Arch: \", runtime.GOOS+\"/\"+runtime.GOARCH)\n\n\treturn nil\n}", "func checkout(p provision.Provisioner, app provision.App, version string) ([]byte, error) {\n\tvar buf bytes.Buffer\n\tpath, err := repository.GetPath()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"tsuru is misconfigured: %s\", err)\n\t}\n\tcmd := fmt.Sprintf(\"cd %s && git checkout %s\", path, version)\n\tif err := p.ExecuteCommand(&buf, &buf, app, cmd); err != nil {\n\t\treturn buf.Bytes(), err\n\t}\n\treturn nil, nil\n}", "func main() {\n\tvar branchesOptions BranchesOptions\n\tvar versionOptions VersionOptions\n\n\tcleanupCmd := &cobra.Command{\n\t\tUse: \"cleanup\",\n\t\tAliases: []string{\"git-cleanupCmd\"},\n\t\tShort: `💫 Remove gone Git branches with ease.`,\n\t\tVersion: version,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn cmd.Help()\n\t\t},\n\t}\n\n\tbranchesCmd := &cobra.Command{\n\t\tUse: \"branches <PATH>\",\n\t\tShort: `Delete local branches that are gone on the remote`,\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn RunBranches(args[0], &branchesOptions, os.Stdout)\n\t\t},\n\t}\n\n\tbranchesCmd.Flags().BoolVarP(&branchesOptions.HasMultipleRepos, \"has-multiple-repos\",\n\t\t\"m\", false, `Delete branches in sub-repositories`)\n\tbranchesCmd.Flags().BoolVarP(&branchesOptions.Force, \"force\",\n\t\t\"f\", false, `Force the deletion, ignoring warnings`)\n\tbranchesCmd.Flags().BoolVarP(&branchesOptions.DryRun, \"dry-run\",\n\t\t\"d\", false, `Preview the branches without deleting them`)\n\tbranchesCmd.Flags().StringVarP(&branchesOptions.Exclude, \"exclude\",\n\t\t\"e\", \"\", `Exclude one or more branches from deletion`)\n\tbranchesCmd.Flags().StringVarP(&branchesOptions.Where, \"where\",\n\t\t\"w\", \"\", `Delete all branches whose output contain a given string`)\n\tbranchesCmd.Flags().StringVar(&branchesOptions.AndWhere, \"and-where\",\n\t\t\"\", `Delete all gone branches whose output contain a given string`)\n\n\tversionCmd := &cobra.Command{\n\t\tUse: \"version\",\n\t\tShort: `Display version information`,\n\t\tArgs: cobra.ExactArgs(0),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn Version(&versionOptions, os.Stdout)\n\t\t},\n\t}\n\n\tversionCmd.Flags().BoolVarP(&versionOptions.Quiet, \"quiet\",\n\t\t\"q\", false, `Only print the version number`)\n\n\tcleanupCmd.AddCommand(branchesCmd)\n\tcleanupCmd.AddCommand(versionCmd)\n\n\tif err := cleanupCmd.Execute(); err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func GitCloneCmd() (bool, error) {\n\tgitRepoDir := filepath.Join(VxCfg.WorkDir, VxCfg.GitRepoName)\n\t_, err := os.Stat(gitRepoDir)\n\tif err == nil {\n\t\tlog.Printf(\"gitRepoDir is clone : %v\\n\", err)\n\t\treturn false, errors.New(\"gitRepoDir: \" + gitRepoDir + \" exists\")\n\t}\n\tcmd := exec.Command(\"git\", \"clone\", VxCfg.GitRepo)\n\tcmd.Dir = VxCfg.WorkDir\n\tresultBytes, err := cmd.Output()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tlog.Printf(\"git clone \" + VxCfg.GitRepo + \" status: \" + string(resultBytes) + \"\\n\")\n\treturn true, nil\n}", "func GitBranch(e *editarea.EditArea) string {\n\tcmdOut, err := exec.Command(\"git\", \"rev-parse\", \"--abbrev-ref\", \"HEAD\").Output()\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn string(cmdOut)\n}" ]
[ "0.7911547", "0.71259296", "0.7000005", "0.69734985", "0.6904909", "0.67912936", "0.666957", "0.6618324", "0.6606666", "0.6528801", "0.6436387", "0.6329439", "0.6293814", "0.6277394", "0.6271859", "0.62361526", "0.6219715", "0.61598617", "0.6095661", "0.60269254", "0.59765816", "0.5968324", "0.5968276", "0.59232473", "0.58171284", "0.58141184", "0.57926965", "0.57915753", "0.57669294", "0.5756688", "0.57363284", "0.57043004", "0.56978625", "0.56959444", "0.5695395", "0.56159836", "0.5610022", "0.5600067", "0.5572816", "0.5570842", "0.5562452", "0.5559535", "0.55494446", "0.5529224", "0.55099666", "0.54834825", "0.54806435", "0.54769075", "0.54603", "0.5452314", "0.5451767", "0.54481965", "0.5442458", "0.54328364", "0.54255253", "0.5420995", "0.53933716", "0.53877777", "0.5378641", "0.5378419", "0.5373912", "0.5371555", "0.5347664", "0.53415567", "0.5332581", "0.532357", "0.53214115", "0.5317328", "0.5316373", "0.5314815", "0.53063226", "0.53005856", "0.52948344", "0.5286819", "0.5263391", "0.52586573", "0.52580905", "0.52359164", "0.5222082", "0.52013737", "0.5200645", "0.51732254", "0.5159816", "0.51590884", "0.5152257", "0.51274264", "0.5106098", "0.50996774", "0.5082732", "0.50625557", "0.5062049", "0.5056793", "0.505351", "0.5046443", "0.50276315", "0.5019425", "0.50164235", "0.50152314", "0.49591017", "0.49545878" ]
0.56290394
35
Call a git command and get the output as string output.
func Get(opts ...string) (string, error) { gotrace.Trace("RUNNING: git %s", strings.Join(opts, " ")) out, err := exec.Command("git", opts...).Output() return string(out), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (gitCommand *GitCommand) RunAndGetOutput() []byte {\n\tstdout, err := gitCommand.Run(false)\n\tif err != nil {\n\t\treturn []byte{}\n\t}\n\n\tdata, err := ioutil.ReadAll(stdout)\n\tif err != nil {\n\t\treturn []byte{}\n\t}\n\n\treturn data\n}", "func Git(cmd string, args ...string) (res *exec.Cmd, stdout, stderr *bytes.Buffer) {\n\tcmdArgs := make([]string, 1)\n\tcmdArgs[0] = cmd\n\tcmdArgs = append(cmdArgs, args...)\n\tres = exec.Command(gitCmd, cmdArgs...)\n\tstdout, stderr = new(bytes.Buffer), new(bytes.Buffer)\n\tres.Stdout, res.Stderr = stdout, stderr\n\treturn\n}", "func Run(args ...string) (output string, err error) {\n\tvar cmd = exec.Command(\"git\", args...)\n\tbts, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn \"\", errors.New(string(bts))\n\t}\n\treturn string(bts), err\n}", "func execGitCommand(ctx context.Context, store store.Store, repositoryID int, args ...string) (string, error) {\n\trepo, err := repositoryIDToRepo(ctx, store, repositoryID)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tcmd := gitserver.DefaultClient.Command(\"git\", args...)\n\tcmd.Repo = repo\n\tout, err := cmd.CombinedOutput(ctx)\n\treturn string(bytes.TrimSpace(out)), errors.Wrap(err, \"gitserver.Command\")\n}", "func (c *config) command(args ...string) (out string, err error) {\n\n\tcwd, _ := os.Getwd()\n\tpathArgs := []string{\"-C\", fmt.Sprintf(\"%v/%v\", cwd, TempDir)}\n\targs = append(pathArgs, args...)\n\n\tcmd := exec.Command(\"git\", args...)\n\n\t// verbose mode\n\tif c.useVerbose {\n\t\tlog.Println(cmd.Args)\n\t}\n\n\t// makes sure the TempDir exist\n\tif _, err := os.Stat(fmt.Sprintf(\"%v/%v\", cwd, TempDir)); os.IsNotExist(err) {\n\t\terr := os.Mkdir(TempDir, 0777)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\n\t// starts command\n\tvar output byteSlice\n\toutput, err = cmd.Output()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tout = output.ToString()\n\treturn out, err\n}", "func (git *Git) Exec(subcmd string, args ...string) (string, error) {\n\tb, err := git.Command(subcmd, args...).CombinedOutput()\n\n\t// Chop last newline\n\tl := len(b)\n\tif l > 0 && b[l-1] == '\\n' {\n\t\tb = b[:l-1]\n\t}\n\n\t// Make output in oneline in error cases\n\tif err != nil {\n\t\tfor i := range b {\n\t\t\tif b[i] == '\\n' {\n\t\t\t\tb[i] = ' '\n\t\t\t}\n\t\t}\n\t}\n\n\treturn string(b), err\n}", "func Run(args ...string) (string, error) {\n\t// TODO: use exex.CommandContext here and refactor.\n\t/* #nosec */\n\tcmd := exec.Command(\"git\", args...)\n\tlog.WithField(\"args\", args).Debug(\"running git\")\n\tbts, err := cmd.CombinedOutput()\n\tlog.WithField(\"output\", string(bts)).\n\t\tDebug(\"git result\")\n\tif err != nil {\n\t\treturn \"\", errors.New(string(bts))\n\t}\n\treturn string(bts), nil\n}", "func commit(msg string) {\n\tgit := exec.Command(\"git\", \"commit\", \"-m\", msg)\n\tgitOut, err := git.Output()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfmt.Println(string(gitOut))\n}", "func (r *Repo) Git(cmd string, args ...string) (res *exec.Cmd, out, err *bytes.Buffer) {\n\tvar path string\n\tif r.WorkDir == \"\" {\n\t\tpath = r.GitDir\n\t} else {\n\t\tpath = r.WorkDir\n\t}\n\tres, out, err = Git(cmd, args...)\n\tres.Dir = path\n\treturn\n}", "func (g *GitUtil) ExecCommand(args ...string) (string, error) {\n\tcmd := exec.Command(\"git\", args...)\n\tcmd.Dir = g.TopLevelPath\n\n\t// Capture output\n\tvar out bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = &stderr\n\n\t// Run\n\terr := cmd.Run()\n\n\tif err != nil {\n\t\tlog.Debugln(stderr.String())\n\t}\n\n\treturn string(bytes.TrimSpace(out.Bytes())), err\n}", "func (gc *GitCommand) Output() (string, error) {\n\targs := append(gc.reqArgs(), gc.addlArgs...)\n\tout, err := exec.Command(gc.binName, args...).Output()\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(out), nil\n}", "func GitCommit() string { return gitCommit }", "func withGit(url, commit, folder string, verbose bool) error {\n\tvar out bytes.Buffer\n\tdata := map[string]string{\n\t\t\"workDir\": folder,\n\t\t\"repoDir\": RepoDir(url),\n\t\t\"remote\": url,\n\t\t\"ref\": commit,\n\t}\n\tif err := gitTemplate.Execute(&out, data); err != nil {\n\t\treturn errors.Wrap(err, \"unable to create git script\")\n\t}\n\tscript := out.String()\n\tif verbose {\n\t\tfmt.Println(script)\n\t}\n\n\tcmd := exec.Command(\"bash\", \"-c\", script)\n\tif verbose {\n\t\tcmdStdout, err := cmd.StdoutPipe()\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"unable to create StdOut pipe for bash\")\n\t\t}\n\t\tstdoutScanner := bufio.NewScanner(cmdStdout)\n\t\tgo func() {\n\t\t\tfor stdoutScanner.Scan() {\n\t\t\t\tfmt.Println(\"bash: \" + stdoutScanner.Text())\n\t\t\t}\n\t\t}()\n\n\t\tcmdStderr, err := cmd.StderrPipe()\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"unable to create StdErr pipe for bash\")\n\t\t}\n\t\tstderrScanner := bufio.NewScanner(cmdStderr)\n\t\tgo func() {\n\t\t\tfor stderrScanner.Scan() {\n\t\t\t\tfmt.Println(\"bash: \" + stderrScanner.Text())\n\t\t\t}\n\t\t}()\n\t}\n\terr := cmd.Start()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"unable to start cloning with git\")\n\t}\n\terr = cmd.Wait()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"unable to clone with git\")\n\t}\n\treturn nil\n}", "func (cu *CmdUtils) execCmd(cmd string, args []string, workingDir string) (string, error) {\n\ttoExecute := exec.Command(\"git\", args...)\n\ttoExecute.Dir = workingDir\n\tstdoutStderr, err := toExecute.CombinedOutput()\n\tif err != nil {\n\t\treturn string(stdoutStderr), fmt.Errorf(\"unable to execute command %s due to %w, %s\", cmd, err, string(stdoutStderr))\n\t}\n\tlog.Debug().Str(\"output\", string(stdoutStderr)).Msg(\"execution finished\")\n\treturn string(stdoutStderr), nil\n}", "func (gitCommand *GitCommand) Run(wait bool) (io.ReadCloser, error) {\n\tcmd := exec.Command(\"git\", gitCommand.Args...)\n\tstdout, err := cmd.StdoutPipe()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif gitCommand.ProcInput != nil {\n\t\tcmd.Stdin = gitCommand.ProcInput\n\t}\n\n\tif err := cmd.Start(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif wait {\n\t\terr = cmd.Wait()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn stdout, nil\n}", "func (ss *Sources) gitFetch(spec v1.SourceSpec) (string, error) {\n\tp := ss.repoPath(spec)\n\t_, err := os.Stat(p)\n\tif os.IsNotExist(err) {\n\t\t// Clone new repo.\n\t\td, _ := filepath.Split(p)\n\t\terr = os.MkdirAll(d, 0750)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\t_, _, err = exe.Run(ss.Log, &exe.Opt{Dir: d}, \"\", \"git\", \"clone\", urlWithToken(spec.URL, spec.Token))\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\t_, _, err = exe.Run(ss.Log, &exe.Opt{Dir: p}, \"\", \"git\", \"checkout\", spec.Ref)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tss.Log.Info(\"GIT-clone\", \"url\", spec.URL, \"ref\", spec.Ref)\n\t} else {\n\t\t// Pull existing repo content.\n\t\t_, _, err = exe.Run(ss.Log, &exe.Opt{Dir: p}, \"\", \"git\", \"pull\", \"origin\", spec.Ref)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tss.Log.V(2).Info(\"GIT-pull\", \"url\", spec.URL, \"ref\", spec.Ref)\n\t}\n\n\t// Get hash.\n\th, _, err := exe.Run(ss.Log, &exe.Opt{Dir: p}, \"\", \"git\", \"rev-parse\", spec.Ref)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\th = strings.TrimRight(h, \"\\n\\r\")\n\tif len(h) == 0 {\n\t\treturn \"\", fmt.Errorf(\"expected git hash\")\n\t}\n\n\treturn h, nil\n}", "func prepareGitCommand(command string, formating ...interface{}) *exec.Cmd {\n\tif formating != nil {\n\t\tcommand = fmt.Sprintf(command, formating...)\n\t}\n\targs := strings.Split(command, \" \")\n\tcmd := exec.Command(\"git\", args...)\n\tcmd.Stderr = os.Stderr\n\treturn cmd\n}", "func (e *osExecutor) RunGitOperation(commands []string, path string) error {\n\tvar args []string\n\targs = append(args, \"git\")\n\targs = append(args, \"-C\")\n\targs = append(args, path)\n\targs = append(args, commands...)\n\n\tcmd := exec.Command(args[0], args[1:]...)\n\n\treturn runLoudly(cmd, e.verboseMode)\n}", "func Git(argv []string, cmdr cmd.Commander) error {\n\tusage := executable.Render(`\nValid commands for git:\n\ngit:remote Adds git remote of application to repository\ngit:remove Removes git remote of application from repository\n\nUse '{{.Name}} help [command]' to learn more.\n`)\n\n\tswitch argv[0] {\n\tcase \"git:remote\":\n\t\treturn gitRemote(argv, cmdr)\n\tcase \"git:remove\":\n\t\treturn gitRemove(argv, cmdr)\n\tcase \"git\":\n\t\tfmt.Print(usage)\n\t\treturn nil\n\tdefault:\n\t\tPrintUsage(cmdr)\n\t\treturn nil\n\t}\n}", "func (c cBuild) getGitCommit() string {\n\tif gproc.SearchBinary(\"git\") == \"\" {\n\t\treturn \"\"\n\t}\n\tvar (\n\t\tcmd = `git log -1 --format=\"%cd %H\" --date=format:\"%Y-%m-%d %H:%M:%S\"`\n\t\ts, _ = gproc.ShellExec(cmd)\n\t)\n\tmlog.Debug(cmd)\n\tif s != \"\" {\n\t\tif !gstr.Contains(s, \"fatal\") {\n\t\t\treturn gstr.Trim(s)\n\t\t}\n\t}\n\treturn \"\"\n}", "func RunGitCommand(command string, formating ...interface{}) error {\n\tcmd := prepareGitCommand(command, formating...)\n\treturn cmd.Run()\n}", "func MocksForFindGit(ctx context.Context, cmd *exec.Command) error {\n\tif strings.Contains(cmd.Name, \"git\") && len(cmd.Args) == 1 && cmd.Args[0] == \"--version\" {\n\t\t_, err := cmd.CombinedOutput.Write([]byte(\"git version 99.99.1\"))\n\t\treturn err\n\t}\n\treturn nil\n}", "func (repo *TestRepo) GitCommand(t *testing.T, args ...string) *exec.Cmd {\n\tt.Helper()\n\n\tgitArgs := []string{\"-C\", repo.Path}\n\tgitArgs = append(gitArgs, args...)\n\n\t//nolint:gosec // The args all come from the test code.\n\tcmd := exec.Command(\"git\", gitArgs...)\n\tcmd.Env = CleanGitEnv()\n\treturn cmd\n}", "func gitVersion() string {\n\tcmd := exec.Command(\"git\", \"rev-list\", \"--max-count=1\", \"--pretty=format:'%ad-%h'\", \"--date=short\", \"HEAD\")\n\tcmd.Dir = camRoot\n\tout, err := cmd.Output()\n\tif err != nil {\n\t\tlog.Fatalf(\"Error running git rev-list in %s: %v\", camRoot, err)\n\t}\n\tv := strings.TrimSpace(string(out))\n\tif m := gitVersionRx.FindStringSubmatch(v); m != nil {\n\t\tv = m[0]\n\t} else {\n\t\tpanic(\"Failed to find git version in \" + v)\n\t}\n\tcmd = exec.Command(\"git\", \"diff\", \"--exit-code\")\n\tcmd.Dir = camRoot\n\tif err := cmd.Run(); err != nil {\n\t\tv += \"+\"\n\t}\n\treturn v\n}", "func CompsCommand(runtime *program.SubRuntime) (retval int, err string) {\n\targv := runtime.Argv\n\n\t// parse the repo uri\n\turi, e := repos.NewRepoURI(argv[0])\n\tif e != nil {\n\t\terr = constants.StringUnparsedRepoName\n\t\tretval = constants.ErrorInvalidRepo\n\t\treturn\n\t}\n\n\t// print each component on one line\n\tfor _, comp := range uri.Components() {\n\t\tfmt.Println(comp)\n\t}\n\n\t// and finish\n\treturn\n}", "func gitInfo(dir string) (string, time.Time) {\n\tfname := dir + \"/.git/HEAD\"\n\tbuf, err := os.ReadFile(fname)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tbuf = bytes.TrimSpace(buf)\n\tvar githash string\n\tif len(buf) == 40 {\n\t\tgithash = string(buf[:40])\n\t} else if bytes.HasPrefix(buf, []byte(\"ref: \")) {\n\t\tfname = dir + \"/.git/\" + string(buf[5:])\n\t\tbuf, err = os.ReadFile(fname)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tgithash = string(buf[:40])\n\t} else {\n\t\tlog.Fatalf(\"githash cannot be recovered from %s\", fname)\n\t}\n\tloadTime := time.Now()\n\treturn githash, loadTime\n}", "func commit() string {\n\thash, _ := sh.Output(\"git\", \"rev-parse\", \"--short\", \"HEAD\")\n\treturn hash\n}", "func (cmd Cmd) Call(ctx context.Context) (string, error) {\n\tbuf := &muxedBuffer{}\n\terr := cmd.Capture(buf, buf).Run(ctx)\n\toutput := strings.TrimSpace(buf.buf.String())\n\treturn output, err\n}", "func GitDiff(options ...string) []string {\n\tcommand := &GitCommand{Args: append([]string{\"diff\"}, options...)}\n\n\toutput := command.RunAndGetOutput()\n\treturn strings.Split(string(output), \"\\x00\")\n}", "func ExecCmdAndGetOutput(cmd string) string {\n\tfields, err := shlex.Split(cmd)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tbytes, err := exec.Command(fields[0], fields[1:]...).Output() // waits for it to complete\n\ts := strings.TrimSpace(string(bytes))\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn s\n}", "func (git *Git) Command(subcmd string, args ...string) *exec.Cmd {\n\t// e.g. 'git diff --cached' -> 'git -C /path/to/repo diff --cached'\n\ta := append([]string{\"-C\", git.root, subcmd}, args...)\n\tcmd := exec.Command(git.bin, a...)\n\treturn cmd\n}", "func RunGitCommandSaveOutputToFile(command string, outputFilename string) error {\n\tcmd := prepareGitCommand(command)\n\toutput, err := cmd.Output()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = os.MkdirAll(os.Args[1]+\"/.git\", os.FileMode(0600))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// remove enclosing quotes if they exist\n\tif len(output) > 0 && output[0] == '\"' {\n\t\toutput = output[1:]\n\t}\n\tif len(output) > 0 && output[len(output)-1] == '\"' {\n\t\toutput = output[:len(output)-1]\n\t}\n\terr = ioutil.WriteFile(os.Args[1]+\"/.git/\"+outputFilename, output, os.FileMode(0600))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Status() string {\n\treturn run.Capture(\"git status\")\n}", "func lisp_command_output(command string) string {\n\tcmd := exec.Command(command)\n\tout, err := cmd.CombinedOutput()\n\tif (err != nil) {\n\t\treturn(\"\")\n\t}\n\toutput := string(out)\n\treturn(output[0:len(output)-1])\n}", "func parseGitCommand(sshOriginalCommand string) (command string, repopath string, err error) {\n\tr, err := regexp.Compile(`git-([a-z-]+) '/?([\\w-+@][\\w-+.@]*/)?([\\w-]+)\\.git'`)\n\tif err != nil {\n\t\treturn \"\", \"\", errors.New(\"parseGitCommand(): could not compile regex\")\n\t}\n\n\tm := r.FindStringSubmatch(sshOriginalCommand)\n\tif len(m) != 4 {\n\t\treturn \"\", \"\", errors.New(\"parseGitCommand(): Invalid GIT command\")\n\t}\n\n\treturn m[1], m[2] + m[3], nil\n}", "func GitDiff(path string) string {\n\targs := []string{\"rev-parse\", \"--abbrev-ref\", \"HEAD\"}\n\tbranch := strings.TrimSuffix(string(common.GitRun(path, args, true)), \"\\n\")\n\n\targs = []string{\"rev-parse\", \"--abbrev-ref\", \"@{upstream}\"}\n\tupstream := strings.TrimSuffix(string(common.GitRun(path, args, false)), \"\\n\")\n\tif upstream == \"\" {\n\t\tupstream = fmt.Sprintf(\"origin/%v\", branch)\n\t}\n\n\targs = []string{\"rev-list\", \"--left-right\", branch + \"...\" + upstream}\n\toutput := common.GitRun(path, args, false)\n\n\tresult := \"\"\n\tif len(output) == 0 {\n\t\tresult = \"✓\"\n\t} else {\n\t\tbehind := 0\n\t\tahead := 0\n\t\tfor _, rev := range strings.Split(string(output), \"\\n\") {\n\t\t\tif rev == \"\" {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\trev = string(rev[0])\n\t\t\tif rev == \">\" {\n\t\t\t\tbehind++\n\t\t\t}\n\t\t\tif rev == \"<\" {\n\t\t\t\tahead++\n\t\t\t}\n\t\t}\n\n\t\tif ahead > 0 {\n\t\t\tresult = fmt.Sprintf(\"▲%d\", ahead)\n\t\t}\n\t\tif behind > 0 {\n\t\t\tresult = fmt.Sprintf(\"%v▼%d\", result, behind)\n\t\t}\n\t}\n\n\treturn result\n}", "func (s *GitTestHelper) RunGit(args ...string) bool {\n\treturn s.Exec(s.Git(), args...)\n}", "func (git *Git) Run() error {\n\tcmd := exec.Command(\"git\", \"clone\", \"--no-checkout\", git.URL, git.SourcePath)\n\n\t_, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"git clone --no-checkout %s %s. %s\", git.URL, git.SourcePath, err)\n\t}\n\n\tcmd = exec.Command(\"git\", \"remote\", \"add\", \"composer\", git.URL)\n\tcmd.Dir = git.SourcePath\n\n\t_, err = cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"git remote add composer %s. %s\", git.URL, err)\n\t}\n\n\tcmd = exec.Command(\"git\", \"fecth\", \"composer\")\n\tcmd.Dir = git.SourcePath\n\tcmd.CombinedOutput()\n\n\tcmd = exec.Command(\"git\", \"checkout\", \"-b\", git.Version, \"composer/\"+git.Version)\n\tcmd.Dir = git.SourcePath\n\tcmd.CombinedOutput()\n\n\tcmd = exec.Command(\"git\", \"reset\", \"--hard\", git.Source[\"reference\"])\n\tcmd.Dir = git.SourcePath\n\tcmd.CombinedOutput()\n\n\t_, err = os.Stat(git.SourcePath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tgit.PathExist = true\n\treturn nil\n}", "func run(command string, args ...string) []byte {\n\n\toutput, _ := exec.Command(command, args...).CombinedOutput()\n\treturn output\n}", "func gitPush(c *cli.Context) error {\n\t// inputs\n\tremote, branch := c.Args().Get(0), c.Args().Get(1)\n\texpected(remote, branch)\n\n\t// execute\n\terr := exec.Command(\"git\", \"push\", remote, branch).Run()\n\tcheckErr(err)\n\n\t// outputs\n\tfmt.Println(\"remote:\" + remote)\n\tfmt.Println(\"branch:\" + branch)\n\treturn nil\n}", "func (c *CmdGitMdget) Run() error {\n\tcli, err := GetGitClient(c.G())\n\tctx := context.Background()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar res []keybase1.GitRepoResult\n\tif len(c.folder) > 0 {\n\t\tfolder, err := ParseTLF(c.folder)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tres, err = cli.GetGitMetadata(ctx, folder)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tres, err = cli.GetAllGitMetadata(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tjsonStr, err := json.MarshalIndent(res, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Println(string(jsonStr))\n\treturn nil\n}", "func run(command string) string {\r\n\tcmd := exec.Command(command)\r\n\tout, err := cmd.CombinedOutput()\r\n\tif err != nil {\r\n\t\treturn \"error\"\r\n\t}\r\n\treturn string(out)\r\n}", "func run(command string, args ...string) []byte {\n\toutput, err := exec.Command(command, args...).Output()\n\tcheck(err)\n\treturn output\n}", "func (j *DSGit) GetGitOps(ctx *Ctx, thrN int) (ch chan error, err error) {\n\tworker := func(c chan error, url string) (e error) {\n\t\tdefer func() {\n\t\t\tif c != nil {\n\t\t\t\tc <- e\n\t\t\t}\n\t\t}()\n\t\tvar (\n\t\t\tsout string\n\t\t\tserr string\n\t\t)\n\t\tcmdLine := []string{GitOpsCommand, url}\n\t\tvar env map[string]string\n\t\tif GitOpsNoCleanup {\n\t\t\tenv = map[string]string{\"SKIP_CLEANUP\": \"1\"}\n\t\t}\n\t\tsout, serr, e = ExecCommand(ctx, cmdLine, \"\", env)\n\t\tif e != nil {\n\t\t\tif GitOpsFailureFatal {\n\t\t\t\tPrintf(\"error executing %v: %v\\n%s\\n%s\\n\", cmdLine, e, sout, serr)\n\t\t\t} else {\n\t\t\t\tPrintf(\"WARNING: error executing %v: %v\\n%s\\n%s\\n\", cmdLine, e, sout, serr)\n\t\t\t\te = nil\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\ttype resultType struct {\n\t\t\tLoc int `json:\"loc\"`\n\t\t\tPls []RawPLS `json:\"pls\"`\n\t\t}\n\t\tvar data resultType\n\t\te = jsoniter.Unmarshal([]byte(sout), &data)\n\t\tif e != nil {\n\t\t\tif GitOpsFailureFatal {\n\t\t\t\tPrintf(\"error unmarshaling from %v\\n\", sout)\n\t\t\t} else {\n\t\t\t\tPrintf(\"WARNING: error unmarshaling from %v\\n\", sout)\n\t\t\t\te = nil\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tj.Loc = data.Loc\n\t\tfor _, f := range data.Pls {\n\t\t\tfiles, _ := strconv.Atoi(f.Files)\n\t\t\tblank, _ := strconv.Atoi(f.Blank)\n\t\t\tcomment, _ := strconv.Atoi(f.Comment)\n\t\t\tcode, _ := strconv.Atoi(f.Code)\n\t\t\tj.Pls = append(\n\t\t\t\tj.Pls,\n\t\t\t\tPLS{\n\t\t\t\t\tLanguage: f.Language,\n\t\t\t\t\tFiles: files,\n\t\t\t\t\tBlank: blank,\n\t\t\t\t\tComment: comment,\n\t\t\t\t\tCode: code,\n\t\t\t\t},\n\t\t\t)\n\t\t}\n\t\treturn\n\t}\n\tif thrN <= 1 {\n\t\treturn nil, worker(nil, j.URL)\n\t}\n\tch = make(chan error)\n\tgo func() { _ = worker(ch, j.URL) }()\n\treturn ch, nil\n}", "func runCommandOutput(dir, command string, args ...string) ([]byte, error) {\n\tcmd := exec.Command(command, args...)\n\tcmd.Dir = dir\n\n\toutput, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, fmt.Sprintf(\"error running '%s':\\n%s\\n\", strings.Join(append([]string{command}, args...), \" \"), output))\n\t}\n\n\treturn output, nil\n}", "func TestGitCommandGetBranchGraph(t *testing.T) {\n\tgitCmd := NewDummyGitCommand()\n\tgitCmd.OSCommand.Command = func(cmd string, args ...string) *exec.Cmd {\n\t\tassert.EqualValues(t, \"git\", cmd)\n\t\tassert.EqualValues(t, []string{\"log\", \"--graph\", \"--color=always\", \"--abbrev-commit\", \"--decorate\", \"--date=relative\", \"--pretty=medium\", \"test\", \"--\"}, args)\n\t\treturn secureexec.Command(\"echo\")\n\t}\n\t_, err := gitCmd.GetBranchGraph(\"test\")\n\tassert.NoError(t, err)\n}", "func (c Command) Run() error {\n\tif c.fs != nil {\n\t\tif err := c.fs.Parse(os.Args[1:]); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif *c.version {\n\t\tfmt.Println(c.b.Version(\"changelog\"))\n\t\treturn nil\n\t}\n\n\t// history is always written to stdout\n\tif *c.history {\n\t\t*c.toStdOut = true\n\t}\n\n\tl := flash.New(flash.WithDebug(*c.debug))\n\n\tgitCmd, err := git.New(l)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttoplevelDir, err := gitCmd.TopLevelDir()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// resolv changelog file path\n\tabsChangelogPath := filepath.Join(toplevelDir, *c.file)\n\tc.file = &absChangelogPath\n\n\tgitCmd.Noop = c.noop\n\n\tif err := c.validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif *c.initConfig {\n\t\treturn c.runWriteConfig(l)\n\t}\n\n\tif !gitCmd.IsRepo() {\n\t\treturn errors.New(\"current folder is not a git repository\")\n\t}\n\n\tcfg, err := config.Load(\".\")\n\tif err != nil {\n\t\tif err != config.ErrEmpty && err != config.ErrNotFound {\n\t\t\treturn err\n\t\t}\n\n\t\tl.Debugw(\"no changelog config file found - using default config\", \"path\", filepath.Join(\".\", config.FileName))\n\n\t\tcfg = &config.Default\n\t}\n\n\tvar dst io.Writer = os.Stdout\n\n\tif !*c.toStdOut {\n\t\tf, err := os.OpenFile(*c.file, os.O_CREATE|os.O_RDWR, 0600)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer f.Close() // nolint: gosec\n\t\tdst = f\n\t}\n\n\tif *c.history {\n\t\treturn c.runHistory(dst, l, *cfg, gitCmd)\n\t}\n\n\thasTags, err := gitCmd.HasTags()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !hasTags {\n\t\treturn c.runInit(dst, l, *cfg, gitCmd)\n\t}\n\n\treturn c.runRelease(dst, l, *cfg, gitCmd)\n}", "func GoCmd(config GoConfig, cwd string, command []string) ([]byte, error) {\n\ttarget := fmt.Sprintf(\"%s/%s\", config.GOOS, config.GOARCH)\n\tif _, ok := validCompilerTargets[target]; !ok {\n\t\treturn nil, fmt.Errorf(fmt.Sprintf(\"Invalid compiler target: %s\", target))\n\t}\n\tgoBinPath := path.Join(config.GOROOT, \"bin\", \"go\")\n\tcmd := exec.Command(goBinPath, command...)\n\tcmd.Dir = cwd\n\tcmd.Env = []string{\n\t\t\"CGO_ENABLED=0\",\n\t\tfmt.Sprintf(\"GOOS=%s\", config.GOOS),\n\t\tfmt.Sprintf(\"GOARCH=%s\", config.GOARCH),\n\t\tfmt.Sprintf(\"GOROOT=%s\", config.GOROOT),\n\t\tfmt.Sprintf(\"GOPATH=%s\", config.GOPATH),\n\t\tfmt.Sprintf(\"PATH=%s/bin\", config.GOROOT),\n\t}\n\tvar stdout bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &stdout\n\tcmd.Stderr = &stderr\n\n\tlog.Printf(\"go cmd: '%v'\", cmd)\n\terr := cmd.Run()\n\tif err != nil {\n\t\tlog.Printf(\"--- stdout ---\\n%s\\n\", stdout.String())\n\t\tlog.Printf(\"--- stderr ---\\n%s\\n\", stderr.String())\n\t\tlog.Print(err)\n\t}\n\n\treturn stdout.Bytes(), err\n}", "func (m *MockIExec) DoGit(dir string, args ...string) (string, error) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{dir}\n\tfor _, a := range args {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"DoGit\", varargs...)\n\tret0, _ := ret[0].(string)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func parseGitCommand(sshcmd string) (command, name string, err error) {\n\t// The following regex validates the git command, which is in the form:\n\t// <git-command> [<namespace>/]<name>\n\t// with namespace being optional. If a namespace is used, we validate it\n\t// according to the following:\n\t// - a namespace is optional\n\t// - a namespace contains only alphanumerics, underlines, @´s, -´s, +´s\n\t// and periods but it does not start with a period (.)\n\t// - one and exactly one slash (/) separates namespace and the actual name\n\tr, err := regexp.Compile(`(git-[a-z-]+) '/?([\\w-+@][\\w-+.@]*/)?([\\w-]+)\\.git'`)\n\t// r, err := regexp.Compile(`git-(upload|receive)-pack '/?([\\w-+@][\\w-+.@]*/)?([\\w-]+)\\.git'`)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tm := r.FindStringSubmatch(sshcmd)\n\tif len(m) != 4 {\n\t\treturn \"\", \"\", errors.New(\"You've tried to execute some weird command, I'm deliberately denying you to do that, get over it.\")\n\t}\n\treturn m[1], m[2] + m[3], nil\n}", "func executeCommand(config *GitXargsConfig, repositoryDir string, repo *github.Repository, worktree *git.Worktree) error {\n\n\tlogger := logging.GetLogger(\"git-xargs\")\n\n\tif len(config.Args) < 1 {\n\t\treturn errors.WithStackTrace(NoCommandSuppliedErr{})\n\t}\n\n\tcmdArgs := config.Args\n\n\tcmd := exec.Command(cmdArgs[0], cmdArgs[1:]...)\n\tcmd.Dir = repositoryDir\n\n\tlogger.WithFields(logrus.Fields{\n\t\t\"Repo\": repo.GetName(),\n\t\t\"Directory\": repositoryDir,\n\t\t\"Command\": config.Args,\n\t}).Debug(\"Executing command against local clone of repo...\")\n\n\tstdoutStdErr, err := cmd.CombinedOutput()\n\n\tif err != nil {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Error\": err,\n\t\t}).Debug(\"Error getting output of command execution\")\n\t\t// Track the command error against the repo\n\t\tconfig.Stats.TrackSingle(CommandErrorOccurredDuringExecution, repo)\n\t\treturn errors.WithStackTrace(err)\n\t}\n\n\tlogger.WithFields(logrus.Fields{\n\t\t\"CombinedOutput\": string(stdoutStdErr),\n\t}).Debug(\"Received output of command run\")\n\n\tstatus, statusErr := worktree.Status()\n\n\tif statusErr != nil {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Error\": statusErr,\n\t\t\t\"Repo\": repo.GetName(),\n\t\t\t\"Dir\": repositoryDir,\n\t\t}).Debug(\"Error looking up worktree status\")\n\n\t\t// Track the status check failure\n\t\tconfig.Stats.TrackSingle(WorktreeStatusCheckFailedCommand, repo)\n\t\treturn errors.WithStackTrace(statusErr)\n\t}\n\n\t// If the supplied command resulted in any changes, we need to stage, add and commit them\n\tif !status.IsClean() {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Repo\": repo.GetName(),\n\t\t}).Debug(\"Local repository worktree no longer clean, will stage and add new files and commit changes\")\n\n\t\t// Track the fact that worktree changes were made following execution\n\t\tconfig.Stats.TrackSingle(WorktreeStatusDirty, repo)\n\n\t\tfor filepath := range status {\n\t\t\tif status.IsUntracked(filepath) {\n\t\t\t\tfmt.Printf(\"Found untracked file %s. Adding to stage\", filepath)\n\t\t\t\t_, addErr := worktree.Add(filepath)\n\t\t\t\tif addErr != nil {\n\t\t\t\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\t\t\t\"Error\": addErr,\n\t\t\t\t\t\t\"Filepath\": filepath,\n\t\t\t\t\t}).Debug(\"Error adding file to git stage\")\n\t\t\t\t\t// Track the file staging failure\n\t\t\t\t\tconfig.Stats.TrackSingle(WorktreeAddFileFailed, repo)\n\t\t\t\t\treturn errors.WithStackTrace(addErr)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t} else {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Repo\": repo.GetName(),\n\t\t}).Debug(\"Local repository status is clean - nothing to stage or commit\")\n\n\t\t// Track the fact that repo had no file changes post command execution\n\t\tconfig.Stats.TrackSingle(WorktreeStatusClean, repo)\n\t}\n\n\treturn nil\n}", "func TestGitCommandGetCommitDifferences(t *testing.T) {\n\ttype scenario struct {\n\t\ttestName string\n\t\tcommand func(string, ...string) *exec.Cmd\n\t\ttest func(string, string)\n\t}\n\n\tscenarios := []scenario{\n\t\t{\n\t\t\t\"Can't retrieve pushable count\",\n\t\t\tfunc(string, ...string) *exec.Cmd {\n\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t},\n\t\t\tfunc(pushableCount string, pullableCount string) {\n\t\t\t\tassert.EqualValues(t, \"?\", pushableCount)\n\t\t\t\tassert.EqualValues(t, \"?\", pullableCount)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Can't retrieve pullable count\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tif args[1] == \"HEAD..@{u}\" {\n\t\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t\t}\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tfunc(pushableCount string, pullableCount string) {\n\t\t\t\tassert.EqualValues(t, \"?\", pushableCount)\n\t\t\t\tassert.EqualValues(t, \"?\", pullableCount)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Retrieve pullable and pushable count\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tif args[1] == \"HEAD..@{u}\" {\n\t\t\t\t\treturn secureexec.Command(\"echo\", \"10\")\n\t\t\t\t}\n\n\t\t\t\treturn secureexec.Command(\"echo\", \"11\")\n\t\t\t},\n\t\t\tfunc(pushableCount string, pullableCount string) {\n\t\t\t\tassert.EqualValues(t, \"11\", pushableCount)\n\t\t\t\tassert.EqualValues(t, \"10\", pullableCount)\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, s := range scenarios {\n\t\tt.Run(s.testName, func(t *testing.T) {\n\t\t\tgitCmd := NewDummyGitCommand()\n\t\t\tgitCmd.OSCommand.Command = s.command\n\t\t\ts.test(gitCmd.GetCommitDifferences(\"HEAD\", \"@{u}\"))\n\t\t})\n\t}\n}", "func (o *RunOptions) Git() gits.Gitter {\n\tif o.Gitter == nil {\n\t\to.Gitter = gits.NewGitCLI()\n\t}\n\treturn o.Gitter\n}", "func TestGitCommandCheckout(t *testing.T) {\n\ttype scenario struct {\n\t\ttestName string\n\t\tcommand func(string, ...string) *exec.Cmd\n\t\ttest func(error)\n\t\tforce bool\n\t}\n\n\tscenarios := []scenario{\n\t\t{\n\t\t\t\"Checkout\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"checkout\", \"test\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t\tfalse,\n\t\t},\n\t\t{\n\t\t\t\"Checkout forced\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"checkout\", \"--force\", \"test\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t\ttrue,\n\t\t},\n\t}\n\n\tfor _, s := range scenarios {\n\t\tt.Run(s.testName, func(t *testing.T) {\n\t\t\tgitCmd := NewDummyGitCommand()\n\t\t\tgitCmd.OSCommand.Command = s.command\n\t\t\ts.test(gitCmd.Checkout(\"test\", CheckoutOptions{Force: s.force}))\n\t\t})\n\t}\n}", "func (h *stiGit) GetInfo(repo string) *SourceInfo {\n\tgit := func(arg ...string) string {\n\t\tcommand := exec.Command(\"git\", arg...)\n\t\tcommand.Dir = repo\n\t\tout, err := command.CombinedOutput()\n\t\tif err != nil {\n\t\t\tlog.V(1).Infof(\"Error executing 'git %#v': %s (%v)\", arg, out, err)\n\t\t\treturn \"\"\n\t\t}\n\t\treturn strings.TrimSpace(string(out))\n\t}\n\treturn &SourceInfo{\n\t\tLocation: git(\"config\", \"--get\", \"remote.origin.url\"),\n\t\tRef: git(\"rev-parse\", \"--abbrev-ref\", \"HEAD\"),\n\t\tCommitID: git(\"rev-parse\", \"--verify\", \"HEAD\"),\n\t\tAuthorName: git(\"--no-pager\", \"show\", \"-s\", \"--format=%an\", \"HEAD\"),\n\t\tAuthorEmail: git(\"--no-pager\", \"show\", \"-s\", \"--format=%ae\", \"HEAD\"),\n\t\tCommitterName: git(\"--no-pager\", \"show\", \"-s\", \"--format=%cn\", \"HEAD\"),\n\t\tCommitterEmail: git(\"--no-pager\", \"show\", \"-s\", \"--format=%ce\", \"HEAD\"),\n\t\tDate: git(\"--no-pager\", \"show\", \"-s\", \"--format=%ad\", \"HEAD\"),\n\t\tMessage: git(\"--no-pager\", \"show\", \"-s\", \"--format=%<(80,trunc)%s\", \"HEAD\"),\n\t}\n}", "func (h *stiGit) Checkout(repo, ref string) error {\n\topts := cmd.CommandOpts{\n\t\tStdout: os.Stdout,\n\t\tStderr: os.Stderr,\n\t\tDir: repo,\n\t}\n\tif log.Is(1) {\n\t\treturn h.RunWithOptions(opts, \"git\", \"checkout\", \"--quiet\", ref)\n\t}\n\treturn h.RunWithOptions(opts, \"git\", \"checkout\", ref)\n}", "func fetch(p provision.Provisioner, app provision.App) ([]byte, error) {\n\tvar buf bytes.Buffer\n\tpath, err := repository.GetPath()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"tsuru is misconfigured: %s\", err)\n\t}\n\tcmd := fmt.Sprintf(\"cd %s && git fetch origin\", path)\n\terr = p.ExecuteCommand(&buf, &buf, app, cmd)\n\tb := buf.Bytes()\n\tlog.Debugf(`\"git fetch\" output: %s`, b)\n\treturn b, err\n}", "func gitPath() (string, error) {\n\tgitOnce.Do(func() {\n\t\tpath, err := exec.LookPath(\"git\")\n\t\tif err != nil {\n\t\t\tgitOnce.err = err\n\t\t\treturn\n\t\t}\n\t\tif runtime.GOOS == \"plan9\" {\n\t\t\tgitOnce.err = errors.New(\"plan9 git does not support the full git command line\")\n\t\t}\n\t\tgitOnce.path = path\n\t})\n\n\treturn gitOnce.path, gitOnce.err\n}", "func GetGitCommit() string {\n\tif Commit != \"\" {\n\t\treturn Commit\n\t}\n\n\tcmd := exec.Command(\"git\", \"rev-parse\", \"--verify\", \"HEAD\")\n\toutput, err := cmd.Output()\n\tif err != nil {\n\t\tklog.Errorf(\"failed to get git commit: %s\", err.Error())\n\t\treturn \"\"\n\t}\n\n\treturn strings.TrimSpace(string(output))\n}", "func GetCommit() (string, string, error) {\n\n\t// Get the current commit hash.\n\tcmd := exec.Command(\"git\", \"rev-parse\", \"HEAD\")\n\tout, _ := cmd.Output()\n\tcommit := strings.Split(string(out), \"\\n\")[0]\n\n\t// Get the git status.\n\tcmd = exec.Command(\"git\", \"status\", \"-s\", \"-uno\")\n\tout, _ = cmd.Output()\n\tlines := strings.Split(string(out), \"\\n\")\n\tstatus := \"\"\n\tif len(lines[0]) > 0 {\n\t\tstatus = \"*\"\n\t}\n\n\treturn commit, status, nil\n\n}", "func getRemote(project string) string {\n\tsession := sh.NewSession()\n\tsession.SetDir(project)\n\tremote, err := session.Command(\"git\", \"remote\").Command(\"head\", \"-n\", \"1\").Output()\n\t// pipe makes this code unreachable\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn strings.TrimSpace(string(remote))\n}", "func checkout(p provision.Provisioner, app provision.App, version string) ([]byte, error) {\n\tvar buf bytes.Buffer\n\tpath, err := repository.GetPath()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"tsuru is misconfigured: %s\", err)\n\t}\n\tcmd := fmt.Sprintf(\"cd %s && git checkout %s\", path, version)\n\tif err := p.ExecuteCommand(&buf, &buf, app, cmd); err != nil {\n\t\treturn buf.Bytes(), err\n\t}\n\treturn nil, nil\n}", "func (s *RepositoriesService) GetCommitRaw(ctx context.Context, owner string, repo string, sha string, opt RawOptions) (string, *Response, error) {\n\tu := fmt.Sprintf(\"repos/%v/%v/commits/%v\", owner, repo, sha)\n\treq, err := s.client.NewRequest(\"GET\", u, nil)\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\n\tswitch opt.Type {\n\tcase Diff:\n\t\treq.Header.Set(\"Accept\", mediaTypeV3Diff)\n\tcase Patch:\n\t\treq.Header.Set(\"Accept\", mediaTypeV3Patch)\n\tdefault:\n\t\treturn \"\", nil, fmt.Errorf(\"unsupported raw type %d\", opt.Type)\n\t}\n\n\tvar buf bytes.Buffer\n\tresp, err := s.client.Do(ctx, req, &buf)\n\tif err != nil {\n\t\treturn \"\", resp, err\n\t}\n\n\treturn buf.String(), resp, nil\n}", "func gitState(repoExisted bool, out string) repoState {\n\tif !repoExisted {\n\t\treturn stateNew\n\t}\n\tif lines := strings.Split(out, \"\\n\"); len(lines) > 2 {\n\t\treturn stateChanged\n\t}\n\treturn stateUnchanged\n}", "func Version() (string, error) {\n\tvar version string\n\n\tcmd := exec.Command(\"git\", \"rev-parse\", \"HEAD\")\n\tcmdReader, err := cmd.StdoutPipe()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tscanner := bufio.NewScanner(cmdReader)\n\tgo func() {\n\t\tfor scanner.Scan() {\n\t\t\tversion = scanner.Text()\n\t\t}\n\t}()\n\n\terr = cmd.Start()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\terr = cmd.Wait()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn version, nil\n}", "func GitVersion() string { return gitVersion }", "func GitCommit(tb testing.TB, msg string) {\n\ttb.Helper()\n\tout, err := fakeGit(\"commit\", \"--allow-empty\", \"-m\", msg)\n\trequire.NoError(tb, err)\n\trequire.Contains(tb, out, \"main\", msg)\n}", "func GetVersionCmd(version string, commit string) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: VersionCmdName,\n\t\tShort: VersionCmdShort,\n\t\tRun: func(*cobra.Command, []string) {\n\t\t\tif commit != \"\" {\n\t\t\t\tfmt.Println(fmt.Sprintf(\"%s (%s)\", version, commit))\n\t\t\t} else {\n\t\t\t\tfmt.Println(version)\n\t\t\t}\n\t\t},\n\t}\n}", "func (c Commit) String() string {\n\treturn c.Sha\n}", "func GitGetCommitHash() error {\n\tout, err := exec.Command(\n\t\t\"git\",\n\t\t\"rev-parse\",\n\t\t\"HEAD\",\n\t).Output()\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"GitGetCommitHash: %s\", err)\n\t}\n\n\tCommitHash = strings.TrimRight(string(out), \"\\n\")\n\treturn nil\n}", "func Version(ctx context.Context, git string) (int, int, error) {\n\tout, err := exec.RunCwd(ctx, \".\", git, \"--version\")\n\tif err != nil {\n\t\treturn -1, -1, err\n\t}\n\tm := gitVersionRegex.FindStringSubmatch(out)\n\tif m == nil {\n\t\treturn -1, -1, fmt.Errorf(\"Failed to parse the git version from output: %q\", out)\n\t}\n\tif len(m) != 3 {\n\t\treturn -1, -1, fmt.Errorf(\"Failed to parse the git version from output: %q\", out)\n\t}\n\tmajor, err := strconv.Atoi(m[1])\n\tif err != nil {\n\t\treturn -1, -1, fmt.Errorf(\"Failed to parse the git version from output: %q\", out)\n\t}\n\tminor, err := strconv.Atoi(m[2])\n\tif err != nil {\n\t\treturn -1, -1, fmt.Errorf(\"Failed to parse the git version from output: %q\", out)\n\t}\n\treturn major, minor, nil\n}", "func cmdCheck(s string) []byte {\n\tfmt.Printf(\" -> %s\\n\", s)\n\tfields := strings.Fields(s)\n\tif len(fields) < 1 {\n\t\tlog.Fatal(`Expected \"command <arguments>\"`)\n\t}\n\tb, err := exec.Command(fields[0], fields[1:]...).Output()\n\tif exiterr, ok := err.(*exec.ExitError); ok {\n\t\tlog.Fatalf(\"%s; stderr: %s\\n\", err, string(exiterr.Stderr))\n\t} else if err != nil {\n\t\tlog.Fatal(\"exec.Command\", err)\n\t}\n\treturn b\n}", "func executeCommandCapturingStdout(args ...string) (string, string) {\n\n\t// We substitute our own pipe for stdout to collect the terminal output\n\t// but must be careful to always restore stadt and close the pripe files.\n\toriginalStdout := os.Stdout\n\treadFile, writeFile, err := os.Pipe()\n\tif err != nil {\n\t\tfmt.Printf(\"Could not capture stdout: %s\", err.Error())\n\t\tos.Exit(1)\n\t}\n\n\t// Be careful to both put stdout back in its proper place, and restore any\n\t// tricks that we played on our child packages to get them to cooperate in our testing.\n\tdefer func() {\n\n\t\t// Restore stdout piping\n\t\tos.Stdout = originalStdout\n\t\twriteFile.Close()\n\t\treadFile.Close()\n\t}()\n\n\t// Set our own pipe as stdout\n\tos.Stdout = writeFile\n\n\t// Run the command with a random token value that does not matter because we won't actually\n\t// be calling AWS and so it won't be able to object\n\toutput := executeCommand(args...)\n\n\t// Restore stdout and close the write end of the pipe so that we can collect the output\n\tos.Stdout = originalStdout\n\twriteFile.Close()\n\n\t// Gather the output into a byte buffer\n\toutputBytes, err := ioutil.ReadAll(readFile)\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to read pipe for stdout: : %s\", err.Error())\n\t\tos.Exit(1)\n\t}\n\n\t// Return the executeCommand output and stdout\n\treturn output, string(outputBytes)\n}", "func checkoutCommand() *exec.Cmd {\n\tscript := strings.Join([]string{\n\t\t\"git fetch --prune\",\n\t\t\"git checkout -f $DEPLOY_COMMIT\",\n\t\t\"git reset --hard $DEPLOY_COMMIT\",\n\t\t\"git clean -fdx\",\n\t\t\"git submodule sync\",\n\t\t\"git submodule init\",\n\t\t\"git submodule update --recursive\",\n\t}, \" && \")\n\n\treturn unbuffered.Command(\"bash -x -c '\" + script + \"'\")\n}", "func executeWithOutput(cmd string, args ...string) string {\n\tout, err := exec.Command(cmd, args...).Output()\n\tif err != nil {\n\t\tlog.Fatalf(\"%s\", err)\n\t}\n\treturn string(out)\n}", "func (g Git) Commit(message string) ([]byte, error) {\n\tcmd := exec.Command(\"git\", \"add\", \".\")\n\tcmd.Dir = g.Work\n\tresult, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tcmd = exec.Command(\"git\", \"commit\", \"-q\", \"-m\", message)\n\tcmd.Dir = g.Work\n\treturn cmd.CombinedOutput()\n}", "func ExecuteShellAndReturnOutput(\n\tcliConfig schema.CliConfiguration,\n\tcommand string,\n\tname string,\n\tdir string,\n\tenv []string,\n\tdryRun bool,\n) (string, error) {\n\tvar b bytes.Buffer\n\n\tu.LogDebug(cliConfig, \"\\nExecuting command:\")\n\tu.LogDebug(cliConfig, command)\n\n\tif dryRun {\n\t\treturn \"\", nil\n\t}\n\n\terr := shellRunner(command, name, dir, env, &b)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn b.String(), nil\n}", "func RunInPath(path string, args ...string) ([]byte, error) {\n\t// is git installed?\n\t_git, _ := Git()\n\tif _git == \"\" {\n\t\treturn nil, MissingGitError\n\t}\n\n\t// if we have a path, attempt to change into it before executing\n\t// the git command\n\tif path != \"\" {\n\t\tvar _dir string\n\n\t\t// do we have a file or a directory?\n\t\t_info, _err := os.Stat(path)\n\t\tif _err != nil {\n\t\t\treturn nil, _err\n\t\t} else if _info.IsDir() {\n\t\t\t_dir = path\n\t\t} else {\n\t\t\t_dir, _ = filepath.Split(path)\n\t\t}\n\n\t\t_cwd, _err := os.Getwd()\n\t\tif _err != nil {\n\t\t\treturn nil, _err\n\t\t}\n\n\t\t// attempt to change into the given path\n\t\t_err = os.Chdir(_dir)\n\t\tif _err != nil {\n\t\t\treturn nil, _err\n\t\t}\n\t\tdefer os.Chdir(_cwd)\n\t}\n\n\t// execute the git command\n\treturn exec.Command(_git, args...).Output()\n}", "func (sess *reconcileStackSession) runCmd(title string, cmd *exec.Cmd) (string, string, error) {\n\t// If not overridden, set the command to run in the working directory.\n\tif cmd.Dir == \"\" {\n\t\tcmd.Dir = sess.workdir\n\t}\n\n\t// If there are extra environment variables, set them.\n\tif sess.extraEnv != nil {\n\t\tif len(cmd.Env) == 0 {\n\t\t\tcmd.Env = os.Environ()\n\t\t}\n\t\tfor k, v := range sess.extraEnv {\n\t\t\tcmd.Env = append(cmd.Env, k+\"=\"+v)\n\t\t}\n\t}\n\n\t// Capture stdout and stderr.\n\tstdoutR, stdoutW := io.Pipe()\n\tstderrR, stderrW := io.Pipe()\n\tcmd.Stdout = stdoutW\n\tcmd.Stderr = stderrW\n\n\t// Start the command asynchronously.\n\terr := cmd.Start()\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\n\t// Kick off some goroutines to stream the output asynchronously. Since Pulumi can take\n\t// a while to run, this helps to debug issues that might be ongoing before a command completes.\n\tvar stdout bytes.Buffer\n\tvar stderr bytes.Buffer\n\tgo func() {\n\t\touts := bufio.NewScanner(stdoutR)\n\t\tfor outs.Scan() {\n\t\t\ttext := outs.Text()\n\t\t\tsess.logger.Info(title, \"Path\", cmd.Path, \"Args\", cmd.Args, \"Stdout\", text)\n\t\t\tstdout.WriteString(text + \"\\n\")\n\t\t}\n\t}()\n\tgo func() {\n\t\terrs := bufio.NewScanner(stderrR)\n\t\tfor errs.Scan() {\n\t\t\ttext := errs.Text()\n\t\t\tsess.logger.Info(title, \"Path\", cmd.Path, \"Args\", cmd.Args, \"Text\", text)\n\t\t\tstderr.WriteString(text + \"\\n\")\n\t\t}\n\t}()\n\n\t// Now wait for the command to finish. No matter what, return everything written to stdout and\n\t// stderr, in addition to the resulting error, if any.\n\terr = cmd.Wait()\n\treturn stdout.String(), stderr.String(), err\n}", "func FindGit(ctx context.Context) (string, int, int, error) {\n\tmtx.Lock()\n\tdefer mtx.Unlock()\n\tif git == \"\" {\n\t\tgitPath, err := osexec.LookPath(\"git\")\n\t\tif err != nil {\n\t\t\treturn \"\", 0, 0, skerr.Wrapf(err, \"Failed to find git\")\n\t\t}\n\t\tmaj, min, err := Version(ctx, gitPath)\n\t\tif err != nil {\n\t\t\treturn \"\", 0, 0, skerr.Wrapf(err, \"Failed to obtain git version\")\n\t\t}\n\t\tsklog.Infof(\"Git is %s; version %d.%d\", gitPath, maj, min)\n\t\tisFromCIPD := IsFromCIPD(gitPath)\n\t\tisFromCIPDVal := 0\n\t\tif isFromCIPD {\n\t\t\tisFromCIPDVal = 1\n\t\t}\n\t\tmetrics2.GetInt64Metric(\"git_from_cipd\").Update(int64(isFromCIPDVal))\n\t\tgit = gitPath\n\t\tgitVersionMajor = maj\n\t\tgitVersionMinor = min\n\t}\n\treturn git, gitVersionMajor, gitVersionMinor, nil\n}", "func main() {\n\n\tfmt.Println(\"Testing Git Access\")\n\n\tfmt.Println(\"Hello World!!\")\n\n}", "func RunCmd(cmd string) (result []byte, err error) {\n\tcmdPip := strings.Split(cmd, \"|\")\n\tif len(cmdPip) < 2 {\n\t\treturn run(cmd).Output()\n\t}\n\n\treturn runPipe(cmdPip)\n}", "func execCmdWithOutput(arg0 string, args ...string) (string, error) {\n\tcmd := exec.Command(arg0, args...)\n\tstdout, err := cmd.StdoutPipe()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif cmd_err := cmd.Start(); cmd_err != nil {\n\t\treturn \"\", cmd_err\n\t}\n\n\toutput, err_out := ioutil.ReadAll(stdout)\n\tif err_out != nil {\n\t\treturn \"\", err_out\n\t}\n\n\treturn string(output), nil\n}", "func gitPath(path string) string {\n\troot := repoRoot()\n\t// git 2.13.0 changed the behavior of --git-path from printing\n\t// a path relative to the repo root to printing a path\n\t// relative to the working directory (issue #19477). Normalize\n\t// both behaviors by running the command from the repo root.\n\tp, err := trimErr(cmdOutputErr(\"git\", \"-C\", root, \"rev-parse\", \"--git-path\", path))\n\tif err != nil {\n\t\t// When --git-path is not available, assume the common case.\n\t\tp = filepath.Join(\".git\", path)\n\t}\n\tif !filepath.IsAbs(p) {\n\t\tp = filepath.Join(root, p)\n\t}\n\treturn p\n}", "func main() {\n\trepoPath := filepath.Join(os.Getenv(\"GOPATH\"), \"src/github.com/libgit2/git2go\")\n\tgitRepo, err := git.OpenRepository(repoPath)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tcommitOid, err := gitRepo.Head()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tblob, _ := gitRepo.LookupBlob(commitOid.Target())\n\tlog.Println(blob)\n\t// commit, err := gitRepo.LookupCommit(commitOid)\n\tcommit, err := gitRepo.LookupCommit(commitOid.Target())\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tcommitTree, err := commit.Tree()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\toptions, err := git.DefaultDiffOptions()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\toptions.IdAbbrev = 40\n\toptions.InterhunkLines = 0\n\toptions.Flags = git.DiffIncludeUntracked\n\tvar parentTree *git.Tree\n\tif commit.ParentCount() > 0 {\n\t\tparentTree, err = commit.Parent(0).Tree()\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}\n\tgitDiff, err := gitRepo.DiffTreeToTree(parentTree, commitTree, &options)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tfindOpts, err := git.DefaultDiffFindOptions()\n\tfindOpts.Flags = git.DiffFindBreakRewrites\n\terr = gitDiff.FindSimilar(&findOpts)\n\n\t// Show all file patch diffs in a commit.\n\tfiles := make([]string, 0)\n\thunks := make([]git.DiffHunk, 0)\n\tlines := make([]git.DiffLine, 0)\n\tpatches := make([]string, 0)\n\terr = gitDiff.ForEach(func(file git.DiffDelta, progress float64) (git.DiffForEachHunkCallback, error) {\n\t\tpatch, err := gitDiff.Patch(len(patches))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tdefer patch.Free()\n\t\tpatchStr, err := patch.String()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpatches = append(patches, patchStr)\n\n\t\tfiles = append(files, file.OldFile.Path)\n\t\treturn func(hunk git.DiffHunk) (git.DiffForEachLineCallback, error) {\n\t\t\thunks = append(hunks, hunk)\n\t\t\treturn func(line git.DiffLine) error {\n\t\t\t\tlines = append(lines, line)\n\t\t\t\treturn nil\n\t\t\t}, nil\n\t\t}, nil\n\t}, git.DiffDetailLines)\n\n\tlog.Println(\"files: \", files, \"\\n\")\n\tlog.Println(\"hunks: \", hunks, \"\\n\")\n\tlog.Println(\"lines: \", lines, \"\\n\")\n\tlog.Println(\"patches: \", patches, \"\\n\")\n}", "func (git Git) getCommitHash() (string, error) {\n\treturn git.exec(\"rev-parse\", \"HEAD\")\n}", "func combCli(cmd string, args ...string) ([]byte, error) {\n\targs = append([]string{cmd}, args...)\n\tfmt.Printf(\"args=%s\\n\", args)\n\treturn exec.Command(\"./comb\", args...).Output()\n}", "func (execImpl *Exec) CommandCombinedOutput(name string, arg ...string) ([]byte, error) {\n\treturn exec.Command(name, arg...).CombinedOutput()\n}", "func runCmd(command string, args ...string) ([]byte, error) {\n\tif *flagVerbose {\n\t\tlog.Printf(\"runCmd: %s %s\", command, strings.Join(args, \" \"))\n\t}\n\n\tout, err := execCommand(command, args...).CombinedOutput()\n\tif err != nil {\n\t\tlog.Printf(\"runCmd: %s\", out)\n\t\treturn nil, err\n\t}\n\n\treturn out, nil\n}", "func Cmd(command string) (r string, e error) {\n\tinputcmd := strings.Split(command, \" \")\n\tcmd := exec.Command(inputcmd[0], inputcmd[1:]...)\n\tstdout, err := cmd.Output()\n\treturn string(stdout), err\n}", "func TestGitCommandCurrentBranchName(t *testing.T) {\n\ttype scenario struct {\n\t\ttestName string\n\t\tcommand func(string, ...string) *exec.Cmd\n\t\ttest func(string, string, error)\n\t}\n\n\tscenarios := []scenario{\n\t\t{\n\t\t\t\"says we are on the master branch if we are\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.Equal(t, \"git\", cmd)\n\t\t\t\treturn secureexec.Command(\"echo\", \"master\")\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"master\", name)\n\t\t\t\tassert.EqualValues(t, \"master\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"falls back to git `git branch --contains` if symbolic-ref fails\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\n\t\t\t\tswitch args[0] {\n\t\t\t\tcase \"symbolic-ref\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"symbolic-ref\", \"--short\", \"HEAD\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t\tcase \"branch\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"branch\", \"--contains\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"echo\", \"* master\")\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"master\", name)\n\t\t\t\tassert.EqualValues(t, \"master\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"handles a detached head\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\n\t\t\t\tswitch args[0] {\n\t\t\t\tcase \"symbolic-ref\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"symbolic-ref\", \"--short\", \"HEAD\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t\tcase \"branch\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"branch\", \"--contains\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"echo\", \"* (HEAD detached at 123abcd)\")\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"123abcd\", name)\n\t\t\t\tassert.EqualValues(t, \"(HEAD detached at 123abcd)\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"bubbles up error if there is one\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.Equal(t, \"git\", cmd)\n\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.Error(t, err)\n\t\t\t\tassert.EqualValues(t, \"\", name)\n\t\t\t\tassert.EqualValues(t, \"\", displayname)\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, s := range scenarios {\n\t\tt.Run(s.testName, func(t *testing.T) {\n\t\t\tgitCmd := NewDummyGitCommand()\n\t\t\tgitCmd.OSCommand.Command = s.command\n\t\t\ts.test(gitCmd.CurrentBranchName())\n\t\t})\n\t}\n}", "func RunCommandAndReturn(host string, command string) (string, error) {\n\tvar output bytes.Buffer\n\tRunCommand(host, command, os.Stdin, &output, os.Stderr)\n\treturn output.String(), nil\n}", "func gitGetTag() (string, error) {\n\tcmd := exec.Command(\"git\", \"describe\", \"--tags\", \"--exact-match\")\n\toutput, err := cmd.CombinedOutput()\n\tlg.dbg(\"git describe --tags --exact-match:\\n%s---\", string(output))\n\tif err != nil {\n\t\tif bytes.HasPrefix(output, []byte(\"fatal: no tag exactly matches\")) {\n\t\t\toutput = nil\n\t\t} else if bytes.HasPrefix(output, []byte(\"fatal: No names found, cannot describe anything\")) {\n\t\t\toutput = nil\n\t\t} else {\n\t\t\treturn \"\", fmt.Errorf(\"git: %v: %s\", err, firstLine(output))\n\t\t}\n\t}\n\ttags := strings.Fields(string(output))\n\tif len(tags) == 0 {\n\t\treturn \"\", nil\n\t}\n\treturn tags[0], nil\n}", "func (r Repository) Commit(msg string, opt *CommitOptions) (Hash, error) {\n\tif opt == nil {\n\t\topt = &CommitOptions{}\n\t}\n\n\targs := []string{\"commit\"}\n\tif opt.Amend {\n\t\targs = append(args, \"--amend\")\n\t} else {\n\t\targs = append(args, \"-m\", msg)\n\t}\n\n\tvar env []string\n\tif opt.AuthorName != \"\" || opt.AuthorEmail != \"\" {\n\t\tenv = []string{\n\t\t\tfmt.Sprintf(\"GIT_AUTHOR_NAME=%v\", opt.AuthorName),\n\t\t\tfmt.Sprintf(\"GIT_AUTHOR_EMAIL=%v\", opt.AuthorEmail),\n\t\t\tfmt.Sprintf(\"GIT_COMMITTER_NAME=%v\", opt.AuthorName),\n\t\t\tfmt.Sprintf(\"GIT_COMMITTER_EMAIL=%v\", opt.AuthorEmail),\n\t\t}\n\t}\n\tif _, err := r.run(env, opt.Timeout, \"commit\", \"-m\", msg); err != nil {\n\t\treturn Hash{}, err\n\t}\n\tout, err := r.run(nil, 0, \"rev-parse\", \"HEAD\")\n\tif err != nil {\n\t\treturn Hash{}, err\n\t}\n\treturn ParseHash(out)\n}", "func (c *Cmd) CombinedOutput() ([]byte, error)", "func RunCmdWithOutput(cmdName string, cmdArgs []string) ([]byte, error) {\n\tcmdLog := fmt.Sprintf(\"%s %s\", cmdName, strings.Join(cmdArgs, \" \"))\n\tout, err := exec.Command(cmdName, cmdArgs...).Output()\n\tif err != nil {\n\t\treturn []byte(\"\"), fmt.Errorf(\"failed execute command; cmd=%s; err=%v\", cmdLog, err)\n\t}\n\treturn out, nil\n}", "func GetGitTag(dir string) (string, error) {\n\treturn runGit(dir, \"describe\", \"--tags\", \"--dirty\", \"--always\")\n}", "func (c *Cmd) CombinedOutput() ([]byte, error) {\n\treturn c.Cmd.CombinedOutput()\n}", "func hash() string {\n\thash, _ := sh.Output(\"git\", \"rev-parse\", \"--short\", \"HEAD\")\n\treturn hash\n}", "func hash() string {\n\thash, _ := sh.Output(\"git\", \"rev-parse\", \"--short\", \"HEAD\")\n\treturn hash\n}" ]
[ "0.74961495", "0.74380195", "0.72340816", "0.7192161", "0.6944512", "0.69108826", "0.68365854", "0.6786665", "0.66299236", "0.6544166", "0.647694", "0.6359478", "0.6296321", "0.62803876", "0.6264853", "0.6264695", "0.6105203", "0.610021", "0.6060656", "0.604493", "0.6029006", "0.597224", "0.5955037", "0.588742", "0.5884492", "0.5858257", "0.5813034", "0.57543725", "0.5729707", "0.57068527", "0.5699893", "0.56935817", "0.56423676", "0.5638367", "0.56210434", "0.56016713", "0.560161", "0.5566007", "0.5555576", "0.55536544", "0.5549999", "0.5545753", "0.55224437", "0.55190986", "0.5512264", "0.5466504", "0.5458705", "0.5441805", "0.54265964", "0.54232967", "0.5422321", "0.54102516", "0.53968817", "0.5395885", "0.5391066", "0.53719604", "0.53500164", "0.5348927", "0.5348899", "0.5338147", "0.5324362", "0.53236127", "0.5315972", "0.53148437", "0.53049815", "0.52767944", "0.52767503", "0.52739495", "0.5260338", "0.52572286", "0.5252883", "0.5252138", "0.52316374", "0.523159", "0.52262414", "0.5226235", "0.5225951", "0.52233607", "0.5223316", "0.52212495", "0.5217669", "0.5206906", "0.52059495", "0.52035964", "0.52009356", "0.51873016", "0.5186354", "0.5184547", "0.5183154", "0.5164148", "0.51601106", "0.51582015", "0.51571757", "0.5139311", "0.51212645", "0.5118382", "0.51171523", "0.5103485", "0.5100131", "0.5100131" ]
0.7203828
3
Commit Do a git commit
func Commit(msg string, errorIfEmpty bool) error { s := Status() if len(s.Added) == 0 { return fmt.Errorf("No files added. Please check.") } if Do("commit", "-m", msg) > 0 { return fmt.Errorf("Unable to commit.") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (h *Host) Commit(msg string) error {\n\th.save()\n\treturn h.maybeGitCommit(h.Serial + \": \" + msg)\n}", "func commit(msg string) {\n\tgit := exec.Command(\"git\", \"commit\", \"-m\", msg)\n\tgitOut, err := git.Output()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfmt.Println(string(gitOut))\n}", "func (machine *Dishwasher) Commit(message string) {\r\n machine.Append(func() (string, error) {\r\n cmd := exec.Command(\"git\", \"commit\")\r\n if len(message) > 0 {\r\n cmd = exec.Command(\"git\", \"commit\", \"-m\", message)\r\n }\r\n cmd.Stdin = os.Stdin\r\n output, oops := cmd.CombinedOutput()\r\n machine.SideEffect(string(output), oops)\r\n return string(output), oops\r\n })\r\n}", "func GitCommit(tb testing.TB, msg string) {\n\ttb.Helper()\n\tout, err := fakeGit(\"commit\", \"--allow-empty\", \"-m\", msg)\n\trequire.NoError(tb, err)\n\trequire.Contains(tb, out, \"main\", msg)\n}", "func (g Git) Commit(message string) ([]byte, error) {\n\tcmd := exec.Command(\"git\", \"add\", \".\")\n\tcmd.Dir = g.Work\n\tresult, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tcmd = exec.Command(\"git\", \"commit\", \"-q\", \"-m\", message)\n\tcmd.Dir = g.Work\n\treturn cmd.CombinedOutput()\n}", "func (s *splicer) commit(msg string, contents map[string]string) error {\n\tfor fname, data := range contents {\n\t\terr := ioutil.WriteFile(s.dir+\"/\"+fname, []byte(data), 0644)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = s.gitCall(\"add\", fname)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn s.gitCall(\"commit\", \"-m\", msg)\n}", "func (d *DeploymentCoreStruct) GitCommit(message string) (_ error) {\n\treturn d.RunInContext(func() (err error) {\n\t\tstatus := git.GetStatus()\n\t\tif status.Ready.CountFiles() > 0 {\n\t\t\tgit.Commit(message, true)\n\t\t}\n\t\treturn\n\t})\n}", "func (s *GitService) commit(w *git.Worktree, r *git.Repository, c *CreateCommit) (*object.Commit, error) {\n\t// Commits the current staging area to the repository, with the new file\n\t// just created. We should provide the object.Signature of Author of the\n\t// gitClient Since version 5.0.1, we can omit the Author signature, being read\n\t// from the git config files.\n\tvar p []plumbing.Hash\n\tif c.Parents != nil && len(c.Parents) > 0 {\n\t\tp = make([]plumbing.Hash, len(c.Parents))\n\t}\n\tif p != nil && len(p) > 0 {\n\t\tfor i, parent := range c.Parents {\n\t\t\tcopy(p[i][:], parent)\n\t\t}\n\t}\n\n\t// calculate time.Time from unix Time\n\tauthorDate := time.Unix(c.Author.Date, 0)\n\tvar committer *object.Signature\n\tif c.Committer != nil {\n\t\tcommitterDate := time.Unix(c.Committer.Date, 0)\n\t\tcommitter = &object.Signature{\n\t\t\tName: c.Committer.Name,\n\t\t\tEmail: c.Committer.Email,\n\t\t\tWhen: committerDate,\n\t\t}\n\t} else {\n\t\tcommitter = &object.Signature{\n\t\t\tName: c.Author.Name,\n\t\t\tEmail: c.Author.Email,\n\t\t\tWhen: authorDate,\n\t\t}\n\t}\n\n\tcommitHash, err := w.Commit(c.Message, &git.CommitOptions{\n\t\tAuthor: &object.Signature{\n\t\t\tName: c.Author.Name,\n\t\t\tEmail: c.Author.Email,\n\t\t\tWhen: authorDate,\n\t\t},\n\t\tCommitter: committer,\n\t\tParents: p,\n\t\tSignKey: c.SignKey,\n\t\tAll: true,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tobj, err := r.CommitObject(commitHash)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn obj, nil\n}", "func commit(words []string, m1 *M1, m2 *M2, trans *Stack) error {\n\tif len(words) != 1 {\n\t\treturn errors.New(\"Invalid COMMIT command. Format: COMMIT\")\n\t}\n\tvar tnew Stack\n\n\t(*trans) = tnew\n\ttrans.Push(*m1, *m2)\n\treturn nil\n}", "func (todo Todo) GitCommit(prefix string) error {\n\t// TODO(#96): there is no way to check that Todo is unreported at compile time\n\tif todo.ID == nil {\n\t\tpanic(fmt.Sprintf(\"Trying to commit an unreported TODO! %v\", todo))\n\t}\n\n\tif err := exec.Command(\"git\", \"add\", todo.Filename).Run(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := exec.Command(\"git\", \"commit\", \"-m\", fmt.Sprintf(\"%s TODO(%s)\", prefix, *todo.ID)).Run(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (git *Git) Commit(msg string) error {\n\tout, err := git.Exec(\"commit\", \"-m\", msg)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"Cannot commit changes to repository at '%s': %s\", git.canonRoot(), out)\n\t}\n\treturn nil\n}", "func GitCommit() string { return gitCommit }", "func GitCommitPush(diretorio string, mensagemCommit string, username string, password string) error {\n\n\t// Abrir um repositório já existente\n\tr, err := git.PlainOpen(diretorio)\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"git.PlainOpen\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\tw, err := r.Worktree()\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.Worktree\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// Adds the new file to the staging area.\n\tlogger.Info.Println(\"git add .\")\n\t_, err = w.Add(\".\")\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Add\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// We can verify the current status of the worktree using the method Status.\n\tlogger.Info.Println(\"git status --porcelain\")\n\tstatus, err := w.Status()\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Status\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\tlogger.Info.Println(status)\n\n\t// Commit dos arquivos\n\tlogger.Info.Printf(\"git commit -m %s\\n\\r\", mensagemCommit)\n\tcommit, err := w.Commit(mensagemCommit, &git.CommitOptions{\n\t\tAuthor: &object.Signature{\n\t\t\tName: \"Marcelo Melo\",\n\t\t\tEmail: \"[email protected]\",\n\t\t\tWhen: time.Now(),\n\t\t},\n\t})\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Commit\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// Prints the current HEAD to verify that all worked well.\n\tlogger.Info.Println(\"git show -s\")\n\tobj, err := r.CommitObject(commit)\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.CommitObject\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\tlogger.Info.Println(obj)\n\n\tlogger.Info.Println(\"git push\")\n\terr = r.Push(&git.PushOptions{\n\t\tRemoteName: \"origin\",\n\t\tProgress: os.Stdout,\n\t\tAuth: &githttp.BasicAuth{\n\t\t\tUsername: username,\n\t\t\tPassword: password,\n\t\t}})\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.Push\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (b *gitBackend) Commit(ctx context.Context) error {\n\n\t// NB(estroz): some form of locking may be necessary so a branch isn't being modified\n\t// by another actor when entering Commit(). No locking is probably fine for now\n\t// since the repo dir is likely a randomized temp dir.\n\n\t// Ensure the currently checked out branch is the one b was configured with.\n\thead, err := b.repo.Head()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error getting git HEAD: %v\", err)\n\t}\n\tif b.pullOpts.ReferenceName.String() != head.Name().String() {\n\t\treturn fmt.Errorf(\"current HEAD %q of git repo is not at expected ref %q\", head.Name().String(), b.pullOpts.ReferenceName.String())\n\t}\n\n\twt, err := b.repo.Worktree()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error getting git working tree: %v\", err)\n\t}\n\n\t// Make sure git state is fresh.\n\tif err := wt.PullContext(ctx, b.pullOpts); err != nil && !errors.Is(err, git.NoErrAlreadyUpToDate) {\n\t\treturn fmt.Errorf(\"error pulling branch %q state: %v\", b.pullOpts.ReferenceName.Short(), err)\n\t}\n\n\t// Stage changes.\n\t// TODO(estroz): staging all changes might be bad if no .gitignore exists\n\t// for temporary mirror artifacts. Clean this up by saving changed files and\n\t// new files created by the backend.\n\taddOpts := &git.AddOptions{\n\t\tAll: true,\n\t}\n\tif err := wt.AddWithOptions(addOpts); err != nil {\n\t\treturn fmt.Errorf(\"error staging git state changes: %v\", err)\n\t}\n\n\t// Commit staged changes.\n\tcommitMsg, err := newCommitMsg(b.cmtOpts.Author)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error creating commit message: %v\", err)\n\t}\n\tcmtHash, err := wt.Commit(commitMsg, b.cmtOpts)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error committing changes: %v\", err)\n\t}\n\n\t// Push committed changes.\n\tpushOpts := &git.PushOptions{\n\t\t// Auth info.\n\t\tAuth: b.pullOpts.Auth,\n\t\tCABundle: b.pullOpts.CABundle,\n\t\tInsecureSkipTLS: b.pullOpts.InsecureSkipTLS,\n\t\t// Push configuration.\n\t\tRemoteName: b.pullOpts.RemoteName,\n\t\tRefSpecs: []gitconfig.RefSpec{gitconfig.RefSpec(b.pullOpts.ReferenceName)},\n\t\tRequireRemoteRefs: []gitconfig.RefSpec{gitconfig.RefSpec(b.pullOpts.ReferenceName)},\n\t\tPrune: false,\n\t\tProgress: b.pullOpts.Progress,\n\t}\n\tif err := b.repo.PushContext(ctx, pushOpts); err != nil {\n\t\treturn fmt.Errorf(\"error pushing %q: %v\", cmtHash, err)\n\t}\n\n\treturn nil\n}", "func (c *Cyclone) commit(msg *erebos.Transport) {\n\tmsg.Commit <- &erebos.Commit{\n\t\tTopic: msg.Topic,\n\t\tPartition: msg.Partition,\n\t\tOffset: msg.Offset,\n\t}\n}", "func (_Editable *EditableTransactor) Commit(opts *bind.TransactOpts, _objectHash string) (*types.Transaction, error) {\n\treturn _Editable.contract.Transact(opts, \"commit\", _objectHash)\n}", "func Commit() string {\n\treturn commit\n}", "func pushCommit(ref *github.Reference, tree *github.Tree) (err error) {\n\t// Get the parent commit to attach the commit to.\n\tparent, _, err := client.Repositories.GetCommit(ctx, *data.sourceOwner, *data.sourceRepo, *ref.Object.SHA)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// This is not always populated, but is needed.\n\tparent.Commit.SHA = parent.SHA\n\n\t// Create the commit using the tree.\n\tdate := time.Now()\n\tauthorEmail := data.authorEmail\n\tauthor := &github.CommitAuthor{Date: &date, Name: data.authorName, Email: authorEmail}\n\tcommit := &github.Commit{Author: author, Message: data.commitMessage, Tree: tree, Parents: []*github.Commit{parent.Commit}}\n\tnewCommit, _, err := client.Git.CreateCommit(ctx, *data.sourceOwner, *data.sourceRepo, commit)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Attach the commit to the master branch.\n\tref.Object.SHA = newCommit.SHA\n\t_, _, err = client.Git.UpdateRef(ctx, *data.sourceOwner, *data.sourceRepo, ref, false)\n\treturn err\n}", "func createCommit(t *testing.T, r *git.Repository, message string) {\n\tw, err := r.Worktree()\n\tif err != nil {\n\t\tt.Errorf(\"Unable to retrieve worktree: %s\", err.Error())\n\t}\n\n\t// Create a file with a random name\n\ts, err := randutil.AlphaString(32)\n\tif err != nil {\n\t\tt.Errorf(\"Unable to generate random string: %s\", err.Error())\n\t}\n\tf, err := w.Filesystem.Create(s)\n\tif err != nil {\n\t\tt.Errorf(\"Unable to create file in worktree: %s\", err.Error())\n\t}\n\txlog.Debugf(\"Created file %s\", f.Name())\n\n\t// Stage the file\n\t_, err = w.Add(f.Name())\n\tif err != nil {\n\t\tt.Errorf(\"Unable to stage file %s: %s\", f.Name(), err.Error())\n\t}\n\n\t_, err = w.Commit(message, &git.CommitOptions{\n\t\tAuthor: &signature,\n\t\tCommitter: &signature,\n\t})\n\tif err != nil {\n\t\tt.Errorf(\"Unable to commit file %s: %s\", f.Name(), err.Error())\n\t}\n}", "func (_BaseLibrary *BaseLibraryTransactor) Commit(opts *bind.TransactOpts, _objectHash string) (*types.Transaction, error) {\n\treturn _BaseLibrary.contract.Transact(opts, \"commit\", _objectHash)\n}", "func (w *Writer) Commit() (err error) {\n\t//var writeOpts opt.WriteOptions\n\treturn w.db.Write(&w.batch, nil)\n}", "func (r Repository) Commit(msg string, opt *CommitOptions) (Hash, error) {\n\tif opt == nil {\n\t\topt = &CommitOptions{}\n\t}\n\n\targs := []string{\"commit\"}\n\tif opt.Amend {\n\t\targs = append(args, \"--amend\")\n\t} else {\n\t\targs = append(args, \"-m\", msg)\n\t}\n\n\tvar env []string\n\tif opt.AuthorName != \"\" || opt.AuthorEmail != \"\" {\n\t\tenv = []string{\n\t\t\tfmt.Sprintf(\"GIT_AUTHOR_NAME=%v\", opt.AuthorName),\n\t\t\tfmt.Sprintf(\"GIT_AUTHOR_EMAIL=%v\", opt.AuthorEmail),\n\t\t\tfmt.Sprintf(\"GIT_COMMITTER_NAME=%v\", opt.AuthorName),\n\t\t\tfmt.Sprintf(\"GIT_COMMITTER_EMAIL=%v\", opt.AuthorEmail),\n\t\t}\n\t}\n\tif _, err := r.run(env, opt.Timeout, \"commit\", \"-m\", msg); err != nil {\n\t\treturn Hash{}, err\n\t}\n\tout, err := r.run(nil, 0, \"rev-parse\", \"HEAD\")\n\tif err != nil {\n\t\treturn Hash{}, err\n\t}\n\treturn ParseHash(out)\n}", "func (t TestRepo) Commit(treeID *git.Oid) *git.Oid {\n\tloc, err := time.LoadLocation(\"America/Chicago\")\n\tCheckFatal(t.test, err)\n\tsig := &git.Signature{\n\t\tName: \"Rand Om Hacker\",\n\t\tEmail: \"[email protected]\",\n\t\tWhen: time.Date(2013, 03, 06, 14, 30, 0, 0, loc),\n\t}\n\n\theadUnborn, err := t.repo.IsHeadUnborn()\n\tCheckFatal(t.test, err)\n\tvar currentTip *git.Commit\n\n\tif !headUnborn {\n\t\tcurrentBranch, err := t.repo.Head()\n\t\tCheckFatal(t.test, err)\n\t\tcurrentTip, err = t.repo.LookupCommit(currentBranch.Target())\n\t\tCheckFatal(t.test, err)\n\t}\n\n\tmessage := \"This is a commit\\n\"\n\ttree, err := t.repo.LookupTree(treeID)\n\tCheckFatal(t.test, err)\n\n\tvar commitID *git.Oid\n\tif headUnborn {\n\t\tcommitID, err = t.repo.CreateCommit(\"HEAD\", sig, sig, message, tree)\n\t} else {\n\t\tcommitID, err = t.repo.CreateCommit(\"HEAD\", sig, sig, message, tree,\n\t\t\tcurrentTip)\n\t}\n\tCheckFatal(t.test, err)\n\n\treturn commitID\n}", "func (_Container *ContainerTransactor) Commit(opts *bind.TransactOpts, _objectHash string) (*types.Transaction, error) {\n\treturn _Container.contract.Transact(opts, \"commit\", _objectHash)\n}", "func Commit(fileToParse, samplesDir, outPath, tmplPath, tmplName, commit string) error {\n\tbuf := &bytes.Buffer{}\n\tif err := executeTemplate(buf, tmplName, tmplPath, commit, nil, nil); err != nil {\n\t\treturn err\n\t}\n\treturn formatedWrite(outPath, buf.Bytes())\n}", "func commit(srvChan chan string, channel, nick, hostname string, args []string) {\n\ttype repoJSON struct {\n\t\tId int\n\t\tOwner map[string]interface{}\n\t\tName string\n\t\tFull_name string\n\t\tDescription string\n\t\tPrivate bool\n\t\tFork bool\n\t\tUrl string\n\t\tHtml_url string\n\t}\n\ttype commitJSON struct {\n\t\tSha string\n\t\tCommit map[string]interface{}\n\t\tUrl string\n\t\tHtml_url string\n\t\tComments_url string\n\t\tAuthor map[string]interface{}\n\t\tCommitter map[string]interface{}\n\t\tParents map[string]interface{}\n\t}\n\ttype urlJSON struct {\n\t\tKind string\n\t\tId string\n\t\tLongUrl string\n\t}\n\tmessage := \"NOTICE \" + channel + \" :\"\n\tsince := rand.Intn(1000000)\n\tres, err := http.Get(\"https://api.github.com/repositories?since=\" + fmt.Sprintf(\"%d\", since))\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tbody, err := ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tvar repos []repoJSON\n\tjson.Unmarshal(body, &repos)\n\tfullName := repos[rand.Intn(len(repos))].Full_name\n\tres, err = http.Get(\"https://api.github.com/repos/\" + fullName + \"/commits\")\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tbody, err = ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tvar commits []commitJSON\n\tjson.Unmarshal(body, &commits)\n\tif len(commits) < 1 {\n\t\tcommit(srvChan, channel, nick, hostname, args) //try again\n\t\treturn\n\t} else {\n\t\tcommitNum := rand.Intn(len(commits))\n\t\tcommitMsg := commits[commitNum].Commit[\"message\"].(string)\n\n\t\tAPIkey, err := ioutil.ReadFile(\"APIkey\")\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\turlReader := strings.NewReader(`{\"longUrl\": \"` + commits[commitNum].Html_url + `\"}`)\n\t\tc := http.Client{}\n\t\tres, err := c.Post(\"https://www.googleapis.com/urlshortener/v1/url?key=\" + string(APIkey), \"application/json\", urlReader)\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tbody, err := ioutil.ReadAll(res.Body)\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tvar googUrl urlJSON\n\t\tjson.Unmarshal(body, &googUrl)\n\t\tmessage += strings.Split(commitMsg, \"\\n\")[0] + \" | \" + googUrl.Id\n\t}\n\tsrvChan <- message\n\tlog.Println(message)\n}", "func (s *SimplePublishNode) Commit() {\n\ts.committed++\n}", "func Commit(r *git.Repository, msg string) bool {\n\t// check for differences.\n\tdiffCmd := fmt.Sprintf(\"diff -U 0 %s %s/pimpminers.conf | grep -v ^@ | wc -l\", stagingFile, localGitRepo)\n\tdiff := RunCommand(diffCmd)\n\tdiff = strings.TrimSpace(diff)\n\tif diff == \"0\" {\n\t\tfmt.Println(\"No changes to commit.\")\n\t} else {\n\t\t// copy file from staging into worktree\n\t\tdiffCmd = fmt.Sprintf(\"diff -U 0 %s %s/pimpminers.conf | grep -v ^@\", stagingFile, localGitRepo)\n\t\tdiff = RunCommand(diffCmd)\n\t\tdiff = strings.TrimSpace(diff)\n\t\tfmt.Println(\"Changes:\")\n\t\tfmt.Println(diff)\n\t\tfmt.Println(\"\\nCommitting changes... \")\n\t\tcopy := fmt.Sprintf(\"cp %s %s/pimpminers.conf\", stagingFile, localGitRepo)\n\t\tRunCommand(copy)\n\t\tw, err := r.Worktree()\n\t\tcheckErr(err)\n\t\t// add files\n\t\t_, err = w.Add(\"pimpminers.conf\")\n\t\tcheckErr(err)\n\t\t// commit\n\t\t_, err = w.Commit(msg, &git.CommitOptions{\n\t\t\tAuthor: &object.Signature{\n\t\t\t\tName: \"pimplabops\",\n\t\t\t\tEmail: \"[email protected]\",\n\t\t\t\tWhen: time.Now(),\n\t\t\t},\n\t\t})\n\t\tcheckErr(err)\n\t\tif err != nil {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (_BaseContent *BaseContentTransactor) Commit(opts *bind.TransactOpts, _objectHash string) (*types.Transaction, error) {\n\treturn _BaseContent.contract.Transact(opts, \"commit\", _objectHash)\n}", "func (c *Client) CommitUpload(request *CommitUploadRequest) (response *CommitUploadResponse, err error) {\n if request == nil {\n request = NewCommitUploadRequest()\n }\n response = NewCommitUploadResponse()\n err = c.Send(request, response)\n return\n}", "func (a v3ioAppender) Commit() error { return nil }", "func (rebase *Rebase) Commit(ID *Oid, author, committer *Signature, message string) error {\n\truntime.LockOSThread()\n\tdefer runtime.UnlockOSThread()\n\n\tauthorSig, err := author.toC()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer C.git_signature_free(authorSig)\n\n\tcommitterSig, err := committer.toC()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer C.git_signature_free(committerSig)\n\n\tcmsg := C.CString(message)\n\tdefer C.free(unsafe.Pointer(cmsg))\n\n\tcerr := C.git_rebase_commit(ID.toC(), rebase.ptr, authorSig, committerSig, nil, cmsg)\n\truntime.KeepAlive(ID)\n\truntime.KeepAlive(rebase)\n\tif cerr < 0 {\n\t\treturn MakeGitError(cerr)\n\t}\n\n\treturn nil\n}", "func (dbi *DB) Commit() {\r\n\tif dbi.status {\r\n\t\tdbi.createOperation(\"DB_COMMIT\")\r\n\t\tdbi.data.commPrepare()\r\n\t\tdbi.data.comm()\r\n\t\tdbi.data.commParse()\r\n\t\tif Show {\r\n\t\t\tmylog.Println(\"[\" + dbi.Sid + \"]SQL=COMMIT;\")\r\n\t\t}\r\n\t}\r\n}", "func (_BaseContentType *BaseContentTypeTransactor) Commit(opts *bind.TransactOpts, _objectHash string) (*types.Transaction, error) {\n\treturn _BaseContentType.contract.Transact(opts, \"commit\", _objectHash)\n}", "func (db *DB) Commit(msg string) error {\n\tif db.parent != nil {\n\t\treturn db.parent.Commit(msg)\n\t}\n\tdb.l.Lock()\n\tdefer db.l.Unlock()\n\tif db.tree == nil {\n\t\t// Nothing to commit\n\t\treturn nil\n\t}\n\tcommit, err := CommitToRef(db.repo, db.tree, db.commit, db.ref, msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif db.commit != nil {\n\t\tdb.commit.Free()\n\t}\n\tdb.commit = commit\n\treturn nil\n}", "func commitEOF() error {\n\treturn cliutil.ActionError(clitypes.CommitEOF)\n}", "func (a *folder_asset) commit() error {\n\treturn (*file_asset)(a).commit()\n}", "func (t *Transacter) Commit(ctx context.Context) error {\r\n\ttx := TxFromContext(ctx)\r\n\tif tx.Tx != nil {\r\n\t\treturn tx.Tx.Commit()\r\n\t}\r\n\treturn nil\r\n}", "func (f CommitFunc) Commit(ctx context.Context, tx *Tx) error {\n\treturn f(ctx, tx)\n}", "func (f CommitFunc) Commit(ctx context.Context, tx *Tx) error {\n\treturn f(ctx, tx)\n}", "func (m *Manager) Commit() error {\n\ttargetRef, err := is.Transport.ParseStoreReference(m.store, m.targetImage)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\topts := buildah.CommitOptions{\n\t\tSquash: true,\n\t\tSystemContext: systemContext,\n\t\tCompression: imagebuildah.Gzip,\n\t}\n\tid, ref, _, err := m.b.Commit(m.ctx, targetRef, opts)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Image-id: '%s'\", id)\n\tlog.Infof(\"Name: '%s'\", ref.Name())\n\tlog.Infof(\"Digest: '%s'\", ref.Digest().String())\n\treturn nil\n}", "func (*txDriver) Commit() error { return nil }", "func (*txDriver) Commit() error { return nil }", "func (d *dockerImageDestination) Commit(ctx context.Context) error {\n\treturn nil\n}", "func (p *PluginClient) GitCommit(meta Meta, secret corev1.Secret) ClientGitCommit {\n\treturn newGitCommit(p, meta, secret)\n}", "func (_BaseAccessWallet *BaseAccessWalletTransactor) Commit(opts *bind.TransactOpts, _objectHash string) (*types.Transaction, error) {\n\treturn _BaseAccessWallet.contract.Transact(opts, \"commit\", _objectHash)\n}", "func (cdt *SqlDBTx) Commit() error {\n\treturn nil\n}", "func (tx *TestTX) Commit() error {\n\targs := tx.Called()\n\treturn args.Error(0)\n}", "func Commit(c git.CommitEntry, rules []Rule) Results {\n\tresults := Results{}\n\tfor _, r := range rules {\n\t\tresults = append(results, r.Run(r, c))\n\t}\n\treturn results\n}", "func (g *Git) Commit(commitAuthor *internalgit.CommitAuthor, commitMessage string) error {\n\tw, err := g.repo.Worktree()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Make sure gitignore is used\n\tpatterns, err := gitignore.ReadPatterns(w.Filesystem, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\tw.Excludes = patterns\n\n\terr = w.AddWithOptions(&git.AddOptions{\n\t\tAll: true,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tstatus, err := w.Status()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// This is a workaround for a bug in go-git where \"add all\" does not add deleted files\n\t// If https://github.com/go-git/go-git/issues/223 is fixed, this can be removed\n\tfor file, s := range status {\n\t\tif s.Worktree == git.Deleted {\n\t\t\t_, err = w.Add(file)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// Get the current hash to be able to diff it with the committed changes later\n\toldHead, err := g.repo.Head()\n\tif err != nil {\n\t\treturn err\n\t}\n\toldHash := oldHead.Hash()\n\n\tvar author *object.Signature\n\tif commitAuthor != nil {\n\t\tauthor = &object.Signature{\n\t\t\tName: commitAuthor.Name,\n\t\t\tEmail: commitAuthor.Email,\n\t\t\tWhen: time.Now(),\n\t\t}\n\t}\n\n\thash, err := w.Commit(commitMessage, &git.CommitOptions{\n\t\tAuthor: author,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcommit, err := g.repo.CommitObject(hash)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_ = g.logDiff(oldHash, commit.Hash)\n\n\treturn nil\n}", "func (t *Table) QueueGitCommit(m *Model, email string) error {\n\tcontent, err := yaml.Marshal(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcomment := fmt.Sprintf(\"Table %s updated.\", t.Name)\n\tname := \"\" // TODO we don't know the user's name\n\n\tm.git.QueueCommit(t.relpath(), content, comment, name, email, time.Now())\n\treturn nil\n}", "func (_BaseContentSpace *BaseContentSpaceTransactor) Commit(opts *bind.TransactOpts, _objectHash string) (*types.Transaction, error) {\n\treturn _BaseContentSpace.contract.Transact(opts, \"commit\", _objectHash)\n}", "func (r *Repository) Commit(h plumbing.Hash) (*Commit, error) {\n\tcommit, err := r.Object(plumbing.CommitObject, h)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn commit.(*Commit), nil\n}", "func (s *PBFTServer) Commit(args CommitArgs, reply *CommitReply) error {\n\t// Verify signature\n\n\ts.lock.Lock()\n\n\ts.stopTimer()\n\n\tif !s.changing && s.view == args.View && s.h <= args.Seq && args.Seq < s.H {\n\t\tent := s.getEntry(entryID{args.View, args.Seq})\n\t\ts.lock.Unlock()\n\n\t\tent.lock.Lock()\n\t\tent.c = append(ent.c, &args)\n\t\tUtil.Dprintf(\"%s[R/Commit]:Args:%+v\", s, args)\n\t\tif !ent.sendReply && ent.sendCommit && s.committed(ent) {\n\t\t\tUtil.Dprintf(\"%s start execute %v @ %v\", s, ent.pp.Message.Op, args.Seq)\n\t\t\t// Execute will make sure there only one execution of one request\n\t\t\tres, _ := s.execute(args.Seq, ent.pp.Message.Op, args.Digest)\n\t\t\tif ent.r == nil {\n\t\t\t\trArgs := ResponseArgs{\n\t\t\t\t\tView: args.View,\n\t\t\t\t\tSeq: ent.pp.Message.Seq,\n\t\t\t\t\tCid: ent.pp.Message.Id,\n\t\t\t\t\tRid: s.id,\n\t\t\t\t\tRes: res,\n\t\t\t\t}\n\t\t\t\tent.r = &rArgs\n\t\t\t}\n\t\t\tent.sendReply = true\n\t\t}\n\t\ts.reply(ent)\n\t\tent.lock.Unlock()\n\t} else {\n\t\ts.lock.Unlock()\n\t}\n\treturn nil\n}", "func (c *ETHController) Commit(opts *bind.TransactOpts, domain string, owner common.Address, secret [32]byte) (*types.Transaction, error) {\n\tname, err := UnqualifiedName(domain, c.domain)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"invalid name %s\", domain)\n\t}\n\n\tcommitment, err := c.Contract.MakeCommitment(nil, name, owner, secret)\n\tif err != nil {\n\t\treturn nil, errors.New(\"failed to create commitment\")\n\t}\n\n\tif opts.Value != nil && opts.Value.Cmp(big.NewInt(0)) != 0 {\n\t\treturn nil, errors.New(\"commitment should have 0 value\")\n\t}\n\n\treturn c.Contract.Commit(opts, commitment)\n}", "func (db *Database) commit(hash common.Hash, batch ethdb.Batch) error {\n\t// If the node does not exist, it's a previously committed node\n\tnode, ok := db.dirties[hash]\n\tif !ok {\n\t\treturn nil\n\t}\n\tfor _, child := range node.childs() {\n\t\tif err := db.commit(child, batch); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif err := batch.Put(hash[:], node.rlp()); err != nil {\n\t\treturn err\n\t}\n\t// If we've reached an optimal batch size, commit and start over\n\tif batch.ValueSize() >= ethdb.IdealBatchSize {\n\t\tif err := batch.Write(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tbatch.Reset()\n\t}\n\treturn nil\n}", "func (w *DiskImage) Commit(ctx context.Context, size int64, expected digest.Digest, opts ...ctrcontent.Opt) error {\n\treturn nil\n}", "func commit(u interface{}, data []byte) error {\n\tvar blockData DataStu\n\terr := json.Unmarshal(data, &blockData)\n\tif err != nil {\n\t\tlogger.Error(\"commit Unmarshal error\", zap.Error(err))\n\t\treturn fmt.Errorf(\"Commit block failed:%v\", err)\n\t}\n\tb := blockData.Block\n\tif b == nil {\n\t\tlogger.Error(\"failed to unmarsal block\")\n\t\treturn errors.New(\"Commit block failed\")\n\t}\n\n\terr = u.(*bftnode).bc.AddBlock(b, []byte(u.(*bftnode).cfg.CountAddr))\n\tif err != nil {\n\t\tlogger.Error(\"Fatal error: commit block failed\", zap.Uint64(\"height\", b.Height), zap.Error(err))\n\t\treturn fmt.Errorf(\"Commit block failed:%v\", err)\n\t}\n\t//update last blockHeight\n\tu.(*bftnode).lastHeight = b.Height\n\tlogger.Info(\"Finished commit block\", zap.Uint64(\"height\", u.(*bftnode).lastHeight), zap.Int(\"data lenght\", len(data)), zap.Int(\"tx lenght\", len(b.Transactions)))\n\treturn nil\n}", "func (b *batcher) Commit(ctx context.Context, commitInfo *files.UploadSessionFinishArg) (entry *files.FileMetadata, err error) {\n\tselect {\n\tcase <-b.closed:\n\t\treturn nil, fserrors.FatalError(errors.New(\"batcher is shutting down\"))\n\tdefault:\n\t}\n\tfs.Debugf(b.f, \"Adding %q to batch\", commitInfo.Commit.Path)\n\tresp := make(chan batcherResponse, 1)\n\tb.in <- batcherRequest{\n\t\tcommitInfo: commitInfo,\n\t\tresult: resp,\n\t}\n\t// If running async then don't wait for the result\n\tif b.async {\n\t\treturn nil, nil\n\t}\n\tresult := <-resp\n\treturn result.entry, result.err\n}", "func (p *Protocol) Commit(ctx context.Context, sm protocol.StateManager) error {\n\theight, err := sm.Height()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcsm, err := NewCandidateStateManager(sm, p.hu.IsPost(config.Greenland, height))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// commit updated view\n\treturn errors.Wrap(csm.Commit(), \"failed to commit candidate change in Commit\")\n}", "func (_BaseAccessControlGroup *BaseAccessControlGroupTransactor) Commit(opts *bind.TransactOpts, _objectHash string) (*types.Transaction, error) {\n\treturn _BaseAccessControlGroup.contract.Transact(opts, \"commit\", _objectHash)\n}", "func (r *RedisStore) Commit(token string, b []byte, expiry time.Time) error {\n\tconn := r.pool.Get()\n\tdefer conn.Close()\n\n\terr := conn.Send(\"MULTI\")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = conn.Send(\"SET\", r.prefix+token, b)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = conn.Send(\"PEXPIREAT\", r.prefix+token, makeMillisecondTimestamp(expiry))\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = conn.Do(\"EXEC\")\n\treturn err\n}", "func (o *qemuOperations) Commit(image string) error {\n\tklog.V(1).Infof(\"Committing %s to backing file...\", image)\n\targs := []string{\"commit\", \"-p\", image}\n\t_, err := qemuExecFunction(nil, reportProgress, \"qemu-img\", args...)\n\treturn err\n}", "func (tc *consumer) Commit(topic string, partition int32, offset int64) error {\n\treturn nil\n}", "func PostCommit(handler func()) {\n\thandler()\n\tos.Exit(0)\n}", "func (mavls *Store) Commit(req *types.ReqHash) ([]byte, error) {\r\n\tbeg := types.Now()\r\n\tdefer func() {\r\n\t\tmlog.Debug(\"Commit\", \"cost\", types.Since(beg))\r\n\t}()\r\n\ttree, ok := mavls.trees.Load(string(req.Hash))\r\n\tif !ok {\r\n\t\tmlog.Error(\"store mavl commit\", \"err\", types.ErrHashNotFound)\r\n\t\treturn nil, types.ErrHashNotFound\r\n\t}\r\n\tif tree == nil {\r\n\t\tmlog.Info(\"store mavl commit,do nothing for kvset is null\")\r\n\t\tmavls.trees.Delete(string(req.Hash))\r\n\t\treturn req.Hash, nil\r\n\t}\r\n\thash := tree.(*mavl.Tree).Save()\r\n\tif hash == nil {\r\n\t\tmlog.Error(\"store mavl commit\", \"err\", types.ErrHashNotFound)\r\n\t\treturn nil, types.ErrDataBaseDamage\r\n\t}\r\n\tmavls.trees.Delete(string(req.Hash))\r\n\treturn req.Hash, nil\r\n}", "func (conn *RealBackend) Commit() {\n}", "func (g *GitLocal) AddCommit(dir string, msg string) error {\n\treturn g.GitCLI.AddCommit(dir, msg)\n}", "func (s *VarlinkInterface) Commit(ctx context.Context, c VarlinkCall, name_ string, image_name_ string, changes_ []string, author_ string, message_ string, pause_ bool, manifestType_ string) error {\n\treturn c.ReplyMethodNotImplemented(ctx, \"io.podman.Commit\")\n}", "func (foo *Foo) Commit() *Foo {\n\tif _, ok := Stage.Foos[foo]; ok {\n\t\tif Stage.BackRepo != nil {\n\t\t\tStage.BackRepo.CommitFoo(foo)\n\t\t}\n\t}\n\treturn foo\n}", "func handleCommit(args ...string) (string, error) {\n\terr := GetCurrentStore().Commit()\n\treturn \"\", err\n}", "func (txn *levelDBTxn) commit() error {\n\t// Check context first to make sure transaction is not cancelled.\n\tselect {\n\tdefault:\n\tcase <-txn.ctx.Done():\n\t\treturn txn.ctx.Err()\n\t}\n\n\ttxn.mu.Lock()\n\tdefer txn.mu.Unlock()\n\n\treturn txn.kv.Write(txn.batch, nil)\n}", "func (w *writer) Commit() error {\n\tif w.closed {\n\t\treturn fmt.Errorf(\"already closed\")\n\t} else if w.committed {\n\t\treturn fmt.Errorf(\"already committed\")\n\t} else if w.cancelled {\n\t\treturn fmt.Errorf(\"already cancelled\")\n\t}\n\tfmt.Println(\"commit:\",\"block size=\",len(w.blocks))\n\n\treturn w.flushBlock()\n}", "func (_m *Repository) CommitTransaction(ctx context.Context) {\n\t_m.Called(ctx)\n}", "func (p *Protocol) Commit(ctx context.Context, sm protocol.StateManager) error {\n\tcenter, err := getCandCenter(sm)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to commit candidate change in Commit\")\n\t}\n\n\tcsm, err := NewCandidateStateManager(sm, center)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// commit updated view\n\treturn errors.Wrap(csm.Commit(), \"failed to commit candidate change in Commit\")\n}", "func (vc *VCSystem) Commit(filename, text string) (err error) {\n\t// TODO add file locking\n\n\t_, filename = filepath.Split(filename)\n\tfilenameExt := filepath.Ext(filename)\n\tfilename = strings.TrimSuffix(filename, filenameExt)\n\n\tfileList := []string{}\n\terr = filepath.Walk(path.Join(vc.BaseFolder, \".vcs\"), func(path string, f os.FileInfo, err error) error {\n\t\tfileList = append(fileList, path)\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfor _, f := range fileList {\n\t\tvar fname string\n\t\tfname, err = filepath.Abs(f)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tbasename := strings.TrimPrefix(fname, vc.BaseFolder)\n\t\tfmt.Println(basename)\n\t}\n\n\tcurrentText := \"\"\n\tn := Block{\n\t\tBranch: vc.CurrentBranch,\n\t\tPatch: reldel.GetPatch([]byte(currentText), []byte(text)),\n\t}\n\th := sha256.New()\n\th.Write([]byte(\"vcf\")) // salt\n\th.Write([]byte(text))\n\th.Write([]byte(n.Patch.Time.String()))\n\tbN, err := json.Marshal(n)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = ioutil.WriteFile(fmt.Sprintf(\"%s.%x%s\", filename, h.Sum(nil), filenameExt), bN, 0755)\n\treturn\n}", "func (r *testRepo) modifyAndCommit(filename string, opts *git.CommitOptions) plumbing.Hash {\n\tvar (\n\t\thash plumbing.Hash\n\t\terr error\n\t\tfile billy.File\n\t)\n\n\tif file, err = r.Source.Filesystem.OpenFile(filename, os.O_RDWR|os.O_CREATE|os.O_APPEND, 0666); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer file.Close()\n\n\tif _, err = file.Write([]byte(fmt.Sprintf(\"commit %d\\n\", r.seqno))); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif _, err = r.Source.Add(filename); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif hash, err = r.Source.Commit(fmt.Sprintf(\"commit %d\", r.seqno), opts); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tr.seqno++\n\n\treturn hash\n}", "func (w *blobWriter) Commit(ctx context.Context, provisional distribution.Descriptor) (canonical distribution.Descriptor, err error) {\n\tspan, ctx := opentracing.StartSpanFromContext(ctx, \"BlobWriter.Commit\")\n\tdefer span.Finish()\n\tspan.LogFields(\n\t\tlog.String(\"digest\", provisional.Digest.String()),\n\t)\n\n\tcanonical, err = w.BlobWriter.Commit(ctx, provisional)\n\tspan.LogFields(\n\t\tlog.String(\"mediaType\", canonical.MediaType),\n\t\tlog.Int64(\"size\", canonical.Size),\n\t)\n\treturn\n}", "func (self *WAL) Commit(requestNumber uint32, serverId uint32) error {\n\tconfirmationChan := make(chan *confirmation)\n\tself.entries <- &commitEntry{confirmationChan, serverId, requestNumber}\n\tconfirmation := <-confirmationChan\n\treturn confirmation.err\n}", "func (w *Wrapper) Commit() error {\n\treturn w.db.Commit()\n}", "func (op *Operation) Commit() error {\n\tif op.wipe {\n\t\tcmd := exec.Command(sgdiskPath, \"--zap-all\", op.dev)\n\t\tif err := cmd.Run(); err != nil {\n\t\t\treturn fmt.Errorf(\"wipe failed: %v\")\n\t\t}\n\t}\n\n\tif len(op.parts) != 0 {\n\t\topts := []string{}\n\t\tfor _, p := range op.parts {\n\t\t\topts = append(opts, fmt.Sprintf(\"--new=%d:%d:+%d\", p.Number, p.Offset, p.Length))\n\t\t\topts = append(opts, fmt.Sprintf(\"--change-name=%d:%s\", p.Number, p.Label))\n\t\t}\n\t\topts = append(opts, op.dev)\n\t\tcmd := exec.Command(sgdiskPath, opts...)\n\t\tif err := cmd.Run(); err != nil {\n\t\t\treturn fmt.Errorf(\"create partitions failed: %v\", err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (gr *GitRepo) addCommit(c *g2g.Commit) bool {\n\tif c == nil {\n\t\treturn false\n\t}\n\n\tvar commit model.Commit\n\n\toID := c.Id()\n\tif oID == nil {\n\t\treturn false\n\t}\n\tcommit.VCSID = oID.String()\n\n\tcommit.Message = c.Message()\n\n\tvar author model.Developer\n\tauthor.Name = c.Author().Name\n\tauthor.Email = c.Author().Email\n\tcommit.Author = author\n\n\tvar committer model.Developer\n\tcommitter.Name = c.Committer().Name\n\tcommitter.Email = c.Committer().Email\n\tcommit.Committer = committer\n\n\tcommit.CommitDate = c.Committer().When\n\tcommit.AuthorDate = c.Author().When\n\n\tparentC := c.Parent(0)\n\tif parentC == nil {\n\t\treturn false\n\t}\n\n\tparentTree, err := parentC.Tree()\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tcTree, err := c.Tree()\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tdiffOpts, err := g2g.DefaultDiffOptions()\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tdiff, err := gr.r.DiffTreeToTree(parentTree, cTree, &diffOpts)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tstats, err := diff.Stats()\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tnDeltas, err := diff.NumDeltas()\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tif gr.cfg.CommitDeltas {\n\t\tfor d := 0; d < nDeltas; d++ {\n\t\t\tvar cdd model.DiffDelta\n\n\t\t\tif gr.cfg.CommitPatches {\n\t\t\t\tpatch, err := diff.Patch(d)\n\t\t\t\tif err != nil || patch == nil {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t\tp, err := patch.String()\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t\tcdd.Patch = &p\n\t\t\t}\n\n\t\t\tdiffDelta, err := diff.GetDelta(d)\n\t\t\tif err != nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tcdd.Status = deltaMap[diffDelta.Status]\n\n\t\t\tvar isBin bool\n\t\t\tif (diffDelta.Flags & g2g.DiffFlagBinary) > 0 {\n\t\t\t\tisBin = true\n\t\t\t}\n\t\t\tcdd.Binary = &isBin\n\n\t\t\t// TODO compute similarity to add to cdd.Similarity\n\n\t\t\tcdd.OldFilePath = &diffDelta.OldFile.Path\n\t\t\tcdd.NewFilePath = &diffDelta.NewFile.Path\n\n\t\t\tcommit.DiffDelta = append(commit.DiffDelta, cdd)\n\t\t}\n\t}\n\n\tcommit.FileChangedCount = stats.FilesChanged()\n\tcommit.InsertionsCount = stats.Insertions()\n\tcommit.DeletionsCount = stats.Deletions()\n\n\tgr.Commits = append(gr.Commits, commit)\n\n\treturn true\n}", "func (w *writer) Commit(ctx context.Context, descriptor distribution.Descriptor) (distribution.Descriptor, error) {\n\tdesc := descriptor\n\tswitch {\n\tcase w.closed:\n\t\treturn desc, fmt.Errorf(\"already closed\")\n\tcase w.committed:\n\t\treturn desc, fmt.Errorf(\"already committed\")\n\tcase w.cancelled:\n\t\treturn desc, fmt.Errorf(\"already cancelled\")\n\t}\n\tw.committed = true\n\treturn desc, nil\n}", "func (ws *workingSet) Commit(ctx context.Context) error {\n\tif err := protocolPreCommit(ctx, ws); err != nil {\n\t\treturn err\n\t}\n\tif err := ws.store.Commit(); err != nil {\n\t\treturn err\n\t}\n\tif err := protocolCommit(ctx, ws); err != nil {\n\t\t// TODO (zhi): wrap the error and eventually panic it in caller side\n\t\treturn err\n\t}\n\tws.Reset()\n\treturn nil\n}", "func commit(ctx context.Context, tx *sqlx.Tx) error {\r\n\tctxx := TxFromContext(ctx)\r\n\tif ctxx != nil {\r\n\t\tif ctxx.Tx != nil {\r\n\t\t\treturn nil\r\n\t\t}\r\n\t}\r\n\treturn tx.Commit()\r\n}", "func (fs *fsMutable) commitImpl(caFs cafs.Fs) error {\n\t/* some sync setup */\n\tif fs.bundle.BundleID == \"\" {\n\t\tif err := fs.bundle.InitializeBundleID(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfs.l = fs.l.With(zap.String(\"bundle\", fs.bundle.BundleID))\n\t}\n\tfs.l.Info(\"Committing changes\")\n\tctx := context.Background() // ??? is this the correct context?\n\t/* `commitChans` includes rules about directionality that apply to threads only,\n\t * so we keep channels without directionality restriction separately.\n\t */\n\tbundleEntryC := make(chan model.BundleEntry)\n\terrorC := make(chan error)\n\tdoneC := make(chan struct{})\n\t/* closing the done channel broadcasts to all threads and is particularly important to prevent\n\t * goroutine leaks in the case of an error from any particular thread.\n\t * see the \"Explicit cancellation\" section of https://blog.golang.org/pipelines\n\t * for more detailed description on using closure of a channel to broadcast errors.\n\t *\n\t * using defer to cleanup concurrency and similar resource usage is preferred throughout,\n\t * both as a stylistic hint that what's going on is cleanup as well as for the practical reason\n\t * that deferred calls still occur, even in the case of a panic(), as described in the blog post\n\t * https://blog.golang.org/defer-panic-and-recover\n\t */\n\tdefer close(doneC)\n\t/* `commitWalkReadDirMap` signals that it's done to the caller by closing the channel containing\n\t * `model.BundleEntry` instances. since the goal of the commit is to produce a sequence of bundle uploads,\n\t * and since the second parameter to reading from a channel is false when the channel is both empty and closed,\n\t * this thread can use reading from the bundle entry channel to detect whether the walk is finished.\n\t */\n\tfs.l.Debug(\"Commit: spinning off goroutines\")\n\t// TODO: internal errors do not bubble up properly when unmounting\n\tgo commitWalkReadDirMap(ctx, fs, commitChans{\n\t\tbundleEntry: bundleEntryC,\n\t\terror: errorC,\n\t\tdone: doneC,\n\t}, caFs)\n\tfileList := make([]model.BundleEntry, 0)\n\tfor {\n\t\tvar bundleEntry model.BundleEntry\n\t\tvar moreBundleEntries bool\n\t\tselect {\n\t\tcase bundleEntry, moreBundleEntries = <-bundleEntryC:\n\t\tcase err := <-errorC:\n\t\t\t// one of the threads has had an error.\n\t\t\treturn err\n\t\t}\n\t\tif !moreBundleEntries {\n\t\t\tbreak\n\t\t}\n\t\tfileList = append(fileList, bundleEntry)\n\t}\n\tfs.l.Debug(\"Commit: goroutines ok. uploading metadata.\")\n\tfs.bundle.BundleEntries = fileList\n\tif err := fs.bundle.UploadBundleEntries(ctx); err != nil {\n\t\treturn err\n\t}\n\tfs.l.Info(\"Commit: ok.\")\n\treturn nil\n}", "func commit() string {\n\thash, _ := sh.Output(\"git\", \"rev-parse\", \"--short\", \"HEAD\")\n\treturn hash\n}", "func (orderTree *OrderTree) Commit() error {\n\terr := orderTree.Save()\n\tif err == nil {\n\t\terr = orderTree.orderDB.Commit()\n\t}\n\treturn err\n}", "func (c *conn) Commit() (err error) {\n\tif atomic.LoadInt32(&c.closed) != 0 {\n\t\treturn driver.ErrBadConn\n\t}\n\tif c.txCtx == nil {\n\t\treturn sql.ErrTxDone\n\t}\n\tctx := c.txCtx\n\tstmts := c.stmts\n\tc.txCtx = nil\n\tc.stmts = stmts[:0]\n\n\tif len(stmts) == 0 {\n\t\treturn nil\n\t}\n\tfor _, stmt := range stmts {\n\t\tc.log(\"commit statement: \", stmt.prefix, stmt.pattern)\n\t\tif err = stmt.commit(ctx); err != nil {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn\n}", "func testCommit(t *testing.T, myApp app.BaseApp, h int64) []byte {\n\t// Commit first block, make sure non-nil hash\n\theader := abci.Header{Height: h}\n\tmyApp.BeginBlock(abci.RequestBeginBlock{Header: header})\n\tmyApp.EndBlock(abci.RequestEndBlock{})\n\tcres := myApp.Commit()\n\thash := cres.Data\n\tassert.NotEmpty(t, hash)\n\treturn hash\n}", "func (p *KVServer) Commit(req CommitRequest, resp *CommitResponse) error {\n\tfmt.Println(\"Received a call to Commit(\", req, \")\")\n\ttx := req.Transaction\n\tmutex.Lock()\n\ttransactions[tx.ID] = tx\n\tmutex.Unlock()\n\tisGenerateNoOps = false\n\tfor isWorkingOnNoOp {\n\t\t// This stopped it from hanging... !\n\t\ttime.Sleep(time.Millisecond)\n\t}\n\tif !isCommitPossible(req.RequiredKeyValues) {\n\t\tmutex.Lock()\n\t\tt := transactions[tx.ID]\n\t\tt.IsAborted = true\n\t\ttransactions[tx.ID] = t\n\t\tmutex.Unlock()\n\t\t*resp = CommitResponse{false, 0, abortedMessage}\n\t\tisGenerateNoOps = true\n\t} else {\n\t\tblockHash := generateCommitBlock(tx.ID, req.RequiredKeyValues)\n\t\tif blockHash == \"\" {\n\t\t\t// a conflicting transaction just commited\n\t\t\tmutex.Lock()\n\t\t\tt := transactions[tx.ID]\n\t\t\tt.IsAborted = true\n\t\t\ttransactions[tx.ID] = t\n\t\t\tmutex.Unlock()\n\t\t\t*resp = CommitResponse{false, 0, abortedMessage + \"Another node committed a conflicting transaction!!\"}\n\t\t\tisGenerateNoOps = true\n\t\t} else {\n\t\t\tisGenerateNoOps = true\n\t\t\tvalidateCommit(req)\n\t\t\tmutex.Lock()\n\t\t\tcommitId := transactions[tx.ID].CommitID\n\t\t\tmutex.Unlock()\n\t\t\t*resp = CommitResponse{true, commitId, \"\"}\n\t\t}\n\t}\n\tprintState()\n\treturn nil\n}", "func (ref *Reference) Commit() (*Commit, error) {\n\treturn ref.repo.getCommit(ref.Object)\n}", "func (ref *Reference) Commit() (*Commit, error) {\n\treturn ref.repo.getCommit(ref.Object)\n}", "func (g *GitHub) PostCommitStatus(service *api.Service, version *api.Version) error {\n\t// Check if github webhook has set.\n\tif service.Repository.Webhook != api.GITHUB {\n\t\treturn fmt.Errorf(\"vcs github webhook hasn't set\")\n\t}\n\n\t// Check if has set commitID.\n\tif version.Commit == \"\" {\n\t\treturn fmt.Errorf(\"commit hasn't set\")\n\t}\n\n\t// Get token.\n\tds := store.NewStore()\n\tdefer ds.Close()\n\ttok, err := ds.FindtokenByUserID(service.UserID, api.GITHUB)\n\tif err != nil {\n\t\tlog.ErrorWithFields(\"find token failed\", log.Fields{\"user_id\": service.UserID, \"error\": err})\n\t\treturn err\n\t}\n\n\tts := oauth2.StaticTokenSource(\n\t\t&oauth2.Token{AccessToken: tok.Vsctoken.AccessToken},\n\t)\n\ttc := oauth2.NewClient(oauth2.NoContext, ts)\n\tclient := github.NewClient(tc)\n\n\t// Post commit status.\n\towner, repo := parseURL(service.Repository.URL)\n\turlHost := osutil.GetStringEnv(cloud.CycloneServer, \"https://fornax-canary.caicloud.io\")\n\n\tvar state string\n\tif version.Status == api.VersionHealthy {\n\t\tstate = api.CISuccess\n\t} else if version.Status == api.VersionFailed || version.Status == api.VersionCancel {\n\t\tstate = api.CIFailure\n\t} else {\n\t\tstate = api.CIPending\n\t}\n\n\tlog.Infof(\"Now, version status is %s, post %s to github\", version.Status, state)\n\turlLog := fmt.Sprintf(\"%s/log?user=%s&service=%s&version=%s\", urlHost, service.UserID,\n\t\tservice.ServiceID, version.VersionID)\n\tlog.Infof(\"Log getting url: %s\", urlLog)\n\tstatus := &github.RepoStatus{\n\t\tState: github.String(state),\n\t\tTargetURL: github.String(urlLog),\n\t\tDescription: github.String(version.ErrorMessage),\n\t\tContext: github.String(\"Cyclone/\" + service.Name),\n\t}\n\n\t_, _, err = client.Repositories.CreateStatus(owner, repo, version.Commit, status)\n\treturn err\n}", "func (tx *PgxTx) Commit(context.Context) error {\n\treturn (*pgx.Tx)(tx).Commit()\n}", "func (bw *SequentialWriter) Commit() {\n\tbw.bio.WriteAt(bw.off, bw.buf.Bytes())\n}", "func gitPush(c *cli.Context) error {\n\t// inputs\n\tremote, branch := c.Args().Get(0), c.Args().Get(1)\n\texpected(remote, branch)\n\n\t// execute\n\terr := exec.Command(\"git\", \"push\", remote, branch).Run()\n\tcheckErr(err)\n\n\t// outputs\n\tfmt.Println(\"remote:\" + remote)\n\tfmt.Println(\"branch:\" + branch)\n\treturn nil\n}", "func (mvccs *KVMVCCStore) Commit(req *types.ReqHash) ([]byte, error) {\n\tbeg := types.Now()\n\tdefer func() {\n\t\tkmlog.Info(\"kvmvcc Commit\", \"cost\", types.Since(beg))\n\t}()\n\t_, ok := mvccs.kvsetmap[string(req.Hash)]\n\tif !ok {\n\t\tkmlog.Error(\"store kvmvcc commit\", \"err\", types.ErrHashNotFound)\n\t\treturn nil, types.ErrHashNotFound\n\t}\n\t//kmlog.Debug(\"KVMVCCStore Commit saveKVSets\", \"hash\", common.ToHex(req.Hash))\n\tmvccs.saveKVSets(mvccs.kvsetmap[string(req.Hash)], mvccs.sync)\n\tdelete(mvccs.kvsetmap, string(req.Hash))\n\treturn req.Hash, nil\n}", "func Commit(db interfaces.DB) error {\n\tif v, ok := db.(interface{ Commit() error }); ok {\n\t\treturn v.Commit()\n\t}\n\treturn errors.New(\"db does not implement commit\")\n}", "func committish(r *build.Rule) string {\n\tcommit := r.AttrString(\"commit\")\n\tif commit != \"\" {\n\t\treturn commit\n\t}\n\treturn r.AttrString(\"tag\")\n}" ]
[ "0.7253733", "0.7246724", "0.72047913", "0.71192944", "0.71116143", "0.6984826", "0.69353336", "0.6924362", "0.68456167", "0.6838363", "0.6761796", "0.67052233", "0.6679999", "0.6579832", "0.65475893", "0.64579743", "0.6450618", "0.6381569", "0.63741416", "0.63672525", "0.6364515", "0.6362611", "0.63328266", "0.6303566", "0.62867254", "0.6268056", "0.62559026", "0.62540424", "0.6251658", "0.6249708", "0.6249549", "0.6240909", "0.6231496", "0.622257", "0.6219305", "0.6218884", "0.6213044", "0.62088287", "0.6193537", "0.6193537", "0.61770517", "0.61685044", "0.61685044", "0.6167041", "0.61585695", "0.61580133", "0.61542696", "0.6151585", "0.6134057", "0.61324316", "0.61187184", "0.61079884", "0.6104068", "0.6093633", "0.6074855", "0.60739285", "0.6072295", "0.6061592", "0.6061274", "0.6058391", "0.604631", "0.60270315", "0.6014425", "0.599887", "0.5972831", "0.59713334", "0.59549904", "0.59512115", "0.5942802", "0.5923833", "0.59173906", "0.5916047", "0.58954686", "0.5895201", "0.58906186", "0.5889267", "0.5889148", "0.5878622", "0.5862201", "0.5845443", "0.58361477", "0.5831402", "0.58309686", "0.58235973", "0.5822297", "0.581741", "0.58160514", "0.58107567", "0.5810313", "0.58033454", "0.57975864", "0.57928747", "0.57928747", "0.5792052", "0.57899266", "0.57847273", "0.57798153", "0.57765293", "0.577506", "0.5772071" ]
0.64495957
17
GitPush Push latest commits
func Push() error { if Do("push") > 0 { return fmt.Errorf("Unable to push commits.") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g Git) Push(branch string) ([]byte, error) {\n\tif branch == \"\" {\n\t\tbranch = \"master\"\n\t}\n\tcmd := exec.Command(\"git\", \"push\", \"-f\", \"-q\", g.Repo, branch)\n\tcmd.Dir = g.Work\n\treturn cmd.CombinedOutput()\n}", "func gitPush(c *cli.Context) error {\n\t// inputs\n\tremote, branch := c.Args().Get(0), c.Args().Get(1)\n\texpected(remote, branch)\n\n\t// execute\n\terr := exec.Command(\"git\", \"push\", remote, branch).Run()\n\tcheckErr(err)\n\n\t// outputs\n\tfmt.Println(\"remote:\" + remote)\n\tfmt.Println(\"branch:\" + branch)\n\treturn nil\n}", "func GitCommitPush(diretorio string, mensagemCommit string, username string, password string) error {\n\n\t// Abrir um repositório já existente\n\tr, err := git.PlainOpen(diretorio)\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"git.PlainOpen\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\tw, err := r.Worktree()\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.Worktree\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// Adds the new file to the staging area.\n\tlogger.Info.Println(\"git add .\")\n\t_, err = w.Add(\".\")\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Add\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// We can verify the current status of the worktree using the method Status.\n\tlogger.Info.Println(\"git status --porcelain\")\n\tstatus, err := w.Status()\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Status\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\tlogger.Info.Println(status)\n\n\t// Commit dos arquivos\n\tlogger.Info.Printf(\"git commit -m %s\\n\\r\", mensagemCommit)\n\tcommit, err := w.Commit(mensagemCommit, &git.CommitOptions{\n\t\tAuthor: &object.Signature{\n\t\t\tName: \"Marcelo Melo\",\n\t\t\tEmail: \"[email protected]\",\n\t\t\tWhen: time.Now(),\n\t\t},\n\t})\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"w.Commit\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\n\t// Prints the current HEAD to verify that all worked well.\n\tlogger.Info.Println(\"git show -s\")\n\tobj, err := r.CommitObject(commit)\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.CommitObject\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\tlogger.Info.Println(obj)\n\n\tlogger.Info.Println(\"git push\")\n\terr = r.Push(&git.PushOptions{\n\t\tRemoteName: \"origin\",\n\t\tProgress: os.Stdout,\n\t\tAuth: &githttp.BasicAuth{\n\t\t\tUsername: username,\n\t\t\tPassword: password,\n\t\t}})\n\tif err != nil {\n\t\tmensagem := fmt.Sprintf(\"%s: %s\", \"r.Push\", err)\n\t\tlogger.Erro.Println(mensagem)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (t TestRepo) Push(name string, refs ...string) {\n\tif len(refs) == 0 {\n\t\trefs = []string{\"refs/heads/master\"}\n\t}\n\terr := t.remote(name).Push(refs, nil)\n\tCheckFatal(t.test, err)\n}", "func (b *Branch) Push() {\n\tpush := exec.Command(\"git\", \"push\", \"--set-upstream\", \"origin\", \"build/version-bump\")\n\tpush.Dir = b.Repo\n\terr := push.Run()\n\tif err != nil {\n\t\tlog.Fatalf(\"an error occured while pushing changes: %v\", err)\n\t}\n}", "func push() {\n\terr := actionutil.GitCommitPush(\n\t\t[]string{action, dockerfile},\n\t\t\"Update action files\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func (d *DeploymentCoreStruct) GitPush(force bool) (_ error) {\n\treturn d.RunInContext(func() (err error) {\n\t\tif d.syncStatus == -2 {\n\t\t\treturn fmt.Errorf(\"Unable to push to an inexistent remote\")\n\t\t}\n\t\tif d.syncStatus == 0 {\n\t\t\treturn fmt.Errorf(\"Unable to push. You need to sync up before\")\n\t\t}\n\t\tpush := make([]string, 1, 4)\n\t\tpush[0] = \"push\"\n\t\tif force {\n\t\t\tpush = append(push, \"-f\")\n\t\t}\n\t\tif d.syncStatus == -1 {\n\t\t\tpush = append(push, \"-u\")\n\t\t\tpush = append(push, strings.Split(d.syncRemoteBranch, \"/\")...)\n\t\t}\n\t\tif git.Do(push...) != 0 {\n\t\t\terr = fmt.Errorf(\"Unable to push\")\n\t\t} else {\n\t\t\td.syncStatus = 1\n\t\t}\n\t\treturn\n\t})\n}", "func (sd *stackediff) syncCommitStackToGitHub(ctx context.Context,\n\tcommits []git.Commit, info *github.GitHubInfo) {\n\n\tvar output string\n\tsd.mustgit(\"status --porcelain --untracked-files=no\", &output)\n\tif output != \"\" {\n\t\tsd.mustgit(\"stash\", nil)\n\t\tdefer sd.mustgit(\"stash pop\", nil)\n\t}\n\n\tcommitUpdated := func(c git.Commit, info *github.GitHubInfo) bool {\n\t\tfor _, pr := range info.PullRequests {\n\t\t\tif pr.Commit.CommitID == c.CommitID {\n\t\t\t\treturn pr.Commit.CommitHash != c.CommitHash\n\t\t\t}\n\t\t}\n\t\treturn true\n\t}\n\n\tvar updatedCommits []git.Commit\n\tfor _, commit := range commits {\n\t\tif commit.WIP {\n\t\t\tbreak\n\t\t}\n\t\tif commitUpdated(commit, info) {\n\t\t\tupdatedCommits = append(updatedCommits, commit)\n\t\t}\n\t}\n\n\tvar refNames []string\n\tfor _, commit := range updatedCommits {\n\t\tbranchName := sd.branchNameFromCommit(info, commit)\n\t\trefNames = append(refNames,\n\t\t\tcommit.CommitHash+\":refs/heads/\"+branchName)\n\t}\n\tif len(updatedCommits) > 0 {\n\t\tsd.mustgit(\"push --force --atomic origin \"+strings.Join(refNames, \" \"), nil)\n\t}\n\tsd.profiletimer.Step(\"SyncCommitStack::PushBranches\")\n}", "func PushAll(dir string, project string, branch string) error {\n\tgithubURL := fmt.Sprintf(\"[email protected]:%s.git\", project)\n\treturn command(dir, \"git\", \"push\", \"-fu\", githubURL, branch)\n}", "func pushCommit(ref *github.Reference, tree *github.Tree) (err error) {\n\t// Get the parent commit to attach the commit to.\n\tparent, _, err := client.Repositories.GetCommit(ctx, *data.sourceOwner, *data.sourceRepo, *ref.Object.SHA)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// This is not always populated, but is needed.\n\tparent.Commit.SHA = parent.SHA\n\n\t// Create the commit using the tree.\n\tdate := time.Now()\n\tauthorEmail := data.authorEmail\n\tauthor := &github.CommitAuthor{Date: &date, Name: data.authorName, Email: authorEmail}\n\tcommit := &github.Commit{Author: author, Message: data.commitMessage, Tree: tree, Parents: []*github.Commit{parent.Commit}}\n\tnewCommit, _, err := client.Git.CreateCommit(ctx, *data.sourceOwner, *data.sourceRepo, commit)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Attach the commit to the master branch.\n\tref.Object.SHA = newCommit.SHA\n\t_, _, err = client.Git.UpdateRef(ctx, *data.sourceOwner, *data.sourceRepo, ref, false)\n\treturn err\n}", "func (a *RepoAPI) push(params interface{}) (resp *rpc.Response) {\n\tm := objx.New(cast.ToStringMap(params))\n\tkeyOrPushToken := m.Get(\"privateKeyOrPushToken\").Str()\n\tpushParams := m.Get(\"params\").MSI()\n\treturn rpc.Success(util.Map{\n\t\t\"data\": a.mods.Repo.Push(pushParams, keyOrPushToken),\n\t})\n}", "func (db *DB) Push(url, ref string) error {\n\tif ref == \"\" {\n\t\tref = db.ref\n\t}\n\t// The '+' prefix sets force=true,\n\t// so the remote ref is created if it doesn't exist.\n\trefspec := fmt.Sprintf(\"+%s:%s\", db.ref, ref)\n\tremote, err := db.repo.CreateAnonymousRemote(url, refspec)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer remote.Free()\n\tpush, err := remote.NewPush()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"git_push_new: %v\", err)\n\t}\n\tdefer push.Free()\n\tif err := push.AddRefspec(refspec); err != nil {\n\t\treturn fmt.Errorf(\"git_push_refspec_add: %v\", err)\n\t}\n\tif err := push.Finish(); err != nil {\n\t\treturn fmt.Errorf(\"git_push_finish: %v\", err)\n\t}\n\treturn nil\n}", "func (c *config) push(remote string, branch string) (output string, err error) {\n\tlog.Printf(\"pushing: %v/%v\\n\", remote, branch)\n\n\tdefaultCommand := []string{\"push\", \"--set-upstream\", remote, branch}\n\n\tif c.useForce {\n\t\tdefaultCommand = append(defaultCommand, \"--force\")\n\t}\n\n\tif c.useTags {\n\t\tdefaultCommand = append(defaultCommand, \"--tags\")\n\t}\n\n\treturn c.command(defaultCommand...)\n}", "func pushUpstream(name string, remote string, giturl string, u interface{}, p interface{}) error {\n\t// Build the local repository path base used for input.\n\tpathBase := path.Clean(viper.Get(\"path\").(string))\n\tin := path.Join(pathBase, name)\n\n\tfmt.Printf(\"Pushing latest %s to remote \\\"%s\\\"\\n\", name, remote)\n\t// Open local git repository.\n\tr, err := git.PlainOpen(in)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Create a named remote (github, gitlab, etc.)\n\t_, err = r.CreateRemote(&config.RemoteConfig{\n\t\tName: remote,\n\t\tURLs: []string{giturl},\n\t})\n\t// Check if the remote was already present. If it was\n\t// remove it and recreate it to ensure integrity.\n\tif err != nil && !strings.Contains(err.Error(), \"remote already exists\") {\n\t\treturn err\n\t} else if err != nil && strings.Contains(err.Error(), \"remote already exists\") {\n\t\terr = r.DeleteRemote(remote)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = r.CreateRemote(&config.RemoteConfig{\n\t\t\tName: remote,\n\t\t\tURLs: []string{giturl},\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Setup push options to be used in the actual push.\n\tpushOptions := git.PushOptions{\n\t\tRemoteName: remote,\n\t\tProgress: os.Stdout,\n\t\tForce: true,\n\t}\n\n\t// If credentials are necessary for the SCM push, add them\n\t// to the pushOptions variable.\n\tif u != nil && p != nil {\n\t\tpushOptions.Auth = &http.BasicAuth{\n\t\t\tUsername: u.(string),\n\t\t\tPassword: p.(string),\n\t\t}\n\t}\n\n\t// Push the code to the upstream SCM.\n\terr = r.Push(&pushOptions)\n\tif err != nil && err != git.NoErrAlreadyUpToDate {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"%s on remote \\\"%s\\\" up to date\\n\", name, remote)\n\n\treturn nil\n}", "func Push() error {\n\tdefer logging.Logger.Sync()\n\tlogging.Logger.Info(\"Open local repository...\")\n\trepository, err := git.PlainOpen(config.Core.Workdir)\n\tif err != nil {\n\t\treturn err\n\t}\n\tlogging.Logger.Debug(\"Get Auth instance...\")\n\tauth, err := GetGitAuth()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlogging.Logger.Info(\"Pushing commits...\")\n\tif err := repository.Push(&git.PushOptions{\n\t\tRemoteName: \"origin\",\n\t\tAuth: auth,\n\t\tProgress: os.Stdout,\n\t}); err != nil {\n\t\treturn err\n\t}\n\n\tlogging.Logger.Info(\"Pushing Tags...\")\n\treturn repository.Push(&git.PushOptions{\n\t\tRemoteName: \"origin\",\n\t\tProgress: os.Stdout,\n\t\tRefSpecs: []gitConfig.RefSpec{\"refs/tags/*:refs/tags/*\"},\n\t\tAuth: auth,\n\t})\n}", "func commit(srvChan chan string, channel, nick, hostname string, args []string) {\n\ttype repoJSON struct {\n\t\tId int\n\t\tOwner map[string]interface{}\n\t\tName string\n\t\tFull_name string\n\t\tDescription string\n\t\tPrivate bool\n\t\tFork bool\n\t\tUrl string\n\t\tHtml_url string\n\t}\n\ttype commitJSON struct {\n\t\tSha string\n\t\tCommit map[string]interface{}\n\t\tUrl string\n\t\tHtml_url string\n\t\tComments_url string\n\t\tAuthor map[string]interface{}\n\t\tCommitter map[string]interface{}\n\t\tParents map[string]interface{}\n\t}\n\ttype urlJSON struct {\n\t\tKind string\n\t\tId string\n\t\tLongUrl string\n\t}\n\tmessage := \"NOTICE \" + channel + \" :\"\n\tsince := rand.Intn(1000000)\n\tres, err := http.Get(\"https://api.github.com/repositories?since=\" + fmt.Sprintf(\"%d\", since))\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tbody, err := ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tvar repos []repoJSON\n\tjson.Unmarshal(body, &repos)\n\tfullName := repos[rand.Intn(len(repos))].Full_name\n\tres, err = http.Get(\"https://api.github.com/repos/\" + fullName + \"/commits\")\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tbody, err = ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn\n\t}\n\tvar commits []commitJSON\n\tjson.Unmarshal(body, &commits)\n\tif len(commits) < 1 {\n\t\tcommit(srvChan, channel, nick, hostname, args) //try again\n\t\treturn\n\t} else {\n\t\tcommitNum := rand.Intn(len(commits))\n\t\tcommitMsg := commits[commitNum].Commit[\"message\"].(string)\n\n\t\tAPIkey, err := ioutil.ReadFile(\"APIkey\")\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\turlReader := strings.NewReader(`{\"longUrl\": \"` + commits[commitNum].Html_url + `\"}`)\n\t\tc := http.Client{}\n\t\tres, err := c.Post(\"https://www.googleapis.com/urlshortener/v1/url?key=\" + string(APIkey), \"application/json\", urlReader)\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tbody, err := ioutil.ReadAll(res.Body)\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tvar googUrl urlJSON\n\t\tjson.Unmarshal(body, &googUrl)\n\t\tmessage += strings.Split(commitMsg, \"\\n\")[0] + \" | \" + googUrl.Id\n\t}\n\tsrvChan <- message\n\tlog.Println(message)\n}", "func push(ctx context.Context, path string, repo *gogit.Repository, branch string, access repoAccess, impl string) error {\n\tswitch impl {\n\tcase sourcev1.LibGit2Implementation:\n\t\tlg2repo, err := libgit2.OpenRepository(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn pushLibgit2(lg2repo, access, branch)\n\tcase sourcev1.GoGitImplementation:\n\t\treturn pushGoGit(ctx, repo, access, branch)\n\tdefault:\n\t\treturn fmt.Errorf(\"unknown git implementation %q\", impl)\n\t}\n}", "func (git *Git) Push(remote, branch string) error {\n\tout, err := git.Exec(\"push\", \"-u\", remote, branch)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"Cannot push changes to %s/%s at '%s': %s\", remote, branch, git.canonRoot(), out)\n\t}\n\treturn nil\n}", "func retagAndPushForGH(tags []string) {\n\tpkgRepo := mageutil.RequireEnv(envGHPackageRepo)\n\treg := regexp.MustCompile(`.*\\:`)\n\tfor _, t := range tags {\n\t\ttag := strings.TrimSpace(t)\n\t\tupdatedTag := reg.ReplaceAllString(tag, fmt.Sprintf(\"%s:\", pkgRepo))\n\t\tfullGHTag := fmt.Sprintf(\"%s/%s\", ghPackagesRegistry, strings.TrimSpace(updatedTag))\n\t\tdockerTag(tag, fullGHTag)\n\t\tfmt.Printf(\"- Pushing image %s\\n\", fullGHTag)\n\t\tdockerutil.Push(fullGHTag).WithCfg(rootBuildDir).ExecVPanic()\n\t}\n}", "func (s *GitService) Push(ctx context.Context, r *git.Repository) error {\n\n\toptions := &git.PushOptions{\n\t\tRemoteName: \"origin\",\n\t\tAuth: &githttp.BasicAuth{Username: s.Client.username, Password: s.Client.token},\n\t}\n\n\terr := r.PushContext(ctx, options)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to push to remote: %w\", err)\n\t}\n\n\treturn nil\n}", "func push() {\n\terr := actionutil.GitConfig(\"goreadme\", email)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\terr = actionutil.GitCommitPush([]string{path}, \"Update readme according to godoc\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func (o TriggerGithubOutput) Push() TriggerGithubPushPtrOutput {\n\treturn o.ApplyT(func(v TriggerGithub) *TriggerGithubPush { return v.Push }).(TriggerGithubPushPtrOutput)\n}", "func (r Repository) Push(localRef, remoteRef string, opt *PushOptions) error {\n\tif opt == nil {\n\t\topt = &PushOptions{}\n\t}\n\tif opt.Remote == \"\" {\n\t\topt.Remote = \"origin\"\n\t}\n\turl, err := r.run(nil, opt.Timeout, \"remote\", \"get-url\", opt.Remote)\n\tif err != nil {\n\t\treturn err\n\t}\n\turl, err = opt.Credentials.addToURL(url)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif _, err := r.run(nil, opt.Timeout, \"push\", url, localRef+\":\"+remoteRef); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c Builds) Push(id uint, branch string) revel.Result {\n\tif branch == \"\" {\n\t\tbranch = \"updates\"\n\t}\n\tcurrentUser := connected(c.RenderArgs, c.Session)\n\tif currentUser == nil {\n\t\tc.Flash.Error(dontPerm)\n\t\treturn c.Redirect(\"/builds\")\n\t}\n\tvar build models.BuildedPackage\n\n\trevel.INFO.Printf(\"Build ID: %d\", id)\n\trevel.INFO.Printf(\"Branch name: %s\", branch)\n\tctx := dbgorm.Db.First(&build, \"build_id=?\", id)\n\tif ctx.Error != nil {\n\t\tc.Flash.Error(\"Error in fetch build by id: %s\", ctx.Error)\n\t\treturn c.Redirect(\"/builds\")\n\t}\n\n\tdbgorm.Db.Model(&build).Related(&build.BuildPackage, \"BuildPackage\")\n\tdbgorm.Db.Model(&build.BuildPackage).Related(&build.BuildPackage.PkgRepo, \"PkgRepo\")\n\tdbgorm.Db.Model(&build).Related(&build.Owner, \"Owner\")\n\tdbgorm.Db.Model(&build).Related(&build.User, \"PushUser\")\n\tdbgorm.Db.Model(&build).Related(&build.PushRepoType, \"PushRepoType\")\n\n\tif (build.Owner.OwnerName != currentUser.UserName) && (currentUser.UserGroup < models.GroupPusher) {\n\t\tc.Flash.Error(dontPerm)\n\t\treturn c.Redirect(\"/builds\")\n\t}\n\n\tvar branchRT models.RepoType\n\tctx = dbgorm.Db.First(&branchRT, \"rt_name=?\", branch)\n\tif ctx.Error != nil {\n\t\trevel.ERROR.Printf(\"Error fetch branch by name: %s\", ctx.Error)\n\t\tc.Flash.Error(ctx.Error.Error())\n\t\treturn c.Redirect(\"/builds\")\n\t}\n\n\tpush := models.PackagesToPush{}\n\tpush.Fill(build, branch)\n\tctx = dbgorm.Db.Create(&push)\n\tif ctx.Error != nil {\n\t\trevel.ERROR.Printf(\"Error in create push: %s\", ctx.Error)\n\t\tc.Flash.Error(ctx.Error.Error())\n\t\treturn c.Redirect(\"/builds\")\n\t}\n\n\tbuild.Pushed = true\n\tbuild.PushUser = *currentUser\n\tbuild.PushUserID = currentUser.ID\n\tbuild.PushRepoType = branchRT\n\tbuild.PushRepoTypeID = branchRT.ID\n\tbuild.User = *currentUser\n\tctx = dbgorm.Db.Save(&build)\n\tif ctx.Error != nil {\n\t\trevel.ERROR.Printf(\"Error in save build package: %s\", ctx.Error)\n\t\tc.Flash.Error(ctx.Error.Error())\n\t\treturn c.Redirect(\"/builds\")\n\t}\n\n\treturn c.Redirect(\"/builds\")\n}", "func (g *GitLocal) Push(dir string, remote string, force bool, refspec ...string) error {\n\treturn g.GitFake.Push(dir, \"origin\", false)\n}", "func storeCommits(ctx context.Context, db *pgxpool.Pool, lastCommitID int64, commits []*vcsinfo.LongCommit) error {\n\tctx, span := trace.StartSpan(ctx, \"gitilesfollower_storeCommits\")\n\tdefer span.End()\n\tcommitID := lastCommitID + 1\n\t// batchSize is only really relevant in the initial load. But we need it to avoid going over\n\t// the 65k limit of placeholder indexes.\n\tconst batchSize = 1000\n\tconst statement = `UPSERT INTO GitCommits (git_hash, commit_id, commit_time, author_email, subject) VALUES `\n\tconst valuesPerRow = 5\n\terr := util.ChunkIter(len(commits), batchSize, func(startIdx int, endIdx int) error {\n\t\tchunk := commits[startIdx:endIdx]\n\t\targuments := make([]interface{}, 0, len(chunk)*valuesPerRow)\n\t\tfor _, c := range chunk {\n\t\t\tcid := fmt.Sprintf(\"%012d\", commitID)\n\t\t\targuments = append(arguments, c.Hash, cid, c.Timestamp, c.Author, c.Subject)\n\t\t\tcommitID++\n\t\t}\n\t\tvp := sqlutil.ValuesPlaceholders(valuesPerRow, len(chunk))\n\t\tif _, err := db.Exec(ctx, statement+vp, arguments...); err != nil {\n\t\t\treturn skerr.Wrap(err)\n\t\t}\n\t\treturn nil\n\t})\n\treturn skerr.Wrap(err)\n}", "func (r *Repo) Push(key string, m proto.Message) error {\n\ttype_usl := reflect.TypeOf(m).String()\n\tvalue, err := proto.Marshal(m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tmsg := &msg.Sih{\n\t\tKey: key,\n\t\tValue: &msg.Any{\n\t\t\tTypeUrl: type_usl,\n\t\t\tValue: value,\n\t\t},\n\t}\n\tdata, err := proto.Marshal(msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\tn, err := r.Write(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = binary.Write(r, binary.LittleEndian, int32(n))\n\treturn err\n}", "func (o TriggerGithubPtrOutput) Push() TriggerGithubPushPtrOutput {\n\treturn o.ApplyT(func(v *TriggerGithub) *TriggerGithubPush {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Push\n\t}).(TriggerGithubPushPtrOutput)\n}", "func Push(pushFile, pkg string) error {\n\t_, err := os.Stat(pkg)\n\tif os.IsNotExist(err) {\n\t\treturn fmt.Errorf(\"couldn't find package %s\", pkg)\n\t}\n\n\tdata, err := ioutil.ReadFile(pushFile)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"couldn't read push file %s: %s\", pushFile, err)\n\t}\n\n\tlines := strings.Split(string(data), \"\\n\")\n\tif len(lines) == 0 || strings.TrimSpace(string(data)) == \"\" {\n\t\treturn fmt.Errorf(\"no targets to push to\")\n\t}\n\n\tsucceeded := 0\n\tattempts := 0\n\tfor _, line := range lines {\n\t\tline = strings.TrimSpace(line)\n\t\tif line == \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tattempts++\n\t\tbits := strings.Split(line, \" \")\n\t\txctbl := bits[0]\n\t\targs := bits[1:]\n\t\txctbl, found := util.FindExec(xctbl)\n\t\tif !found {\n\t\t\ttell.Errorf(\"couldn't find location of %s\", xctbl)\n\t\t\tcontinue\n\t\t}\n\n\t\t// check for inline string replacement in the args\n\t\tinline := false\n\t\tfor i, arg := range args {\n\t\t\tif strings.Contains(arg, \"$PKG\") {\n\t\t\t\targs[i] = strings.Replace(arg, \"$PKG\", pkg, -1)\n\t\t\t\tinline = true\n\t\t\t}\n\t\t}\n\n\t\t// if none inline, then make it the last argument\n\t\tif !inline {\n\t\t\targs = append(args, pkg)\n\t\t}\n\n\t\tcmd := exec.Command(xctbl, args...)\n\t\tcmd.Stdout = os.Stdout\n\t\tcmd.Stderr = os.Stderr\n\t\terr = cmd.Run()\n\t\ttell.IfErrorf(err, \"running %s %+v failed\", xctbl, args)\n\t\tif err == nil {\n\t\t\tsucceeded++\n\t\t}\n\t}\n\n\tif attempts != succeeded {\n\t\ttell.Errorf(\"pushed to %d / %d targets\", succeeded, attempts)\n\t\treturn fmt.Errorf(\"some lines failed to execute\")\n\t}\n\n\ttell.Infof(\"pushed to %d / %d targets\", succeeded, attempts)\n\treturn nil\n}", "func (d *Docker) Push(uri string) error {\n\tcommand := strings.Join([]string{\n\t\t\"push\", uri,\n\t}, \" \")\n\treturn d.exec(command)\n}", "func commitLocalChanges(config *GitXargsConfig, worktree *git.Worktree, remoteRepository *github.Repository, localRepository *git.Repository) error {\n\n\tlogger := logging.GetLogger(\"git-xargs\")\n\n\t// With all our untracked files staged, we can now create a commit, passing the All\n\t// option when configuring our commit option so that all modified and deleted files\n\t// will have their changes committed\n\tcommitOps := &git.CommitOptions{\n\t\tAll: true,\n\t}\n\n\t_, commitErr := worktree.Commit(config.CommitMessage, commitOps)\n\n\tif commitErr != nil {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Error\": commitErr,\n\t\t\t\"Repo\": remoteRepository.GetName(),\n\t\t})\n\n\t\t// If we reach this point, we were unable to commit our changes, so we'll\n\t\t// continue rather than attempt to push an empty branch and open an empty PR\n\t\tconfig.Stats.TrackSingle(CommitChangesFailed, remoteRepository)\n\t\treturn errors.WithStackTrace(commitErr)\n\t}\n\n\t// If --skip-pull-requests was passed, track the repos whose changes were committed directly to the main branch\n\tif config.SkipPullRequests {\n\t\tconfig.Stats.TrackSingle(CommitsMadeDirectlyToBranch, remoteRepository)\n\t}\n\n\treturn nil\n}", "func (d *Release) Push(only ...string) error {\n\tif _ = d.merge(); len(d.src) == 0 {\n\t\treturn ErrMissing\n\t}\n\tif d.rebase(only); len(d.src) == 0 {\n\t\treturn ErrMissing\n\t}\n\tvar g errgroup.Group\n\tfor _, server := range d.to {\n\t\tc := server\n\t\tg.Go(func() error {\n\t\t\treturn c.Bulk(d.src)\n\t\t})\n\t}\n\td.err = g.Wait()\n\treturn d.err\n}", "func (g *Git) Push(ctx context.Context, file []byte, filePath string) error {\n\t// when multiple resources are created we need to update the git refs head after push\n\tg.mu.Lock()\n\tdefer g.mu.Unlock()\n\n\tnewFile, err := g.fs.Create(filePath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to create file: %w\", err)\n\t}\n\n\t_, err = newFile.Write(file)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to write to file: %w\", err)\n\t}\n\terr = newFile.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tw, err := g.repo.Worktree()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = w.Add(filePath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to add: %w\", err)\n\t}\n\t_, err = w.Commit(createCommitMsg(\"created\", filePath), commitOpts())\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to commit: %w\", err)\n\t}\n\n\tif err := g.repo.FetchContext(ctx, &git.FetchOptions{RemoteName: remoteName, Auth: g.auth}); err != nil && !errors.Is(err, git.NoErrAlreadyUpToDate) {\n\t\treturn fmt.Errorf(\"unable to fetch: %w\", err)\n\t}\n\n\tif err := g.repo.PushContext(ctx, &git.PushOptions{RemoteName: remoteName, Auth: g.auth, Force: true}); err != nil {\n\t\treturn fmt.Errorf(\"unable to push: %w\", err)\n\t}\n\n\treturn nil\n}", "func (repo *Workdir) Push() error {\n\tfor _, ext := range repo.Versions {\n\t\tlogrus.Infof(\"push extension to Erda from %s\", ext.Dirname)\n\t\tif err := ext.Push(); err != nil {\n\t\t\treturn errors.Wrapf(err, \"failed to Push: %s\", ext.Dirname)\n\t\t}\n\t}\n\n\treturn nil\n}", "func PushRepo() bool {\n\t// var cmdOut []byte\n\tsuccess := false\n\tvar err error\n\n\tcmdArgs := []string{\"push\"}\n\tif _, err = exec.Command(\"git\", cmdArgs...).Output(); err != nil {\n\t\tfmt.Fprintln(os.Stderr, \"There was an error staging all changes\", err)\n\t\treturn success\n\t}\n\tsuccess = true\n\treturn success\n}", "func (g *Git) Push(ctx context.Context, remoteName string, force bool) error {\n\treturn g.repo.PushContext(ctx, &git.PushOptions{\n\t\tRemoteName: remoteName,\n\t\tForce: force,\n\t})\n}", "func (self *WtfPush) Push(blob []byte) {\n\tselect {\n\tcase self.pushch <- blob:\n\tdefault:\n\t}\n}", "func GitCommit() string { return gitCommit }", "func TestIntegrtionForGitHooks(t *testing.T) {\n\tassert := assert.New(t)\n\n\tserver, requestsChan := helpers.CreateRequestRecorder(t)\n\n\tupstream := setupBareGitRepo(t)\n\tdefer helpers.CleanupRepository(t, upstream.Path)\n\n\tcfgDir, cfg := setupConfig(t, upstream)\n\tdefer os.RemoveAll(cfgDir)\n\n\thook := setupStore(t, server.URL, &cfg)\n\n\tassert.Nil(upstream.InstallHooks(filepath.Join(cfgDir, \"config.json\"), false))\n\n\trepo, gitRepo := helpers.CreateGitRepo(t, \"clone\")\n\tdefer helpers.CleanupRepository(t, repo.Path)\n\n\tfmt.Printf(\"Upstream: %s\\ncfgDir: %s\\nrepo: %s\\n\", upstream.Path, cfgDir, repo.Path)\n\n\t_, err := gitRepo.CreateRemote(&git_config.RemoteConfig{\n\t\tName: \"origin\",\n\t\tURLs: []string{upstream.Path},\n\t})\n\tassert.Nil(err)\n\n\tworktree, err := gitRepo.Worktree()\n\tassert.Nil(err)\n\n\tprogressBuffer := new(bytes.Buffer)\n\tpushOptions := &git.PushOptions{\n\t\tRemoteName: \"origin\",\n\t\tRefSpecs: []git_config.RefSpec{\"refs/heads/master:refs/heads/master\"},\n\t\tProgress: progressBuffer,\n\t}\n\n\torigHead, err := worktree.Commit(\"Initial commit\", &git.CommitOptions{\n\t\tAuthor: &object.Signature{\n\t\t\tName: \"Author\",\n\t\t\tEmail: \"[email protected]\",\n\t\t\tWhen: time.Now(),\n\t\t},\n\t})\n\n\tassert.Nil(gitRepo.Push(pushOptions))\n\tfmt.Printf(\"Response from first push:\\n %s\\n\", progressBuffer.String())\n\n\tprogressBuffer.Reset()\n\n\tnewHead, err := worktree.Commit(\"New commit\", &git.CommitOptions{\n\t\tAuthor: &object.Signature{\n\t\t\tName: \"Author\",\n\t\t\tEmail: \"[email protected]\",\n\t\t\tWhen: time.Now(),\n\t\t},\n\t})\n\n\tassert.Nil(gitRepo.Push(pushOptions))\n\tfmt.Printf(\"Response from second push:\\n %s\\n\", progressBuffer.String())\n\n\trequests := helpers.AssertNumRequests(t, 2, requestsChan)\n\n\tcases := []testCase{\n\t\t{\n\t\t\trecorded: &requests[0],\n\t\t\tmessage: \"Initial commit\",\n\t\t\tcommitId: origHead.String(),\n\t\t\ttarget: events.PushPayloadCommitTarget{\n\t\t\t\tBranch: \"master\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\trecorded: &requests[1],\n\t\t\tmessage: \"New commit\",\n\t\t\tcommitId: newHead.String(),\n\t\t\ttarget: events.PushPayloadCommitTarget{\n\t\t\t\tBranch: \"master\",\n\t\t\t},\n\t\t},\n\t}\n\n\trunTests(t, cases, upstream, hook)\n}", "func (s *TagStore) Push(localName string, imagePushConfig *ImagePushConfig) error {\n\tvar (\n\t\tlocalRepo Repository\n\t\tsf = streamformatter.NewJSONStreamFormatter()\n\t)\n\n\t// Resolve the Repository name from fqn to RepositoryInfo\n\trepoInfo, err := s.registryService.ResolveRepository(localName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// If we're not using a custom registry, we know the restrictions\n\t// applied to repository names and can warn the user in advance.\n\t// Custom repositories can have different rules, and we must also\n\t// allow pushing by image ID.\n\tif repoInfo.Official {\n\t\tusername := imagePushConfig.AuthConfig.Username\n\t\tif username == \"\" {\n\t\t\tusername = \"<user>\"\n\t\t}\n\t\tname := localName\n\t\tparts := strings.Split(repoInfo.LocalName, \"/\")\n\t\tif len(parts) > 0 {\n\t\t\tname = parts[len(parts)-1]\n\t\t}\n\t\treturn fmt.Errorf(\"You cannot push a \\\"root\\\" repository. Please rename your repository to <user>/<repo> (ex: %s/%s)\", username, name)\n\t}\n\n\tif repoInfo.Index.Official && s.ConfirmDefPush && !imagePushConfig.Force {\n\t\treturn fmt.Errorf(\"Error: Status 403 trying to push repository %s to official registry: needs to be forced\", localName)\n\t} else if repoInfo.Index.Official && !s.ConfirmDefPush && imagePushConfig.Force {\n\t\tlogrus.Infof(\"Push of %s to official registry has been forced\", localName)\n\t}\n\n\tendpoints, err := s.registryService.LookupPushEndpoints(repoInfo.CanonicalName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treposLen := 1\n\tif imagePushConfig.Tag == \"\" {\n\t\treposLen = len(s.Repositories[repoInfo.LocalName])\n\t}\n\n\timagePushConfig.OutStream.Write(sf.FormatStatus(\"\", \"The push refers to a repository [%s] (len: %d)\", repoInfo.CanonicalName, reposLen))\n\tmatching := s.getRepositoryList(localName)\nLoop:\n\tfor _, namedRepo := range matching {\n\t\tfor _, localRepo = range namedRepo {\n\t\t\tbreak Loop\n\t\t}\n\t}\n\tif localRepo == nil {\n\t\treturn fmt.Errorf(\"Repository does not exist: %s\", localName)\n\t}\n\n\tvar lastErr error\n\tfor _, endpoint := range endpoints {\n\t\tlogrus.Debugf(\"Trying to push %s to %s %s\", repoInfo.CanonicalName, endpoint.URL, endpoint.Version)\n\n\t\tpusher, err := s.NewPusher(endpoint, localRepo, repoInfo, imagePushConfig, sf)\n\t\tif err != nil {\n\t\t\tlastErr = err\n\t\t\tcontinue\n\t\t}\n\t\tif fallback, err := pusher.Push(); err != nil {\n\t\t\tif fallback {\n\t\t\t\tlastErr = err\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tlogrus.Debugf(\"Not continuing with error: %v\", err)\n\t\t\treturn err\n\n\t\t}\n\n\t\ts.eventsService.Log(\"push\", repoInfo.LocalName, \"\")\n\t\treturn nil\n\t}\n\n\tif lastErr == nil {\n\t\tlastErr = fmt.Errorf(\"no endpoints found for %s\", repoInfo.CanonicalName)\n\t}\n\treturn lastErr\n}", "func (c *Corpus) syncGitCommits(ctx context.Context, conf polledGitCommits, loop bool) error {\n\tcmd := exec.CommandContext(ctx, \"git\", \"show-ref\", \"refs/remotes/origin/master\")\n\tenvutil.SetDir(cmd, conf.dir)\n\tout, err := cmd.Output()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\touts := strings.TrimSpace(string(out))\n\tif outs == \"\" {\n\t\treturn fmt.Errorf(\"no remote found for refs/remotes/origin/master\")\n\t}\n\tref := strings.Fields(outs)[0]\n\tc.mu.Lock()\n\trefHash := c.gitHashFromHexStr(ref)\n\tc.enqueueCommitLocked(refHash)\n\tc.mu.Unlock()\n\n\tidle := false\n\tfor {\n\t\thash := c.gitCommitToIndex()\n\t\tif hash == \"\" {\n\t\t\tif !loop {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tif !idle {\n\t\t\t\tlog.Printf(\"All git commits index for %v; idle.\", conf.repo)\n\t\t\t\tidle = true\n\t\t\t}\n\t\t\ttime.Sleep(5 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tif err := c.indexCommit(conf, hash); err != nil {\n\t\t\tlog.Printf(\"Error indexing %v: %v\", hash, err)\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn ctx.Err()\n\t\t\t\t// TODO: temporary vs permanent failure? reschedule? fail hard?\n\t\t\t\t// For now just loop with a sleep.\n\t\t\tcase <-time.After(5 * time.Second):\n\t\t\t}\n\t\t}\n\t}\n}", "func (m *MockIExec) DoGitPush(dir string, args ...string) (string, error) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{dir}\n\tfor _, a := range args {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"DoGitPush\", varargs...)\n\tret0, _ := ret[0].(string)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *gitlabClient) PushEvents(context.Context, string, []interface{}) ([]sdk.VCSPushEvent, error) {\n\treturn nil, fmt.Errorf(\"Not implemented on Gitlab\")\n}", "func main() {\n\n/*\n\tCheckArgs(\"<repository-path>\")\n\tpath := os.Args[1]\n\tfmt.Println(path)\n\tr, err := git.PlainOpen(path)\n\tCheckIfError(err)\n\n\tInfo(\"git push\")\n\t// push using default options\n\terr = r.Push(&git.PushOptions{})\n\tCheckIfError(err)\n*/\n\topenFile()\n}", "func Push(repo Repo, target string, config dvid.Config) error {\n\tif target == \"\" {\n\t\ttarget = message.DefaultAddress\n\t\tdvid.Infof(\"No target specified for push, defaulting to %q\\n\", message.DefaultAddress)\n\t}\n\n\t// Get the push configuration\n\troiname, err := getROI(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdata, err := getDataInstances(repo, config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Establish connection with target, which may be itself\n\ts, err := message.NewPushSocket(target)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to create new push socket: %s\", err.Error())\n\t}\n\n\t// Send PUSH command start\n\tif err = s.SendCommand(CommandPushStart); err != nil {\n\t\treturn err\n\t}\n\n\t// Send the repo metadata\n\t// TODO -- add additional information indicating origin and push configuration\n\tdvid.Infof(\"Sending repo %s data to %q\\n\", repo.RootUUID(), target)\n\trepoSerialization, err := repo.GobEncode()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err = s.SendBinary(\"repo\", repoSerialization); err != nil {\n\t\treturn err\n\t}\n\n\t// For each data instance, send the data delimited by the roi\n\tfor _, instance := range data {\n\t\tdvid.Infof(\"Sending instance %q data to %q\\n\", instance.DataName(), target)\n\t\tif err := instance.Send(s, roiname, repo.RootUUID()); err != nil {\n\t\t\tdvid.Errorf(\"Aborting send of instance %q data\\n\", instance.DataName())\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Send PUSH command end\n\tdvid.Debugf(\"Sending PUSH STOP command to %q\\n\", target)\n\tif err = s.SendCommand(CommandPushStop); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (g *Gist) GetGitPushURL() string {\n\tif g == nil || g.GitPushURL == nil {\n\t\treturn \"\"\n\t}\n\treturn *g.GitPushURL\n}", "func buildPush(ctx context.Context, appName, wasmProductsDir, checkoutDir, skiaRevision string, topic *pubsub.Topic) error {\n\terr := td.Do(ctx, td.Props(\"Build \"+appName+\" image\").Infra(), func(ctx context.Context) error {\n\t\trunCmd := &sk_exec.Command{\n\t\t\tName: \"make\",\n\t\t\tArgs: []string{\"bazel_release_ci\"},\n\t\t\tInheritEnv: true,\n\t\t\tEnv: []string{\n\t\t\t\t\"COPY_FROM_DIR=\" + wasmProductsDir,\n\t\t\t\t\"STABLE_DOCKER_TAG=\" + skiaRevision,\n\t\t\t},\n\t\t\tDir: filepath.Join(checkoutDir, appName),\n\t\t\tLogStdout: true,\n\t\t\tLogStderr: true,\n\t\t}\n\t\t_, err := sk_exec.RunCommand(ctx, runCmd)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn docker.PublishToTopic(ctx, \"gcr.io/skia-public/\"+appName, skiaRevision, common.REPO_SKIA, topic)\n}", "func (fbo *folderBranchOps) PushStatusChange() {\n\tfbo.config.KBFSOps().PushStatusChange()\n}", "func (p *Pusher) Push(ctx context.Context) error {\n\tif p.PushFormat == \"\" {\n\t\tp.PushFormat = expfmt.FmtText\n\t}\n\n\tresps := make(chan (error))\n\tgo func() {\n\t\tresps <- p.push(ctx)\n\t}()\n\n\tselect {\n\tcase err := <-resps:\n\t\treturn err\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\t}\n}", "func (c *Client) PushChangeset(o Object, chng func(Object) error, opts *WriteOpts) error {\n\terr := chng(o)\n\tif err != nil {\n\t\treturn err\n\t}\n\tnmerge := 0\npush:\n\terr = c.Push(o, opts)\n\tif err == ErrModified {\n\t\tvar upd bool\n\t\tnmerge++\n\t\tif nmerge > maxMerges {\n\t\t\treturn fmt.Errorf(\"exceeded max merges: %s\", err)\n\t\t}\n\t\tupd, err = c.Update(o, nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif !upd {\n\t\t\treturn errors.New(\"failure updating...\")\n\t\t}\n\t\terr = chng(o)\n\t\tif err == ErrDone {\n\t\t\treturn nil\n\t\t}\n\t\tgoto push\n\t}\n\treturn err\n}", "func (s *GitService) commit(w *git.Worktree, r *git.Repository, c *CreateCommit) (*object.Commit, error) {\n\t// Commits the current staging area to the repository, with the new file\n\t// just created. We should provide the object.Signature of Author of the\n\t// gitClient Since version 5.0.1, we can omit the Author signature, being read\n\t// from the git config files.\n\tvar p []plumbing.Hash\n\tif c.Parents != nil && len(c.Parents) > 0 {\n\t\tp = make([]plumbing.Hash, len(c.Parents))\n\t}\n\tif p != nil && len(p) > 0 {\n\t\tfor i, parent := range c.Parents {\n\t\t\tcopy(p[i][:], parent)\n\t\t}\n\t}\n\n\t// calculate time.Time from unix Time\n\tauthorDate := time.Unix(c.Author.Date, 0)\n\tvar committer *object.Signature\n\tif c.Committer != nil {\n\t\tcommitterDate := time.Unix(c.Committer.Date, 0)\n\t\tcommitter = &object.Signature{\n\t\t\tName: c.Committer.Name,\n\t\t\tEmail: c.Committer.Email,\n\t\t\tWhen: committerDate,\n\t\t}\n\t} else {\n\t\tcommitter = &object.Signature{\n\t\t\tName: c.Author.Name,\n\t\t\tEmail: c.Author.Email,\n\t\t\tWhen: authorDate,\n\t\t}\n\t}\n\n\tcommitHash, err := w.Commit(c.Message, &git.CommitOptions{\n\t\tAuthor: &object.Signature{\n\t\t\tName: c.Author.Name,\n\t\t\tEmail: c.Author.Email,\n\t\t\tWhen: authorDate,\n\t\t},\n\t\tCommitter: committer,\n\t\tParents: p,\n\t\tSignKey: c.SignKey,\n\t\tAll: true,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tobj, err := r.CommitObject(commitHash)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn obj, nil\n}", "func (g *GitLocal) StashPush(dir string) error {\n\treturn g.GitCLI.StashPush(dir)\n}", "func TestCreatePushRemote(t *testing.T) {\n\tresetFlags()\n\tdefer resetFlags()\n\n\tfake, cleanup := jiritest.NewFakeJiriRoot(t)\n\tdefer cleanup()\n\n\tfake.EnableRemoteManifestPush()\n\tdefer fake.DisableRemoteManifestPush()\n\n\tmanifestDir := filepath.Join(fake.X.Root, \"manifest\")\n\tsnapshotDir := filepath.Join(manifestDir, \"snapshot\")\n\tlabel := \"test\"\n\n\tgit := gitutil.New(fake.X.NewSeq(), gitutil.RootDirOpt(manifestDir))\n\tcommitCount, err := git.CountCommits(\"master\", \"\")\n\tif err != nil {\n\t\tt.Fatalf(\"git.CountCommits(\\\"master\\\", \\\"\\\") failed: %v\", err)\n\t}\n\n\t// Create snapshot with -push-remote flag set to true.\n\tsnapshotDirFlag = snapshotDir\n\tpushRemoteFlag = true\n\tif err := runSnapshotCreate(fake.X, []string{label}); err != nil {\n\t\tt.Fatalf(\"%v\", err)\n\t}\n\n\t// Check that repo has one new commit.\n\tnewCommitCount, err := git.CountCommits(\"master\", \"\")\n\tif err != nil {\n\t\tt.Fatalf(\"git.CountCommits(\\\"master\\\", \\\"\\\") failed: %v\", err)\n\t}\n\tif got, want := newCommitCount, commitCount+1; got != want {\n\t\tt.Errorf(\"unexpected commit count: got %v want %v\", got, want)\n\t}\n\n\t// Check that new label is commited.\n\tlabelFile := filepath.Join(snapshotDir, \"labels\", label)\n\tif !git.IsFileCommitted(labelFile) {\n\t\tt.Errorf(\"expected file %v to be committed but it was not\", labelFile)\n\t}\n}", "func (wp *workPool) push(r *remote.RepoT) {\n\toop := *wp\n\toop = append(oop, r)\n\t*wp = oop\n}", "func (s *githubWebhookService) OnPush(ctx context.Context, req interface{}) (interface{}, error) {\n\tgithubPushEvent := req.(*github.PushEvent)\n\tpushEvent := MapGithubPushEventToVcsPushEvent(githubPushEvent)\n\ts.queue.Enqueue(pushEvent)\n\treturn nil, nil\n}", "func scanGitCommits(sess *session.Session, repo *gitprovider.Repository, clone *git.Repository, dir, checkpoint string, targetPathMap map[string]string) {\n\tcommitHistories, err := gitHandler.GetRepositoryHistory(clone)\n\tif err != nil {\n\t\tsess.Out.Error(\"[THREAD][%s] Error getting commit history: %s\\n\", repo.FullName, err)\n\t\treturn\n\t}\n\tsess.Out.Debug(\"[THREAD][%s] Number of commits: %d\\n\", repo.FullName, len(commitHistories))\n\n\tfor _, commit := range commitHistories {\n\t\tif strings.TrimSpace(commit.Hash.String()) == strings.TrimSpace(checkpoint) {\n\t\t\tsess.Out.Debug(\"\\nCheckpoint Reached !!\\n\")\n\t\t\tbreak\n\t\t}\n\t\tsess.Out.Debug(\"[THREAD][%s] Analyzing commit: %s\\n\", repo.FullName, commit.Hash)\n\t\tchanges, _ := gitHandler.GetChanges(commit, clone)\n\t\tsess.Out.Debug(\"[THREAD][%s] Changes in %s: %d\\n\", repo.FullName, commit.Hash, len(changes))\n\t\tfor _, change := range changes {\n\t\t\tp := gitHandler.GetChangePath(change)\n\n\t\t\t_, exists := targetPathMap[path.Join(dir, p)]\n\t\t\tif len(targetPathMap) > 0 && !exists {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tallContent := \"\"\n\t\t\tsess.Out.Debug(\"FILE: %s/%s\\n\", dir, p)\n\t\t\tsess.Out.Debug(\"Repo URL: %s/commit/%s\\n\", repo.URL, commit.Hash.String())\n\t\t\tpatch, _ := gitHandler.GetPatch(change)\n\t\t\tdiffs := patch.FilePatches()\n\t\t\tfor _, diff := range diffs {\n\t\t\t\tchunks := diff.Chunks()\n\t\t\t\tfor _, chunk := range chunks {\n\t\t\t\t\tif chunk.Type() == 1 {\n\t\t\t\t\t\tallContent += chunk.Content()\n\t\t\t\t\t\tallContent += \"\\n\\n\"\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tmatchFile := signatures.NewMatchFile(p, allContent)\n\t\t\tif matchFile.IsSkippable() {\n\t\t\t\tsess.Out.Debug(\"[THREAD][%s] Skipping %s\\n\", repo.FullName, matchFile.Path)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tisTestContext := matchFile.IsTestContext()\n\t\t\tif isTestContext && *sess.Options.SkipTestContexts {\n\t\t\t\tsess.Out.Debug(\"[THREAD][%s] Skipping %s\\n\", repo.FullName, matchFile.Path)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tsess.Out.Debug(\"[THREAD][%s] Matching: %s...\\n\", repo.FullName, matchFile.Path)\n\t\t\tfor _, signature := range sess.Signatures {\n\t\t\t\tmatches := signature.Match(matchFile)\n\n\t\t\t\tfor _, match := range matches {\n\t\t\t\t\tfinding := &findings.Finding{\n\t\t\t\t\t\tFilePath: p,\n\t\t\t\t\t\tAction: signature.Part(),\n\t\t\t\t\t\tDescription: signature.Description(),\n\t\t\t\t\t\tComment: signature.Comment(),\n\t\t\t\t\t\tRepositoryName: repo.Name,\n\t\t\t\t\t\tRepositoryURL: repo.URL,\n\t\t\t\t\t\tFileURL: fmt.Sprintf(\"%s/blob/%s/%s\", repo.URL, repo.DefaultBranch, p),\n\t\t\t\t\t\tLine: match.Line,\n\t\t\t\t\t\tIsTestContext: isTestContext,\n\t\t\t\t\t}\n\n\t\t\t\t\tif *sess.Options.LogSecret {\n\t\t\t\t\t\tfinding.LineContent = match.LineContent\n\t\t\t\t\t\tfinding.TruncateLineContent(findings.MaxLineChar)\n\t\t\t\t\t}\n\n\t\t\t\t\thashID, err := finding.GenerateHashID()\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tsess.Out.Error(\"Unable to generate hash ID for %v, skipping...\", finding.FileURL)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tfinding.ID = hashID\n\n\t\t\t\t\tsess.AddFinding(finding)\n\n\t\t\t\t\tsess.Out.Warn(\" %s: %s\\n\", strings.ToUpper(session.PathScan), finding.Description)\n\t\t\t\t\tsess.Out.Info(\" Path........: %s\\n\", finding.FilePath)\n\t\t\t\t\tsess.Out.Info(\" Repo........: %s\\n\", repo.FullName)\n\t\t\t\t\tsess.Out.Info(\" Author......: %s\\n\", finding.CommitAuthor)\n\t\t\t\t\tsess.Out.Info(\" Comment.....: %s\\n\", finding.Comment)\n\t\t\t\t\tsess.Out.Info(\" File URL....: %s\\n\", finding.FileURL)\n\t\t\t\t\tsess.Out.Info(\" Line........: %v\\n\", finding.Line)\n\t\t\t\t\tsess.Out.Info(\" ------------------------------------------------\\n\\n\")\n\t\t\t\t\tsess.Stats.IncrementFindings()\n\t\t\t\t}\n\n\t\t\t\t//if signature.Match(matchFile) {\n\t\t\t\t//\tlatestContent, err := ioutil.ReadFile(path.Join(dir, p))\n\t\t\t\t//\tif err != nil {\n\t\t\t\t//\t\tsess.Out.Info(\"[LATEST FILE NOT FOUND]: %s/%s\\n\", dir, p)\n\t\t\t\t//\t\tcontinue\n\t\t\t\t//\t}\n\t\t\t\t//\tmatchFile = signatures.NewMatchFile(p, string(latestContent))\n\t\t\t\t//\tif signature.Match(matchFile) {\n\t\t\t\t//\t\tfinding := &findings.Finding{\n\t\t\t\t//\t\t\tFilePath: p,\n\t\t\t\t//\t\t\tAction: session.ContentScan,\n\t\t\t\t//\t\t\tDescription: signature.Description(),\n\t\t\t\t//\t\t\tComment: signature.Comment(),\n\t\t\t\t//\t\t\tRepositoryName: repo.Name,\n\t\t\t\t//\t\t\tCommitHash: commit.Hash.String(),\n\t\t\t\t//\t\t\tCommitMessage: strings.TrimSpace(commit.Message),\n\t\t\t\t//\t\t\tCommitAuthor: commit.Author.String(),\n\t\t\t\t//\t\t\tRepositoryURL: repo.URL,\n\t\t\t\t//\t\t\tFileURL: fmt.Sprintf(\"%s/blob/%s/%s\", repo.URL, repo.DefaultBranch, p),\n\t\t\t\t//\t\t\tCommitURL: fmt.Sprintf(\"%s/commit/%s\", repo.URL, commit.Hash.String()),\n\t\t\t\t//\t\t\tIsTestContext: isTestContext,\n\t\t\t\t//\t\t}\n\t\t\t\t//\n\t\t\t\t//\t\thashID, err := finding.GenerateHashID()\n\t\t\t\t//\t\tif err != nil {\n\t\t\t\t//\t\t\tsess.Out.Error(\"Unable to generate hash ID for %v, skipping...\", finding.FileURL)\n\t\t\t\t//\t\t\tcontinue\n\t\t\t\t//\t\t}\n\t\t\t\t//\t\tfinding.ID = hashID\n\t\t\t\t//\n\t\t\t\t//\t\tsess.AddFinding(finding)\n\t\t\t\t//\n\t\t\t\t//\t\tsess.Out.Warn(\" %s: %s\\n\", strings.ToUpper(session.ContentScan), finding.Description)\n\t\t\t\t//\t\tsess.Out.Info(\" Path.......: %s\\n\", finding.FilePath)\n\t\t\t\t//\t\tsess.Out.Info(\" Repo.......: %s\\n\", repo.FullName)\n\t\t\t\t//\t\tsess.Out.Info(\" Message....: %s\\n\", TruncateString(finding.CommitMessage, 100))\n\t\t\t\t//\t\tsess.Out.Info(\" Author.....: %s\\n\", finding.CommitAuthor)\n\t\t\t\t//\t\tsess.Out.Info(\" Comment....: %s\\n\", finding.Comment)\n\t\t\t\t//\t\tsess.Out.Info(\" File URL...: %s\\n\", finding.FileURL)\n\t\t\t\t//\t\tsess.Out.Info(\" Commit URL.: %s\\n\", finding.CommitURL)\n\t\t\t\t//\t\tsess.Out.Info(\" ------------------------------------------------\\n\\n\")\n\t\t\t\t//\t\tsess.Stats.IncrementFindings()\n\t\t\t\t//\t}\n\t\t\t\t//}\n\t\t\t}\n\t\t\tsess.Stats.IncrementFiles()\n\t\t}\n\t\tsess.Stats.IncrementCommits()\n\t\tsess.Out.Debug(\"[THREAD][%s] Done analyzing changes in %s\\n\", repo.FullName, commit.Hash)\n\t}\n}", "func (t *BuildTree) Push() error {\n\terr := t.Build()\n\tif err != nil {\n\t\treturn err\n\t}\n\tutils.Info(\"Logging into registry\")\n\tfor _, credential := range t.credentials {\n\t\terr = utils.DockerLogin(credential.Registry, credential.Username, credential.Password)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tutils.Info(\"Pushing new images\")\n\tfor _, node := range t.rootNodes {\n\t\terr = t.pushNodeAndChildren(node)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func pullByCommitWithRollback(config *gctsDeployOptions, telemetryData *telemetry.CustomData, command command.ExecRunner,\n\thttpClient piperhttp.Sender, repoState string, repoMetadataInitState *getRepositoryResponseBody,\n\tcurrentBranch string, targetBranch string, branchRollbackRequired bool) error {\n\n\tlog.Entry().Infof(\"gCTS Deploy: Pull by Commit step execution to commit %v\", config.Commit)\n\tpullByCommitErr := pullByCommit(config, telemetryData, command, httpClient)\n\tif pullByCommitErr != nil {\n\t\tlog.Entry().WithError(pullByCommitErr).Error(\"step execution failed at Pull By Commit. Trying to rollback to last commit\")\n\t\tif config.Rollback {\n\t\t\t//Rollback to last commit.\n\t\t\trollbackOptions := gctsRollbackOptions{\n\t\t\t\tUsername: config.Username,\n\t\t\t\tPassword: config.Password,\n\t\t\t\tRepository: config.Repository,\n\t\t\t\tHost: config.Host,\n\t\t\t\tClient: config.Client,\n\t\t\t\tSkipSSLVerification: config.SkipSSLVerification,\n\t\t\t}\n\t\t\trollbackErr := rollback(&rollbackOptions, telemetryData, command, httpClient)\n\t\t\tif rollbackErr != nil {\n\t\t\t\tlog.Entry().WithError(rollbackErr).Error(\"step execution failed while rolling back commit\")\n\t\t\t\treturn rollbackErr\n\t\t\t}\n\t\t\tif repoState == repoStateNew && branchRollbackRequired {\n\t\t\t\t// Rollback branch\n\t\t\t\t// Rollback branch. Resetting branches\n\t\t\t\ttargetBranch = repoMetadataInitState.Result.Branch\n\t\t\t\tcurrentBranch = config.Branch\n\t\t\t\tlog.Entry().Errorf(\"Rolling Back from %v to %v\", currentBranch, targetBranch)\n\t\t\t\tswitchBranch(config, httpClient, currentBranch, targetBranch)\n\t\t\t}\n\t\t}\n\t\treturn pullByCommitErr\n\t}\n\treturn nil\n\n}", "func (b *binding) Push(ctx context.Context, local, remote string) error {\n\treturn b.Command(\"push\", local, remote).Run(ctx)\n}", "func (hg *Hg) Commits() []string {\n\tout, err := Run(hg.Dir, \"hg\", \"incoming\", \"-n\", \"-q\", \"--template\", \"{node|short} {desc|strip|firstline}\\n\")\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn out\n}", "func parseHookPush(payload githubWebhook.PushPayload, client *authentication.Client) (*GeneralizedResult, error) {\n\tvar result = GeneralizedResult{}\n\n\tresult.AddedOrModifiedFiles = lookForYamlInArray(payload.HeadCommit.Added)\n\tmodifiedFiles := lookForYamlInArray(payload.HeadCommit.Modified)\n\n\tresult.AddedOrModifiedFiles = append(result.AddedOrModifiedFiles, modifiedFiles...)\n\n\tif len(result.AddedOrModifiedFiles) == 0 {\n\t\treturn nil, nil\n\t} else {\n\t\tcommitSha := payload.HeadCommit.ID\n\t\tbranchRef := payload.Ref\n\n\t\tresult.RepoName = payload.Repository.Name\n\t\tresult.OwnerName = payload.Repository.Owner.Login\n\t\tresult.UserName = payload.Pusher.Name\n\t\tresult.Sha = commitSha\n\t\tresult.Branch = branchRef\n\n\t\treturn &result, nil\n\t}\n}", "func GitMerge(pr *github.PullRequest, message string) error {\n\n\tfilepath := git.GetRepositoryFilePath(pr.Head.Repository.FullName)\n\tremoteRepositoryURL := git.GenerateCloneURL(pr.Head.Repository.FullName)\n\n\tif !git.Exists(filepath) {\n\t\tif _, err := git.Clone(remoteRepositoryURL); err != nil {\n\t\t\tpr.PostComment(\"I could not pull \" + pr.Head.Repository.FullName + \" from GitHub.\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err := git.Fetch(filepath); err != nil {\n\t\tgit.Prune(filepath)\n\t\tpr.PostComment(\"I could not fetch the latest changes from GitHub. Please try again in a few minutes.\")\n\t\treturn err\n\t}\n\n\tif err := git.Checkout(filepath, pr.Head.Ref); err != nil {\n\t\tpr.PostComment(\"I could not checkout \" + pr.Head.Ref + \" locally.\")\n\t\treturn err\n\t}\n\n\tif err := git.Reset(filepath, path.Join(\"origin\", pr.Head.Ref)); err != nil {\n\t\tpr.PostComment(\"I could not checkout \" + pr.Head.Ref + \" locally.\")\n\t\treturn err\n\t}\n\n\tif err := git.Config(filepath, \"user.name\", git.GetName()); err != nil {\n\t\tpr.PostComment(\"I could run git config for user.name on the server.\")\n\t\treturn err\n\t}\n\n\tif err := git.Config(filepath, \"user.email\", git.GetEmail()); err != nil {\n\t\tpr.PostComment(\"I could run git config for user.email on the server.\")\n\t\treturn err\n\t}\n\n\tif err := git.Rebase(filepath, path.Join(\"origin\", pr.Base.Ref)); err != nil {\n\t\tpr.PostComment(\"I could not rebase \" + pr.Head.Ref + \" with \" + pr.Base.Ref + \". There are conflicts.\")\n\t\treturn err\n\t}\n\n\tif err := git.Push(filepath, pr.Head.Ref); err != nil {\n\t\tpr.PostComment(\"I could not push the changes to \" + pr.Base.Ref + \".\")\n\t\treturn err\n\t}\n\n\tif err := git.Checkout(filepath, pr.Base.Ref); err != nil {\n\t\tpr.PostComment(\"I could not checkout \" + pr.Base.Ref + \" locally.\")\n\t\treturn err\n\t}\n\n\tif err := git.Fetch(filepath); err != nil {\n\t\tgit.Prune(filepath)\n\t\tpr.PostComment(\"I could not fetch the latest changes from GitHub. Please try again in a few minutes.\")\n\t\treturn err\n\t}\n\n\tif err := git.Reset(filepath, path.Join(\"origin\", pr.Base.Ref)); err != nil {\n\t\tpr.PostComment(\"I could not checkout \" + pr.Base.Ref + \" locally.\")\n\t\treturn err\n\t}\n\n\tif err := git.Merge(filepath, pr.Head.Ref, message); err != nil {\n\n\t\tpr.PostComment(\"I could not merge \" + pr.Head.Ref + \" into \" + pr.Base.Ref + \".\" + \"\\nNext time use your fingers for more than just picking your nose.\")\n\t\treturn err\n\t}\n\n\tif err := git.Push(filepath, pr.Base.Ref); err != nil {\n\t\tpr.PostComment(\"I could not push the changes to \" + pr.Base.Ref + \".\")\n\t\treturn err\n\t}\n\n\tpr.PostComment(\"I just merged \" + pr.Head.Ref + \" into \" + pr.Base.Ref+ \"\\nProbably you heard this before:\\nI did it, but I did not enjoy it...\")\n\treturn nil\n}", "func (gr *GitRepo) FetchCommits() error {\n\tgr.Commits = make([]model.Commit, 0) // give number of commits\n\n\trw, err := gr.r.Walk()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = rw.PushHead()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = rw.Iterate(gr.addCommit)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func commit(msg string) {\n\tgit := exec.Command(\"git\", \"commit\", \"-m\", msg)\n\tgitOut, err := git.Output()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfmt.Println(string(gitOut))\n}", "func pushLocalBranch(config *GitXargsConfig, remoteRepository *github.Repository, localRepository *git.Repository) error {\n\tlogger := logging.GetLogger(\"git-xargs\")\n\n\tif config.DryRun {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Repo\": remoteRepository.GetName(),\n\t\t}).Debug(\"Skipping branch push to remote origin because --dry-run flag is set\")\n\n\t\tconfig.Stats.TrackSingle(PushBranchSkipped, remoteRepository)\n\t\treturn nil\n\t}\n\t// Push the changes to the remote repo\n\tpo := &git.PushOptions{\n\t\tRemoteName: \"origin\",\n\t\tAuth: &http.BasicAuth{\n\t\t\tUsername: remoteRepository.GetOwner().GetLogin(),\n\t\t\tPassword: os.Getenv(\"GITHUB_OAUTH_TOKEN\"),\n\t\t},\n\t}\n\tpushErr := localRepository.Push(po)\n\n\tif pushErr != nil {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Error\": pushErr,\n\t\t\t\"Repo\": remoteRepository.GetName(),\n\t\t}).Debug(\"Error pushing new branch to remote origin\")\n\n\t\t// Track the push failure\n\t\tconfig.Stats.TrackSingle(PushBranchFailed, remoteRepository)\n\t\treturn errors.WithStackTrace(pushErr)\n\t}\n\n\tlogger.WithFields(logrus.Fields{\n\t\t\"Repo\": remoteRepository.GetName(),\n\t}).Debug(\"Successfully pushed local branch to remote origin\")\n\n\t// If --skip-pull-requests was passed, track the fact that these changes were pushed directly to the main branch\n\tif config.SkipPullRequests {\n\t\tconfig.Stats.TrackSingle(DirectCommitsPushedToRemoteBranch, remoteRepository)\n\t}\n\n\treturn nil\n}", "func Update(repo *git.Repository, auth Auth, msg string) (hash string, err error) {\n\t// worktree of the project using the go standard library\n\tw, err := repo.Worktree()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Adds the new file to the staging area\n\t_, err = w.Add(\".\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Commits the current staging area to the repository, with the new file\n\tcmt, err := w.Commit(msg, &git.CommitOptions{})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Push using default options\n\treturn cmt.String(), repo.Push(&git.PushOptions{\n\t\tAuth: auth.Method(),\n\t})\n}", "func (api *API) PushTags(cn *collection.Collection, push PushConfig) error {\n\tlog.Debugf(\n\t\t\"%s 'push' collection: %+v (%d repos / %d tags)\",\n\t\tfn(), cn, cn.RepoCount(), cn.TagCount(),\n\t)\n\tlog.Debugf(\"%s push config: %+v\", fn(), push)\n\n\tdone := make(chan error, cn.TagCount())\n\n\tif cn.TagCount() == 0 {\n\t\tlog.Infof(\"%s No tags to push\", fn())\n\t\treturn nil\n\t}\n\n\tfor _, ref := range cn.Refs() {\n\t\trepo := cn.Repo(ref)\n\t\ttags := cn.Tags(ref)\n\n\t\tlog.Debugf(\"%s repository: %+v\", fn(), repo)\n\t\tfor _, tg := range tags {\n\t\t\tlog.Debugf(\"%s tag: %+v\", fn(), tg)\n\t\t}\n\n\t\tgo func(repo *repository.Repository, tags []*tag.Tag, done chan error) {\n\t\t\tfor _, tg := range tags {\n\t\t\t\tsrcRef := repo.Name() + \":\" + tg.Name()\n\t\t\t\tdstRef := push.Registry + push.Prefix + \"/\" + repo.Path() + \":\" + tg.Name()\n\n\t\t\t\tlog.Infof(\"[PULL/PUSH] PULLING %s\", srcRef)\n\t\t\t\tif err := api.dockerClient.Pull(srcRef); err != nil {\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tlog.Infof(\"[PULL/PUSH] PUSHING %s => %s\", srcRef, dstRef)\n\t\t\t\tif err := api.dockerClient.Tag(srcRef, dstRef); err != nil {\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif err := api.dockerClient.Push(dstRef); err != nil {\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tdone <- nil\n\t\t\t}\n\t\t}(repo, tags, done)\n\t}\n\n\treturn wait.Until(done)\n}", "func (u *Update) Push(obj utils.M) *Update {\n\tu.update[\"$push\"] = obj\n\treturn u\n}", "func (j *JPush) Push(req *PushRequest) (*PushResponse, error) {\n\turl := j.GetURL(\"push\") + \"push\"\n\tif req.Audience.Aud.File != nil {\n\t\turl += \"/file\"\n\t}\n\tbuf, err := json.Marshal(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := j.request(\"POST\", url, bytes.NewReader(buf), nil)\n\tret := new(PushResponse)\n\terr2 := json.Unmarshal(resp, ret)\n\tif err2 != nil {\n\t\treturn nil, err2\n\t}\n\treturn ret, err\n}", "func Commit(r *git.Repository, msg string) bool {\n\t// check for differences.\n\tdiffCmd := fmt.Sprintf(\"diff -U 0 %s %s/pimpminers.conf | grep -v ^@ | wc -l\", stagingFile, localGitRepo)\n\tdiff := RunCommand(diffCmd)\n\tdiff = strings.TrimSpace(diff)\n\tif diff == \"0\" {\n\t\tfmt.Println(\"No changes to commit.\")\n\t} else {\n\t\t// copy file from staging into worktree\n\t\tdiffCmd = fmt.Sprintf(\"diff -U 0 %s %s/pimpminers.conf | grep -v ^@\", stagingFile, localGitRepo)\n\t\tdiff = RunCommand(diffCmd)\n\t\tdiff = strings.TrimSpace(diff)\n\t\tfmt.Println(\"Changes:\")\n\t\tfmt.Println(diff)\n\t\tfmt.Println(\"\\nCommitting changes... \")\n\t\tcopy := fmt.Sprintf(\"cp %s %s/pimpminers.conf\", stagingFile, localGitRepo)\n\t\tRunCommand(copy)\n\t\tw, err := r.Worktree()\n\t\tcheckErr(err)\n\t\t// add files\n\t\t_, err = w.Add(\"pimpminers.conf\")\n\t\tcheckErr(err)\n\t\t// commit\n\t\t_, err = w.Commit(msg, &git.CommitOptions{\n\t\t\tAuthor: &object.Signature{\n\t\t\t\tName: \"pimplabops\",\n\t\t\t\tEmail: \"[email protected]\",\n\t\t\t\tWhen: time.Now(),\n\t\t\t},\n\t\t})\n\t\tcheckErr(err)\n\t\tif err != nil {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (dc *DockerClient) Push(ref string) (io.ReadCloser, error) {\n\tregistryAuth := dc.cnf.GetRegistryAuth(\n\t\trepository.GetRegistry(ref),\n\t)\n\n\tpushOptions := types.ImagePushOptions{RegistryAuth: registryAuth}\n\tif registryAuth == \"\" {\n\t\tpushOptions = types.ImagePushOptions{RegistryAuth: \"IA==\"}\n\t}\n\n\treturn dc.cli.ImagePush(context.Background(), ref, pushOptions)\n}", "func (c Repository) RPush(value string) {\n\tc.Client.RPush(\"alist\", value)\n}", "func PushRepoAndCreatePullRequest(dir string, upstreamRepo *GitRepository, forkRepo *GitRepository, base string, prDetails *PullRequestDetails, filter *PullRequestFilter, commit bool, commitMessage string, push bool, dryRun bool, gitter Gitter, provider GitProvider) (*PullRequestInfo, error) {\n\tuserAuth := provider.UserAuth()\n\tif commit {\n\t\terr := gitter.Add(dir, \"-A\")\n\t\tif err != nil {\n\t\t\treturn nil, errors.WithStack(err)\n\t\t}\n\t\tchanged, err := gitter.HasChanges(dir)\n\t\tif err != nil {\n\t\t\treturn nil, errors.WithStack(err)\n\t\t}\n\t\tif !changed {\n\t\t\tlog.Logger().Warnf(\"No changes made to the source code in %s. Code must be up to date!\", dir)\n\t\t\treturn nil, nil\n\t\t}\n\t\tif commitMessage == \"\" {\n\t\t\tcommitMessage = prDetails.Message\n\t\t}\n\t\terr = gitter.CommitDir(dir, commitMessage)\n\t\tif err != nil {\n\t\t\treturn nil, errors.WithStack(err)\n\t\t}\n\t}\n\n\theadPrefix := \"\"\n\n\tusername := upstreamRepo.Organisation\n\tcloneURL := upstreamRepo.CloneURL\n\tif forkRepo != nil {\n\t\tusername = forkRepo.Organisation\n\t\tcloneURL = forkRepo.CloneURL\n\t}\n\n\tif upstreamRepo.Organisation != username {\n\t\theadPrefix = username + \":\"\n\t}\n\n\tgha := &GitPullRequestArguments{\n\t\tGitRepository: upstreamRepo,\n\t\tTitle: prDetails.Title,\n\t\tBody: prDetails.Message,\n\t\tBase: base,\n\t\tLabels: prDetails.Labels,\n\t}\n\tvar existingPr *GitPullRequest\n\n\tforkPushURL, err := gitter.CreateAuthenticatedURL(cloneURL, &userAuth)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"creating push URL for %s\", cloneURL)\n\t}\n\n\tif filter != nil && push {\n\t\t// lets rebase an existing PR\n\t\texistingPrs, err := FilterOpenPullRequests(provider, upstreamRepo.Organisation, upstreamRepo.Name, *filter)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"finding existing PRs using filter %s on repo %s/%s\", filter.String(), upstreamRepo.Organisation, upstreamRepo.Name)\n\t\t}\n\n\t\tif len(existingPrs) > 1 {\n\t\t\tsort.SliceStable(existingPrs, func(i, j int) bool {\n\t\t\t\t// sort in descending order of PR numbers (assumes PRs numbers increment!)\n\t\t\t\treturn util.DereferenceInt(existingPrs[j].Number) < util.DereferenceInt(existingPrs[i].Number)\n\t\t\t})\n\t\t\tprs := make([]string, 0)\n\t\t\tfor _, pr := range existingPrs {\n\t\t\t\tprs = append(prs, pr.URL)\n\t\t\t}\n\t\t\tlog.Logger().Debugf(\"Found more than one PR %s using filter %s on repo %s/%s so rebasing latest PR %s\", strings.Join(prs, \", \"), filter.String(), upstreamRepo.Organisation, upstreamRepo.Name, existingPrs[:1][0].URL)\n\t\t\texistingPr = existingPrs[0]\n\t\t} else if len(existingPrs) == 1 {\n\t\t\texistingPr = existingPrs[0]\n\t\t}\n\t}\n\tremoteBranch := prDetails.BranchName\n\tif existingPr != nil {\n\t\tif util.DereferenceString(existingPr.HeadOwner) == username && existingPr.HeadRef != nil && existingPr.Number != nil {\n\t\t\tremote := \"origin\"\n\t\t\tif forkRepo != nil && forkRepo.Fork {\n\t\t\t\tremote = \"upstream\"\n\t\t\t}\n\t\t\tchangeBranch, err := gitter.Branch(dir)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.WithStack(err)\n\t\t\t}\n\t\t\tlocalBranchUUID, err := uuid.NewUUID()\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrapf(err, \"creating UUID for local branch\")\n\t\t\t}\n\t\t\t// We use this \"dummy\" local branch to pull into to avoid having to work with FETCH_HEAD as our local\n\t\t\t// representation of the remote branch. This is an oddity of the pull/%d/head remote.\n\t\t\tlocalBranch := localBranchUUID.String()\n\t\t\tremoteBranch = *existingPr.HeadRef\n\t\t\tfetchRefSpec := fmt.Sprintf(\"pull/%d/head:%s\", *existingPr.Number, localBranch)\n\t\t\terr = gitter.FetchBranch(dir, remote, fetchRefSpec)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrapf(err, \"fetching %s for merge\", fetchRefSpec)\n\t\t\t}\n\n\t\t\terr = gitter.CreateBranchFrom(dir, prDetails.BranchName, localBranch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrapf(err, \"creating branch %s from %s\", prDetails.BranchName, fetchRefSpec)\n\t\t\t}\n\t\t\terr = gitter.Checkout(dir, prDetails.BranchName)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrapf(err, \"checking out branch %s\", prDetails.BranchName)\n\t\t\t}\n\t\t\terr = gitter.MergeTheirs(dir, changeBranch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrapf(err, \"merging %s into %s\", changeBranch, fetchRefSpec)\n\t\t\t}\n\t\t\terr = gitter.RebaseTheirs(dir, fmt.Sprintf(localBranch), \"\", true)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.WithStack(err)\n\t\t\t}\n\t\t\tchangedFiles, err := gitter.ListChangedFilesFromBranch(dir, localBranch)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrap(err, \"failed to list changed files\")\n\t\t\t}\n\t\t\tif changedFiles == \"\" {\n\t\t\t\tlog.Logger().Info(\"No file changes since the existing PR. Nothing to push.\")\n\t\t\t\treturn nil, nil\n\t\t\t}\n\t\t} else {\n\t\t\t// We can only update an existing PR if the owner of that PR is this user, so we clear the existingPr\n\t\t\texistingPr = nil\n\t\t}\n\t}\n\tvar pr *GitPullRequest\n\tif !dryRun && existingPr != nil {\n\t\tgha.Head = headPrefix + remoteBranch\n\t\t// work out the minimal similar title\n\t\tif strings.HasPrefix(existingPr.Title, \"chore(deps): bump \") {\n\t\t\torigWords := strings.Split(existingPr.Title, \" \")\n\t\t\tnewWords := strings.Split(prDetails.Title, \" \")\n\t\t\tanswer := make([]string, 0)\n\t\t\tfor i, w := range newWords {\n\t\t\t\tif len(origWords) > i && origWords[i] == w {\n\t\t\t\t\tanswer = append(answer, w)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif answer[len(answer)-1] == \"bump\" {\n\t\t\t\t// if there are no similarities in the actual dependency, then add a generic form of words\n\t\t\t\tanswer = append(answer, \"dependency\", \"versions\")\n\t\t\t}\n\t\t\tif answer[len(answer)-1] == \"to\" || answer[len(answer)-1] == \"from\" {\n\t\t\t\t// remove trailing prepositions\n\t\t\t\tanswer = answer[:len(answer)-1]\n\t\t\t}\n\t\t\tgha.Title = strings.Join(answer, \" \")\n\t\t} else {\n\t\t\tgha.Title = prDetails.Title\n\t\t}\n\t\tgha.Body = fmt.Sprintf(\"%s\\n<hr />\\n\\n%s\", prDetails.Message, existingPr.Body)\n\t\tvar err error\n\t\tpr, err = provider.UpdatePullRequest(gha, *existingPr.Number)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"updating pull request %s\", existingPr.URL)\n\t\t}\n\t\tlog.Logger().Infof(\"Updated Pull Request: %s\", util.ColorInfo(pr.URL))\n\t}\n\tif dryRun {\n\t\tlog.Logger().Infof(\"Commit created but not pushed; would have updated pull request %s with %s and used commit message %s. Please manually delete %s when you are done\", util.ColorInfo(existingPr.URL), prDetails.String(), commitMessage, util.ColorInfo(dir))\n\t\treturn nil, nil\n\t} else if push {\n\t\terr := gitter.Push(dir, forkPushURL, true, fmt.Sprintf(\"%s:%s\", \"HEAD\", remoteBranch))\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"pushing merged branch %s\", remoteBranch)\n\t\t}\n\t}\n\tif existingPr == nil {\n\t\tgha.Head = headPrefix + prDetails.BranchName\n\n\t\tpr, err = provider.CreatePullRequest(gha)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"creating pull request with arguments %v\", gha.String())\n\t\t}\n\t\tlog.Logger().Infof(\"Created Pull Request: %s\", util.ColorInfo(pr.URL))\n\t}\n\n\tprInfo := &PullRequestInfo{\n\t\tGitProvider: provider,\n\t\tPullRequest: pr,\n\t\tPullRequestArguments: gha,\n\t}\n\n\terr = addLabelsToPullRequest(prInfo, prDetails.Labels)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to add labels %+v to PR %s\", prDetails.Labels, pr.URL)\n\t}\n\n\treturn prInfo, nil\n}", "func (t *Table) QueueGitCommit(m *Model, email string) error {\n\tcontent, err := yaml.Marshal(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcomment := fmt.Sprintf(\"Table %s updated.\", t.Name)\n\tname := \"\" // TODO we don't know the user's name\n\n\tm.git.QueueCommit(t.relpath(), content, comment, name, email, time.Now())\n\treturn nil\n}", "func (cli *Client) Push(c context.Context, p *Payload) error {\n\treturn cli.do(http.MethodPost, \"/push\", p)\n}", "func TestGitCommandPush(t *testing.T) {\n\ttype scenario struct {\n\t\ttestName string\n\t\tgetGitConfigValue func(string) (string, error)\n\t\tcommand func(string, ...string) *exec.Cmd\n\t\topts PushOpts\n\t\ttest func(error)\n\t}\n\n\tprompt := func(passOrUname string) string {\n\t\treturn \"\\n\"\n\t}\n\n\tscenarios := []scenario{\n\t\t{\n\t\t\t\"Push with force disabled, follow-tags on\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"--follow-tags\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{Force: false, PromptUserForCredential: prompt},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with force enabled, follow-tags on\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"--follow-tags\", \"--force-with-lease\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{Force: true, PromptUserForCredential: prompt},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with force disabled, follow-tags off\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"false\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{Force: false, PromptUserForCredential: prompt},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with an error occurring, follow-tags on\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"--follow-tags\"}, args)\n\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t},\n\t\t\tPushOpts{Force: false, PromptUserForCredential: prompt},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.Error(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with force disabled, follow-tags off, upstream supplied\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"false\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"origin\", \"master\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{\n\t\t\t\tForce: false,\n\t\t\t\tUpstreamRemote: \"origin\",\n\t\t\t\tUpstreamBranch: \"master\",\n\t\t\t\tPromptUserForCredential: prompt,\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with force disabled, follow-tags off, setting upstream\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"false\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"--set-upstream\", \"origin\", \"master\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{\n\t\t\t\tForce: false,\n\t\t\t\tUpstreamRemote: \"origin\",\n\t\t\t\tUpstreamBranch: \"master\",\n\t\t\t\tPromptUserForCredential: prompt,\n\t\t\t\tSetUpstream: true,\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with force enabled, follow-tags off, setting upstream\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"false\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"--force-with-lease\", \"--set-upstream\", \"origin\", \"master\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{\n\t\t\t\tForce: true,\n\t\t\t\tUpstreamRemote: \"origin\",\n\t\t\t\tUpstreamBranch: \"master\",\n\t\t\t\tPromptUserForCredential: prompt,\n\t\t\t\tSetUpstream: true,\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with remote branch but no origin\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"false\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tPushOpts{\n\t\t\t\tForce: true,\n\t\t\t\tUpstreamRemote: \"\",\n\t\t\t\tUpstreamBranch: \"master\",\n\t\t\t\tPromptUserForCredential: prompt,\n\t\t\t\tSetUpstream: true,\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.Error(t, err)\n\t\t\t\tassert.EqualValues(t, \"Must specify a remote if specifying a branch\", err.Error())\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with force disabled, follow-tags off, upstream supplied\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"false\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"origin\", \"master\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{\n\t\t\t\tForce: false,\n\t\t\t\tUpstreamRemote: \"origin\",\n\t\t\t\tUpstreamBranch: \"master\",\n\t\t\t\tPromptUserForCredential: prompt,\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with force disabled, follow-tags off, setting upstream\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"false\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"--set-upstream\", \"origin\", \"master\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{\n\t\t\t\tForce: false,\n\t\t\t\tUpstreamRemote: \"origin\",\n\t\t\t\tUpstreamBranch: \"master\",\n\t\t\t\tPromptUserForCredential: prompt,\n\t\t\t\tSetUpstream: true,\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"Push with force enabled, follow-tags off, setting upstream\",\n\t\t\tfunc(string) (string, error) {\n\t\t\t\treturn \"false\", nil\n\t\t\t},\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\t\t\t\tassert.EqualValues(t, []string{\"push\", \"--force-with-lease\", \"--set-upstream\", \"origin\", \"master\"}, args)\n\n\t\t\t\treturn secureexec.Command(\"echo\")\n\t\t\t},\n\t\t\tPushOpts{\n\t\t\t\tForce: true,\n\t\t\t\tUpstreamRemote: \"origin\",\n\t\t\t\tUpstreamBranch: \"master\",\n\t\t\t\tPromptUserForCredential: prompt,\n\t\t\t\tSetUpstream: true,\n\t\t\t},\n\t\t\tfunc(err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, s := range scenarios {\n\t\tt.Run(s.testName, func(t *testing.T) {\n\t\t\tgitCmd := NewDummyGitCommand()\n\t\t\tgitCmd.OSCommand.Command = s.command\n\t\t\tgitCmd.getGitConfigValue = s.getGitConfigValue\n\t\t\terr := gitCmd.Push(s.opts)\n\t\t\ts.test(err)\n\t\t})\n\t}\n}", "func PushArgs(tag string) []string {\n\treturn []string{\"push\", tag}\n}", "func RunPush(cmd Runner, tags []string) error {\n\tfmt.Println(\"Pushing image\", tags)\n\tfor _, tag := range tags {\n\t\targs := PushArgs(tag)\n\t\tif err := cmd.Run(\"docker\", args...); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (g *git) gitPull() {\n\n\twTree, err := g.Repo.Worktree()\n\tif err != nil {\n\t\tlog.Fatalf(\"[Pull] Failed get work tree: %s\\n\", err.Error())\n\t}\n\n\terr = wTree.Pull(&go_git.PullOptions{\n\t\tReferenceName: plumbing.ReferenceName(g.Branch),\n\t\tAuth: g.Auth,\n\t\tSingleBranch: true,\n\t\tProgress: os.Stdout,\n\t\t//Force: true,\n\t})\n\n\tif err == nil {\n\t\tlog.Printf(\"[Pull] Success!\\n\")\n\t} else {\n\t\tswitch err {\n\t\tcase go_git.ErrUnstagedChanges:\n\t\t\tif g.Force {\n\t\t\t\tlog.Printf(\"[Pull] Info: %s (local repository changed).\\nRule for forced pull - %v. Force pulling...\", err.Error(), g.Force)\n\t\t\t\terr := wTree.Reset(&go_git.ResetOptions{\n\t\t\t\t\tMode: go_git.ResetMode(1),\n\t\t\t\t})\n\t\t\t\tif err == nil {\n\t\t\t\t\tlog.Printf(\"[Pull] Success!\\n\")\n\t\t\t\t} else {\n\t\t\t\t\tlog.Printf(\"[Pull] Error: %s\\n\", err.Error())\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tlog.Printf(\"[Pull] Error: %s (local repository changed).\\nRule for forced pull - %v. Can`t force pull.\\n\", err.Error(), g.Force)\n\t\t\t}\n\t\tcase go_git.NoErrAlreadyUpToDate:\n\t\t\tlog.Printf(\"[Pull] Nothing to pull: %s\\n\", err.Error())\n\t\tdefault:\n\t\t\tlog.Printf(\"[Pull] Error: %s\\n\", err.Error())\n\t\t}\n\t}\n\n\t// show last commit\n\tref, err := g.Repo.Head()\n\tif err != nil {\n\t\tlog.Printf(\"Failed get reference where HEAD is pointing to: %s\\n\", err.Error())\n\t}\n\tcommit, err := g.Repo.CommitObject(ref.Hash())\n\tif err != nil {\n\t\tlog.Printf(\"[Pull] Can`t show last commit: %s\\n\", err.Error())\n\t} else {\n\t\tlog.Printf(\"[Pull] Last commit: %s\\n\", commit)\n\t}\n}", "func (m *MilestoneTasks) PushIssues(ctx *wf.TaskContext, milestones ReleaseMilestones, version string, kind ReleaseKind) error {\n\tissues, err := m.Client.FetchMilestoneIssues(ctx, m.RepoOwner, m.RepoName, milestones.Current)\n\tif err != nil {\n\t\treturn err\n\t}\n\tctx.Printf(\"Processing %d open issues in milestone %d.\", len(issues), milestones.Current)\n\tfor issueNumber, labels := range issues {\n\t\tvar newLabels *[]string\n\t\tvar newMilestone *int\n\t\tvar actions []string // A short description of actions taken, for the log line.\n\t\tremoveLabel := func(name string) {\n\t\t\tif !labels[name] {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tnewLabels = new([]string)\n\t\t\tfor label := range labels {\n\t\t\t\tif label == name {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\t*newLabels = append(*newLabels, label)\n\t\t\t}\n\t\t\tactions = append(actions, fmt.Sprintf(\"removed label %q\", name))\n\t\t}\n\t\tif kind == KindBeta && strings.HasSuffix(version, \"beta1\") {\n\t\t\tremoveLabel(\"okay-after-beta1\")\n\t\t} else if kind == KindRC && strings.HasSuffix(version, \"rc1\") {\n\t\t\tremoveLabel(\"okay-after-rc1\")\n\t\t} else if kind == KindMajor || kind == KindMinor {\n\t\t\tnewMilestone = &milestones.Next\n\t\t\tactions = append(actions, fmt.Sprintf(\"pushed to milestone %d\", milestones.Next))\n\t\t}\n\t\tif newMilestone == nil && newLabels == nil {\n\t\t\tctx.Printf(\"Nothing to do for issue %d.\", issueNumber)\n\t\t\tcontinue\n\t\t}\n\t\t_, _, err := m.Client.EditIssue(ctx, m.RepoOwner, m.RepoName, issueNumber, &github.IssueRequest{\n\t\t\tMilestone: newMilestone,\n\t\t\tLabels: newLabels,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tctx.Printf(\"Updated issue %d: %s.\", issueNumber, strings.Join(actions, \", \"))\n\t}\n\tif kind == KindMajor || kind == KindMinor {\n\t\t_, _, err := m.Client.EditMilestone(ctx, m.RepoOwner, m.RepoName, milestones.Current, &github.Milestone{\n\t\t\tState: github.String(\"closed\"),\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tctx.Printf(\"Closed milestone %d.\", milestones.Current)\n\t}\n\treturn nil\n}", "func Commits(res http.ResponseWriter, req *http.Request) {\n\tvar commits []git.CommitJSON\n\tvar commitJSON git.CommitJSON\n\tvar i int\n\n\t// Ignore this error as we fall back to page 0 when p is empty\n\tpageNumber, _ := strconv.Atoi(req.FormValue(\"p\"))\n\n\t// Get the number of commits to display per page\n\tcommitsPerPage, err := strconv.Atoi(req.FormValue(\"commits\"))\n\tif err != nil {\n\t\tcommitsPerPage = defaultCommitsPerPage\n\t}\n\n\tlog.Println(pageNumber, commitsPerPage)\n\n\t// Get the repository details\n\tparams := mux.Vars(req)\n\trepoDB, err := models.GetRepository(params[\"pid\"])\n\tif err != nil {\n\t\tpanic(\"missing\")\n\t}\n\n\trepo, err := git2go.OpenRepository(repoDB.Location)\n\tif err != nil {\n\t\tpanic(\"repository folder missing\")\n\t}\n\n\todb, err := repo.Odb()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\todb.ForEach(func(id *git2go.Oid) error {\n\t\tif i > commitsPerPage || id.IsZero() {\n\t\t\treturn nil\n\t\t}\n\n\t\t// Retrieve a commit\n\t\tcommit, err := repo.LookupCommit(id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\ti++\n\t\tcommitJSON.Generate(commit)\n\t\tcommits = append(commits, commitJSON)\n\t\treturn nil\n\t})\n\n\tmarshaledCommitJSON, _ := json.Marshal(commits)\n\tres.Write(marshaledCommitJSON)\n}", "func (u *Update) Push(obj types.M) *Update {\n\tu.update[\"$push\"] = obj\n\treturn u\n}", "func (wa *WebAPI) ProcessPush(p Push, b *dinghyfile.PipelineBuilder) error {\n\t// Ensure dinghyfile was changed.\n\tif !p.ContainsFile(wa.Config.DinghyFilename) {\n\t\tb.Logger.Infof(\"Push does not include %s, skipping.\", wa.Config.DinghyFilename)\n\t\terrstat, status, _ := p.GetCommitStatus()\n\t\tif errstat == nil && status == \"\" {\n\t\t\tp.SetCommitStatus(git.StatusSuccess, fmt.Sprintf(\"No changes in %v.\", wa.Config.DinghyFilename))\n\t\t}\n\t\treturn nil\n\t}\n\n\tb.Logger.Info(\"Dinghyfile found in commit for repo \" + p.Repo())\n\n\t// Set commit status to the pending yellow dot.\n\tp.SetCommitStatus(git.StatusPending, git.DefaultMessagesByBuilderAction[b.Action][git.StatusPending])\n\n\tfor _, filePath := range p.Files() {\n\t\tcomponents := strings.Split(filePath, \"/\")\n\t\tif components[len(components)-1] == wa.Config.DinghyFilename {\n\t\t\t// Process the dinghyfile.\n\t\t\terr := b.ProcessDinghyfile(p.Org(), p.Repo(), filePath, p.Branch())\n\t\t\t// Set commit status based on result of processing.\n\t\t\tif err != nil {\n\t\t\t\tif err == dinghyfile.ErrMalformedJSON {\n\t\t\t\t\tb.Logger.Errorf(\"Error processing Dinghyfile (malformed JSON): %s\", err.Error())\n\t\t\t\t\tp.SetCommitStatus(git.StatusFailure, \"Error processing Dinghyfile (malformed JSON)\")\n\t\t\t\t} else {\n\t\t\t\t\tb.Logger.Errorf(\"Error processing Dinghyfile: %s\", err.Error())\n\t\t\t\t\tp.SetCommitStatus(git.StatusError, fmt.Sprintf(\"%s\", err.Error()))\n\t\t\t\t}\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tp.SetCommitStatus(git.StatusSuccess, git.DefaultMessagesByBuilderAction[b.Action][git.StatusSuccess])\n\t\t}\n\t}\n\treturn nil\n}", "func SyncPush(w http.ResponseWriter, req *http.Request) {\n\tqueryString := req.URL.Query()\n\tuid := queryString.Get(\"uid\")\n\n\tdata, err := ioutil.ReadAll(req.Body)\n\tif err != nil {\n\t\thttp.Error(w, \"bad request\", http.StatusBadRequest)\n\t\treturn\n\t}\n\tgo PushMessage(uid, data)\n\n\tif DEBUG {\n\t\t// echo\n\t\tw.Write(data)\n\t}\n}", "func (g Git) Commit(message string) ([]byte, error) {\n\tcmd := exec.Command(\"git\", \"add\", \".\")\n\tcmd.Dir = g.Work\n\tresult, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tcmd = exec.Command(\"git\", \"commit\", \"-q\", \"-m\", message)\n\tcmd.Dir = g.Work\n\treturn cmd.CombinedOutput()\n}", "func (remote *LocalRemote) Push(image, imageRoot string) error {\n\tlog.Println(\"pushing local\", remote.Url.Path)\n\n\treturn remote.rsyncTo(imageRoot, \"\")\n}", "func ShouldBranchBePushed(branchName string) bool {\n\ttrackingBranchName := GetTrackingBranchName(branchName)\n\tcmd := command.New(\"git\", \"rev-list\", \"--left-right\", branchName+\"...\"+trackingBranchName)\n\treturn cmd.Output() != \"\"\n}", "func (api *API) CollectPushTags(cn *collection.Collection, push PushConfig) (*collection.Collection, error) {\n\tlog.Debugf(\n\t\t\"%s collection: %+v (%d repos / %d tags)\",\n\t\tfn(), cn, cn.RepoCount(), cn.TagCount(),\n\t)\n\tlog.Debugf(\"%s push config: %+v\", fn(), push)\n\n\trefs := make([]string, len(cn.Refs()))\n\tdone := make(chan error, len(cn.Refs()))\n\ttags := make(map[string][]*tag.Tag)\n\n\tfor i, repo := range cn.Repos() {\n\t\tgo func(repo *repository.Repository, i int, done chan error) {\n\t\t\trefs[i] = repo.Ref()\n\n\t\t\tpushPrefix := push.Prefix\n\t\t\tif pushPrefix == \"\" {\n\t\t\t\tpushPrefix = repo.PushPrefix()\n\t\t\t}\n\n\t\t\tvar pushRepoPath string\n\t\t\tpushRepoPath = pushPrefix + \"/\" + repo.Path()\n\t\t\tpushRepoPath = pushRepoPath[1:] // Leading \"/\" in prefix should be removed!\n\n\t\t\tpushRef := fmt.Sprintf(\"%s/%s~/.*/\", push.Registry, pushRepoPath)\n\n\t\t\tlog.Debugf(\"%s 'push' reference: %+v\", fn(repo.Ref()), pushRef)\n\n\t\t\tpushRepo, err := repository.ParseRef(pushRef)\n\t\t\tif err != nil {\n\t\t\t\tdone <- err\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tlog.Infof(\"[PULL/PUSH] ANALYZE %s => %s\", repo.Ref(), pushRef)\n\n\t\t\tusername, password, _ := api.dockerClient.Config().GetCredentials(push.Registry)\n\n\t\t\tpushedTags, err := remote.FetchTags(pushRepo, username, password)\n\t\t\tif err != nil {\n\t\t\t\tif !strings.Contains(err.Error(), \"404 Not Found\") {\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tlog.Warnf(\"%s repo not found: %+s\", fn(repo.Ref()), pushRef)\n\n\t\t\t\tpushedTags = make(map[string]*tag.Tag)\n\t\t\t}\n\t\t\tlog.Debugf(\"%s pushed tags: %+v\", fn(repo.Ref()), pushedTags)\n\n\t\t\tremoteTags := cn.TagMap(repo.Ref())\n\t\t\tlog.Debugf(\"%s remote tags: %+v\", fn(repo.Ref()), remoteTags)\n\n\t\t\tsortedKeys, tagNames, joinedTags := tag.Join(\n\t\t\t\tremoteTags,\n\t\t\t\tpushedTags,\n\t\t\t\trepo.Tags(),\n\t\t\t)\n\t\t\tlog.Debugf(\"%s joined tags: %+v\", fn(repo.Ref()), joinedTags)\n\n\t\t\ttagsToPush := make([]*tag.Tag, 0)\n\t\t\tfor _, key := range sortedKeys {\n\t\t\t\tname := tagNames[key]\n\t\t\t\ttg := joinedTags[name]\n\n\t\t\t\tif tg.NeedsPush(push.UpdateChanged) {\n\t\t\t\t\ttagsToPush = append(tagsToPush, tg)\n\t\t\t\t}\n\t\t\t}\n\t\t\tlog.Debugf(\"%s tags to push: %+v\", fn(repo.Ref()), tagsToPush)\n\n\t\t\ttags[repo.Ref()] = tagsToPush\n\n\t\t\tdone <- nil\n\n\t\t\treturn\n\t\t}(repo, i, done)\n\t}\n\n\tif err := wait.Until(done); err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Debugf(\"%s 'push' tags: %+v\", fn(), tags)\n\n\tpn, err := collection.New(refs, tags)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Debugf(\n\t\t\"%s 'push' collection: %+v (%d repos / %d tags)\",\n\t\tfn(), cn, cn.RepoCount(), cn.TagCount(),\n\t)\n\n\treturn pn, nil\n}", "func (r *repo) log(old, cur string) ([]*commit, error) {\n\tout, err := exec.Command(\"git\", \"log\", \"--oneline\", fmt.Sprintf(\"%s...%s\", old, cur)).CombinedOutput() // TODO(adam): -- $dir\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar commits []*commit\n\trdr := bufio.NewScanner(bytes.NewReader(out))\n\tfor rdr.Scan() {\n\t\tline := strings.TrimSpace(rdr.Text())\n\t\tif err := rdr.Err(); err != nil {\n\t\t\tif err == io.EOF {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\treturn nil, err\n\t\t}\n\t\tif line == \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\t// split into shortRef and message\n\t\tidx := strings.Index(line, \" \")\n\t\tif idx > 0 {\n\t\t\tshortRef := line[:idx-1]\n\t\t\tmessage := line[idx:]\n\t\t\tcommits = append(commits, &commit{\n\t\t\t\tshortRef: shortRef,\n\t\t\t\tmessage: message,\n\t\t\t})\n\t\t}\n\t}\n\treturn commits, nil\n}", "func (s *SimplePublishNode) Commit() {\n\ts.committed++\n}", "func (g *GitLocal) PushMaster(dir string) error {\n\treturn g.GitFake.PushMaster(dir)\n}", "func (p *Pipe) Push(values ...phono.Param) {\n\tif len(values) == 0 {\n\t\treturn\n\t}\n\tparams := params(make(map[string][]phono.ParamFunc))\n\tp.events <- eventMessage{\n\t\tevent: push,\n\t\tparams: params.add(values...),\n\t}\n}", "func (r *Repository) GetPushedAt() Timestamp {\n\tif r == nil || r.PushedAt == nil {\n\t\treturn Timestamp{}\n\t}\n\treturn *r.PushedAt\n}", "func (git *Git) Commits() []string {\n\tout, err := Run(git.Dir, \"git\", \"log\", \"HEAD..origin/master\", \"--oneline\")\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn out\n}", "func (c ProwlClient) Push(n Notification) error {\n\n\tkeycsv := strings.Join(n.apikeys, \",\")\n\n\tvals := url.Values{\n\t\t\"apikey\": []string{keycsv},\n\t\t\"application\": []string{n.Application},\n\t\t\"description\": []string{n.Description},\n\t\t\"event\": []string{n.Event},\n\t\t\"priority\": []string{string(n.Priority)},\n\t}\n\n\tif n.URL != \"\" {\n\t\tvals[\"url\"] = []string{n.URL}\n\t}\n\n\tif c.ProviderKey != \"\" {\n\t\tvals[\"providerkey\"] = []string{c.ProviderKey}\n\t}\n\n\tr, err := http.PostForm(apiURL+\"/add\", vals)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer r.Body.Close()\n\n\tif r.StatusCode != 200 {\n\t\terr = decodeError(r.Status, r.Body)\n\t}\n\n\treturn err\n}", "func (m *Manager) Push() error {\n\ttargetImage := m.ensureTargetImageTransport()\n\ttargetRef, err := alltransports.ParseImageName(targetImage)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\topts := buildah.PushOptions{\n\t\tCompression: imagebuildah.Gzip,\n\t\tReportWriter: os.Stderr,\n\t\tStore: m.store,\n\t\tSystemContext: systemContext,\n\t}\n\n\tlog.Infof(\"Pushing image: '%s'\", targetImage)\n\tref, _, err := buildah.Push(m.ctx, m.targetImage, targetRef, opts)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog.Infof(\"%s, digest: '%s'\", ref.String(), ref.Digest().String())\n\treturn nil\n}", "func (t *Transport) Push(ctx msg.Context, subject string, data []byte) error {\n\tif len(data) > t.change {\n\t\treturn t.blob.Push(ctx, subject, data)\n\t}\n\treturn t.pubsub.Push(ctx, subject, data)\n}", "func RunOktetoPush(oktetoPath, workdir string) error {\n\tcmd := exec.Command(oktetoPath, \"push\")\n\tif workdir != \"\" {\n\t\tcmd.Dir = workdir\n\t\tcmd.Env = append(cmd.Env, fmt.Sprintf(\"%s=%s\", constants.OktetoHomeEnvVar, workdir))\n\t}\n\tif v := os.Getenv(model.OktetoURLEnvVar); v != \"\" {\n\t\tcmd.Env = append(cmd.Env, fmt.Sprintf(\"%s=%s\", model.OktetoURLEnvVar, v))\n\t}\n\n\tlog.Printf(\"Running '%s'\", cmd.String())\n\n\to, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"okteto stack deploy failed: %s - %s\", string(o), err)\n\t}\n\tlog.Printf(\"okteto stack deploy success\")\n\treturn nil\n}", "func pushTestImages(ctx context.Context, repoURL string, tags []string) error {\n\tfor _, tag := range tags {\n\t\tremoteImg := fmt.Sprintf(\"%s/podinfo:%s\", repoURL, tag)\n\t\terr := tftestenv.RetagAndPush(ctx, fmt.Sprintf(\"%s:%s\", localImg, tag), remoteImg)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\treturn nil\n}", "func (cli *DockerClient) ImagePush(ctx context.Context, registry, repo, tag string) error {\n\tv := url.Values{}\n\timage := repo\n\tif registry != \"\" {\n\t\timage = fmt.Sprintf(\"%s/%s\", registry, repo)\n\t}\n\tv.Set(\"tag\", tag)\n\n\theaders := map[string][]string{\n\t\t\"X-Registry-Auth\": {base64.URLEncoding.EncodeToString([]byte(\"{\\\"username\\\":\\\"\\\",\\\"password\\\":\\\"\\\", \\\"auth\\\":\\\"\\\",\\\"email\\\":\\\"\\\"}\"))},\n\t}\n\treturn cli.post(ctx, fmt.Sprintf(\"/images/%s/push\", image), v, nil, headers, true)\n}" ]
[ "0.7113327", "0.7086695", "0.69694716", "0.66233176", "0.6561084", "0.6558431", "0.6540989", "0.6505147", "0.6372477", "0.6327776", "0.6313685", "0.6308401", "0.6247429", "0.62335414", "0.62103194", "0.6168965", "0.61616284", "0.6116616", "0.61156374", "0.61033505", "0.60790366", "0.5956845", "0.5808478", "0.58023256", "0.5795065", "0.5790139", "0.5749945", "0.57374644", "0.57355356", "0.57353187", "0.57314956", "0.570606", "0.57048523", "0.5694087", "0.569143", "0.5663358", "0.56530845", "0.5645572", "0.56315506", "0.55290914", "0.5511041", "0.55015874", "0.549402", "0.54928", "0.5452873", "0.5449968", "0.54309815", "0.54283917", "0.5425529", "0.54129213", "0.53601956", "0.5352568", "0.534777", "0.53329504", "0.53313017", "0.5330214", "0.5315785", "0.5308105", "0.52961296", "0.5283601", "0.5281677", "0.5262512", "0.5260133", "0.52406573", "0.52379775", "0.52366614", "0.5235274", "0.52307856", "0.52256215", "0.52244973", "0.52219075", "0.5221543", "0.52191806", "0.52129513", "0.52099735", "0.5206782", "0.5200735", "0.5189994", "0.5171595", "0.51541495", "0.5151285", "0.51460797", "0.51345533", "0.5127511", "0.51231384", "0.51159304", "0.5114619", "0.5098608", "0.508222", "0.50806314", "0.5077503", "0.5070847", "0.5061759", "0.5045653", "0.5043022", "0.5030633", "0.5029899", "0.5028353", "0.5020066", "0.50164956" ]
0.6263923
12
NewLimitedWriter will automatically Close the writer on certain conditions defined by options. Since some options include async events the writes will be mutex protected meaning only 1 thread will be able to call Write/Close at the time
func NewLimitedWriter(w io.WriteCloser, options ...LimitedWriterOption) io.WriteCloser { for _, o := range options { w = o(w) } return NewSyncedWriteCloser(w) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewLimitedWriter(w io.Writer, limit int) io.Writer {\n\treturn &limitedWriter{w, limit}\n}", "func ConcurrentWriteLimit(n int) LoggerOption { return concurrentWriteLimit(n) }", "func RateLimitedWriter(w io.Writer, config *Configuration) io.Writer {\n\tif !config.isRateLimitedWr() {\n\t\treturn w\n\t}\n\treturn ratelimit.Writer(w,\n\t\tratelimit.NewBucketWithRate(\n\t\t\tconfig.rateLimitWr(),\n\t\t\tconfig.maxBurstCapWr()))\n}", "func NewLimitWriter(w io.Writer, limiters ...Limiter) *LimitWriter {\n\treturn &LimitWriter{\n\t\tw: w,\n\t\tlimiters: limiters,\n\t}\n}", "func NewWriter(w io.Writer, maxDelay time.Duration) *Writer {\n\treturn &Writer{\n\t\tw: bufio.NewWriter(w),\n\t\td: maxDelay,\n\t}\n}", "func New(w io.Writer, n int64) io.Writer {\n\treturn &limitWriter{\n\t\tw: w,\n\t\tn: n,\n\t}\n}", "func WithMaxBytes(maxBytes int) LimitedWriterOption {\n\tbytesWritten := 0\n\treturn func(w io.WriteCloser) io.WriteCloser {\n\t\tpreCheck := NewPreWriteCallbacks(w, func(p []byte) error {\n\t\t\tif bytesWritten+len(p) > maxBytes {\n\t\t\t\tif err := w.Close(); err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"failed to close WriteCloser writing maxBytes; Close error was: %w\", err)\n\t\t\t\t}\n\t\t\t\treturn ErrTooLargeWrite\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\n\t\treturn NewPostWriteCallbacks(preCheck, func(p []byte, n int, err error) {\n\t\t\tbytesWritten += n\n\t\t})\n\t}\n}", "func NewWriter(limit int) *Wrap {\n\treturn &Wrap{\n\t\tLimit: limit,\n\t\tNewline: defaultNewline,\n\t\tKeepNewlines: true,\n\t\t// Keep whitespaces following a forceful line break. If disabled,\n\t\t// leading whitespaces in a line are only kept if the line break\n\t\t// was not forceful, meaning a line break that was already present\n\t\t// in the input\n\t\tPreserveSpace: false,\n\t\tTabWidth: defaultTabWidth,\n\n\t\tbuf: &bytes.Buffer{},\n\t}\n}", "func NewCapWriter(limit int) *CapWriter {\n\tcw := make(CapWriter, 0, limit)\n\treturn &cw\n}", "func NewWithWriter(threads, length int, w PanicWriter) *Queue {\n\tvar q Queue\n\tq.j = make(jobs, length)\n\tq.w = w\n\tq.spawnThreads(threads)\n\treturn &q\n}", "func (th *Writer) ApplyLimit(l *Limiter) {\n\tth.limiter = l.Limiter\n}", "func WriterOpt(w io.Writer) Option {\n\treturn func(o *options) {\n\t\to.writer = w\n\t}\n}", "func newWriter(w io.Writer) *bufio.Writer {\n\t// TODO(nickng): use sync.Pool to reduce allocation per new connection.\n\treturn bufio.NewWriter(w)\n}", "func newWriter(filename string, verbose bool, out io.Writer) (io.Writer, error) {\n\twriters := make([]io.Writer, 0)\n\tif len(filename) > 0 || !verbose {\n\t\tfile, err := os.OpenFile(filename, os.O_RDWR|os.O_APPEND|os.O_CREATE, 0660)\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\twriters = append(writers, file)\n\t}\n\n\tif verbose {\n\t\twriters = append(writers, out)\n\t}\n\n\treturn io.MultiWriter(writers...), nil\n}", "func NewFileWriter(w io.Writer, limit Size) File {\n\treturn &FileWriter{Writer: w, Limit: limit}\n}", "func NewWriterSize(w io.Writer, maxDelay time.Duration, size int) *Writer {\n\treturn &Writer{\n\t\tw: bufio.NewWriterSize(w, size),\n\t\td: maxDelay,\n\t}\n}", "func (m *Nitro) NewWriter() *Writer {\n\tw := m.newWriter()\n\tw.next = m.wlist\n\tm.wlist = w\n\tw.dwrCtx.Init()\n\n\tm.shutdownWg1.Add(1)\n\tgo m.collectionWorker(w)\n\tif m.useMemoryMgmt {\n\t\tm.shutdownWg2.Add(1)\n\t\tgo m.freeWorker(w)\n\t}\n\n\treturn w\n}", "func NewWriterOnce(setters ...optionSetter) (WriterOnce, error) {\n\to := options{\n\t\tmemBytes: DefaultMemBytes,\n\t\tmaxBytes: DefaultMaxBytes,\n\t}\n\tfor _, s := range setters {\n\t\tif err := s(&o); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tif o.memBytes == 0 {\n\t\to.memBytes = DefaultMemBytes\n\t}\n\treturn &writerOnce{o: o}, nil\n}", "func NewSpanLimitWriter(spanWriter spanstore.Writer, options SpanLimitOptions) (*SpanLimitWriter, error) {\n\tcache, err := lru.New(options.cacheSize)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &SpanLimitWriter{\n\t\tspanWriter: spanWriter,\n\t\ttraceIDtoSpanCount: cache,\n\t\tmaxSpansPerTrace: options.maxSpansPerTrace,\n\t}, nil\n}", "func newWriter(config *config.Compaction) (flush.Writer, error) {\n\tvar writers []multi.SubWriter\n\n\t// Configure S3 writer if present\n\tif config.S3 != nil {\n\t\tw, err := s3.New(config.S3.Bucket, config.S3.Prefix, config.S3.Region, config.S3.Endpoint, config.S3.SSE, config.S3.AccessKey, config.S3.SecretKey, config.S3.Concurrency)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\twriters = append(writers, w)\n\t}\n\n\t// Configure Azure writer if present\n\tif config.Azure != nil {\n\t\tw, err := azure.New(config.Azure.Container, config.Azure.Prefix)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\twriters = append(writers, w)\n\t}\n\n\t// Configure GCS writer if present\n\tif config.GCS != nil {\n\t\tw, err := gcs.New(config.GCS.Bucket, config.GCS.Prefix)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\twriters = append(writers, w)\n\t}\n\n\t// Configure BigQuery writer if present\n\tif config.BigQuery != nil {\n\t\tw, err := bigquery.New(config.BigQuery.Project, config.BigQuery.Dataset, config.BigQuery.Table)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\twriters = append(writers, w)\n\t}\n\n\t// Configure File writer if present\n\tif config.File != nil {\n\t\tw, err := file.New(config.File.Directory)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\twriters = append(writers, w)\n\t}\n\n\t// Configure Talaria writer if present\n\tif config.Talaria != nil {\n\t\tw, err := talaria.New(config.Talaria.Endpoint, config.Talaria.CircuitTimeout, config.Talaria.MaxConcurrent, config.Talaria.ErrorPercentThreshold)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\twriters = append(writers, w)\n\t}\n\n\t// If no writers were configured, error out\n\tif len(writers) == 0 {\n\t\treturn noop.New(), errors.New(\"compact: writer was not configured\")\n\t}\n\n\t// Setup a multi-writer for all configured writers\n\treturn multi.New(writers...), nil\n}", "func (t *ThrottledWriteCloser) SetLimit(r rate.Limit, b int) error {\n\treturn t.pool.SetLimitByID(r, b, t.id)\n}", "func NewWriter(opt ...interface{}) *Writer {\r\n\tif len(opt) > 0 {\r\n\t\tswitch opt[0].(type) {\r\n\t\tcase []byte:\r\n\t\t\treturn &Writer{Bytes: opt[0].([]byte), c: 0, isFixed: true}\r\n\t\t}\r\n\t}\r\n\treturn &Writer{Bytes: []byte{}, c: 0, isFixed: false}\r\n}", "func NewWriter(fname string, bufferLength int) *Writer {\n\tw := &Writer{\n\t\trec: make(chan []byte, bufferLength),\n\t\trot: make(chan bool),\n\t\tfilename: fname,\n\t\trotate: true,\n\t}\n\n\t// open the file for the first time\n\tif err := w._Rotate(); err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"Writer(%q): %s\\n\", w.filename, err)\n\t\treturn nil\n\t}\n\n\tgo func() {\n\t\tdefer func() {\n\t\t\tif w.file != nil {\n\t\t\t\tw.file.Close()\n\t\t\t}\n\t\t}()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-w.rot:\n\t\t\t\tif err := w._Rotate(); err != nil {\n\t\t\t\t\tfmt.Fprintf(os.Stderr, \"Writer(%q): %s\\n\", w.filename, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\tcase rec, ok := <-w.rec:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tnow := time.Now()\n\t\t\t\tif (w.maxsize > 0 && w.maxsize_cursize >= w.maxsize) ||\n\t\t\t\t\t(w.daily && now.Day() != w.daily_opendate) {\n\t\t\t\t\tif err := w._Rotate(); err != nil {\n\t\t\t\t\t\tfmt.Fprintf(os.Stderr, \"Writer(%q): %s\\n\", w.filename, err)\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// Perform the write\n\t\t\t\tn, err := w.file.Write(rec)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Fprintf(os.Stderr, \"Writer(%q): %s\\n\", w.filename, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t// Update the counts\n\t\t\t\tw.maxsize_cursize += n\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn w\n}", "func NewWriter(conn *Connection, subscriber eventbus.Subscriber, keepAlive ...time.Duration) *Writer {\n\tkas := 30 * time.Second\n\tif len(keepAlive) > 0 {\n\t\tkas = keepAlive[0]\n\t}\n\n\tpw := &Writer{\n\t\tConnection: conn,\n\t\tsubscriber: subscriber,\n\t\tkeepAlive: kas,\n\t}\n\n\treturn pw\n}", "func New(w io.Writer, opts ...Option) *Writer {\n\twriter := &Writer{\n\t\tw: w,\n\t\ttimeFormat: \"2006-01-02 15:04:05\",\n\t\tlvl: logging.LevelInformational,\n\t}\n\n\tfor _, opt := range opts {\n\t\topt(writer)\n\t}\n\n\treturn writer\n}", "func newWriter(w http.ResponseWriter) *responseWriter {\n\trw := writerPool.Get().(*responseWriter)\n\trw.status = http.StatusOK\n\trw.ResponseWriter = w\n\trw.wroteHeader = false\n\trw.n = 0\n\n\treturn rw\n}", "func New(fpath string, options ...WriteOption) (*Writer, error) {\n\tf, err := os.OpenFile(fpath, os.O_CREATE+os.O_WRONLY+syscall.O_DIRECT, 0666)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdio, err := directio.New(f)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tw := bufio.NewWriterSize(dio, 67108864)\n\theader := [64]byte{}\n\t_, err = w.Write(header[:])\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\twr := &Writer{\n\t\tname: fpath,\n\t\tfile: f,\n\t\tbuf: w,\n\t\tdio: dio,\n\t\toffset: reservedHeader,\n\t\tindex: make(index, 0, 1000),\n\t\tmu: sync.Mutex{},\n\t}\n\tfor _, option := range options {\n\t\toption(wr)\n\t}\n\treturn wr, nil\n}", "func NewWrappedWriter(w io.Writer) (*WrappedWriter, error) {\n\treturn &WrappedWriter{wrapped: w, bw: bufio.NewWriterSize(w, 1024*1024)}, nil\n}", "func New(w io.Writer, size int) *Logger {\n\n\t// Create a value of type logger and init the channel and timer value.\n\tl := Logger{\n\t\twrite: make(chan string, size), // Buffered channel if size > 0.\n\t}\n\n\t// Add one to the waitgroup to track the write goroutine.\n\tl.wg.Add(1)\n\n\t// Create the write goroutine that performs the actual writes to disk.\n\tgo func() {\n\n\t\t// Range over the channel and write each data received to disk.\n\t\t// Once the channel is close and flushed the loop will terminate.\n\t\tfor d := range l.write {\n\n\t\t\t// Simulate write to disk.\n\t\t\tfmt.Fprintf(w, d)\n\t\t}\n\n\t\t// Mark that we are done and termianted.\n\t\tl.wg.Done()\n\t}()\n\treturn &l\n}", "func WithMaxConcurrentWriters(count int) WarcFileWriterOption {\n\treturn newFuncWarcFileOption(func(o *warcFileWriterOptions) {\n\t\to.maxConcurrentWriters = count\n\t})\n}", "func NewAsyncWriter(ctx context.Context, w io.Writer) io.Writer {\r\n\tctx, cancel := context.WithCancel(ctx)\r\n\ta := asyncWriter{\r\n\t\tctx: ctx,\r\n\t\twriter: w,\r\n\t\tlock: sync.Mutex{},\r\n\t\tdata: make(chan []byte),\r\n\t}\r\n\tgo func() {\r\n\t\tdefer func() {\r\n\t\t\tcancel()\r\n\t\t\t<-ctx.Done()\r\n\t\t\tclose(a.data)\r\n\t\t}()\r\n\t\tfor {\r\n\t\t\tselect {\r\n\t\t\tcase <-a.ctx.Done():\r\n\t\t\t\treturn\r\n\t\t\tdefault:\r\n\t\t\t}\r\n\t\t\tvar data []byte\r\n\t\t\ta.lock.Lock()\r\n\t\t\tswitch len(a.buffer) {\r\n\t\t\tcase 0:\r\n\t\t\t\ta.lock.Unlock()\r\n\t\t\t\tselect {\r\n\t\t\t\tcase <-a.ctx.Done():\r\n\t\t\t\t\treturn\r\n\t\t\t\tcase data = <-a.data:\r\n\t\t\t\t}\r\n\t\t\tdefault:\r\n\t\t\t\tdata = a.buffer[0]\r\n\t\t\t\ta.buffer = a.buffer[1:]\r\n\t\t\t\ta.lock.Unlock()\r\n\t\t\t}\r\n\t\t\tbuf := make([]byte, len(data))\r\n\t\t\tcopy(buf, data)\r\n\t\t\tif _, err := a.writer.Write(buf); err != nil {\r\n\t\t\t\ta.err = err\r\n\t\t\t\treturn\r\n\t\t\t}\r\n\t\t}\r\n\t}()\r\n\treturn &a\r\n}", "func (p *ioThrottlerPool) NewThrottledWriteCloser(writer io.WriteCloser, r rate.Limit, b int, id string) *ThrottledWriteCloser {\n\tp.mu.Lock()\n\tdefer p.mu.Unlock()\n\tthrottler := ioThrottler{\n\t\tlimiter: rate.NewLimiter(r, b),\n\t}\n\tp.connections[id] = &throttler\n\tp.updateBufferSize()\n\treturn &ThrottledWriteCloser{\n\t\torigWriteCloser: writer,\n\t\tid: id,\n\t\tpool: p,\n\t}\n\n}", "func NewWriter(opts WriterOptions) *Writer {\n\treturn &Writer{opts: opts}\n}", "func Writer(out io.WriteCloser) Option {\n\treturn func(o *options) {\n\t\to.out = out\n\t}\n}", "func TestShardWriter_Write_PoolMax(t *testing.T) {\n\tts := newTestWriteService(writeShardSlow)\n\ts := coordinator.NewService(coordinator.Config{\n\t\tShardWriterTimeout: toml.Duration(100 * time.Millisecond),\n\t})\n\ts.Listener = ts.muxln\n\ts.TSDBStore = &ts.TSDBStore\n\tif err := s.Open(); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tdefer s.Close()\n\tdefer ts.Close()\n\n\tw := coordinator.NewShardWriter(100*time.Millisecond, 1)\n\tw.MetaClient = &metaClient{host: ts.ln.Addr().String()}\n\tnow := time.Now()\n\n\tshardID := uint64(1)\n\townerID := uint64(2)\n\tvar points []models.Point\n\tpoints = append(points, models.MustNewPoint(\n\t\t\"cpu\", models.Tags{\"host\": \"server01\"}, map[string]interface{}{\"value\": int64(100)}, now,\n\t))\n\n\tgo w.WriteShard(shardID, ownerID, points)\n\ttime.Sleep(time.Millisecond)\n\tif err := w.WriteShard(shardID, ownerID, points); err == nil || err.Error() != \"timed out waiting for free connection\" {\n\t\tt.Fatalf(\"unexpected error: %v\", err)\n\t}\n}", "func (r *ThrottledWriteCloser) Close() error {\n\tr.pool.mu.Lock()\n\tdefer r.pool.mu.Unlock()\n\tdelete(r.pool.connections, r.id)\n\tr.pool.updateBufferSize()\n\treturn r.origWriteCloser.Close()\n}", "func NewWriter(w io.Writer, opts WriterOpts) (deprecated.LegacyPackedWriter, error) {\n\twr := &writer{opts: opts}\n\tsubopts := deprecated.LegacyPackedWriterOpts{\n\t\tMarshal: deprecated.MarshalFunc(opts.Marshal),\n\t\tIndex: opts.Index,\n\t\tMaxItems: opts.MaxItems,\n\t\tMaxBytes: opts.MaxBytes,\n\t\tFlushed: opts.Flushed,\n\t}\n\n\tcompress := false\n\tswitch opts.FlateLevel {\n\tcase flate.BestSpeed,\n\t\tflate.BestCompression,\n\t\tflate.DefaultCompression,\n\t\tflate.HuffmanOnly:\n\t\tcompress = true\n\t}\n\n\tif compress {\n\t\twr.compressor = NewFlateTransform(opts.FlateLevel)\n\t\tsubopts.Transform = wr.compressor.CompressTransform\n\t}\n\twr.LegacyPackedWriter = deprecated.NewLegacyPackedWriter(w, subopts)\n\treturn wr, nil\n}", "func (f *File) Writer(done chan struct{}, wg *sync.WaitGroup) (client.Writer, error) {\n\treturn newWriter(), nil\n}", "func NewLimited(workers uint) Pool {\n\tif workers == 0 {\n\t\tpanic(\"invalid workers '0'\")\n\t}\n\n\tp := &limitedPool{\n\t\tworkers: workers,\n\t}\n\n\tp.initialize()\n\n\treturn p\n}", "func NewWriter(w io.Writer) *Writer {\n\treturn NewWriterSize(w, defaultBufSize)\n}", "func NewFileWriter(w io.Writer, options ...FileWriterOption) *FileWriter {\n\tbw := bufio.NewWriter(w)\n\tfw := &FileWriter{\n\t\tw: &writePosStruct{\n\t\t\tw: bw,\n\t\t\tpos: 0,\n\t\t},\n\t\tbw: bw,\n\t\tversion: 1,\n\t\tschemaWriter: &schema{}, // no allocTracker is set here because we're creating a writer. We assume for the moment that writers have enough control over input that they're trusted.\n\t\tkvStore: make(map[string]string),\n\t\trowGroups: []*parquet.RowGroup{},\n\t\tcreatedBy: \"parquet-go\",\n\t\tnewPageFunc: newDataPageV1Writer,\n\t\tctx: context.Background(),\n\t}\n\n\tfor _, opt := range options {\n\t\topt(fw)\n\t}\n\n\t// if a WithSchemaDefinition option was provided, the schema needs to be set after everything else\n\t// as other options can change settings on the schemaWriter (such as the maximum page size).\n\tif fw.schemaDef != nil {\n\t\tif err := fw.schemaWriter.SetSchemaDefinition(fw.schemaDef); err != nil {\n\t\t\tpanic(err) // TODO: this shouldn't happen, but still isn't great. We messed up the API design for options and NewFileWriter.\n\t\t}\n\t}\n\n\treturn fw\n}", "func NewWriter(out io.Writer, opts *Options) (*Writer, error) {\n\to := *opts\n\to.setDefaults()\n\tw := &Writer{\n\t\topts: o,\n\t\tblock: make([]byte, o.BlockSize),\n\t}\n\n\tif opts.BlockSize >= (1 << 24) {\n\t\treturn nil, errors.New(\"reftable: invalid blocksize\")\n\t}\n\n\tw.paddedWriter.out = out\n\tw.Stats.BlockStats = map[byte]*BlockStats{}\n\tfor _, c := range \"rgoi\" {\n\t\tw.Stats.BlockStats[byte(c)] = new(BlockStats)\n\t}\n\n\tif opts.IndexObjects {\n\t\tw.objIndex = map[string][]uint64{}\n\t}\n\n\tw.blockWriter = w.newBlockWriter(BlockTypeRef)\n\treturn w, nil\n}", "func (t *ThrottledReadWriteCloser) SetLimit(rRead, rWrite rate.Limit, bRead, bWrite int) error {\n\terr := t.ThrottledReadCloser.SetLimit(rRead, bRead)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn t.ThrottledWriteCloser.SetLimit(rWrite, bWrite)\n}", "func WithWriteBufSize(s int) core.BuildOption {\n\treturn func(o interface{}) {\n\t\to.(*serverOptions).WriteBufSize = s\n\t}\n}", "func Writer(w io.Writer) Option {\n\treturn func(c *Config) Option {\n\t\tprevious := c.ioWriter\n\t\tc.ioWriter = w\n\t\treturn Writer(previous)\n\t}\n}", "func (th *Writer) Write(p []byte) (int, error) {\n\tif th.limiter == nil {\n\t\treturn th.w.Write(p)\n\t}\n\tn, err := th.w.Write(p)\n\tif err != nil {\n\t\treturn n, err\n\t}\n\n\tif err := th.limiter.WaitN(th.ctx, n); err != nil {\n\t\treturn n, err\n\t}\n\n\treturn n, err\n}", "func (c *LimitedConnection) Write(b []byte) (written int, err error) {\n\treturn c.rateLimitLoop(&c.writeNotBefore, &c.writeDeadline, c.inner.Write, b)\n}", "func NewWriter(path, label string, transformers []string, bufFreePool *WriteBufPool, errReporter *errors.Once) *Writer {\n\tmu := &sync.Mutex{}\n\tfw := &Writer{\n\t\tlabel: label,\n\t\tbufFreePool: bufFreePool,\n\t\tmu: mu,\n\t\tcond: sync.NewCond(mu),\n\t\tlastBlockFlushed: -1,\n\t\terr: errReporter,\n\t}\n\tfw.NewBuf()\n\t// Create a recordio file\n\tctx := vcontext.Background()\n\tout, err := file.Create(ctx, path)\n\tif err != nil {\n\t\tfw.err.Set(errors.E(err, fmt.Sprintf(\"fieldio newwriter %s\", path)))\n\t\treturn fw\n\t}\n\tfw.out = out\n\tfw.wout = out.Writer(ctx)\n\tfw.rio = recordio.NewWriter(fw.wout, recordio.WriterOpts{\n\t\tTransformers: transformers,\n\t\tMarshal: fw.marshalBlock,\n\t\tIndex: fw.indexCallback,\n\t\tMaxFlushParallelism: 2,\n\t})\n\tfw.rio.AddHeader(recordio.KeyTrailer, true)\n\treturn fw\n}", "func (logger *Logger) NewWriter(level event.Level) io.WriteCloser {\n\tif level < 0 {\n\t\tlevel = event.Info\n\t}\n\n\tpipeReader, pipeWriter := io.Pipe()\n\n\t//TODO(.) I don't like goroutines, and there's no technical reason why this requires one.\n\t// The current implementation is just simpler.\n\tgo logger.writerScanner(pipeReader, level)\n\truntime.SetFinalizer(pipeWriter, writerFinalizer)\n\n\treturn pipeWriter\n}", "func (t *ThrottledConn) SetLimit(rRead, rWrite rate.Limit, bRead, bWrite int) error {\n\treturn t.ThrottledReadWriteCloser.SetLimit(rRead, rWrite, bRead, bWrite)\n}", "func NewWriter(w io.Writer) *Writer {\n\tvar bw Writer\n\tbw.Reset(w)\n\treturn &bw\n}", "func (w *LimitWriter) Write(p []byte) (n int, err error) {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\tfor i := 0; i < len(w.limiters); i++ {\n\t\tif !w.limiters[i].AllowN(int64(len(p))) {\n\t\t\tfor j := i - 1; j >= 0; j-- {\n\t\t\t\tw.limiters[j].AllowN(-int64(len(p))) // Revert limiters limits if not allowed\n\t\t\t}\n\t\t\treturn 0, ErrLimitReached\n\t\t}\n\t}\n\tn, err = w.w.Write(p)\n\tw.written += int64(n)\n\treturn\n}", "func NewWriter(w io.Writer) io.WriteCloser {\n\treturn NewWriterSizeLevel(w, -1, DefaultCompression)\n}", "func NewWrappedWriter(writer store.Writer, onFinalize FinalizeFunc) *WrappedWriter {\n\treturn &WrappedWriter{writer: writer, onFinalize: onFinalize}\n}", "func WriteTimeout(v time.Duration) Option {\n\treturn optionSetter(func(opt *Options) {\n\t\topt.WriteTimeout = v\n\t})\n}", "func WriteTimeout(t time.Duration) Option {\n\treturn func(ws *websocket) {\n\t\tws.options.writeTimeout = t\n\t}\n}", "func NewWriter(f File, o *db.Options) *Writer {\n\tw := &Writer{\n\t\tcloser: f,\n\t\tblockRestartInterval: o.GetBlockRestartInterval(),\n\t\tblockSize: o.GetBlockSize(),\n\t\tcmp: o.GetComparer(),\n\t\tcompression: o.GetCompression(),\n\t\tprevKey: make([]byte, 0, 256),\n\t\trestarts: make([]uint32, 0, 256),\n\t}\n\tif f == nil {\n\t\tw.err = errors.New(\"leveldb/table: nil file\")\n\t\treturn w\n\t}\n\t// If f does not have a Flush method, do our own buffering.\n\ttype flusher interface {\n\t\tFlush() error\n\t}\n\tif _, ok := f.(flusher); ok {\n\t\tw.writer = f\n\t} else {\n\t\tw.bufWriter = bufio.NewWriter(f)\n\t\tw.writer = w.bufWriter\n\t}\n\treturn w\n}", "func OptionSetWriter(w io.Writer) Option {\n\treturn func(p *ProgressBar) {\n\t\tp.config.writer = w\n\t}\n}", "func NewChunkedWriter(w io.Writer) io.WriteCloser", "func NewWriter(filename string, maxsize int64) *Writer {\n\treturn &Writer{\n\t\tName: filename,\n\t\tMaxsize: maxsize,\n\t}\n}", "func NewBufferedWriter(client *driver.Client, db, collection string, size int) *BufferedWriterMongo {\n\tif size == 0 {\n\t\tsize = 1\n\t}\n\treturn &BufferedWriterMongo{\n\t\tbufferSize: size,\n\t\tdb: db,\n\t\tcollection: collection,\n\t\tbuffer: make([]interface{}, 0, size+1),\n\t\tclient: client,\n\t}\n}", "func newSeriesLimiter(maxSeries int) queryrange.Middleware {\n\treturn seriesLimiterMiddleware(maxSeries)\n}", "func NewWarcFileWriter(opts ...WarcFileWriterOption) *WarcFileWriter {\n\to := defaultwarcFileWriterOptions()\n\tfor _, opt := range opts {\n\t\topt.apply(&o)\n\t}\n\tw := &WarcFileWriter{opts: &o,\n\t\tclosing: make(chan struct{}), // signal channel\n\t\tclosed: make(chan struct{}),\n\t\tmiddleCh: make(chan *job),\n\t\tjobs: make(chan *job),\n\t\tshutWriters: &sync.WaitGroup{},\n\t}\n\tw.shutWriters.Add(o.maxConcurrentWriters)\n\n\t// the middle layer\n\tgo func() {\n\t\texit := func(v *job, needSend bool) {\n\t\t\tclose(w.closed)\n\t\t\tif needSend {\n\t\t\t\tw.jobs <- v\n\t\t\t}\n\t\t\tclose(w.jobs)\n\t\t}\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-w.closing:\n\t\t\t\texit(nil, false)\n\t\t\t\treturn\n\t\t\tcase v := <-w.middleCh:\n\t\t\t\tselect {\n\t\t\t\tcase <-w.closing:\n\t\t\t\t\texit(v, true)\n\t\t\t\t\treturn\n\t\t\t\tcase w.jobs <- v:\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tfor i := 0; i < o.maxConcurrentWriters; i++ {\n\t\twriter := &singleWarcFileWriter{opts: &o, shutWriters: w.shutWriters}\n\t\tif o.compress {\n\t\t\twriter.gz, _ = gzip.NewWriterLevel(nil, o.gzipLevel)\n\t\t}\n\t\tw.writers = append(w.writers, writer)\n\t\tgo worker(writer, w.jobs)\n\t}\n\treturn w\n}", "func NewWriter(w io.Writer) *csv.Writer", "func newWriteOptions() writeOptions {\n\treturn writeOptions{CreateDisposition: bigquery.CreateIfNeeded}\n}", "func (w *Writer) Close() error {}", "func New() *Writer {\n\ttermWidth, _ = getTermSize()\n\tif termWidth != 0 {\n\t\toverFlowHandled = true\n\t}\n\n\treturn &Writer{\n\t\tOut: Out,\n\t\tRefreshInterval: RefreshInterval,\n\n\t\tmtx: &sync.Mutex{},\n\t}\n}", "func New(size int) *Logger {\n\n\t// Create a value of type logger and init the channel\n\t// and timer value.\n\tl := Logger{\n\t\twrite: make(chan string, size), // Buffered channel if size > 0.\n\t\ttimer: time.NewTimer(time.Hour), // Some abitrary large value.\n\t}\n\n\t// Add one to the waitgroup to track\n\t// the write goroutine.\n\tl.wg.Add(1)\n\n\t// Create the write goroutine that performs the actual\n\t// writes to disk.\n\tgo func() {\n\n\t\t// Range over the channel and write each data received to disk.\n\t\t// Once the channel is close and flushed the loop will terminate.\n\t\tfor d := range l.write {\n\n\t\t\t// Help to simulate disk latency issues.\n\t\t\t// WOULD NOT NEED THIS IN PRODUCTION CODE.\n\t\t\tl.pretendDiskFull()\n\n\t\t\t// Write to disk and decrement the pendingWrites counter.\n\t\t\tlog.Println(d)\n\t\t\tatomic.AddInt32(&l.pendingWrites, -1)\n\t\t}\n\n\t\t// Mark that we are done and terminated.\n\t\tl.wg.Done()\n\t}()\n\n\treturn &l\n}", "func WithLimit(l int) Opts {\n\treturn func(r *retryable) {\n\t\tif l > 0 {\n\t\t\tr.limit = l\n\t\t}\n\t}\n}", "func WithWriter(w io.Writer) Option {\n\treturn func(g *Generator) error {\n\t\tg.getWriter = func() (io.Writer, error) {\n\t\t\treturn w, nil\n\t\t}\n\t\treturn nil\n\t}\n}", "func NewQueuedWriter(w OutputWriter) OutputWriter {\n\tqw := &queuedWriter{\n\t\tOutputWriter: w,\n\t\tnextBufSz: 8,\n\t\t// 20 is more than enough to reach the max size of 1024*1024.\n\t\tqueues: make(chan chan queuedMessage, 20),\n\t\tdone: make(chan error),\n\t}\n\tgo qw.run()\n\treturn qw\n}", "func NewLimitedRoundTripper(next http.RoundTripper, codec queryrange.Codec, limits Limits, middlewares ...queryrange.Middleware) http.RoundTripper {\n\ttransport := limitedRoundTripper{\n\t\tnext: next,\n\t\tcodec: codec,\n\t\tlimits: limits,\n\t\tmiddleware: queryrange.MergeMiddlewares(middlewares...),\n\t}\n\treturn transport\n}", "func NewWriter(restartInterval int) *Writer {\n\treturn &Writer{\n\t\trestartInterval: restartInterval,\n\t\tbuf: new(bytes.Buffer),\n\t}\n}", "func newWriteBackoff() wait.Backoff {\n\t// Return a exponential backoff configuration which returns durations for a total time of ~40s.\n\t// Example: 0, .5s, 1.2s, 2.3s, 4s, 6s, 10s, 16s, 24s, 37s\n\t// Jitter is added as a random fraction of the duration multiplied by the jitter factor.\n\treturn wait.Backoff{\n\t\tDuration: 500 * time.Millisecond,\n\t\tFactor: 1.5,\n\t\tSteps: 5,\n\t\tJitter: 0.4,\n\t}\n}", "func NewStatsWriter(cfg *config.AgentConfig, in <-chan *pb.StatsPayload, telemetryCollector telemetry.TelemetryCollector) *StatsWriter {\n\tsw := &StatsWriter{\n\t\tin: in,\n\t\tstats: &info.StatsWriterInfo{},\n\t\tstop: make(chan struct{}),\n\t\tflushChan: make(chan chan struct{}),\n\t\tsyncMode: cfg.SynchronousFlushing,\n\t\teasylog: log.NewThrottled(5, 10*time.Second), // no more than 5 messages every 10 seconds\n\t\tconf: cfg,\n\t}\n\tclimit := cfg.StatsWriter.ConnectionLimit\n\tif climit == 0 {\n\t\t// Allow 1% of the connection limit to outgoing sends. The original\n\t\t// connection limit was removed and used to be 2000 (1% = 20)\n\t\tclimit = 20\n\t}\n\tqsize := cfg.StatsWriter.QueueSize\n\tif qsize == 0 {\n\t\tpayloadSize := float64(maxEntriesPerPayload * bytesPerEntry)\n\t\t// default to 25% of maximum memory.\n\t\tmaxmem := cfg.MaxMemory / 4\n\t\tif maxmem == 0 {\n\t\t\t// or 250MB if unbound\n\t\t\tmaxmem = 250 * 1024 * 1024\n\t\t}\n\t\tqsize = int(math.Max(1, maxmem/payloadSize))\n\t}\n\tlog.Debugf(\"Stats writer initialized (climit=%d qsize=%d)\", climit, qsize)\n\tsw.senders = newSenders(cfg, sw, pathStats, climit, qsize, telemetryCollector)\n\treturn sw\n}", "func wrapWriter(w http.ResponseWriter) writerProxy {\n\tbw := basicWriter{ResponseWriter: w}\n\treturn &bw\n}", "func NewMaxWriter() MaxWriter {\n\treturn MaxWriter{}\n}", "func (d *driver) Writer(ctx context.Context, path string, append bool) (storagedriver.FileWriter, error) {\n\tkey := d.obsPath(path)\n\tif !append {\n\t\t// TODO (brianbland): cancel other uploads at this path\n\t//\tobs.InitLog(\"/obs_log/OBS-SDK.log\", 1024*1024*100, 10, obs.LEVEL_INFO, false)\n\t//\tobs.SyncLog()\n\t\toutput, err := d.Client.InitiateMultipartUpload(&obs.InitiateMultipartUploadInput{\n\t\t\tObjectOperationInput: obs.ObjectOperationInput{\n\t\t\t\tBucket: d.Bucket,\n\t\t\t\tKey: key,\n\t\t\t\tACL: d.getACL(),\n\t\t\t\tStorageClass: d.getStorageClass(),\n\t\t\t\t//SseHeader: obs.SseKmsHeader{Encryption: obs.DEFAULT_SSE_KMS_ENCRYPTION},\n\t\t\t},\n\t\t\tContentType: d.getContentType(),\n\t\t})\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn d.newWriter(key, output.UploadId, nil), nil\n\t}\n\n\toutput, err := d.Client.ListMultipartUploads(&obs.ListMultipartUploadsInput{\n\t\tBucket: d.Bucket,\n\t\tPrefix: key,\n\t})\n\tif err != nil {\n\t\treturn nil, parseError(path, err)\n\t}\n\n\tfor _, multi := range output.Uploads {\n\t\tif key != multi.Key {\n\t\t\tcontinue\n\t\t}\n\t\toutput, err := d.Client.ListParts(&obs.ListPartsInput{\n\t\t\tBucket: d.Bucket,\n\t\t\tKey: key,\n\t\t\tUploadId: multi.UploadId,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn nil, parseError(path, err)\n\t\t}\n\n\t\tvar multiSize int64\n\t\tfor _, part := range output.Parts {\n\t\t\tmultiSize += part.Size\n\t\t}\n\t\treturn d.newWriter(key, multi.UploadId, output.Parts), nil\n\t}\n\treturn nil, storagedriver.PathNotFoundError{Path: path}\n}", "func (q *T) NewWriter(id bqueue.ID, p bqueue.Priority, bytes int) (bqueue.Writer, error) {\n\tw := &writer{\n\t\tid: id,\n\t\tpriority: p,\n\t\tq: q,\n\t\tfree: vsync.NewSemaphore(),\n\t\tisActive: idle,\n\t}\n\tw.free.IncN(uint(bytes))\n\n\tq.mutex.Lock()\n\tdefer q.mutex.Unlock()\n\tif q.isClosed {\n\t\treturn nil, bqueue.ErrBQueueIsClosed\n\t}\n\tq.addPriorityLocked(p)\n\tif _, ok := q.writers[w.id]; ok {\n\t\treturn nil, bqueue.ErrWriterAlreadyExists\n\t}\n\tq.writers[w.id] = w\n\treturn w, nil\n}", "func NewWriterLevel(w io.Writer, level int) io.WriteCloser {\n\treturn NewWriterSizeLevel(w, -1, level)\n}", "func NewBufioWriter(file string, sec int) *BufioWriter {\n\tvar buf bytes.Buffer\n\n\tbw := &BufioWriter{file: file, ticker: sec, buf: &buf}\n\tbw.cmd = make(chan int)\n\tbw.cmdWait = make(chan int)\n\n\tticker := time.NewTicker(time.Duration(sec) * time.Second)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tbw.flush()\n\t\t\tcase c := <-bw.cmd:\n\t\t\t\tswitch c {\n\t\t\t\tcase 1: // flush\n\t\t\t\t\tbw.flush()\n\t\t\t\tcase 2:\n\t\t\t\t\tbw.flush()\n\t\t\t\t\tbw.cmdWait <- 1\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn bw\n}", "func NewWriterSizeLevel(w io.Writer, size int64, level int) io.WriteCloser {\n\t// the reason for which size is an argument is that lzma, unlike gzip,\n\t// stores the size before any compressed data. gzip appends the size and\n\t// the checksum at the end of the stream, thus it can compute the size\n\t// while reading data from pipe.\n\tvar z encoder\n\tpr, pw := syncPipe()\n\tgo func() {\n\t\terr := z.encoder(pr, w, size, level)\n\t\tpr.CloseWithError(err)\n\t}()\n\treturn pw\n}", "func NewWriter(w io.Writer) *Writer {\n\tif bw, ok := w.(*Writer); ok {\n\t\treturn bw\n\t}\n\treturn &Writer{\n\t\tfree: 8,\n\t\tw: w,\n\t}\n}", "func WithForcedWriter(w io.Writer) LoggersOption {\n\treturn func(l *Loggers) {\n\t\tl.writer = w\n\t}\n}", "func newWriter(config *config.Stream, fs afs.Service, rotationURL string, index int, created time.Time, emitter *emitter.Service) (*writer, error) {\n\twriterCloser, err := fs.NewWriter(context.Background(), config.URL, file.DefaultFileOsMode)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresult := &writer{\n\t\tfs: fs,\n\t\tindex: index,\n\t\tdestURL: config.URL,\n\t\trotationURL: rotationURL,\n\t\tcloser: writerCloser,\n\t\tcreated: created,\n\t}\n\tresult.config = config\n\n\tif rotation := config.Rotation; rotation != nil {\n\t\tinitRotation(result, rotation, created, emitter)\n\t}\n\tif config.IsGzip() {\n\t\tgzWriter := gzip.NewWriter(writerCloser)\n\t\tresult.writer = gzWriter\n\t\tresult.flusher = gzWriter\n\n\t} else {\n\t\twriter := bufio.NewWriter(writerCloser)\n\t\tresult.writer = writer\n\t\tresult.flusher = writer\n\t}\n\treturn result, nil\n}", "func NewWriter(w io.Writer) *Writer {\n\treturn NewWriterWidth(w, 16)\n}", "func (w *FlushingWriter) Write(data []byte) (written int, err error) {\n\tw.writeMutex.Lock()\n\tdefer w.writeMutex.Unlock()\n\tif w.closed {\n\t\treturn 0, io.EOF\n\t}\n\tw.wrote = true\n\twritten, err = w.WriterFlusher.Write(data)\n\tif err != nil {\n\t\treturn\n\t}\n\tif w.MaxLatency == 0 {\n\t\tw.WriterFlusher.Flush()\n\t\treturn\n\t}\n\tif w.flushPending {\n\t\treturn\n\t}\n\tw.flushPending = true\n\tif w.timer == nil {\n\t\tw.timer = time.AfterFunc(w.MaxLatency, w.delayedFlush)\n\t} else {\n\t\tw.timer.Reset(w.MaxLatency)\n\t}\n\treturn\n}", "func newFlushWriter(w io.Writer) flushWriter {\n\tfw := flushWriter{writer: w}\n\tif f, ok := w.(http.Flusher); ok {\n\t\tfw.flusher = f\n\t}\n\n\treturn fw\n}", "func Writer(c LineChan, exitChan chan int, client relp.Client) error {\n\tdefer func() { exitChan <- 1 }()\n\n\t// lush every 10 seconds\n\tticker := time.NewTicker(10 * time.Second)\n\n\tmore := true\n\tfor {\n\t\t// read new string and flush the output channel periodically\n\t\tselect {\n\t\tcase line, more := <-c:\n\t\t\tif more {\n\t\t\t\tclient.SendString(*line)\n\t\t\t}\n\t\t\tbreak\n\t\tcase <-ticker.C:\n\t\t\tclient.Flush()\n\t\t\tbreak\n\t\t}\n\n\t\tif !more {\n\t\t\tif debug {\n\t\t\t\tlog.Debug(\"Writer exiting due to empty channel\")\n\t\t\t}\n\t\t\tif *doDrain {\n\t\t\t\t_ = client.Drain(30 * time.Second)\n\t\t\t}\n\t\t\treturn client.Close()\n\t\t}\n\t}\n}", "func NewRateLimitedLogger(logger Interface, logsPerSecond rate.Limit, burstSize int, reg prometheus.Registerer) Interface {\n\tdiscardedLogLinesCounter := promauto.With(reg).NewCounterVec(prometheus.CounterOpts{\n\t\tName: \"logger_rate_limit_discarded_log_lines_total\",\n\t\tHelp: \"Total number of discarded log lines per level.\",\n\t}, []string{\"level\"})\n\n\treturn &RateLimitedLogger{\n\t\tnext: logger,\n\t\tlimiter: rate.NewLimiter(logsPerSecond, burstSize),\n\t\tdiscardedInfoLogLinesCounter: discardedLogLinesCounter.WithLabelValues(infoLevel),\n\t\tdiscardedDebugLogLinesCounter: discardedLogLinesCounter.WithLabelValues(debugLevel),\n\t\tdiscardedWarnLogLinesCounter: discardedLogLinesCounter.WithLabelValues(warnLevel),\n\t\tdiscardedErrorLogLinesCounter: discardedLogLinesCounter.WithLabelValues(errorLevel),\n\t}\n}", "func NewTimedWriter(bw *BufferedWriter, d time.Duration) *TimedWriter {\n\tw := &TimedWriter{\n\t\tBufferedWriter: bw,\n\t\tticker: time.NewTicker(d),\n\t\tdone: make(chan struct{}),\n\t}\n\tgo w.loop()\n\treturn w\n}", "func NewLimitedConnection(inner net.Conn, limiter *MultiLimiter) *LimitedConnection {\n\tbufSize := limiter.Burst()\n\tif bufSize > MaxBurstSize {\n\t\tbufSize = MaxBurstSize\n\t}\n\treturn &LimitedConnection{\n\t\tinner: inner,\n\n\t\tlimiterMu: new(sync.RWMutex),\n\t\tlimiter: limiter,\n\t\tabortWait: make(chan struct{}),\n\n\t\tclose: make(chan struct{}),\n\t\tupdateLimiter: make(chan *MultiLimiter),\n\t}\n}", "func NewWriter(t mockConstructorTestingTNewWriter) *Writer {\n\tmock := &Writer{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func WithWriteTimeout(timeout time.Duration) OptsFunc {\n\treturn func(o *redis.Options) {\n\t\to.WriteTimeout = timeout\n\t}\n}", "func NewWriter(cfg WriterConfig) (w *Writer, e error) {\n\tcfg.applyDefaults()\n\tif e := cfg.validate(); e != nil {\n\t\treturn nil, e\n\t}\n\n\tw = &Writer{\n\t\tfilename: cfg.Filename,\n\t\tc: eal.Zmalloc[C.HrlogWriter](\"HrlogWriter\", C.sizeof_HrlogWriter, cfg.Socket),\n\t}\n\tw.c.filename = C.CString(cfg.Filename)\n\tw.c.count = C.int64_t(cfg.Count)\n\n\tw.ThreadWithCtrl = ealthread.NewThreadWithCtrl(\n\t\tcptr.Func0.C(C.HrlogWriter_Run, w.c),\n\t\tunsafe.Pointer(&w.c.ctrl),\n\t)\n\tdefer func() {\n\t\tif e != nil {\n\t\t\tw.Close()\n\t\t}\n\t}()\n\n\tw.queue, e = ringbuffer.New(cfg.RingCapacity, cfg.Socket, ringbuffer.ProducerMulti, ringbuffer.ConsumerSingle)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\tw.c.queue = (*C.struct_rte_ring)(w.queue.Ptr())\n\n\tlogger.Info(\"Writer open\",\n\t\tzap.String(\"filename\", cfg.Filename),\n\t\tzap.Uintptr(\"queue\", uintptr(unsafe.Pointer(w.c.queue))),\n\t)\n\treturn w, nil\n}", "func newMetricsWriter(w http.ResponseWriter, r *http.Request, collector collector) *metricWriter {\n\tinfo := &Info{TimeStart: time.Now(), Request: r, Header: w.Header()}\n\treturn &metricWriter{w: w, info: info, collector: collector}\n}", "func NewLimitMW(lc LimitCond) endpoint.Middleware {\n\treturn func(next endpoint.EndPoint) endpoint.EndPoint {\n\t\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\t\tif lc.Current() > lc.Limit() {\n\t\t\t\treturn nil, fmt.Errorf(\"over limit qps, current is %d, limit is %d\", lc.Current(), lc.Limit())\n\t\t\t}\n\t\t\tlc.Incr()\n\t\t\treturn next(ctx, request)\n\t\t}\n\t}\n}", "func wrapWriter(w http.ResponseWriter) writerProxy {\n\tvar _, cn = w.(http.CloseNotifier) // nolint\n\tvar _, fl = w.(http.Flusher)\n\tvar _, hj = w.(http.Hijacker)\n\tvar _, rf = w.(io.ReaderFrom)\n\n\tvar bw = basicWriter{ResponseWriter: w}\n\tif cn && fl && hj && rf {\n\t\treturn &fancyWriter{&bw}\n\t}\n\tif fl {\n\t\treturn &flushWriter{&bw}\n\t}\n\treturn &bw\n}", "func NewBufferedWriter(w io.Writer, size int, mode IOMode) (bw *BufferedWriter) {\n\n\tbw = &BufferedWriter{\n\t\twriter: w,\n\t\tmode: mode,\n\t\tbuffer: make([]byte, size),\n\t\tbuffered: 0,\n\t\tmustFlush: false,\n\t}\n\n\treturn bw\n}", "func DialWriteTimeout(d time.Duration) DialOpt {\n\treturn func(do *dialOpts) {\n\t\tdo.writeTimeout = d\n\t}\n}" ]
[ "0.716447", "0.69312716", "0.66316056", "0.6544516", "0.59969217", "0.59826624", "0.59427977", "0.5693132", "0.5654763", "0.56334114", "0.5568591", "0.5514889", "0.54932827", "0.54898715", "0.54849005", "0.5482889", "0.5476103", "0.54604113", "0.5444612", "0.541179", "0.53988415", "0.53875554", "0.53770167", "0.53318614", "0.53305024", "0.53114223", "0.53052926", "0.52880085", "0.5280019", "0.52727246", "0.5240763", "0.5240085", "0.52271855", "0.522329", "0.5217432", "0.51898056", "0.51773983", "0.516838", "0.5164119", "0.51618433", "0.5155341", "0.5154218", "0.5150873", "0.51262903", "0.5120834", "0.5110608", "0.50890493", "0.50823534", "0.5078621", "0.5070471", "0.50678974", "0.5064291", "0.5057263", "0.50421095", "0.50414276", "0.5039587", "0.50240827", "0.5015426", "0.4997617", "0.4981385", "0.49782813", "0.49726254", "0.49708295", "0.4967046", "0.4965111", "0.49617636", "0.4958977", "0.49470237", "0.4945208", "0.49401855", "0.49384055", "0.49298438", "0.49257135", "0.49113935", "0.490904", "0.48941243", "0.48935562", "0.4887718", "0.48873612", "0.48830628", "0.48770687", "0.48684704", "0.48545107", "0.48418868", "0.4836449", "0.4831138", "0.48309737", "0.48258668", "0.48139495", "0.48122093", "0.48115778", "0.48114377", "0.48102447", "0.48094904", "0.48071918", "0.47939712", "0.4789481", "0.47885314", "0.478754", "0.47845843" ]
0.79731476
0
WithMaxBytes will block writes which would make the total stream larger than maxBytes.
func WithMaxBytes(maxBytes int) LimitedWriterOption { bytesWritten := 0 return func(w io.WriteCloser) io.WriteCloser { preCheck := NewPreWriteCallbacks(w, func(p []byte) error { if bytesWritten+len(p) > maxBytes { if err := w.Close(); err != nil { return fmt.Errorf("failed to close WriteCloser writing maxBytes; Close error was: %w", err) } return ErrTooLargeWrite } return nil }) return NewPostWriteCallbacks(preCheck, func(p []byte, n int, err error) { bytesWritten += n }) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func WithMaxInflightBytes(n int) WriterOption {\n\treturn func(ms *ManagedStream) {\n\t\tms.streamSettings.MaxInflightBytes = n\n\t}\n}", "func MaxBytes(m int64) optionSetter {\n\treturn func(o *options) error {\n\t\to.maxBytes = m\n\t\treturn nil\n\t}\n}", "func WithMaxBytes(maxBytes int) ProducerOption {\n\treturn func(p *ProducerConfiguration) {\n\t\t_ = p.KafkaConfig.SetKey(\"message.max.bytes\", maxBytes)\n\t}\n}", "func WithMaxBatchSizeBytes(n int64) WriteHandlerOption {\n\treturn func(w *WriteHandler) {\n\t\tw.maxBatchSizeBytes = n\n\t}\n}", "func (s *Server) SetMaxHeaderBytes(b int) {\n\ts.config.MaxHeaderBytes = b\n}", "func MaxDataBytes(maxBytes, evidenceBytes int64, valsCount int) int64 {\n\tmaxDataBytes := maxBytes -\n\t\tMaxOverheadForBlock -\n\t\tMaxHeaderBytes -\n\t\tMaxCommitBytes(valsCount) -\n\t\tevidenceBytes\n\n\tif maxDataBytes < 0 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"Negative MaxDataBytes. Block.MaxBytes=%d is too small to accommodate header&lastCommit&evidence=%d\",\n\t\t\tmaxBytes,\n\t\t\t-(maxDataBytes - maxBytes),\n\t\t))\n\t}\n\n\treturn maxDataBytes\n}", "func (s *settings) SetMaxWriteSize(size uint) {\n\ts.wMaxSize = size\n}", "func MaxBytesHandler(h Handler, n int64) Handler {\n\treturn HandlerFunc(func(w ResponseWriter, r *Request) {\n\t\tr2 := *r\n\t\tr2.Body = MaxBytesReader(w, r.Body, n)\n\t\th.ServeHTTP(w, &r2)\n\t})\n}", "func MaxDataBytes(maxBytes int64, keyType crypto.KeyType, evidenceBytes int64, valsCount int) int64 {\n\tmaxDataBytes := maxBytes -\n\t\tMaxOverheadForBlock -\n\t\tMaxHeaderBytes -\n\t\tMaxCoreChainLockSize -\n\t\tMaxCommitOverheadBytes -\n\t\tevidenceBytes\n\n\tif maxDataBytes < 0 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"Negative MaxDataBytes. Block.MaxBytes=%d is too small to accommodate header&lastCommit&evidence=%d\",\n\t\t\tmaxBytes,\n\t\t\t-(maxDataBytes - maxBytes),\n\t\t))\n\t}\n\n\treturn maxDataBytes\n}", "func BufferedByteLimit(n int) LoggerOption { return bufferedByteLimit(n) }", "func MaxBytesReader(w ResponseWriter, r io.ReadCloser, n int64) io.ReadCloser {\n\treturn &maxBytesReader{respWriter: w, readCloser: r, bytesRemaining: n}\n}", "func MaxHeaderBytes(v int) Option {\n\treturn optionSetter(func(opt *Options) {\n\t\topt.MaxHeaderBytes = v\n\t})\n}", "func (mm *BytesMonitor) MaximumBytes() int64 {\n\tmm.mu.Lock()\n\tdefer mm.mu.Unlock()\n\treturn mm.mu.maxAllocated\n}", "func (e SszNetworkEncoder) EncodeWithMaxLength(w io.Writer, msg interface{}) (int, error) {\n\tif msg == nil {\n\t\treturn 0, nil\n\t}\n\tb, err := e.doEncode(msg)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tif uint64(len(b)) > MaxChunkSize {\n\t\treturn 0, fmt.Errorf(\n\t\t\t\"size of encoded message is %d which is larger than the provided max limit of %d\",\n\t\t\tlen(b),\n\t\t\tMaxChunkSize,\n\t\t)\n\t}\n\t// write varint first\n\t_, err = w.Write(proto.EncodeVarint(uint64(len(b))))\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tif e.UseSnappyCompression {\n\t\treturn writeSnappyBuffer(w, b)\n\t}\n\treturn w.Write(b)\n}", "func makeBuf(max int) []byte {\n\tif max > BufferSize {\n\t\tmax = BufferSize\n\t}\n\treturn make([]byte, max)\n}", "func (c *BaseConn) SetMaxWriteDelay(delay time.Duration) {\n\tc.stream.SetMaxWriteDelay(delay)\n}", "func MaxDataBytesNoEvidence(maxBytes int64, valsCount int) int64 {\n\tmaxDataBytes := maxBytes -\n\t\tMaxOverheadForBlock -\n\t\tMaxHeaderBytes -\n\t\tMaxCommitBytes(valsCount)\n\n\tif maxDataBytes < 0 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"Negative MaxDataBytesUnknownEvidence. Block.MaxBytes=%d is too small to accommodate header&lastCommit&evidence=%d\",\n\t\t\tmaxBytes,\n\t\t\t-(maxDataBytes - maxBytes),\n\t\t))\n\t}\n\n\treturn maxDataBytes\n}", "func MaxRequestMaxBytes(max int) ConsumerOption {\n\treturn func(o *api.ConsumerConfig) error {\n\t\to.MaxRequestMaxBytes = max\n\t\treturn nil\n\t}\n}", "func (br *BinaryReader) ReadVarBytesWithMaxLimit(max int) []byte {\n\tif max > 0x1000000 {\n\t\tbr.Err = fmt.Errorf(\"max length exceeded\")\n\t\treturn nil\n\t}\n\tn := br.ReadVarUIntWithMaxLimit(uint64(max))\n\tb := make([]byte, n)\n\tbr.ReadLE(b)\n\treturn b\n}", "func MaxDataBytesNoEvidence(maxBytes int64, keyType crypto.KeyType, valsCount int) int64 {\n\tmaxDataBytes := maxBytes -\n\t\tMaxOverheadForBlock -\n\t\tMaxHeaderBytes -\n\t\tMaxCoreChainLockSize -\n\t\tMaxCommitOverheadBytes\n\n\tif maxDataBytes < 0 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"Negative MaxDataBytesUnknownEvidence. Block.MaxBytes=%d is too small to accommodate header&lastCommit&evidence=%d\",\n\t\t\tmaxBytes,\n\t\t\t-(maxDataBytes - maxBytes),\n\t\t))\n\t}\n\n\treturn maxDataBytes\n}", "func maxBodySize(sz int64) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tr.Body = http.MaxBytesReader(w, r.Body, sz)\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func (builder *Builder) MaxSizeInKb(maxSizeInKb uint64) *Builder {\n\tbuilder.maxSizeInKb = maxSizeInKb\n\treturn builder\n}", "func WithMaxSortedBuffer(maxBufferSize int) func(*redisDAL) {\n\treturn func(r *redisDAL) {\n\t\tr.maxSetSize = maxBufferSize\n\t}\n}", "func (o *FifoCreateReqWeb) SetMaxLengthBytes(v int64) {\n\to.MaxLengthBytes = v\n}", "func WithMaxInflightRequests(n int) WriterOption {\n\treturn func(ms *ManagedStream) {\n\t\tms.streamSettings.MaxInflightRequests = n\n\t}\n}", "func (self *IoRate) TakeMax(other *IoRate) {\n\tself.Read.TakeMax(other.Read)\n\tself.Write.TakeMax(other.Write)\n}", "func (e SszNetworkEncoder) GetMaxChunkSize() uint64 {\n\treturn MaxChunkSize\n}", "func (s *settings) GetMaxWriteSize() uint {\n\treturn s.wMaxSize\n}", "func (r *Request) LimitBytes(n int64) *Request {\n\tpanic(\"TODO\")\n\treturn r\n}", "func (c *Config) MaxSize(stream string) (uint, error) {\n\tkey, err := keyName(stream, \"maxsize\")\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn c.v.GetSizeInBytes(key), nil\n}", "func Bytes(b []byte, limit int) []byte {\n\tf := NewWriter(limit)\n\t_, _ = f.Write(b)\n\n\treturn f.buf.Bytes()\n}", "func MaxBandwidth(maxBw opus.Bandwidth) Option {\n\treturn func(args *Options) {\n\t\targs.MaxBandwidth = maxBw\n\t}\n}", "func (us *awsStream) Write(data []byte) (err error) {\n\tif _, err = us.gzw.Write(data); err != nil {\n\t\tlog.Print(\"failed to write to stream\")\n\t\treturn err\n\t}\n\tif us.buf.Len() >= utils.MinDataChunk {\n\t\tpayload := make([]byte, us.buf.Len())\n\t\tif _, err = us.buf.Read(payload); err != nil {\n\t\t\tlog.Print(\"failed to get compressed payload\")\n\t\t\treturn err\n\t\t}\n\t\tif err = us.uploadPart(payload); err != nil {\n\t\t\tlog.Print(\"failed to upload compressed payload\")\n\t\t\treturn err\n\t\t}\n\t\tif len(us.completedParts) >= 819 { // 4Gb ~ 5Mb * 819.2\n\t\t\t//log.Print(\"Maximum file size reached, flushing...\")\n\t\t\t//err = us.Flush()\n\t\t\t//if err != nil {\n\t\t\t//\treturn err\n\t\t\t//}\n\t\t\treturn errs.ErrMaxFileSize\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Cache) SizeMaxBytes() int {\n\tn := 0\n\tfor _, shard := range c.shards {\n\t\tn += shard.SizeMaxBytes()\n\t}\n\treturn n\n}", "func (q *Queue) SetMaxLen(maxLen int) {\n\tq.lock.Lock()\n\tdefer q.lock.Unlock()\n\n\tq.maxLen = maxLen\n}", "func NewBodyByteLimit(limit int64) BodyByteLimit {\n\treturn BodyByteLimit{limit}\n}", "func (msg *MsgTx) MaxPayloadLength(pver uint32) uint32 {\n\treturn 1024 * 10\n}", "func WriteFixedBytes(w io.Writer, data []byte) error {\n\tif len(data) >= math.MaxUint32 {\n\t\treturn ErrorDataSizeOutOfRange\n\t}\n\t_, err := w.Write(data)\n\treturn err\n}", "func MaxPageSize(m int) func(*ParquetWriter) error {\n\treturn func(p *ParquetWriter) error {\n\t\tp.max = m\n\t\treturn nil\n\t}\n}", "func BodyLimitMW(size int64) func(http.Handler) http.Handler {\n\tbodyLimit := size\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tr.Body = http.MaxBytesReader(w, r.Body, bodyLimit)\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func (w *Writer) Write(b []byte) (n int, err error) {\n\tif w.Maxsize > 0 && w.size > w.Maxsize {\n\t\tw.rolling()\n\t}\n\n\tif w.file == nil {\n\t\tif err := w.open(); err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tif w.bw != nil {\n\t\tn, err = w.bw.Write(b)\n\t} else {\n\t\tn, err = w.file.Write(b)\n\t}\n\n\tif err == nil {\n\t\tw.size += int64(n)\n\t}\n\treturn n, err\n}", "func WithMaxMinByteConsumerOption(min, max int) ConsumerOption {\n\treturn func(c *Consumer) {\n\t\tc.config.MinBytes = min\n\t\tc.config.MaxBytes = max\n\t}\n}", "func WithMax(m uint) Option {\n\treturn func(opt *options) {\n\t\topt.max = m\n\t}\n}", "func (m *SlimBlock) MaxPayloadLength(pver uint32) uint32 {\n\treturn MaxBlockPayload\n}", "func (m *MailTips) SetMaxMessageSize(value *int32)() {\n err := m.GetBackingStore().Set(\"maxMessageSize\", value)\n if err != nil {\n panic(err)\n }\n}", "func (cc *ContinueCompress) MaxMessageSize() int {\n\treturn cc.maxMessageSize\n}", "func WithMaxSize(maxMB int) Option {\n\treturn WithDialOpts(grpc.WithDefaultCallOptions(\n\t\tgrpc.MaxCallRecvMsgSize(maxMB*MB),\n\t\tgrpc.MaxCallSendMsgSize(maxMB*MB),\n\t))\n}", "func (cd *ContinueDecompress) MaxMessageSize() int {\n\treturn cd.maxMessageSize\n}", "func (e SszNetworkEncoder) MaxLength(length int) int {\n\tif e.UseSnappyCompression {\n\t\treturn snappy.MaxEncodedLen(length)\n\t}\n\treturn length\n}", "func NewMaxWriter() MaxWriter {\n\treturn MaxWriter{}\n}", "func (m Logon) SetMaxMessageSize(v int) {\n\tm.Set(field.NewMaxMessageSize(v))\n}", "func WithMaxSize(s uint) LogBufferOption {\n\treturn func(o *logBufferOptions) {\n\t\to.maxSize = s\n\t}\n}", "func MaxBlockLen(ct CompressionType) uint64 {\n\tif ct == Snappy {\n\t\t// https://github.com/golang/snappy/blob/2a8bb927dd31d8daada140a5d09578521ce5c36a/encode.go#L76\n\t\treturn 6 * (0xffffffff - 32) / 7\n\t}\n\treturn math.MaxUint64\n}", "func (o *Replication) SetMaxQueueSizeBytes(v int64) {\n\to.MaxQueueSizeBytes = v\n}", "func defaultMaxInflightBytes(n int) option.ClientOption {\n\treturn &defaultInflightBytesSetting{maxBytes: n}\n}", "func (msg *MsgTx) MaxPayloadLength(pver uint32) uint32 {\n\treturn MaxBlockPayload\n}", "func WriteBytes(w io.Writer, lenBuf []byte, m *Message) {\n\trawData := m.Bytes()\n\tsize := len([]byte(rawData))\n\t// println(\"write size:\", size, string(rawData), \".\")\n\tUint32toBytes(lenBuf, uint32(size))\n\tw.Write(lenBuf)\n\tw.Write(rawData)\n}", "func (st *Settings) SetMaxConcurrentStreams(streams uint32) {\n\tst.maxStreams = streams\n}", "func (st *Settings) SetMaxConcurrentStreams(streams uint32) {\n\tst.maxStreams = streams\n}", "func (params *KeyParameters) MaxMsgBytes() int {\n\treturn (params.P.BitLen() / 8) - 4\n}", "func (w *Whisper) SetMaxMessageSize(size uint32) error {\n\tif size > MaxMessageSize {\n\t\treturn fmt.Errorf(\"message size too large [%d>%d]\", size, MaxMessageSize)\n\t}\n\tw.settings.Store(maxMsgSizeIdx, size)\n\treturn nil\n}", "func (sc *simpleLRUStatsCache) SetBytesLimit(BytesLimit int64) {\n\tsc.mu.Lock()\n\tdefer sc.mu.Unlock()\n\tsc.memTracker.SetBytesLimit(BytesLimit)\n\tsc.memCapacity = BytesLimit\n}", "func (b *B) SetBytes(n int64) {}", "func (d *Decoder) SetMaxArraySize(size uint) {\n\td.maxArraySize = int(size)\n}", "func (this *FeedableBuffer) Maximize() {\n\tthis.ExpandTo(this.maxByteCount)\n}", "func (o *FifoCreateReqWeb) GetMaxLengthBytes() int64 {\n\tif o == nil {\n\t\tvar ret int64\n\t\treturn ret\n\t}\n\n\treturn o.MaxLengthBytes\n}", "func (b *B) SetBytes(n int64)", "func (msg *Block) MaxPayloadLength(pver uint32) uint32 {\n\t// Block header at 80 bytes + transaction count + max transactions which can vary up to the MaxBlockPayload\n\t// (including the block header and transaction count).\n\treturn MaxBlockPayload\n}", "func WithMaxInboxSize(size int) ConvergerOpt {\n\treturn func(converger *converger) {\n\t\tconverger.inbox = make(chan convergeRequest, size)\n\t}\n}", "func (mw *MaxWriter) Max() int {\n\treturn mw.max\n}", "func (msg *MsgGetCFilterV2) MaxPayloadLength(pver uint32) uint32 {\n\t// Block hash.\n\treturn chainhash.HashSize\n}", "func WithMaxFileSize(size int64) WarcFileWriterOption {\n\treturn newFuncWarcFileOption(func(o *warcFileWriterOptions) {\n\t\to.maxFileSize = size\n\t})\n}", "func EntryByteLimit(n int) LoggerOption { return entryByteLimit(n) }", "func (s *Static) ResizeBytes(n int) []byte {\n\tif n > s.max {\n\t\tpanic(ErrStaticSize)\n\t}\n\ts.buf = s.buf[:n]\n\treturn s.buf\n}", "func (ch *clientSecureChannel) MaxChunkCount() uint32 {\n\treturn ch.maxChunkCount\n}", "func PoolMaxStreams(n int) client.Option {\n\treturn func(o *client.Options) {\n\t\tif o.Context == nil {\n\t\t\to.Context = context.Background()\n\t\t}\n\t\to.Context = context.WithValue(o.Context, poolMaxStreams{}, n)\n\t}\n}", "func (this SnappyCodec) MaxEncodedLen(srcLen int) int {\n\treturn 32 + srcLen + srcLen/6\n}", "func (ctx *Context) BodyBytes() []byte {\r\n\tif ctx.limitedRequestBody != nil {\r\n\t\treturn ctx.limitedRequestBody\r\n\t}\r\n\tif ctx.R.Body == nil {\r\n\t\tctx.limitedRequestBody = []byte{}\r\n\t\treturn ctx.limitedRequestBody\r\n\t}\r\n\tsafe := &io.LimitedReader{R: ctx.R.Body, N: ctx.frame.config.multipartMaxMemory}\r\n\tlimitedRequestBody, _ := ioutil.ReadAll(safe)\r\n\tctx.R.Body.Close()\r\n\tbf := bytes.NewBuffer(limitedRequestBody)\r\n\tctx.R.Body = ioutil.NopCloser(bf)\r\n\tctx.limitedRequestBody = limitedRequestBody\r\n\treturn limitedRequestBody\r\n}", "func (c Config) MaxHeaderBytesOrDefault() int {\n\tif c.MaxHeaderBytes > 0 {\n\t\treturn c.MaxHeaderBytes\n\t}\n\treturn DefaultMaxHeaderBytes\n}", "func TestV3Curl_MaxStreams_BelowLimit_NoTLS_Small(t *testing.T) {\n\ttestV3CurlMaxStream(t, false, withCfg(*e2e.NewConfigNoTLS()), withMaxConcurrentStreams(3))\n}", "func (st *Settings) SetMaxFrameSize(size uint32) {\n\tst.frameSize = size\n}", "func (st *Settings) SetMaxFrameSize(size uint32) {\n\tst.frameSize = size\n}", "func BufferedWrite(writer *bufio.Writer, data []byte) error {\n\tvar lengthBuf [8]byte\n\tbinary.BigEndian.PutUint64(lengthBuf[:], uint64(len(data)))\n\twriter.Write(lengthBuf[:])\n\twriter.Flush()\n\t_, err := writer.Write(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\twriter.Flush()\n\treturn nil\n}", "func (msg *MsgVersion) MaxPayloadSize(pver uint32) uint32 {\n\treturn 48\n}", "func (msg *MsgFilterClear) MaxPayloadLength(pver uint32) uint32 {\n\treturn 0\n}", "func (opts *FIFOCompactionOptions) SetMaxTableFilesSize(value uint64) {\n\tC.rocksdb_fifo_compaction_options_set_max_table_files_size(opts.c, C.uint64_t(value))\n}", "func (c *LimitedConnection) Write(b []byte) (written int, err error) {\n\treturn c.rateLimitLoop(&c.writeNotBefore, &c.writeDeadline, c.inner.Write, b)\n}", "func (t *http2Client) Write(s *Stream, hdr []byte, data []byte, opts *Options) error {\n\tselect {\n\tcase <-s.ctx.Done():\n\t\treturn ContextErr(s.ctx.Err())\n\tcase <-t.ctx.Done():\n\t\treturn ErrConnClosing\n\tdefault:\n\t}\n\n\tif hdr == nil && data == nil && opts.Last {\n\n\t\tt.controlBuf.put(&dataFrame{streamID: s.id, endStream: true, f: func() {}})\n\t\treturn nil\n\t}\n\n\temptyLen := http2MaxFrameLen - len(hdr)\n\tif emptyLen > len(data) {\n\t\temptyLen = len(data)\n\t}\n\thdr = append(hdr, data[:emptyLen]...)\n\tdata = data[emptyLen:]\n\tvar (\n\t\tstreamQuota int\n\t\tstreamQuotaVer uint32\n\t\terr error\n\t)\n\tfor idx, r := range [][]byte{hdr, data} {\n\t\tfor len(r) > 0 {\n\t\t\tsize := http2MaxFrameLen\n\t\t\tif size > len(r) {\n\t\t\t\tsize = len(r)\n\t\t\t}\n\t\t\tif streamQuota == 0 { // Used up all the locally cached stream quota.\n\n\t\t\t\tstreamQuota, streamQuotaVer, err = s.sendQuotaPool.get(math.MaxInt32, s.waiters)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t\tif size > streamQuota {\n\t\t\t\tsize = streamQuota\n\t\t\t}\n\n\t\t\ttq, _, err := t.sendQuotaPool.get(size, s.waiters)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif tq < size {\n\t\t\t\tsize = tq\n\t\t\t}\n\t\t\tltq, _, err := t.localSendQuota.get(size, s.waiters)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tstreamQuota -= size\n\t\t\tp := r[:size]\n\t\t\tvar endStream bool\n\n\t\t\tif opts.Last {\n\t\t\t\tif len(r)-size == 0 { // No more data in r after this iteration.\n\t\t\t\t\tif idx == 0 { // We're writing data header.\n\t\t\t\t\t\tif len(data) == 0 { // There's no data to follow.\n\t\t\t\t\t\t\tendStream = true\n\t\t\t\t\t\t}\n\t\t\t\t\t} else { // We're writing data.\n\t\t\t\t\t\tendStream = true\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tsuccess := func() {\n\t\t\t\tltq := ltq\n\t\t\t\tt.controlBuf.put(&dataFrame{streamID: s.id, endStream: endStream, d: p, f: func() { t.localSendQuota.add(ltq) }})\n\t\t\t\tr = r[size:]\n\t\t\t}\n\t\t\tfailure := func() { // The stream quota version must have changed.\n\n\t\t\t\ts.sendQuotaPool.lockedAdd(streamQuota + size)\n\t\t\t}\n\t\t\tif !s.sendQuotaPool.compareAndExecute(streamQuotaVer, success, failure) {\n\n\t\t\t\tt.sendQuotaPool.add(size)\n\t\t\t\tt.localSendQuota.add(ltq)\n\t\t\t\tstreamQuota = 0\n\t\t\t}\n\t\t}\n\t}\n\tif streamQuota > 0 { // Add the left over quota back to stream.\n\t\ts.sendQuotaPool.add(streamQuota)\n\t}\n\tif !opts.Last {\n\t\treturn nil\n\t}\n\ts.mu.Lock()\n\tif s.state != streamDone {\n\t\ts.state = streamWriteDone\n\t}\n\ts.mu.Unlock()\n\treturn nil\n}", "func (s *settings) SetMaxReadSize(size uint) {\n\ts.rMaxSize = size\n}", "func (p *Policy) setMaxBlockSize(ic *interop.Context, args []stackitem.Item) stackitem.Item {\n\tvalue := uint32(toBigInt(args[0]).Int64())\n\tif value > payload.MaxSize {\n\t\tpanic(fmt.Errorf(\"MaxBlockSize cannot be more than the maximum payload size = %d\", payload.MaxSize))\n\t}\n\tok, err := p.checkValidators(ic)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif !ok {\n\t\treturn stackitem.NewBool(false)\n\t}\n\tp.lock.Lock()\n\tdefer p.lock.Unlock()\n\terr = p.setUint32WithKey(ic.DAO, maxBlockSizeKey, value)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tp.isValid = false\n\treturn stackitem.NewBool(true)\n}", "func (b *LimitedBuffer) Write(p []byte) (n int, err error) {\n\tb.writeMutex.Lock()\n\tdefer b.writeMutex.Unlock()\n\n\tgotLen := len(p)\n\tif gotLen >= b.limit {\n\t\tb.buf = p[gotLen-b.limit:]\n\t} else if gotLen > 0 {\n\t\tnewLength := len(b.buf) + gotLen\n\t\tif newLength <= b.limit {\n\t\t\tb.buf = append(b.buf, p...)\n\t\t} else {\n\t\t\ttruncateIndex := newLength - b.limit\n\t\t\tb.buf = append(b.buf[truncateIndex:], p...)\n\t\t}\n\t}\n\treturn gotLen, nil\n}", "func MaxMessageSize(size int64) Option {\n\tif size < 0 {\n\t\tpanic(\"size must be non-negative\")\n\t}\n\treturn func(ws *websocket) {\n\t\tws.options.maxMessageSize = size\n\t}\n}", "func (o NetworkPacketCaptureOutput) MaximumBytesPerSession() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *NetworkPacketCapture) pulumi.IntPtrOutput { return v.MaximumBytesPerSession }).(pulumi.IntPtrOutput)\n}", "func MaxBufferSize(size int) Options {\n\treturn func(c *config) {\n\t\tc.maxBufferSize = size\n\t}\n}", "func (transaction *FileCreateTransaction) SetMaxBackoff(max time.Duration) *FileCreateTransaction {\n\tif max.Nanoseconds() < 0 {\n\t\tpanic(\"maxBackoff must be a positive duration\")\n\t} else if max.Nanoseconds() < transaction.minBackoff.Nanoseconds() {\n\t\tpanic(\"maxBackoff must be greater than or equal to minBackoff\")\n\t}\n\ttransaction.maxBackoff = &max\n\treturn transaction\n}", "func WithMaxConnections(maxConnsPerHost int) configurer {\n\treturn func(conf *config) {\n\t\tconf.maxConnsPerHost = maxConnsPerHost\n\t}\n}", "func addMaxRSS(w io.Writer) {\n}", "func (w *Whisper) MaxMessageSize() uint32 {\n\tval, _ := w.settings.Load(maxMsgSizeIdx)\n\treturn val.(uint32)\n}", "func WithMaxFileSZ(sz int) Option {\n\treturn func(opts *options) {\n\t\topts.maxFileSZ = sz\n\t}\n}", "func (qw *QueueWorker) WithMaxWork(maxWork int) *QueueWorker {\n\tqw.maxWork = maxWork\n\treturn qw\n}" ]
[ "0.64424664", "0.63611335", "0.63254035", "0.6157941", "0.6022508", "0.58639663", "0.5862048", "0.585846", "0.57976556", "0.579048", "0.5726838", "0.5713632", "0.5687121", "0.56670904", "0.56304896", "0.5605672", "0.5575398", "0.5557216", "0.5519068", "0.5500462", "0.5431439", "0.5420619", "0.538901", "0.5350175", "0.5344632", "0.5314837", "0.52914935", "0.52700776", "0.52305746", "0.52102983", "0.52051026", "0.5184235", "0.51789343", "0.5144351", "0.50717545", "0.5067734", "0.50633544", "0.5059732", "0.5046853", "0.50373065", "0.50371337", "0.50253934", "0.5018548", "0.4997968", "0.49928507", "0.49878561", "0.49867696", "0.49841926", "0.4983736", "0.49822783", "0.4979329", "0.49782565", "0.4967753", "0.49556953", "0.49544618", "0.49530786", "0.4949693", "0.49423754", "0.49423754", "0.4933306", "0.4923735", "0.491702", "0.49158365", "0.48985595", "0.48775142", "0.4875027", "0.48729274", "0.4867289", "0.48467723", "0.48329866", "0.48244753", "0.48233527", "0.4820533", "0.48134458", "0.48130286", "0.47849628", "0.478262", "0.4773941", "0.47686875", "0.4765148", "0.47637403", "0.47637403", "0.47543243", "0.47516426", "0.47507986", "0.47481954", "0.47460935", "0.4737327", "0.47352773", "0.47318926", "0.47294652", "0.47251725", "0.47237697", "0.47234818", "0.47163647", "0.47119254", "0.47101134", "0.47090647", "0.47089976", "0.47081485" ]
0.73057044
0
Deprecated: Use Orden.ProtoReflect.Descriptor instead.
func (*Orden) Descriptor() ([]byte, []int) { return file_chat_proto_rawDescGZIP(), []int{0} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ProtoFromDescriptor(d protoreflect.Descriptor) proto.Message {\n\tswitch d := d.(type) {\n\tcase protoreflect.FileDescriptor:\n\t\treturn ProtoFromFileDescriptor(d)\n\tcase protoreflect.MessageDescriptor:\n\t\treturn ProtoFromMessageDescriptor(d)\n\tcase protoreflect.FieldDescriptor:\n\t\treturn ProtoFromFieldDescriptor(d)\n\tcase protoreflect.OneofDescriptor:\n\t\treturn ProtoFromOneofDescriptor(d)\n\tcase protoreflect.EnumDescriptor:\n\t\treturn ProtoFromEnumDescriptor(d)\n\tcase protoreflect.EnumValueDescriptor:\n\t\treturn ProtoFromEnumValueDescriptor(d)\n\tcase protoreflect.ServiceDescriptor:\n\t\treturn ProtoFromServiceDescriptor(d)\n\tcase protoreflect.MethodDescriptor:\n\t\treturn ProtoFromMethodDescriptor(d)\n\tdefault:\n\t\t// WTF??\n\t\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\t\treturn res.AsProto()\n\t\t}\n\t\treturn nil\n\t}\n}", "func (*Deprecation) Descriptor() ([]byte, []int) {\n\treturn file_external_cfgmgmt_response_nodes_proto_rawDescGZIP(), []int{8}\n}", "func (*Decl) Descriptor() ([]byte, []int) {\n\treturn file_google_api_expr_v1alpha1_checked_proto_rawDescGZIP(), []int{2}\n}", "func (*AddPeerRequest) Descriptor() ([]byte, []int) {\n\treturn file_github_com_yahuizhan_dappley_metrics_go_api_rpc_pb_rpc_proto_rawDescGZIP(), []int{8}\n}", "func (*TraceProto) Descriptor() ([]byte, []int) {\n\treturn file_internal_tracing_extended_extended_trace_proto_rawDescGZIP(), []int{0}\n}", "func (*PatchCollectorsRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_clarifai_api_service_proto_rawDescGZIP(), []int{161}\n}", "func (*MyProto) Descriptor() ([]byte, []int) {\n\treturn file_my_proto_proto_rawDescGZIP(), []int{0}\n}", "func (*TokenProto) Descriptor() ([]byte, []int) {\n\treturn file_Security_proto_rawDescGZIP(), []int{0}\n}", "func (*NetProtoTalker) Descriptor() ([]byte, []int) {\n\treturn file_pkg_smgrpc_smgrpc_proto_rawDescGZIP(), []int{1}\n}", "func (*AddPeerResponse) Descriptor() ([]byte, []int) {\n\treturn file_github_com_yahuizhan_dappley_metrics_go_api_rpc_pb_rpc_proto_rawDescGZIP(), []int{30}\n}", "func (*ValidatorUpdate) Descriptor() ([]byte, []int) {\n\treturn file_tm_replay_proto_rawDescGZIP(), []int{9}\n}", "func (*StandardProtocols) Descriptor() ([]byte, []int) {\n\treturn file_org_apache_beam_model_pipeline_v1_beam_runner_api_proto_rawDescGZIP(), []int{54}\n}", "func (*Listen) Descriptor() ([]byte, []int) {\n\treturn file_pkg_smgrpc_smgrpc_proto_rawDescGZIP(), []int{4}\n}", "func ProtoFromFileDescriptor(d protoreflect.FileDescriptor) *descriptorpb.FileDescriptorProto {\n\tif imp, ok := d.(protoreflect.FileImport); ok {\n\t\td = imp.FileDescriptor\n\t}\n\ttype canProto interface {\n\t\tFileDescriptorProto() *descriptorpb.FileDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.FileDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif fd, ok := res.AsProto().(*descriptorpb.FileDescriptorProto); ok {\n\t\t\treturn fd\n\t\t}\n\t}\n\treturn protodesc.ToFileDescriptorProto(d)\n}", "func (*Type) Descriptor() ([]byte, []int) {\n\treturn file_google_api_expr_v1alpha1_checked_proto_rawDescGZIP(), []int{1}\n}", "func (StandardPTransforms_DeprecatedPrimitives) EnumDescriptor() ([]byte, []int) {\n\treturn file_org_apache_beam_model_pipeline_v1_beam_runner_api_proto_rawDescGZIP(), []int{4, 1}\n}", "func (x *fastReflection_MsgUpdateParams) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_MsgUpdateParams\n}", "func (x *fastReflection_FlagOptions) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_FlagOptions\n}", "func (*GetPeerInfoRequest) Descriptor() ([]byte, []int) {\n\treturn file_github_com_yahuizhan_dappley_metrics_go_api_rpc_pb_rpc_proto_rawDescGZIP(), []int{6}\n}", "func (*Module) Descriptor() ([]byte, []int) {\n\treturn file_google_devtools_cloudtrace_v2_trace_proto_rawDescGZIP(), []int{3}\n}", "func (*Instance) Descriptor() ([]byte, []int) {\n\treturn file_cloudprovider_externalgrpc_protos_externalgrpc_proto_rawDescGZIP(), []int{28}\n}", "func (*Reference) Descriptor() ([]byte, []int) {\n\treturn file_google_api_expr_v1alpha1_checked_proto_rawDescGZIP(), []int{3}\n}", "func (*TelemetryParams) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{62}\n}", "func (x *fastReflection_Metadata) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Metadata\n}", "func ProtoFromMethodDescriptor(d protoreflect.MethodDescriptor) *descriptorpb.MethodDescriptorProto {\n\ttype canProto interface {\n\t\tMethodDescriptorProto() *descriptorpb.MethodDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.MethodDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif md, ok := res.AsProto().(*descriptorpb.MethodDescriptorProto); ok {\n\t\t\treturn md\n\t\t}\n\t}\n\treturn protodesc.ToMethodDescriptorProto(d)\n}", "func (x *fastReflection_Params) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Params\n}", "func (*GetDelegationTokenResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_Security_proto_rawDescGZIP(), []int{4}\n}", "func (*DirectiveUndelegate) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{10}\n}", "func ProtoFromFieldDescriptor(d protoreflect.FieldDescriptor) *descriptorpb.FieldDescriptorProto {\n\ttype canProto interface {\n\t\tFieldDescriptorProto() *descriptorpb.FieldDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.FieldDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif fd, ok := res.AsProto().(*descriptorpb.FieldDescriptorProto); ok {\n\t\t\treturn fd\n\t\t}\n\t}\n\treturn protodesc.ToFieldDescriptorProto(d)\n}", "func (*Description) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{4}\n}", "func (*Embed) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{2}\n}", "func (*PlanChange_Removed) Descriptor() ([]byte, []int) {\n\treturn edgelq_limits_proto_v1alpha2_plan_change_proto_rawDescGZIP(), []int{0, 3}\n}", "func (*Example) Descriptor() ([]byte, []int) {\n\treturn file_google_ai_generativelanguage_v1beta2_discuss_service_proto_rawDescGZIP(), []int{4}\n}", "func (*Modifier) Descriptor() ([]byte, []int) {\n\treturn file_FillerGame_proto_rawDescGZIP(), []int{6}\n}", "func (*Message6024) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{26}\n}", "func (*Undefined) Descriptor() ([]byte, []int) {\n\treturn file_rpc_rpc_proto_rawDescGZIP(), []int{3}\n}", "func (*Name) Descriptor() ([]byte, []int) {\n\treturn file_examples_documents_example_proto_rawDescGZIP(), []int{25}\n}", "func (*CodeLens) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{164}\n}", "func (*SemanticTokensDelta) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{223}\n}", "func (*GetDatanodeInfoResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_ClientDatanodeProtocol_proto_rawDescGZIP(), []int{13}\n}", "func (*CancelPlanResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_ClientDatanodeProtocol_proto_rawDescGZIP(), []int{23}\n}", "func (x *fastReflection_ModuleOptions) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_ModuleOptions\n}", "func (*KnownTypes) Descriptor() ([]byte, []int) {\n\treturn file_jsonpb_proto_test2_proto_rawDescGZIP(), []int{8}\n}", "func (*Embed_EmbedField) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{2, 1}\n}", "func (*QueryPlanStatusResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_ClientDatanodeProtocol_proto_rawDescGZIP(), []int{25}\n}", "func (x *fastReflection_Evidence) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Evidence\n}", "func ToDescriptorProto(message protoreflect.MessageDescriptor) *descriptorpb.DescriptorProto {\n\tp := &descriptorpb.DescriptorProto{\n\t\tName: proto.String(string(message.Name())),\n\t\tOptions: proto.Clone(message.Options()).(*descriptorpb.MessageOptions),\n\t}\n\tfor i, fields := 0, message.Fields(); i < fields.Len(); i++ {\n\t\tp.Field = append(p.Field, ToFieldDescriptorProto(fields.Get(i)))\n\t}\n\tfor i, exts := 0, message.Extensions(); i < exts.Len(); i++ {\n\t\tp.Extension = append(p.Extension, ToFieldDescriptorProto(exts.Get(i)))\n\t}\n\tfor i, messages := 0, message.Messages(); i < messages.Len(); i++ {\n\t\tp.NestedType = append(p.NestedType, ToDescriptorProto(messages.Get(i)))\n\t}\n\tfor i, enums := 0, message.Enums(); i < enums.Len(); i++ {\n\t\tp.EnumType = append(p.EnumType, ToEnumDescriptorProto(enums.Get(i)))\n\t}\n\tfor i, xranges := 0, message.ExtensionRanges(); i < xranges.Len(); i++ {\n\t\txrange := xranges.Get(i)\n\t\tp.ExtensionRange = append(p.ExtensionRange, &descriptorpb.DescriptorProto_ExtensionRange{\n\t\t\tStart: proto.Int32(int32(xrange[0])),\n\t\t\tEnd: proto.Int32(int32(xrange[1])),\n\t\t\tOptions: proto.Clone(message.ExtensionRangeOptions(i)).(*descriptorpb.ExtensionRangeOptions),\n\t\t})\n\t}\n\tfor i, oneofs := 0, message.Oneofs(); i < oneofs.Len(); i++ {\n\t\tp.OneofDecl = append(p.OneofDecl, ToOneofDescriptorProto(oneofs.Get(i)))\n\t}\n\tfor i, ranges := 0, message.ReservedRanges(); i < ranges.Len(); i++ {\n\t\trrange := ranges.Get(i)\n\t\tp.ReservedRange = append(p.ReservedRange, &descriptorpb.DescriptorProto_ReservedRange{\n\t\t\tStart: proto.Int32(int32(rrange[0])),\n\t\t\tEnd: proto.Int32(int32(rrange[1])),\n\t\t})\n\t}\n\tfor i, names := 0, message.ReservedNames(); i < names.Len(); i++ {\n\t\tp.ReservedName = append(p.ReservedName, string(names.Get(i)))\n\t}\n\treturn p\n}", "func (*DirectiveCreateValidator) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{7}\n}", "func (*DescribeRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_engine_proto_rawDescGZIP(), []int{4}\n}", "func (*APILevel) Descriptor() ([]byte, []int) {\n\treturn file_Notify_proto_rawDescGZIP(), []int{4}\n}", "func (*GetDelegationTokenRequestProto) Descriptor() ([]byte, []int) {\n\treturn file_Security_proto_rawDescGZIP(), []int{3}\n}", "func (*LabelDescriptor) Descriptor() ([]byte, []int) {\n\treturn edgelq_logging_proto_v1alpha2_common_proto_rawDescGZIP(), []int{0}\n}", "func (*Code) Descriptor() ([]byte, []int) {\n\treturn file_internal_pkg_pb_ports_proto_rawDescGZIP(), []int{2}\n}", "func (*ApiListener) Descriptor() ([]byte, []int) {\n\treturn file_envoy_config_listener_v2_api_listener_proto_rawDescGZIP(), []int{0}\n}", "func (*ModifyRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_engine_proto_rawDescGZIP(), []int{10}\n}", "func (*MetadataUpdateEventProto) Descriptor() ([]byte, []int) {\n\treturn file_inotify_proto_rawDescGZIP(), []int{7}\n}", "func (*ValidatorUpdates) Descriptor() ([]byte, []int) {\n\treturn file_core_abci_v1alpha1_abci_proto_rawDescGZIP(), []int{6}\n}", "func (*FormatMessage) Descriptor() ([]byte, []int) {\n\treturn file_google_devtools_clouddebugger_v2_data_proto_rawDescGZIP(), []int{0}\n}", "func (*Person) Descriptor() ([]byte, []int) {\n\treturn file_protomessage_proto_rawDescGZIP(), []int{0}\n}", "func (*WatchRequestTypeProto) Descriptor() ([]byte, []int) {\n\treturn file_raft_proto_rawDescGZIP(), []int{25}\n}", "func (*RelationTupleDelta) Descriptor() ([]byte, []int) {\n\treturn file_ory_keto_acl_v1alpha1_write_service_proto_rawDescGZIP(), []int{1}\n}", "func (*Validator) Descriptor() ([]byte, []int) {\n\treturn file_tm_replay_proto_rawDescGZIP(), []int{13}\n}", "func (*AnalysisMessageWeakSchema_ArgType) Descriptor() ([]byte, []int) {\n\treturn file_analysis_v1alpha1_message_proto_rawDescGZIP(), []int{1, 0}\n}", "func (*PlanChange_Modified) Descriptor() ([]byte, []int) {\n\treturn edgelq_limits_proto_v1alpha2_plan_change_proto_rawDescGZIP(), []int{0, 1}\n}", "func (StandardProtocols_Enum) EnumDescriptor() ([]byte, []int) {\n\treturn file_org_apache_beam_model_pipeline_v1_beam_runner_api_proto_rawDescGZIP(), []int{54, 0}\n}", "func (x *fastReflection_RpcCommandOptions) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_RpcCommandOptions\n}", "func (x *fastReflection_ServiceCommandDescriptor) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_ServiceCommandDescriptor\n}", "func (*PlanChange_Added) Descriptor() ([]byte, []int) {\n\treturn edgelq_limits_proto_v1alpha2_plan_change_proto_rawDescGZIP(), []int{0, 0}\n}", "func (*MetadataProto) Descriptor() ([]byte, []int) {\n\treturn file_raft_proto_rawDescGZIP(), []int{7}\n}", "func (*Primitive) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{15}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_google_ai_generativelanguage_v1beta2_discuss_service_proto_rawDescGZIP(), []int{2}\n}", "func (*Message12796) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{1}\n}", "func (*AnalysisMessageWeakSchema) Descriptor() ([]byte, []int) {\n\treturn file_analysis_v1alpha1_message_proto_rawDescGZIP(), []int{1}\n}", "func (x *fastReflection_LightClientAttackEvidence) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_LightClientAttackEvidence\n}", "func (*FeedbackMetrics) Descriptor() ([]byte, []int) {\n\treturn file_ssn_dataservice_v1_dataservice_proto_rawDescGZIP(), []int{12}\n}", "func (*DirectiveDelegate) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{9}\n}", "func (*DirectiveEditValidator) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{8}\n}", "func (*DeleteTeam) Descriptor() ([]byte, []int) {\n\treturn file_uac_Team_proto_rawDescGZIP(), []int{6}\n}", "func (*CreateAlterRequest) Descriptor() ([]byte, []int) {\n\treturn file_grpc_exercicio_proto_rawDescGZIP(), []int{1}\n}", "func (SocketAddress_Protocol) EnumDescriptor() ([]byte, []int) {\n\treturn file_envoy_config_core_v3_address_proto_rawDescGZIP(), []int{1, 0}\n}", "func (*Trickle) Descriptor() ([]byte, []int) {\n\treturn file_cmd_server_grpc_proto_sfu_proto_rawDescGZIP(), []int{4}\n}", "func (*Message6127) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{24}\n}", "func (*Message12818) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{5}\n}", "func (*Metadata) Descriptor() ([]byte, []int) {\n\treturn file_authzed_api_v0_namespace_proto_rawDescGZIP(), []int{0}\n}", "func (*Real) Descriptor() ([]byte, []int) {\n\treturn file_jsonpb_proto_test2_proto_rawDescGZIP(), []int{6}\n}", "func ToFileDescriptorProto(file protoreflect.FileDescriptor) *descriptorpb.FileDescriptorProto {\n\tp := &descriptorpb.FileDescriptorProto{\n\t\tName: proto.String(file.Path()),\n\t\tOptions: proto.Clone(file.Options()).(*descriptorpb.FileOptions),\n\t}\n\tif file.Package() != \"\" {\n\t\tp.Package = proto.String(string(file.Package()))\n\t}\n\tfor i, imports := 0, file.Imports(); i < imports.Len(); i++ {\n\t\timp := imports.Get(i)\n\t\tp.Dependency = append(p.Dependency, imp.Path())\n\t\tif imp.IsPublic {\n\t\t\tp.PublicDependency = append(p.PublicDependency, int32(i))\n\t\t}\n\t\tif imp.IsWeak {\n\t\t\tp.WeakDependency = append(p.WeakDependency, int32(i))\n\t\t}\n\t}\n\tfor i, locs := 0, file.SourceLocations(); i < locs.Len(); i++ {\n\t\tloc := locs.Get(i)\n\t\tl := &descriptorpb.SourceCodeInfo_Location{}\n\t\tl.Path = append(l.Path, loc.Path...)\n\t\tif loc.StartLine == loc.EndLine {\n\t\t\tl.Span = []int32{int32(loc.StartLine), int32(loc.StartColumn), int32(loc.EndColumn)}\n\t\t} else {\n\t\t\tl.Span = []int32{int32(loc.StartLine), int32(loc.StartColumn), int32(loc.EndLine), int32(loc.EndColumn)}\n\t\t}\n\t\tl.LeadingDetachedComments = append([]string(nil), loc.LeadingDetachedComments...)\n\t\tif loc.LeadingComments != \"\" {\n\t\t\tl.LeadingComments = proto.String(loc.LeadingComments)\n\t\t}\n\t\tif loc.TrailingComments != \"\" {\n\t\t\tl.TrailingComments = proto.String(loc.TrailingComments)\n\t\t}\n\t\tif p.SourceCodeInfo == nil {\n\t\t\tp.SourceCodeInfo = &descriptorpb.SourceCodeInfo{}\n\t\t}\n\t\tp.SourceCodeInfo.Location = append(p.SourceCodeInfo.Location, l)\n\n\t}\n\tfor i, messages := 0, file.Messages(); i < messages.Len(); i++ {\n\t\tp.MessageType = append(p.MessageType, ToDescriptorProto(messages.Get(i)))\n\t}\n\tfor i, enums := 0, file.Enums(); i < enums.Len(); i++ {\n\t\tp.EnumType = append(p.EnumType, ToEnumDescriptorProto(enums.Get(i)))\n\t}\n\tfor i, services := 0, file.Services(); i < services.Len(); i++ {\n\t\tp.Service = append(p.Service, ToServiceDescriptorProto(services.Get(i)))\n\t}\n\tfor i, exts := 0, file.Extensions(); i < exts.Len(); i++ {\n\t\tp.Extension = append(p.Extension, ToFieldDescriptorProto(exts.Get(i)))\n\t}\n\tif syntax := file.Syntax(); syntax != protoreflect.Proto2 {\n\t\tp.Syntax = proto.String(file.Syntax().String())\n\t}\n\treturn p\n}", "func (*Person) Descriptor() ([]byte, []int) {\n\treturn file_api_proto_rawDescGZIP(), []int{8}\n}", "func (*GetTeamByName) Descriptor() ([]byte, []int) {\n\treturn file_uac_Team_proto_rawDescGZIP(), []int{2}\n}", "func (*ExternalGrpcNode) Descriptor() ([]byte, []int) {\n\treturn file_cloudprovider_externalgrpc_protos_externalgrpc_proto_rawDescGZIP(), []int{1}\n}", "func (*Message5903) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{34}\n}", "func (StatusMessage_Reference) EnumDescriptor() ([]byte, []int) {\n\treturn file_google_devtools_clouddebugger_v2_data_proto_rawDescGZIP(), []int{1, 0}\n}", "func (*Instant) Descriptor() ([]byte, []int) {\n\treturn file_proto_google_fhir_proto_stu3_datatypes_proto_rawDescGZIP(), []int{7}\n}", "func (*SafetyFeedback) Descriptor() ([]byte, []int) {\n\treturn file_google_ai_generativelanguage_v1beta2_safety_proto_rawDescGZIP(), []int{1}\n}", "func (*Message7921) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{19}\n}", "func (*Version) Descriptor() ([]byte, []int) {\n\treturn file_google_protobuf_compiler_plugin_proto_rawDescGZIP(), []int{0}\n}", "func (*CancelDelegationTokenResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_Security_proto_rawDescGZIP(), []int{8}\n}", "func (*Name) Descriptor() ([]byte, []int) {\n\treturn file_api_proto_rawDescGZIP(), []int{2}\n}", "func (x *fastReflection_MsgUpdateParamsResponse) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_MsgUpdateParamsResponse\n}", "func (*EvictWritersResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_ClientDatanodeProtocol_proto_rawDescGZIP(), []int{11}\n}", "func (*PlanChange) Descriptor() ([]byte, []int) {\n\treturn edgelq_limits_proto_v1alpha2_plan_change_proto_rawDescGZIP(), []int{0}\n}", "func (*Message12817) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{22}\n}" ]
[ "0.7074704", "0.705885", "0.7006217", "0.69509065", "0.69417596", "0.68928474", "0.6831902", "0.68113196", "0.68061244", "0.6795427", "0.6787744", "0.6785139", "0.67669904", "0.6762595", "0.6745581", "0.6745484", "0.6744348", "0.674337", "0.6737492", "0.6733155", "0.6728783", "0.6723494", "0.67177033", "0.6709404", "0.670806", "0.67031544", "0.6693836", "0.6693549", "0.66795", "0.6677782", "0.6672672", "0.6660984", "0.66490436", "0.6646653", "0.6646021", "0.6642092", "0.66419566", "0.6639115", "0.6637826", "0.6633814", "0.6624338", "0.6624038", "0.6622644", "0.6622484", "0.6619947", "0.661852", "0.6617372", "0.66161114", "0.66138995", "0.6611475", "0.66047436", "0.6602263", "0.6599819", "0.6599465", "0.6595738", "0.6592787", "0.6591075", "0.65868104", "0.65856874", "0.6584318", "0.65837055", "0.6581771", "0.6579176", "0.6577613", "0.65753216", "0.6573828", "0.6570509", "0.6569227", "0.6569119", "0.6566453", "0.65620345", "0.6561265", "0.65608454", "0.6558176", "0.65561795", "0.65540797", "0.6552998", "0.65488863", "0.6548523", "0.65465945", "0.6543094", "0.6539853", "0.6539424", "0.6537544", "0.6535517", "0.65350443", "0.653422", "0.6534063", "0.6533881", "0.6530695", "0.65293914", "0.6528861", "0.65274316", "0.6523805", "0.6523055", "0.6521592", "0.6521534", "0.65200984", "0.6518772", "0.6517754", "0.65172565" ]
0.0
-1
Deprecated: Use Paquete.ProtoReflect.Descriptor instead.
func (*Paquete) Descriptor() ([]byte, []int) { return file_chat_proto_rawDescGZIP(), []int{1} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ProtoFromDescriptor(d protoreflect.Descriptor) proto.Message {\n\tswitch d := d.(type) {\n\tcase protoreflect.FileDescriptor:\n\t\treturn ProtoFromFileDescriptor(d)\n\tcase protoreflect.MessageDescriptor:\n\t\treturn ProtoFromMessageDescriptor(d)\n\tcase protoreflect.FieldDescriptor:\n\t\treturn ProtoFromFieldDescriptor(d)\n\tcase protoreflect.OneofDescriptor:\n\t\treturn ProtoFromOneofDescriptor(d)\n\tcase protoreflect.EnumDescriptor:\n\t\treturn ProtoFromEnumDescriptor(d)\n\tcase protoreflect.EnumValueDescriptor:\n\t\treturn ProtoFromEnumValueDescriptor(d)\n\tcase protoreflect.ServiceDescriptor:\n\t\treturn ProtoFromServiceDescriptor(d)\n\tcase protoreflect.MethodDescriptor:\n\t\treturn ProtoFromMethodDescriptor(d)\n\tdefault:\n\t\t// WTF??\n\t\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\t\treturn res.AsProto()\n\t\t}\n\t\treturn nil\n\t}\n}", "func (*Deprecation) Descriptor() ([]byte, []int) {\n\treturn file_external_cfgmgmt_response_nodes_proto_rawDescGZIP(), []int{8}\n}", "func (*AddPeerRequest) Descriptor() ([]byte, []int) {\n\treturn file_github_com_yahuizhan_dappley_metrics_go_api_rpc_pb_rpc_proto_rawDescGZIP(), []int{8}\n}", "func ProtoFromFileDescriptor(d protoreflect.FileDescriptor) *descriptorpb.FileDescriptorProto {\n\tif imp, ok := d.(protoreflect.FileImport); ok {\n\t\td = imp.FileDescriptor\n\t}\n\ttype canProto interface {\n\t\tFileDescriptorProto() *descriptorpb.FileDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.FileDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif fd, ok := res.AsProto().(*descriptorpb.FileDescriptorProto); ok {\n\t\t\treturn fd\n\t\t}\n\t}\n\treturn protodesc.ToFileDescriptorProto(d)\n}", "func (*Decl) Descriptor() ([]byte, []int) {\n\treturn file_google_api_expr_v1alpha1_checked_proto_rawDescGZIP(), []int{2}\n}", "func (*TraceProto) Descriptor() ([]byte, []int) {\n\treturn file_internal_tracing_extended_extended_trace_proto_rawDescGZIP(), []int{0}\n}", "func ProtoFromMethodDescriptor(d protoreflect.MethodDescriptor) *descriptorpb.MethodDescriptorProto {\n\ttype canProto interface {\n\t\tMethodDescriptorProto() *descriptorpb.MethodDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.MethodDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif md, ok := res.AsProto().(*descriptorpb.MethodDescriptorProto); ok {\n\t\t\treturn md\n\t\t}\n\t}\n\treturn protodesc.ToMethodDescriptorProto(d)\n}", "func (*MyProto) Descriptor() ([]byte, []int) {\n\treturn file_my_proto_proto_rawDescGZIP(), []int{0}\n}", "func (*PatchCollectorsRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_clarifai_api_service_proto_rawDescGZIP(), []int{161}\n}", "func (x *fastReflection_Params) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Params\n}", "func (*NetProtoTalker) Descriptor() ([]byte, []int) {\n\treturn file_pkg_smgrpc_smgrpc_proto_rawDescGZIP(), []int{1}\n}", "func (x *fastReflection_FlagOptions) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_FlagOptions\n}", "func (*StandardProtocols) Descriptor() ([]byte, []int) {\n\treturn file_org_apache_beam_model_pipeline_v1_beam_runner_api_proto_rawDescGZIP(), []int{54}\n}", "func (x *fastReflection_MsgUpdateParams) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_MsgUpdateParams\n}", "func (*Listen) Descriptor() ([]byte, []int) {\n\treturn file_pkg_smgrpc_smgrpc_proto_rawDescGZIP(), []int{4}\n}", "func (x *fastReflection_Metadata) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Metadata\n}", "func (*Example) Descriptor() ([]byte, []int) {\n\treturn file_google_ai_generativelanguage_v1beta2_discuss_service_proto_rawDescGZIP(), []int{4}\n}", "func (*AddPeerResponse) Descriptor() ([]byte, []int) {\n\treturn file_github_com_yahuizhan_dappley_metrics_go_api_rpc_pb_rpc_proto_rawDescGZIP(), []int{30}\n}", "func (*Instance) Descriptor() ([]byte, []int) {\n\treturn file_cloudprovider_externalgrpc_protos_externalgrpc_proto_rawDescGZIP(), []int{28}\n}", "func ProtoFromFieldDescriptor(d protoreflect.FieldDescriptor) *descriptorpb.FieldDescriptorProto {\n\ttype canProto interface {\n\t\tFieldDescriptorProto() *descriptorpb.FieldDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.FieldDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif fd, ok := res.AsProto().(*descriptorpb.FieldDescriptorProto); ok {\n\t\t\treturn fd\n\t\t}\n\t}\n\treturn protodesc.ToFieldDescriptorProto(d)\n}", "func (x *fastReflection_Evidence) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Evidence\n}", "func (StandardPTransforms_DeprecatedPrimitives) EnumDescriptor() ([]byte, []int) {\n\treturn file_org_apache_beam_model_pipeline_v1_beam_runner_api_proto_rawDescGZIP(), []int{4, 1}\n}", "func (*TokenProto) Descriptor() ([]byte, []int) {\n\treturn file_Security_proto_rawDescGZIP(), []int{0}\n}", "func (*ValidatorUpdate) Descriptor() ([]byte, []int) {\n\treturn file_tm_replay_proto_rawDescGZIP(), []int{9}\n}", "func (*GetPeerInfoRequest) Descriptor() ([]byte, []int) {\n\treturn file_github_com_yahuizhan_dappley_metrics_go_api_rpc_pb_rpc_proto_rawDescGZIP(), []int{6}\n}", "func (*Type) Descriptor() ([]byte, []int) {\n\treturn file_google_api_expr_v1alpha1_checked_proto_rawDescGZIP(), []int{1}\n}", "func (*Embed) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{2}\n}", "func (*PlanChange_Removed) Descriptor() ([]byte, []int) {\n\treturn edgelq_limits_proto_v1alpha2_plan_change_proto_rawDescGZIP(), []int{0, 3}\n}", "func ToDescriptorProto(message protoreflect.MessageDescriptor) *descriptorpb.DescriptorProto {\n\tp := &descriptorpb.DescriptorProto{\n\t\tName: proto.String(string(message.Name())),\n\t\tOptions: proto.Clone(message.Options()).(*descriptorpb.MessageOptions),\n\t}\n\tfor i, fields := 0, message.Fields(); i < fields.Len(); i++ {\n\t\tp.Field = append(p.Field, ToFieldDescriptorProto(fields.Get(i)))\n\t}\n\tfor i, exts := 0, message.Extensions(); i < exts.Len(); i++ {\n\t\tp.Extension = append(p.Extension, ToFieldDescriptorProto(exts.Get(i)))\n\t}\n\tfor i, messages := 0, message.Messages(); i < messages.Len(); i++ {\n\t\tp.NestedType = append(p.NestedType, ToDescriptorProto(messages.Get(i)))\n\t}\n\tfor i, enums := 0, message.Enums(); i < enums.Len(); i++ {\n\t\tp.EnumType = append(p.EnumType, ToEnumDescriptorProto(enums.Get(i)))\n\t}\n\tfor i, xranges := 0, message.ExtensionRanges(); i < xranges.Len(); i++ {\n\t\txrange := xranges.Get(i)\n\t\tp.ExtensionRange = append(p.ExtensionRange, &descriptorpb.DescriptorProto_ExtensionRange{\n\t\t\tStart: proto.Int32(int32(xrange[0])),\n\t\t\tEnd: proto.Int32(int32(xrange[1])),\n\t\t\tOptions: proto.Clone(message.ExtensionRangeOptions(i)).(*descriptorpb.ExtensionRangeOptions),\n\t\t})\n\t}\n\tfor i, oneofs := 0, message.Oneofs(); i < oneofs.Len(); i++ {\n\t\tp.OneofDecl = append(p.OneofDecl, ToOneofDescriptorProto(oneofs.Get(i)))\n\t}\n\tfor i, ranges := 0, message.ReservedRanges(); i < ranges.Len(); i++ {\n\t\trrange := ranges.Get(i)\n\t\tp.ReservedRange = append(p.ReservedRange, &descriptorpb.DescriptorProto_ReservedRange{\n\t\t\tStart: proto.Int32(int32(rrange[0])),\n\t\t\tEnd: proto.Int32(int32(rrange[1])),\n\t\t})\n\t}\n\tfor i, names := 0, message.ReservedNames(); i < names.Len(); i++ {\n\t\tp.ReservedName = append(p.ReservedName, string(names.Get(i)))\n\t}\n\treturn p\n}", "func (*Message6024) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{26}\n}", "func (*CancelPlanResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_ClientDatanodeProtocol_proto_rawDescGZIP(), []int{23}\n}", "func (*Modifier) Descriptor() ([]byte, []int) {\n\treturn file_FillerGame_proto_rawDescGZIP(), []int{6}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_google_ai_generativelanguage_v1beta2_discuss_service_proto_rawDescGZIP(), []int{2}\n}", "func (*Reference) Descriptor() ([]byte, []int) {\n\treturn file_google_api_expr_v1alpha1_checked_proto_rawDescGZIP(), []int{3}\n}", "func ToFileDescriptorProto(file protoreflect.FileDescriptor) *descriptorpb.FileDescriptorProto {\n\tp := &descriptorpb.FileDescriptorProto{\n\t\tName: proto.String(file.Path()),\n\t\tOptions: proto.Clone(file.Options()).(*descriptorpb.FileOptions),\n\t}\n\tif file.Package() != \"\" {\n\t\tp.Package = proto.String(string(file.Package()))\n\t}\n\tfor i, imports := 0, file.Imports(); i < imports.Len(); i++ {\n\t\timp := imports.Get(i)\n\t\tp.Dependency = append(p.Dependency, imp.Path())\n\t\tif imp.IsPublic {\n\t\t\tp.PublicDependency = append(p.PublicDependency, int32(i))\n\t\t}\n\t\tif imp.IsWeak {\n\t\t\tp.WeakDependency = append(p.WeakDependency, int32(i))\n\t\t}\n\t}\n\tfor i, locs := 0, file.SourceLocations(); i < locs.Len(); i++ {\n\t\tloc := locs.Get(i)\n\t\tl := &descriptorpb.SourceCodeInfo_Location{}\n\t\tl.Path = append(l.Path, loc.Path...)\n\t\tif loc.StartLine == loc.EndLine {\n\t\t\tl.Span = []int32{int32(loc.StartLine), int32(loc.StartColumn), int32(loc.EndColumn)}\n\t\t} else {\n\t\t\tl.Span = []int32{int32(loc.StartLine), int32(loc.StartColumn), int32(loc.EndLine), int32(loc.EndColumn)}\n\t\t}\n\t\tl.LeadingDetachedComments = append([]string(nil), loc.LeadingDetachedComments...)\n\t\tif loc.LeadingComments != \"\" {\n\t\t\tl.LeadingComments = proto.String(loc.LeadingComments)\n\t\t}\n\t\tif loc.TrailingComments != \"\" {\n\t\t\tl.TrailingComments = proto.String(loc.TrailingComments)\n\t\t}\n\t\tif p.SourceCodeInfo == nil {\n\t\t\tp.SourceCodeInfo = &descriptorpb.SourceCodeInfo{}\n\t\t}\n\t\tp.SourceCodeInfo.Location = append(p.SourceCodeInfo.Location, l)\n\n\t}\n\tfor i, messages := 0, file.Messages(); i < messages.Len(); i++ {\n\t\tp.MessageType = append(p.MessageType, ToDescriptorProto(messages.Get(i)))\n\t}\n\tfor i, enums := 0, file.Enums(); i < enums.Len(); i++ {\n\t\tp.EnumType = append(p.EnumType, ToEnumDescriptorProto(enums.Get(i)))\n\t}\n\tfor i, services := 0, file.Services(); i < services.Len(); i++ {\n\t\tp.Service = append(p.Service, ToServiceDescriptorProto(services.Get(i)))\n\t}\n\tfor i, exts := 0, file.Extensions(); i < exts.Len(); i++ {\n\t\tp.Extension = append(p.Extension, ToFieldDescriptorProto(exts.Get(i)))\n\t}\n\tif syntax := file.Syntax(); syntax != protoreflect.Proto2 {\n\t\tp.Syntax = proto.String(file.Syntax().String())\n\t}\n\treturn p\n}", "func (*Description) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{4}\n}", "func (*Embed_EmbedField) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{2, 1}\n}", "func (*Trickle) Descriptor() ([]byte, []int) {\n\treturn file_cmd_server_grpc_proto_sfu_proto_rawDescGZIP(), []int{4}\n}", "func (*CreateAlterRequest) Descriptor() ([]byte, []int) {\n\treturn file_grpc_exercicio_proto_rawDescGZIP(), []int{1}\n}", "func (x *fastReflection_ModuleOptions) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_ModuleOptions\n}", "func (*GetDatanodeInfoResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_ClientDatanodeProtocol_proto_rawDescGZIP(), []int{13}\n}", "func (*QueryPlanStatusResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_ClientDatanodeProtocol_proto_rawDescGZIP(), []int{25}\n}", "func (*AnalysisMessageWeakSchema_ArgType) Descriptor() ([]byte, []int) {\n\treturn file_analysis_v1alpha1_message_proto_rawDescGZIP(), []int{1, 0}\n}", "func (*DirectiveUndelegate) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{10}\n}", "func (*Name) Descriptor() ([]byte, []int) {\n\treturn file_examples_documents_example_proto_rawDescGZIP(), []int{25}\n}", "func (*Module) Descriptor() ([]byte, []int) {\n\treturn file_google_devtools_cloudtrace_v2_trace_proto_rawDescGZIP(), []int{3}\n}", "func (*Code) Descriptor() ([]byte, []int) {\n\treturn file_internal_pkg_pb_ports_proto_rawDescGZIP(), []int{2}\n}", "func (x *fastReflection_LightClientAttackEvidence) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_LightClientAttackEvidence\n}", "func (*SafetyFeedback) Descriptor() ([]byte, []int) {\n\treturn file_google_ai_generativelanguage_v1beta2_safety_proto_rawDescGZIP(), []int{1}\n}", "func (*GetDelegationTokenResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_Security_proto_rawDescGZIP(), []int{4}\n}", "func (*PlanChange_Added) Descriptor() ([]byte, []int) {\n\treturn edgelq_limits_proto_v1alpha2_plan_change_proto_rawDescGZIP(), []int{0, 0}\n}", "func (*AnalysisMessageWeakSchema) Descriptor() ([]byte, []int) {\n\treturn file_analysis_v1alpha1_message_proto_rawDescGZIP(), []int{1}\n}", "func ProtoFromMessageDescriptor(d protoreflect.MessageDescriptor) *descriptorpb.DescriptorProto {\n\ttype canProto interface {\n\t\tMessageDescriptorProto() *descriptorpb.DescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.MessageDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif md, ok := res.AsProto().(*descriptorpb.DescriptorProto); ok {\n\t\t\treturn md\n\t\t}\n\t}\n\treturn protodesc.ToDescriptorProto(d)\n}", "func (*KnownTypes) Descriptor() ([]byte, []int) {\n\treturn file_jsonpb_proto_test2_proto_rawDescGZIP(), []int{8}\n}", "func (*TelemetryParams) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{62}\n}", "func (*Message5903) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{34}\n}", "func (x *fastReflection_ServiceCommandDescriptor) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_ServiceCommandDescriptor\n}", "func (*EvictWritersResponseProto) Descriptor() ([]byte, []int) {\n\treturn file_ClientDatanodeProtocol_proto_rawDescGZIP(), []int{11}\n}", "func (*ApiListener) Descriptor() ([]byte, []int) {\n\treturn file_envoy_config_listener_v2_api_listener_proto_rawDescGZIP(), []int{0}\n}", "func (*Message6127) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{24}\n}", "func (x *fastReflection_RpcCommandOptions) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_RpcCommandOptions\n}", "func (*CodeLens) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{164}\n}", "func (*DeleteTeam) Descriptor() ([]byte, []int) {\n\treturn file_uac_Team_proto_rawDescGZIP(), []int{6}\n}", "func (*Version) Descriptor() ([]byte, []int) {\n\treturn file_google_protobuf_compiler_plugin_proto_rawDescGZIP(), []int{0}\n}", "func (*Person) Descriptor() ([]byte, []int) {\n\treturn file_protomessage_proto_rawDescGZIP(), []int{0}\n}", "func (*RelationTupleDelta) Descriptor() ([]byte, []int) {\n\treturn file_ory_keto_acl_v1alpha1_write_service_proto_rawDescGZIP(), []int{1}\n}", "func (*FormatMessage) Descriptor() ([]byte, []int) {\n\treturn file_google_devtools_clouddebugger_v2_data_proto_rawDescGZIP(), []int{0}\n}", "func (*PlanChange_Modified) Descriptor() ([]byte, []int) {\n\treturn edgelq_limits_proto_v1alpha2_plan_change_proto_rawDescGZIP(), []int{0, 1}\n}", "func (*DirectiveCreateValidator) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{7}\n}", "func (*SemanticTokensDelta) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{223}\n}", "func (*StrategyChange) Descriptor() ([]byte, []int) {\n\treturn file_rpc_proto_rawDescGZIP(), []int{6}\n}", "func (*Preferences) Descriptor() ([]byte, []int) {\n\treturn file_google_ads_googleads_v2_services_reach_plan_service_proto_rawDescGZIP(), []int{8}\n}", "func (*Message12818) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{5}\n}", "func (*MetadataProto) Descriptor() ([]byte, []int) {\n\treturn file_raft_proto_rawDescGZIP(), []int{7}\n}", "func (*APILevel) Descriptor() ([]byte, []int) {\n\treturn file_Notify_proto_rawDescGZIP(), []int{4}\n}", "func (*ModifyRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_engine_proto_rawDescGZIP(), []int{10}\n}", "func (*ExternalGrpcNode) Descriptor() ([]byte, []int) {\n\treturn file_cloudprovider_externalgrpc_protos_externalgrpc_proto_rawDescGZIP(), []int{1}\n}", "func (*Message12796) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{1}\n}", "func (*PlanChange) Descriptor() ([]byte, []int) {\n\treturn edgelq_limits_proto_v1alpha2_plan_change_proto_rawDescGZIP(), []int{0}\n}", "func (*Message7511) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{16}\n}", "func (*PedidoPyme) Descriptor() ([]byte, []int) {\n\treturn file_helloworld_helloworld_proto_rawDescGZIP(), []int{1}\n}", "func (*ValidatorUpdates) Descriptor() ([]byte, []int) {\n\treturn file_core_abci_v1alpha1_abci_proto_rawDescGZIP(), []int{6}\n}", "func (*DescribeRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_engine_proto_rawDescGZIP(), []int{4}\n}", "func (*Validator) Descriptor() ([]byte, []int) {\n\treturn file_tm_replay_proto_rawDescGZIP(), []int{13}\n}", "func (*Primitive) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{15}\n}", "func (*Message12817) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{22}\n}", "func (*MetadataUpdateEventProto) Descriptor() ([]byte, []int) {\n\treturn file_inotify_proto_rawDescGZIP(), []int{7}\n}", "func (*Message7928) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{18}\n}", "func (*PrivateApiCF) Descriptor() ([]byte, []int) {\n\treturn file_pkg_kascfg_kascfg_proto_rawDescGZIP(), []int{24}\n}", "func (*GetTeamByName) Descriptor() ([]byte, []int) {\n\treturn file_uac_Team_proto_rawDescGZIP(), []int{2}\n}", "func (*DirectiveDelegate) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{9}\n}", "func (*Undefined) Descriptor() ([]byte, []int) {\n\treturn file_rpc_rpc_proto_rawDescGZIP(), []int{3}\n}", "func (*Persistent) Descriptor() ([]byte, []int) {\n\treturn file_msgs_msgs_proto_rawDescGZIP(), []int{2}\n}", "func (*DirectiveEditValidator) Descriptor() ([]byte, []int) {\n\treturn file_Harmony_proto_rawDescGZIP(), []int{8}\n}", "func (*WatchRequestTypeProto) Descriptor() ([]byte, []int) {\n\treturn file_raft_proto_rawDescGZIP(), []int{25}\n}", "func (*Message7921) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{19}\n}", "func (StandardProtocols_Enum) EnumDescriptor() ([]byte, []int) {\n\treturn file_org_apache_beam_model_pipeline_v1_beam_runner_api_proto_rawDescGZIP(), []int{54, 0}\n}", "func (*Message6108) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{31}\n}", "func (*Message7920) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{20}\n}", "func (*FeedbackMetrics) Descriptor() ([]byte, []int) {\n\treturn file_ssn_dataservice_v1_dataservice_proto_rawDescGZIP(), []int{12}\n}" ]
[ "0.71970433", "0.69474417", "0.68986475", "0.6868476", "0.68605566", "0.68491507", "0.6848551", "0.683634", "0.683079", "0.6795646", "0.6790544", "0.6761866", "0.67528534", "0.67488086", "0.67405385", "0.6737484", "0.6721846", "0.6719707", "0.6711924", "0.6711404", "0.67067426", "0.67058504", "0.66942346", "0.66883254", "0.6683996", "0.6682925", "0.66825414", "0.6682167", "0.6681464", "0.6678643", "0.6674312", "0.6673499", "0.66479105", "0.66358757", "0.6632766", "0.662395", "0.66226655", "0.6620012", "0.6615848", "0.6614465", "0.66129106", "0.6611672", "0.66094255", "0.66020155", "0.6599583", "0.659804", "0.6594994", "0.658557", "0.6581239", "0.657991", "0.65774405", "0.65737087", "0.6572288", "0.6569808", "0.65675974", "0.6564263", "0.65642565", "0.65564865", "0.6555307", "0.65545934", "0.6554501", "0.6552104", "0.65520346", "0.6551727", "0.6545552", "0.6545249", "0.6542646", "0.6534937", "0.65328085", "0.6532233", "0.6531081", "0.65310436", "0.652948", "0.6527702", "0.6527491", "0.6525363", "0.65234715", "0.65228415", "0.6519201", "0.6518768", "0.65178335", "0.65141386", "0.650686", "0.65051657", "0.6503247", "0.6503196", "0.64977974", "0.64966935", "0.6494067", "0.64928", "0.64921236", "0.64914787", "0.6486291", "0.6485987", "0.6485658", "0.6485321", "0.6485161", "0.64785784", "0.6474949", "0.6474619" ]
0.6704001
22
Deprecated: Use Message.ProtoReflect.Descriptor instead.
func (*Message) Descriptor() ([]byte, []int) { return file_chat_proto_rawDescGZIP(), []int{2} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ProtoFromMessageDescriptor(d protoreflect.MessageDescriptor) *descriptorpb.DescriptorProto {\n\ttype canProto interface {\n\t\tMessageDescriptorProto() *descriptorpb.DescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.MessageDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif md, ok := res.AsProto().(*descriptorpb.DescriptorProto); ok {\n\t\t\treturn md\n\t\t}\n\t}\n\treturn protodesc.ToDescriptorProto(d)\n}", "func ProtoFromDescriptor(d protoreflect.Descriptor) proto.Message {\n\tswitch d := d.(type) {\n\tcase protoreflect.FileDescriptor:\n\t\treturn ProtoFromFileDescriptor(d)\n\tcase protoreflect.MessageDescriptor:\n\t\treturn ProtoFromMessageDescriptor(d)\n\tcase protoreflect.FieldDescriptor:\n\t\treturn ProtoFromFieldDescriptor(d)\n\tcase protoreflect.OneofDescriptor:\n\t\treturn ProtoFromOneofDescriptor(d)\n\tcase protoreflect.EnumDescriptor:\n\t\treturn ProtoFromEnumDescriptor(d)\n\tcase protoreflect.EnumValueDescriptor:\n\t\treturn ProtoFromEnumValueDescriptor(d)\n\tcase protoreflect.ServiceDescriptor:\n\t\treturn ProtoFromServiceDescriptor(d)\n\tcase protoreflect.MethodDescriptor:\n\t\treturn ProtoFromMethodDescriptor(d)\n\tdefault:\n\t\t// WTF??\n\t\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\t\treturn res.AsProto()\n\t\t}\n\t\treturn nil\n\t}\n}", "func ToDescriptorProto(message protoreflect.MessageDescriptor) *descriptorpb.DescriptorProto {\n\tp := &descriptorpb.DescriptorProto{\n\t\tName: proto.String(string(message.Name())),\n\t\tOptions: proto.Clone(message.Options()).(*descriptorpb.MessageOptions),\n\t}\n\tfor i, fields := 0, message.Fields(); i < fields.Len(); i++ {\n\t\tp.Field = append(p.Field, ToFieldDescriptorProto(fields.Get(i)))\n\t}\n\tfor i, exts := 0, message.Extensions(); i < exts.Len(); i++ {\n\t\tp.Extension = append(p.Extension, ToFieldDescriptorProto(exts.Get(i)))\n\t}\n\tfor i, messages := 0, message.Messages(); i < messages.Len(); i++ {\n\t\tp.NestedType = append(p.NestedType, ToDescriptorProto(messages.Get(i)))\n\t}\n\tfor i, enums := 0, message.Enums(); i < enums.Len(); i++ {\n\t\tp.EnumType = append(p.EnumType, ToEnumDescriptorProto(enums.Get(i)))\n\t}\n\tfor i, xranges := 0, message.ExtensionRanges(); i < xranges.Len(); i++ {\n\t\txrange := xranges.Get(i)\n\t\tp.ExtensionRange = append(p.ExtensionRange, &descriptorpb.DescriptorProto_ExtensionRange{\n\t\t\tStart: proto.Int32(int32(xrange[0])),\n\t\t\tEnd: proto.Int32(int32(xrange[1])),\n\t\t\tOptions: proto.Clone(message.ExtensionRangeOptions(i)).(*descriptorpb.ExtensionRangeOptions),\n\t\t})\n\t}\n\tfor i, oneofs := 0, message.Oneofs(); i < oneofs.Len(); i++ {\n\t\tp.OneofDecl = append(p.OneofDecl, ToOneofDescriptorProto(oneofs.Get(i)))\n\t}\n\tfor i, ranges := 0, message.ReservedRanges(); i < ranges.Len(); i++ {\n\t\trrange := ranges.Get(i)\n\t\tp.ReservedRange = append(p.ReservedRange, &descriptorpb.DescriptorProto_ReservedRange{\n\t\t\tStart: proto.Int32(int32(rrange[0])),\n\t\t\tEnd: proto.Int32(int32(rrange[1])),\n\t\t})\n\t}\n\tfor i, names := 0, message.ReservedNames(); i < names.Len(); i++ {\n\t\tp.ReservedName = append(p.ReservedName, string(names.Get(i)))\n\t}\n\treturn p\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_google_ai_generativelanguage_v1beta2_discuss_service_proto_rawDescGZIP(), []int{2}\n}", "func (x *fastReflection_MsgUpdateParams) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_MsgUpdateParams\n}", "func (*Message6024) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{26}\n}", "func ProtoFromMethodDescriptor(d protoreflect.MethodDescriptor) *descriptorpb.MethodDescriptorProto {\n\ttype canProto interface {\n\t\tMethodDescriptorProto() *descriptorpb.MethodDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.MethodDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif md, ok := res.AsProto().(*descriptorpb.MethodDescriptorProto); ok {\n\t\t\treturn md\n\t\t}\n\t}\n\treturn protodesc.ToMethodDescriptorProto(d)\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{7}\n}", "func (*Embed) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{2}\n}", "func (SVC_Messages) EnumDescriptor() ([]byte, []int) {\n\treturn file_netmessages_proto_rawDescGZIP(), []int{4}\n}", "func (SVC_Messages) EnumDescriptor() ([]byte, []int) {\n\treturn file_csgo_netmessages_proto_rawDescGZIP(), []int{4}\n}", "func (*Message12796) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{1}\n}", "func ProtoFromFieldDescriptor(d protoreflect.FieldDescriptor) *descriptorpb.FieldDescriptorProto {\n\ttype canProto interface {\n\t\tFieldDescriptorProto() *descriptorpb.FieldDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.FieldDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif fd, ok := res.AsProto().(*descriptorpb.FieldDescriptorProto); ok {\n\t\t\treturn fd\n\t\t}\n\t}\n\treturn protodesc.ToFieldDescriptorProto(d)\n}", "func (*Message12818) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{5}\n}", "func (*Message7928) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{18}\n}", "func (*Message7920) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{20}\n}", "func (*Message7511) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{16}\n}", "func (*AddPeerRequest) Descriptor() ([]byte, []int) {\n\treturn file_github_com_yahuizhan_dappley_metrics_go_api_rpc_pb_rpc_proto_rawDescGZIP(), []int{8}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_toit_model_pubsub_message_proto_rawDescGZIP(), []int{2}\n}", "func (*GetMessageRequest) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{11}\n}", "func (*DownlinkMessage) Descriptor() ([]byte, []int) {\n\treturn file_ttn_lorawan_v3_messages_proto_rawDescGZIP(), []int{1}\n}", "func (*Message7921) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{19}\n}", "func (*Message5903) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{34}\n}", "func (*Message3920) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{17}\n}", "func (*GenerateMessageRequest) Descriptor() ([]byte, []int) {\n\treturn file_google_ai_generativelanguage_v1beta2_discuss_service_proto_rawDescGZIP(), []int{0}\n}", "func (*DeleteMessageRequest) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{13}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_helloworld_helloworld_proto_rawDescGZIP(), []int{0}\n}", "func (*Message6127) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{24}\n}", "func (*Message12817) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{22}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_pkg_proto_l3_proto_rawDescGZIP(), []int{0}\n}", "func (*Embed_EmbedField) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{2, 1}\n}", "func ProtoFromFileDescriptor(d protoreflect.FileDescriptor) *descriptorpb.FileDescriptorProto {\n\tif imp, ok := d.(protoreflect.FileImport); ok {\n\t\td = imp.FileDescriptor\n\t}\n\ttype canProto interface {\n\t\tFileDescriptorProto() *descriptorpb.FileDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.FileDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif fd, ok := res.AsProto().(*descriptorpb.FileDescriptorProto); ok {\n\t\t\treturn fd\n\t\t}\n\t}\n\treturn protodesc.ToFileDescriptorProto(d)\n}", "func (*AnalysisMessageWeakSchema) Descriptor() ([]byte, []int) {\n\treturn file_analysis_v1alpha1_message_proto_rawDescGZIP(), []int{1}\n}", "func (x *fastReflection_Metadata) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Metadata\n}", "func (CLC_Messages) EnumDescriptor() ([]byte, []int) {\n\treturn file_csgo_netmessages_proto_rawDescGZIP(), []int{1}\n}", "func (*Message5867) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{29}\n}", "func (CLC_Messages) EnumDescriptor() ([]byte, []int) {\n\treturn file_netmessages_proto_rawDescGZIP(), []int{1}\n}", "func (*Message6578) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{7}\n}", "func (*MessageWithId) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{8}\n}", "func (*Message6108) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{31}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_toit_api_data_proto_rawDescGZIP(), []int{1}\n}", "func (*Message7919) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{21}\n}", "func (*FormatMessage) Descriptor() ([]byte, []int) {\n\treturn file_google_devtools_clouddebugger_v2_data_proto_rawDescGZIP(), []int{0}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{2}\n}", "func (*MyProto) Descriptor() ([]byte, []int) {\n\treturn file_my_proto_proto_rawDescGZIP(), []int{0}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_proto_message_message_proto_rawDescGZIP(), []int{0}\n}", "func (*Message5908) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{13}\n}", "func (x *fastReflection_ServiceCommandDescriptor) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_ServiceCommandDescriptor\n}", "func (x *fastReflection_EventReceive) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_EventReceive\n}", "func (*EnumMessage) Descriptor() ([]byte, []int) {\n\treturn file_enum_enum_example_proto_rawDescGZIP(), []int{0}\n}", "func (*Proto3Message) Descriptor() ([]byte, []int) {\n\treturn file_runtime_internal_examplepb_proto3_proto_rawDescGZIP(), []int{0}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_Forum_proto_rawDescGZIP(), []int{2}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_chat_proto_rawDescGZIP(), []int{0}\n}", "func (x *fastReflection_Params) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Params\n}", "func (*Primitive) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{15}\n}", "func (*Message5907) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{32}\n}", "func (*GetChannelMessagesRequest) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{9}\n}", "func (*Message12821) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{2}\n}", "func (*TraceProto) Descriptor() ([]byte, []int) {\n\treturn file_internal_tracing_extended_extended_trace_proto_rawDescGZIP(), []int{0}\n}", "func (*MsgWithRequiredBytes) Descriptor() ([]byte, []int) {\n\treturn file_jsonpb_proto_test2_proto_rawDescGZIP(), []int{11}\n}", "func (*Messages) Descriptor() ([]byte, []int) {\n\treturn file_proto_model_v1_archive_proto_rawDescGZIP(), []int{1}\n}", "func (x *fastReflection_FlagOptions) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_FlagOptions\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_proto_service_proto_rawDescGZIP(), []int{1}\n}", "func (*Message7865) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{15}\n}", "func (*Message12774) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{0}\n}", "func (*Msg) Descriptor() ([]byte, []int) {\n\treturn file_chatMsg_msg_proto_rawDescGZIP(), []int{1}\n}", "func LegacyLoadMessageDesc(t reflect.Type) protoreflect.MessageDescriptor {\n\treturn legacyLoadMessageDesc(t, \"\")\n}", "func (x *fastReflection_Evidence) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_Evidence\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_uni_proto_rawDescGZIP(), []int{9}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{0}\n}", "func (*Message6107) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{11}\n}", "func (*NetProtoTalker) Descriptor() ([]byte, []int) {\n\treturn file_pkg_smgrpc_smgrpc_proto_rawDescGZIP(), []int{1}\n}", "func (*Message6126) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{8}\n}", "func (NET_Messages) EnumDescriptor() ([]byte, []int) {\n\treturn file_netmessages_proto_rawDescGZIP(), []int{0}\n}", "func (x *fastReflection_MsgUpdateParamsResponse) Descriptor() protoreflect.MessageDescriptor {\n\treturn md_MsgUpdateParamsResponse\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_api_protobuf_spec_example_example_proto_rawDescGZIP(), []int{0}\n}", "func (*Message12820) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{3}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_api_proto_rawDescGZIP(), []int{4}\n}", "func (*Message12819) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{4}\n}", "func (*MessageInfo) Descriptor() ([]byte, []int) {\n\treturn file_TSPArchiveMessages_proto_rawDescGZIP(), []int{1}\n}", "func (NET_Messages) EnumDescriptor() ([]byte, []int) {\n\treturn file_csgo_netmessages_proto_rawDescGZIP(), []int{0}\n}", "func (*BroadcastMsg) Descriptor() ([]byte, []int) {\n\treturn file_proto_rpc_rpc_proto_rawDescGZIP(), []int{1}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_proto_threads_proto_rawDescGZIP(), []int{1}\n}", "func (*MsgWithIndirectRequired) Descriptor() ([]byte, []int) {\n\treturn file_jsonpb_proto_test2_proto_rawDescGZIP(), []int{10}\n}", "func (*Messages) Descriptor() ([]byte, []int) {\n\treturn file_Forum_proto_rawDescGZIP(), []int{3}\n}", "func (*CMsgClientToGCPlayerStatsRequest) Descriptor() ([]byte, []int) {\n\treturn file_dota_gcmessages_client_proto_rawDescGZIP(), []int{143}\n}", "func (*Message6052) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{25}\n}", "func (*Listen) Descriptor() ([]byte, []int) {\n\treturn file_pkg_smgrpc_smgrpc_proto_rawDescGZIP(), []int{4}\n}", "func (*ListMessagesRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_threads_proto_rawDescGZIP(), []int{14}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_protos_addressbook_proto_rawDescGZIP(), []int{0}\n}", "func (*Message10319) Descriptor() ([]byte, []int) {\n\treturn file_datasets_google_message4_benchmark_message4_2_proto_rawDescGZIP(), []int{6}\n}", "func (*CMsgInspectElement) Descriptor() ([]byte, []int) {\n\treturn file_steam_htmlmessages_proto_rawDescGZIP(), []int{145}\n}", "func (*Msg) Descriptor() ([]byte, []int) {\n\treturn file_msgs_msgs_proto_rawDescGZIP(), []int{10}\n}", "func ToFieldDescriptorProto(field protoreflect.FieldDescriptor) *descriptorpb.FieldDescriptorProto {\n\tp := &descriptorpb.FieldDescriptorProto{\n\t\tName: proto.String(string(field.Name())),\n\t\tNumber: proto.Int32(int32(field.Number())),\n\t\tLabel: descriptorpb.FieldDescriptorProto_Label(field.Cardinality()).Enum(),\n\t\tOptions: proto.Clone(field.Options()).(*descriptorpb.FieldOptions),\n\t}\n\tif field.IsExtension() {\n\t\tp.Extendee = fullNameOf(field.ContainingMessage())\n\t}\n\tif field.Kind().IsValid() {\n\t\tp.Type = descriptorpb.FieldDescriptorProto_Type(field.Kind()).Enum()\n\t}\n\tif field.Enum() != nil {\n\t\tp.TypeName = fullNameOf(field.Enum())\n\t}\n\tif field.Message() != nil {\n\t\tp.TypeName = fullNameOf(field.Message())\n\t}\n\tif field.HasJSONName() {\n\t\t// A bug in older versions of protoc would always populate the\n\t\t// \"json_name\" option for extensions when it is meaningless.\n\t\t// When it did so, it would always use the camel-cased field name.\n\t\tif field.IsExtension() {\n\t\t\tp.JsonName = proto.String(strs.JSONCamelCase(string(field.Name())))\n\t\t} else {\n\t\t\tp.JsonName = proto.String(field.JSONName())\n\t\t}\n\t}\n\tif field.Syntax() == protoreflect.Proto3 && field.HasOptionalKeyword() {\n\t\tp.Proto3Optional = proto.Bool(true)\n\t}\n\tif field.HasDefault() {\n\t\tdef, err := defval.Marshal(field.Default(), field.DefaultEnumValue(), field.Kind(), defval.Descriptor)\n\t\tif err != nil && field.DefaultEnumValue() != nil {\n\t\t\tdef = string(field.DefaultEnumValue().Name()) // occurs for unresolved enum values\n\t\t} else if err != nil {\n\t\t\tpanic(fmt.Sprintf(\"%v: %v\", field.FullName(), err))\n\t\t}\n\t\tp.DefaultValue = proto.String(def)\n\t}\n\tif oneof := field.ContainingOneof(); oneof != nil {\n\t\tp.OneofIndex = proto.Int32(int32(oneof.Index()))\n\t}\n\treturn p\n}", "func (*PatchCollectorsRequest) Descriptor() ([]byte, []int) {\n\treturn file_proto_clarifai_api_service_proto_rawDescGZIP(), []int{161}\n}", "func (*Persistent) Descriptor() ([]byte, []int) {\n\treturn file_msgs_msgs_proto_rawDescGZIP(), []int{2}\n}", "func (*PatchMessageParams) Descriptor() ([]byte, []int) {\n\treturn file_message_proto_rawDescGZIP(), []int{3}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_api_proto_rawDescGZIP(), []int{2}\n}", "func (*SendMessageRequest) Descriptor() ([]byte, []int) {\n\treturn file_chat_v1_messages_proto_rawDescGZIP(), []int{17}\n}", "func (*Message) Descriptor() ([]byte, []int) {\n\treturn file_Notification_proto_rawDescGZIP(), []int{0}\n}" ]
[ "0.7343522", "0.7332535", "0.72937775", "0.71048754", "0.7050007", "0.6987258", "0.6960036", "0.691296", "0.68633527", "0.6859684", "0.68459857", "0.6845709", "0.68352896", "0.68222934", "0.6815316", "0.68140835", "0.6813839", "0.68118334", "0.6800345", "0.68002284", "0.6799499", "0.6796825", "0.67926", "0.67908096", "0.6789334", "0.67887735", "0.67856336", "0.67687345", "0.6758159", "0.67567915", "0.67553294", "0.6754251", "0.67511344", "0.67430514", "0.67410934", "0.6740679", "0.6739927", "0.6733324", "0.6731579", "0.6729524", "0.6722112", "0.6720688", "0.67200774", "0.6719126", "0.67151237", "0.67124426", "0.6710721", "0.67040044", "0.67013395", "0.67009544", "0.6700674", "0.66970474", "0.6696509", "0.66904116", "0.6690397", "0.6689474", "0.6688465", "0.66880924", "0.6680005", "0.6678234", "0.66777873", "0.66735333", "0.6672438", "0.666546", "0.6663106", "0.66629595", "0.66620827", "0.6661886", "0.66608155", "0.6655773", "0.66515446", "0.6651183", "0.6649967", "0.6646376", "0.6644945", "0.6638674", "0.66365266", "0.66338146", "0.6632549", "0.6631053", "0.6631018", "0.66247624", "0.66224045", "0.6620985", "0.66124934", "0.66123223", "0.6607028", "0.6605048", "0.66043353", "0.66025764", "0.6599376", "0.65979433", "0.65965575", "0.6596149", "0.65944296", "0.6593448", "0.65857214", "0.658271", "0.6579932", "0.6579099" ]
0.67455775
33
Path Type implements the starlark.Value.Type() method.
func (p Path) Type() string { return fmt.Sprintf("Path") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p Path) Type(reference, value interface{}) Path { return p.with(Type, reference, value) }", "func Type(value r.Value) r.Type {\n\tif !value.IsValid() || value == None {\n\t\treturn nil\n\t}\n\treturn value.Type()\n}", "func (value *Value) Type() Type {\n\treturn value.valueType\n}", "func (a ValueNode) GetType() string {\n\treturn \"ValueNode\"\n}", "func (o FioSpecVolumeVolumeSourceHostPathOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceHostPath) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (this *Value) Type() int {\n\treturn this.parsedType\n}", "func (val Value) Type() Type {\n\treturn val.typ\n}", "func (v *Value) Type() kiwi.ValueType {\n\treturn Type\n}", "func (this *Self) Type() value.Type { return value.JSON }", "func (o IopingSpecVolumeVolumeSourceHostPathOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceHostPath) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (v Value) Type() Type {\n\treturn v.typ\n}", "func (v Value) Type() Type {\n\treturn v.Typ\n}", "func (this *Element) Type() value.Type { return value.JSON }", "func (s *StringPointerValue) Type() string {\n\treturn \"string\"\n}", "func (v Value) Type() Type {\n\tif !v.v.IsValid() {\n\t\treturn TypeUndefined\n\t}\n\n\tif v.v.CanInterface() {\n\t\ti := v.v.Interface()\n\t\tswitch i.(type) {\n\t\tcase Function:\n\t\t\treturn TypeFunction\n\t\tcase Object:\n\t\t\tif _, ok := i.(stringObject); ok {\n\t\t\t\treturn TypeString\n\t\t\t}\n\t\t\treturn TypeObject\n\t\t}\n\t}\n\n\tswitch v.v.Kind() {\n\tcase reflect.Ptr:\n\t\treturn TypeNull\n\tcase reflect.Bool:\n\t\treturn TypeBoolean\n\tcase reflect.Float64:\n\t\treturn TypeNumber\n\tdefault:\n\t\treturn TypeUndefined\n\t}\n\n}", "func (sv *SupernodesValue) Type() string {\n\treturn \"supernodes\"\n}", "func (v Value) Type() querypb.Type {\n\treturn v.typ\n}", "func (v *Value) Type() *JSONType {\n\tt := C.zj_Type(v.V)\n\tif t == nil {\n\t\treturn nil\n\t}\n\tret := JSONType(*t)\n\treturn &ret\n}", "func (element *Element) Type(value string) *Element {\n\treturn element.Attr(\"type\", value)\n}", "func (o FioSpecVolumeVolumeSourceHostPathPtrOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FioSpecVolumeVolumeSourceHostPath) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Type\n\t}).(pulumi.StringPtrOutput)\n}", "func (j *SearchHighlight) Type(v string) *SearchHighlight {\n\tj.Json.SetPath([]string{\"fields\", j.field, \"type\"}, v)\n\treturn j\n}", "func (object Object) Type(value interface{}) Object {\n\treturn object.Property(as.PropertyType, value)\n}", "func (o JavaScriptFunctionBindingOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v JavaScriptFunctionBinding) string { return v.Type }).(pulumi.StringOutput)\n}", "func (s Schemas) Type() Type {\n\tif s.TF != nil {\n\t\tswitch s.TF.Type {\n\t\tcase schema.TypeBool:\n\t\t\treturn TypeBool\n\t\tcase schema.TypeInt, schema.TypeFloat:\n\t\t\treturn TypeNumber\n\t\tcase schema.TypeString:\n\t\t\treturn TypeString\n\t\tcase schema.TypeList, schema.TypeSet:\n\t\t\treturn s.ElemSchemas().Type().ListOf()\n\t\tcase schema.TypeMap:\n\t\t\treturn TypeMap\n\t\tdefault:\n\t\t\treturn TypeUnknown\n\t\t}\n\t}\n\n\treturn TypeUnknown\n}", "func (o ApiOperationRequestQueryParameterOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ApiOperationRequestQueryParameter) string { return v.Type }).(pulumi.StringOutput)\n}", "func (e REnv) Type() Type { return e.Value().Type() }", "func (o IopingSpecVolumeVolumeSourceHostPathPtrOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceHostPath) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Type\n\t}).(pulumi.StringPtrOutput)\n}", "func (r Resource) Type() string {\n\treturn r.typ\n}", "func (ro ReadOnlyPathMap) TypeForPath(path string) (devicechange.ValueType, error) {\n\tfor k, subPaths := range ro {\n\t\tfor k1, sp := range subPaths {\n\t\t\tif k1 == \"/\" {\n\t\t\t\tif k == path {\n\t\t\t\t\treturn sp.ValueType, nil\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tif k+k1 == path {\n\t\t\t\t\treturn sp.ValueType, nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn devicechange.ValueType_EMPTY, fmt.Errorf(\"path %s not found in RO paths of model\", path)\n}", "func (rw ReadWritePathMap) TypeForPath(path string) (devicechange.ValueType, error) {\n\tfor k, elem := range rw {\n\t\tif k == path {\n\t\t\treturn elem.ValueType, nil\n\t\t}\n\t}\n\treturn devicechange.ValueType_EMPTY, fmt.Errorf(\"path %s not found in RW paths of model\", path)\n}", "func (op *OperationProperty) Type() string {\n\treturn \"github.com/wunderkraut/radi-api/operation.Operation\"\n}", "func (node *GoValueNode) GetType() (reflect.Type, error) {\n\n\treturn node.thisValue.Type(), nil\n}", "func (this *NowStr) Type() value.Type { return value.STRING }", "func (md *MetaData) Type(key ...string) string {\n\tif ki, ok := md.keyInfo[Key(key).String()]; ok {\n\t\treturn ki.tomlType.typeString()\n\t}\n\treturn \"\"\n}", "func (NamespaceNode) Type() string { return TypeNamespaceNode }", "func (n *piName) Type() Type {\n\treturn n.t\n}", "func (v Value) Type() Type {\n\tpanic(message)\n}", "func (o JavaScriptFunctionBindingResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v JavaScriptFunctionBindingResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (NamespaceStar) Type() string { return TypeNamespaceStar }", "func (p RProc) Type() Type { return p.Value().Type() }", "func (o CSharpFunctionBindingOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CSharpFunctionBinding) string { return v.Type }).(pulumi.StringOutput)\n}", "func (n Name) Type() string {\n\t_, t := n.GetLookupAndType()\n\treturn t\n}", "func (o CSharpFunctionBindingResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CSharpFunctionBindingResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o MonitoredResourceDescriptorOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v MonitoredResourceDescriptor) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o *PairAnyValueAnyValue) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (o MetricStatusPatchOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MetricStatusPatch) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (o MetricStatusPatchOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MetricStatusPatch) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (data *Instance) Type() Value {\n\treturn data.TypeTag\n}", "func (l *DirName) Type() string {\n\treturn \"string\"\n}", "func (t *jsonDataType) Type() interface{} {\n\treturn \"\"\n}", "func (i ID) Type() string {\n\tstuff := strings.Split(i.String(), \"/\")\n\treturn stuff[0]\n}", "func (obj *SObject) Type() string {\n\tattributes := obj.AttributesField()\n\tif attributes == nil {\n\t\treturn \"\"\n\t}\n\treturn attributes.Type\n}", "func (a AttributeValue) Type() AttributeValueType {\n\tif a.orig.Value == nil {\n\t\treturn AttributeValueNULL\n\t}\n\tswitch a.orig.Value.(type) {\n\tcase *otlpcommon.AnyValue_StringValue:\n\t\treturn AttributeValueSTRING\n\tcase *otlpcommon.AnyValue_BoolValue:\n\t\treturn AttributeValueBOOL\n\tcase *otlpcommon.AnyValue_IntValue:\n\t\treturn AttributeValueINT\n\tcase *otlpcommon.AnyValue_DoubleValue:\n\t\treturn AttributeValueDOUBLE\n\tcase *otlpcommon.AnyValue_KvlistValue:\n\t\treturn AttributeValueMAP\n\tcase *otlpcommon.AnyValue_ArrayValue:\n\t\treturn AttributeValueARRAY\n\t}\n\treturn AttributeValueNULL\n}", "func (t Type) Type() string {\n\treturn t.typeName\n}", "func (v Value) Type() ValueType {\n\tif v.iface == nil {\n\t\treturn NilType\n\t}\n\tswitch v.iface.(type) {\n\tcase int64:\n\t\treturn IntType\n\tcase float64:\n\t\treturn FloatType\n\tcase bool:\n\t\treturn BoolType\n\tcase string:\n\t\treturn StringType\n\tcase *Table:\n\t\treturn TableType\n\tcase *Code:\n\t\treturn CodeType\n\tcase Callable:\n\t\treturn FunctionType\n\tcase *Thread:\n\t\treturn ThreadType\n\tcase *UserData:\n\t\treturn UserDataType\n\tdefault:\n\t\treturn UnknownType\n\t}\n}", "func (s *String) Type() ObjectType {\n\treturn STRING\n}", "func (ts *TypeSet) Type(s string) Type {\n\tts.RLock()\n\tdefer ts.RUnlock()\n\treturn ts.types[s]\n}", "func (n *Node) Type(ctx context.Context) provider.ResourceType {\n\tif n.nodeType != nil {\n\t\treturn *n.nodeType\n\t}\n\n\tt := provider.ResourceType_RESOURCE_TYPE_INVALID\n\n\t// Try to read from xattrs\n\ttypeAttr, err := n.XattrInt32(ctx, prefixes.TypeAttr)\n\tif err == nil {\n\t\tt = provider.ResourceType(typeAttr)\n\t\tn.nodeType = &t\n\t\treturn t\n\t}\n\n\t// Fall back to checking on disk\n\tfi, err := os.Lstat(n.InternalPath())\n\tif err != nil {\n\t\treturn t\n\t}\n\n\tswitch {\n\tcase fi.IsDir():\n\t\tif _, err = n.Xattr(ctx, prefixes.ReferenceAttr); err == nil {\n\t\t\tt = provider.ResourceType_RESOURCE_TYPE_REFERENCE\n\t\t} else {\n\t\t\tt = provider.ResourceType_RESOURCE_TYPE_CONTAINER\n\t\t}\n\tcase fi.Mode().IsRegular():\n\t\tt = provider.ResourceType_RESOURCE_TYPE_FILE\n\tcase fi.Mode()&os.ModeSymlink != 0:\n\t\tt = provider.ResourceType_RESOURCE_TYPE_SYMLINK\n\t\t// TODO reference using ext attr on a symlink\n\t\t// nodeType = provider.ResourceType_RESOURCE_TYPE_REFERENCE\n\t}\n\tn.nodeType = &t\n\treturn t\n}", "func (d UserData) Type() string {\n\tval := d.ModelData.Get(models.NewFieldName(\"Type\", \"type\"))\n\tif !d.Has(models.NewFieldName(\"Type\", \"type\")) {\n\t\treturn *new(string)\n\t}\n\treturn val.(string)\n}", "func (_this *Report) Type() string {\n\tvar ret string\n\tvalue := _this.Value_JS.Get(\"type\")\n\tret = (value).String()\n\treturn ret\n}", "func (j *JSONData) Type() JSONType {\n\tif j == nil || j.value == nil { // no data\n\t\treturn JSONnil\n\t}\n\tvalue := *j.value\n\tif value == nil {\n\t\treturn JSONnull\n\t}\n\tif _, ok := value.(bool); ok {\n\t\treturn JSONboolean\n\t}\n\tif _, ok := value.(float64); ok {\n\t\treturn JSONnumber\n\t}\n\tif _, ok := value.(string); ok {\n\t\treturn JSONstring\n\t}\n\tif _, ok := value.([]interface{}); ok {\n\t\treturn JSONarray\n\t}\n\tif _, ok := value.(map[string]interface{}); ok {\n\t\treturn JSONobject\n\t}\n\tpanic(errors.New(\"JSONData corrupt\"))\n}", "func (o MetricSpecPatchOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MetricSpecPatch) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (o MetricSpecPatchOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MetricSpecPatch) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (s *StringStep) Type() interface{} {\n\treturn TypeString\n}", "func (e *Entry) Type() tree.Type {\n\tif e.notFound {\n\t\treturn tree.TYPE_NOTFOUND\n\t}\n\tswitch e.st.Mode() & os.ModeType {\n\tcase 0:\n\t\treturn tree.TYPE_REGULAR\n\tcase os.ModeDir:\n\t\treturn tree.TYPE_DIRECTORY\n\tcase os.ModeSymlink:\n\t\treturn tree.TYPE_SYMLINK\n\tdefault:\n\t\treturn tree.TYPE_UNKNOWN\n\t}\n}", "func (o MonitoredResourceDescriptorResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v MonitoredResourceDescriptorResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o ScalarFunctionPropertiesResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ScalarFunctionPropertiesResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o ApiOperationRequestRepresentationFormParameterOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ApiOperationRequestRepresentationFormParameter) string { return v.Type }).(pulumi.StringOutput)\n}", "func (myOperatingSystemType *OperatingSystemType) Type() (param string) {\n\treturn myOperatingSystemType.Typevar\n}", "func (attr *Attribute) Type() Type {\n\treturn attr.typ\n}", "func (s *Service) Type() interface{} {\n\treturn (*s)[jsonldType]\n}", "func (b baseValue) Type() string {\n\treturn string(b.flagType)\n}", "func (l *NullLiteral) Type() Type {\n\treturn UnknownType\n}", "func (o MetricTargetPatchOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v MetricTargetPatch) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (o ScalarFunctionPropertiesOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ScalarFunctionProperties) string { return v.Type }).(pulumi.StringOutput)\n}", "func (s S) Type() Type {\n\treturn s.typ\n}", "func (obj GoObject) Type() Type {\n\treturn GetGoType(reflect.TypeOf(obj.val))\n}", "func (o StorageSettingOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v StorageSetting) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (n *node) Type() string {\n\tif n.nodeType == \"\" {\n\t\treturn \"node\"\n\t}\n\treturn n.nodeType\n}", "func (flg Var) Type() string {\n\treturn \"@filename or JSON literal\"\n}", "func (p *Predicate) Type() Type {\n\tif p.anchor == nil {\n\t\treturn Immutable\n\t}\n\treturn Temporal\n}", "func (v *VInteger) Type() string {\n\treturn \"integer\"\n}", "func (s *String) Type() Type {\n\treturn STRING_OBJ\n}", "func (r *VarRef) Type() Type {\n\treturn r.ExprType\n}", "func (nsfd NamespaceFd) Type() (species.NamespaceType, error) {\n\tt, err := ioctl(int(nsfd), _NS_GET_NSTYPE)\n\treturn species.NamespaceType(t), err\n}", "func (k *Key) Type() string {\n\treturn k.base.Type()\n}", "func (d *Driver) Type() (t string) {\n\treturn \"go\"\n}", "func (*StringSchema) Type() int {\n\treturn String\n}", "func (o StorageSettingResponseOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v StorageSettingResponse) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (o AzureMachineLearningServiceFunctionBindingResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AzureMachineLearningServiceFunctionBindingResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (m *Property) GetType() (val string, set bool) {\n\tif m.Type == nil {\n\t\treturn\n\t}\n\n\treturn *m.Type, true\n}", "func (o ApiOperationResponseRepresentationFormParameterOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ApiOperationResponseRepresentationFormParameter) string { return v.Type }).(pulumi.StringOutput)\n}", "func (r *Rdispatch) Type() int8 {\n\treturn RdispatchTpe\n}", "func (s *Smpval) Type() reflect.Type {\n\treturn s.val.Type()\n}", "func (o AzureMachineLearningServiceFunctionBindingOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AzureMachineLearningServiceFunctionBinding) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o ReferenceInputPropertiesResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ReferenceInputPropertiesResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (s *Swift) Type() Type {\n\tif hasBranchCode(s.value) {\n\t\treturn Type11\n\t}\n\treturn Type8\n}", "func (m Shape) GetType() string {\n\treturn m.Type\n}", "func (o AzureMachineLearningStudioFunctionBindingOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AzureMachineLearningStudioFunctionBinding) string { return v.Type }).(pulumi.StringOutput)\n}", "func (nsf NamespaceFile) Type() (species.NamespaceType, error) {\n\tt, err := ioctl(int(nsf.Fd()), _NS_GET_NSTYPE)\n\treturn species.NamespaceType(t), err\n}" ]
[ "0.7426185", "0.67868197", "0.6723539", "0.6706437", "0.657168", "0.6492181", "0.64873666", "0.6484821", "0.6484744", "0.64651024", "0.64247006", "0.64185387", "0.64053804", "0.6343937", "0.633024", "0.6304487", "0.6299285", "0.62893945", "0.62660074", "0.62631536", "0.6236115", "0.6228473", "0.6228036", "0.621171", "0.6171286", "0.6161899", "0.61560816", "0.6152463", "0.6151192", "0.61489195", "0.61472046", "0.6144614", "0.6137164", "0.6135511", "0.6130519", "0.6127005", "0.6115918", "0.61037475", "0.6097738", "0.6097477", "0.60909516", "0.60891604", "0.6082147", "0.6066978", "0.6061715", "0.60596466", "0.60596466", "0.6050206", "0.6045077", "0.60393983", "0.6038654", "0.60345083", "0.60259473", "0.6025006", "0.60242534", "0.6012637", "0.6010242", "0.6008266", "0.6007219", "0.6003113", "0.5988961", "0.59849876", "0.59849876", "0.5976634", "0.59652686", "0.59544754", "0.5941195", "0.59396446", "0.5934535", "0.59327835", "0.5932149", "0.5930332", "0.5929747", "0.5921442", "0.5919957", "0.59029424", "0.58956367", "0.58943003", "0.58915114", "0.5888488", "0.5879263", "0.58791757", "0.58764684", "0.5864796", "0.5860231", "0.58440197", "0.5843489", "0.5842129", "0.58375883", "0.58342963", "0.5834156", "0.5832169", "0.58291703", "0.5827353", "0.58258593", "0.58198917", "0.58174056", "0.581541", "0.5814113", "0.5808484" ]
0.7162587
1
Freeze implements the starlark.Value.Freeze() method.
func (p Path) Freeze() {}
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (i *Index) Freeze() {\n\ti.frozen = true\n}", "func (p *Poly) freeze() {\n\tfor i := 0; i < n; i++ {\n\t\tp[i] = freeze(p[i])\n\t}\n}", "func freeze(o *goja.Object) {\n\tfor _, key := range o.Keys() {\n\t\to.DefineDataProperty(key, o.Get(key), goja.FLAG_FALSE, goja.FLAG_FALSE, goja.FLAG_TRUE)\n\t}\n}", "func (df *DataFrame) Freeze() { df.frozen = true }", "func (i *Instance) Freeze() {\n}", "func (t *Target) Freeze() {}", "func (v *ReadCloserValue) Freeze() {}", "func Freeze(x int32) int32 {\n\tx -= 9829 * ((13*x) >> 17)\n\tx -= 9829 * ((427*x + 2097152) >> 22)\n\ty := x + 9829\n\tv := subtle.ConstantTimeLessOrEq(int(x), -1)\n\treturn int32(subtle.ConstantTimeSelect(v, int(y), int(x)))\n}", "func (f *chainFreezer) freeze(db database.KeyValueStore) {\n\tnfdb := &nofreezedb{KeyValueStore: db}\n\n\tvar (\n\t\tbackoff bool\n\t\ttriggered chan struct{} // Used in tests\n\t)\n\tfor {\n\t\tselect {\n\t\tcase <-f.quit:\n\t\t\tlog.Info(\"Freezer shutting down\")\n\t\t\treturn\n\t\tdefault:\n\t\t}\n\t\tif backoff {\n\t\t\t// If we were doing a manual trigger, notify it\n\t\t\tif triggered != nil {\n\t\t\t\ttriggered <- struct{}{}\n\t\t\t\ttriggered = nil\n\t\t\t}\n\t\t\tselect {\n\t\t\tcase <-time.NewTimer(freezerRecheckInterval).C:\n\t\t\t\tbackoff = false\n\t\t\tcase triggered = <-f.trigger:\n\t\t\t\tbackoff = false\n\t\t\tcase <-f.quit:\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\t// Retrieve the freezing threshold.\n\t\thash := ReadHeadBlockHash(nfdb)\n\t\tif hash == (common.Hash{}) {\n\t\t\tlog.Debug(\"Current full block hash unavailable\") // new chain, empty database\n\t\t\tbackoff = true\n\t\t\tcontinue\n\t\t}\n\t\tnumber := ReadHeaderNumber(nfdb, hash)\n\t\tthreshold := atomic.LoadUint64(&f.threshold)\n\t\tfrozen := atomic.LoadUint64(&f.frozen)\n\t\tswitch {\n\t\tcase number == nil:\n\t\t\tlog.Error(\"Current full block number unavailable\", \"hash\", hash)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\n\t\tcase *number < threshold:\n\t\t\tlog.Debug(\"Current full block not old enough\", \"number\", *number, \"hash\", hash, \"delay\", threshold)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\n\t\tcase *number-threshold <= frozen:\n\t\t\tlog.Debug(\"Ancient blocks frozen already\", \"number\", *number, \"hash\", hash, \"frozen\", frozen)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\t\t}\n\t\thead := ReadHeader(nfdb, hash, *number)\n\t\tif head == nil {\n\t\t\tlog.Error(\"Current full block unavailable\", \"number\", *number, \"hash\", hash)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\t\t}\n\n\t\t// Seems we have data ready to be frozen, process in usable batches\n\t\tvar (\n\t\t\tstart = time.Now()\n\t\t\tfirst, _ = f.Ancients()\n\t\t\tlimit = *number - threshold\n\t\t)\n\t\tif limit-first > freezerBatchLimit {\n\t\t\tlimit = first + freezerBatchLimit\n\t\t}\n\t\tancients, err := f.freezeRange(nfdb, first, limit)\n\t\tif err != nil {\n\t\t\tlog.Error(\"Error in block freeze operation\", \"err\", err)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\t\t}\n\n\t\t// Batch of blocks have been frozen, flush them before wiping from leveldb\n\t\tif err := f.Sync(); err != nil {\n\t\t\tlog.Critical(\"Failed to flush frozen tables\", \"err\", err)\n\t\t}\n\n\t\t// Wipe out all data from the active database\n\t\tbatch := db.NewBatch()\n\t\tfor i := 0; i < len(ancients); i++ {\n\t\t\t// Always keep the genesis block in active database\n\t\t\tif first+uint64(i) != 0 {\n\t\t\t\tDeleteBlockWithoutNumber(batch, ancients[i], first+uint64(i))\n\t\t\t\tDeleteCanonicalHash(batch, first+uint64(i))\n\t\t\t}\n\t\t}\n\t\tif err := batch.Write(); err != nil {\n\t\t\tlog.Critical(\"Failed to delete frozen canonical blocks\", \"err\", err)\n\t\t}\n\t\tbatch.Reset()\n\n\t\t// Wipe out side chains also and track dangling side chains\n\t\tvar dangling []common.Hash\n\t\tfrozen = atomic.LoadUint64(&f.frozen) // Needs reload after during freezeRange\n\t\tfor number := first; number < frozen; number++ {\n\t\t\t// Always keep the genesis block in active database\n\t\t\tif number != 0 {\n\t\t\t\tdangling = ReadAllHashes(db, number)\n\t\t\t\tfor _, hash := range dangling {\n\t\t\t\t\tlog.Debug(\"Deleting side chain\", \"number\", number, \"hash\", hash)\n\t\t\t\t\tDeleteBlock(batch, hash, number)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif err := batch.Write(); err != nil {\n\t\t\tlog.Critical(\"Failed to delete frozen side blocks\", \"err\", err)\n\t\t}\n\t\tbatch.Reset()\n\n\t\t// Step into the future and delete and dangling side chains\n\t\tif frozen > 0 {\n\t\t\ttip := frozen\n\t\t\tfor len(dangling) > 0 {\n\t\t\t\tdrop := make(map[common.Hash]struct{})\n\t\t\t\tfor _, hash := range dangling {\n\t\t\t\t\tlog.Debug(\"Dangling parent from Freezer\", \"number\", tip-1, \"hash\", hash)\n\t\t\t\t\tdrop[hash] = struct{}{}\n\t\t\t\t}\n\t\t\t\tchildren := ReadAllHashes(db, tip)\n\t\t\t\tfor i := 0; i < len(children); i++ {\n\t\t\t\t\t// Dig up the child and ensure it's dangling\n\t\t\t\t\tchild := ReadHeader(nfdb, children[i], tip)\n\t\t\t\t\tif child == nil {\n\t\t\t\t\t\tlog.Error(\"Missing dangling header\", \"number\", tip, \"hash\", children[i])\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tif _, ok := drop[child.ParentHash]; !ok {\n\t\t\t\t\t\tchildren = append(children[:i], children[i+1:]...)\n\t\t\t\t\t\ti--\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\t// Delete all block data associated with the child\n\t\t\t\t\tlog.Debug(\"Deleting dangling block\", \"number\", tip, \"hash\", children[i], \"parent\", child.ParentHash)\n\t\t\t\t\tDeleteBlock(batch, children[i], tip)\n\t\t\t\t}\n\t\t\t\tdangling = children\n\t\t\t\ttip++\n\t\t\t}\n\t\t\tif err := batch.Write(); err != nil {\n\t\t\t\tlog.Critical(\"Failed to delete dangling side blocks\", \"err\", err)\n\t\t\t}\n\t\t}\n\n\t\t// Log something friendly for the user\n\t\tcontext := []interface{}{\n\t\t\t\"blocks\", frozen - first, \"elapsed\", common.PrettyDuration(time.Since(start)), \"number\", frozen - 1,\n\t\t}\n\t\tif n := len(ancients); n > 0 {\n\t\t\tcontext = append(context, []interface{}{\"hash\", ancients[n-1]}...)\n\t\t}\n\t\tlog.Infof(\"Deep froze chain segment: %+v\", context...)\n\n\t\t// Avoid database thrashing with tiny writes\n\t\tif frozen-first < freezerBatchLimit {\n\t\t\tbackoff = true\n\t\t}\n\t}\n}", "func (c *Container) Freeze() *Container {\n\tif c == nil {\n\t\treturn nil\n\t}\n\tif c.flags&flagDirty != 0 {\n\t\tif roaringParanoia {\n\t\t\tpanic(\"freezing dirty container\")\n\t\t}\n\t\t// c.Repair won't work if this is already frozen, but in\n\t\t// theory that can't happen?\n\t\tc.Repair()\n\t}\n\t// don't need to freeze\n\tif c.flags&flagFrozen != 0 {\n\t\treturn c\n\t}\n\tc.flags |= flagFrozen\n\treturn c\n}", "func (me TxsdAnimTimingAttrsFill) IsFreeze() bool { return me.String() == \"freeze\" }", "func (g *metadataGraph) freeze(ctx context.Context) {\n\tg.root.freeze(ctx)\n}", "func (d *Dam) Unlock() {\n\td.freeze.Unlock()\n}", "func (s *Sub) Freeze() {}", "func main() {\n\tx := new(int)\n\t*x++ // ok\n\n\tx = freeze(x)\n\n\tfmt.Println(*x) // ok; prints 1\n\t//*x++ // not ok; panics!\n}", "func (recv *Object) FreezeNotify() {\n\tC.g_object_freeze_notify((*C.GObject)(recv.native))\n\n\treturn\n}", "func (sa *SuffixArray) Freeze() error { return sa.ba.Freeze() }", "func (s *inMemSpannerServer) Unfreeze() {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tclose(s.freezed)\n}", "func (iv *writeOnlyInterval) freeze(s *Schema) (*Interval, error) {\n\tif err := iv.closeCurrentSegment(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tiv.Segments = make([]*Segment, iv.NumSegments)\n\tfor i := 0; i < iv.NumSegments; i++ {\n\t\tif !iv.DiskBacked {\n\t\t\tiv.Segments[i] = &Segment{Bytes: iv.buffers[i].Bytes()}\n\t\t\tcontinue\n\t\t}\n\t\tfilename := iv.SegmentFilename(s, i)\n\t\tf, err := os.Open(filename)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tmapped, err := mmap.Map(f, mmap.RDONLY, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tiv.Segments[i] = &Segment{File: f, Bytes: mapped}\n\t}\n\treturn &iv.Interval, nil\n}", "func (d *Director) FreezeMonkey(rng *rand.Rand, intensity float64) {\n\tif intensity < 0.1 {\n\t\treturn\n\t}\n\ttarget := d.randomAgent(rng)\n\tduration := d.makeDuration(rng, 1000, intensity)\n\tlog.Printf(\"[monkey] Freezing %v for %v\", target, duration)\n\tgo target.Stop(duration)\n}", "func NewFreezeParams() *FreezeParams {\n\treturn &FreezeParams{\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (_OracleMgr *OracleMgrCaller) FreezePeriod(opts *bind.CallOpts) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _OracleMgr.contract.Call(opts, out, \"freezePeriod\")\n\treturn *ret0, err\n}", "func (*CapturedStacktrace) Freeze() {}", "func (s *Client) Freeze(username string) error {\n\tuser := s.Init(username)\n\tdata, err := json.Marshal(user)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\ttx, err := s.db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tstmt, err := tx.Prepare(`INSERT OR REPLACE INTO frozen_user (user_id, data)VALUES((SELECT id FROM users WHERE username = ?), ?);`)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\t_, err = stmt.Exec(username, string(data))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (n *SoupNode) Freeze() {}", "func (kb *KubeBackend) Unfreeze() error {\n\tlogrus.Infof(\"set deployment %s replica=1\", kb.ID())\n\tdeployment, err := kb.manager.di.Lister().Deployments(kb.manager.namespace).Get(kb.ID())\n\tif err != nil {\n\t\tif err := kb.manager.syncBackend(kb.ID()); err != nil {\n\t\t\tlogrus.Warnf(\"sycn app with error: %v\", err)\n\t\t}\n\t\treturn err\n\t}\n\tif *deployment.Spec.Replicas != 0 {\n\t\treturn nil\n\t}\n\tvar targetReplica int32 = 1\n\tdeployment.Spec.Replicas = &targetReplica\n\t_, err = kb.manager.client.AppsV1().Deployments(kb.manager.namespace).\n\t\tUpdate(context.Background(), deployment, metav1.UpdateOptions{})\n\tif err != nil {\n\t\treturn err\n\t}\n\tkb.Lock()\n\tkb.stateStarting = true\n\tkb.stateHealthy = false\n\tkb.Unlock()\n\treturn nil\n}", "func (n *metadataNode) freeze(ctx context.Context) {\n\tn.assertNonFrozen()\n\n\t// md may be already non-nil for the root, this is fine.\n\tif n.md == nil {\n\t\tn.md = mergeIntoPrefixMetadata(ctx, n.prefix, n.acls)\n\t}\n\tn.acls = nil // mark as frozen, release unnecessary memory\n\n\tfor _, child := range n.children {\n\t\tchild.freeze(ctx)\n\t}\n}", "func (o *FreezeParams) WithTimeout(timeout time.Duration) *FreezeParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (d *Dam) Lock() {\n\td.freeze.Lock()\n}", "func updateFrozenState(db *IndexerDb, assetID uint64, closedAt *uint64, creator, freeze, holder types.Address) error {\n\t// Semi-blocking migration.\n\t// Hold accountingLock for the duration of the Transaction search + account_asset update.\n\tdb.accountingLock.Lock()\n\tdefer db.accountingLock.Unlock()\n\n\tminRound := uint64(0)\n\tif closedAt != nil {\n\t\tminRound = *closedAt\n\t}\n\n\tholderb64 := encoding.Base64(holder[:])\n\trow := db.db.QueryRow(freezeTransactionsQuery, freeze[:], holderb64, assetID, minRound)\n\tvar found uint64\n\terr := row.Scan(&found)\n\tif err != nil && err != sql.ErrNoRows {\n\t\treturn err\n\t}\n\n\t// If there are any freeze transactions then the default no longer applies.\n\t// Exit early if the asset was frozen\n\tif found != 0 {\n\t\treturn nil\n\t}\n\n\t// If there were no freeze transactions, re-initialize the frozen value.\n\tfrozen := !bytes.Equal(creator[:], holder[:])\n\tdb.db.Exec(`UPDATE account_asset SET frozen = $1 WHERE assetid = $2 and addr = $3`, frozen, assetID, holder[:])\n\n\treturn nil\n}", "func (_OracleMgr *OracleMgrCallerSession) FreezePeriod() (*big.Int, error) {\n\treturn _OracleMgr.Contract.FreezePeriod(&_OracleMgr.CallOpts)\n}", "func (r *RunCtx) Freeze() {\n}", "func (o *FreezeParams) WithContext(ctx context.Context) *FreezeParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (r *AccountDIDRegistry) UnFreeze(did DID) error {\n\texist := r.HasAccountDID(did)\n\tif !exist {\n\t\treturn fmt.Errorf(\"did %s not existed\", did)\n\t}\n\treturn r.auditStatus(did, Normal)\n}", "func (transaction *TokenUpdateTransaction) SetFreezeKey(publicKey Key) *TokenUpdateTransaction {\n\ttransaction._RequireNotFrozen()\n\ttransaction.freezeKey = publicKey\n\treturn transaction\n}", "func (c *TickMap) Freeze() []string {\n\tc.mlock.Lock()\n\tdefer c.mlock.Unlock()\n\ts := make([]string, len(c.m))\n\ti := 0\n\tfor key, _ := range c.m {\n\t\ts[i] = key\n\t\ti++\n\t}\n\treturn s\n}", "func (_OracleMgr *OracleMgrSession) FreezePeriod() (*big.Int, error) {\n\treturn _OracleMgr.Contract.FreezePeriod(&_OracleMgr.CallOpts)\n}", "func (o *FreezeParams) WithDefaults() *FreezeParams {\n\to.SetDefaults()\n\treturn o\n}", "func (r *AccountDIDRegistry) Freeze(did DID) error {\n\texist := r.HasAccountDID(did)\n\tif !exist {\n\t\treturn fmt.Errorf(\"did %s not existed\", did)\n\t}\n\treturn r.auditStatus(did, Frozen)\n}", "func (m *ClusterService) clusterFreeze(ctx context.Context, args struct {\n\tStatus bool\n}) (*proto.GeneralResp, error) {\n\tif _, _, err := permissions(ctx, ADMIN); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := m.cluster.setDisableAutoAllocate(args.Status); err != nil {\n\t\treturn nil, err\n\t}\n\treturn proto.Success(\"success\"), nil\n}", "func (gg GlobGroup) Freeze() {}", "func freezetheworld() {\n\tatomic.Store(&freezing, 1)\n\t// stopwait and preemption requests can be lost\n\t// due to races with concurrently executing threads,\n\t// so try several times\n\tfor i := 0; i < 5; i++ {\n\t\t// this should tell the scheduler to not start any new goroutines\n\t\tsched.stopwait = freezeStopWait\n\t\tatomic.Store(&sched.gcwaiting, 1)\n\t\t// this should stop running goroutines\n\t\tif !preemptall() {\n\t\t\tbreak // no running goroutines\n\t\t}\n\t\tusleep(1000)\n\t}\n\t// to be sure\n\tusleep(1000)\n\tpreemptall()\n\tusleep(1000)\n}", "func (p *BailServiceClient) FreezeBail(dealerId int64, amount float64, orderId int64) (r *Bail, err error) {\n\tif err = p.sendFreezeBail(dealerId, amount, orderId); err != nil {\n\t\treturn\n\t}\n\treturn p.recvFreezeBail()\n}", "func (s *Service) Frozen(ctx context.Context, req *pb.FrozenRequest) (*pb.FrozenResponse, error) {\n\tif !strings.HasPrefix(strings.Title(req.Address), \"Mx\") {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"invalid address\")\n\t}\n\n\tcState := s.blockchain.CurrentState()\n\tcState.RLock()\n\tdefer cState.RUnlock()\n\n\tvar reqCoin *coins.Model\n\n\tif req.CoinId != nil {\n\t\tcoinID := types.CoinID(req.CoinId.GetValue())\n\t\treqCoin = cState.Coins().GetCoin(coinID)\n\t\tif reqCoin == nil {\n\t\t\treturn nil, s.createError(status.New(codes.NotFound, \"Coin not found\"), transaction.EncodeError(code.NewCoinNotExists(\"\", coinID.String())))\n\t\t}\n\t}\n\tvar frozen []*pb.FrozenResponse_Frozen\n\n\tcState.FrozenFunds().GetFrozenFunds(s.blockchain.Height())\n\n\tfor i := s.blockchain.Height(); i <= s.blockchain.Height()+candidates.UnbondPeriod; i++ {\n\n\t\tif timeoutStatus := s.checkTimeout(ctx); timeoutStatus != nil {\n\t\t\treturn nil, timeoutStatus.Err()\n\t\t}\n\n\t\tfunds := cState.FrozenFunds().GetFrozenFunds(i)\n\t\tif funds == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, fund := range funds.List {\n\t\t\tif fund.Address.String() != req.Address {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tcoin := reqCoin\n\t\t\tif coin == nil {\n\t\t\t\tcoin = cState.Coins().GetCoin(fund.Coin)\n\t\t\t} else {\n\t\t\t\tif coin.ID() != fund.Coin {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\tfrozen = append(frozen, &pb.FrozenResponse_Frozen{\n\t\t\t\tHeight: funds.Height(),\n\t\t\t\tAddress: fund.Address.String(),\n\t\t\t\tCandidateKey: fund.CandidateKey.String(),\n\t\t\t\tCoin: &pb.Coin{\n\t\t\t\t\tId: uint64(fund.Coin),\n\t\t\t\t\tSymbol: coin.GetFullSymbol(),\n\t\t\t\t},\n\t\t\t\tValue: fund.Value.String(),\n\t\t\t})\n\t\t}\n\t}\n\n\treturn &pb.FrozenResponse{Frozen: frozen}, nil\n}", "func (px *Paxos) freeMemory() {\n // Assertion: px is already locked by the callee\n\n // reproduction of Min() without requesting a lock\n // Question: Can I do this without duplciating code?\n min := px.done[px.me]\n for i := 0; i < len(px.done); i++ {\n if px.done[i] < min {\n min = px.done[i]\n }\n }\n min += 1\n\n for i, _ := range px.Instances {\n if i < min {\n delete(px.Instances, i)\n }\n }\n}", "func (*NoCopy) Lock() {}", "func (_Token *TokenFilterer) FilterFreeze(opts *bind.FilterOpts, from []common.Address) (*TokenFreezeIterator, error) {\n\n\tvar fromRule []interface{}\n\tfor _, fromItem := range from {\n\t\tfromRule = append(fromRule, fromItem)\n\t}\n\n\tlogs, sub, err := _Token.contract.FilterLogs(opts, \"Freeze\", fromRule)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &TokenFreezeIterator{contract: _Token.contract, event: \"Freeze\", logs: logs, sub: sub}, nil\n}", "func (_Token *TokenFilterer) WatchFreeze(opts *bind.WatchOpts, sink chan<- *TokenFreeze, from []common.Address) (event.Subscription, error) {\n\n\tvar fromRule []interface{}\n\tfor _, fromItem := range from {\n\t\tfromRule = append(fromRule, fromItem)\n\t}\n\n\tlogs, sub, err := _Token.contract.WatchLogs(opts, \"Freeze\", fromRule)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(TokenFreeze)\n\t\t\t\tif err := _Token.contract.UnpackLog(event, \"Freeze\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (s EncryptedChatWaitingArray) Retain(keep func(x EncryptedChatWaiting) bool) EncryptedChatWaitingArray {\n\tn := 0\n\tfor _, x := range s {\n\t\tif keep(x) {\n\t\t\ts[n] = x\n\t\t\tn++\n\t\t}\n\t}\n\ts = s[:n]\n\n\treturn s\n}", "func (p *BailServiceClient) UnfreezeBail(dealerId int64, amount float64, orderId int64) (r *Bail, err error) {\n\tif err = p.sendUnfreezeBail(dealerId, amount, orderId); err != nil {\n\t\treturn\n\t}\n\treturn p.recvUnfreezeBail()\n}", "func (stackEntry *valuePayloadPropagationStackEntry) Retain() *valuePayloadPropagationStackEntry {\n\treturn &valuePayloadPropagationStackEntry{\n\t\tCachedPayload: stackEntry.CachedPayload.Retain(),\n\t\tCachedPayloadMetadata: stackEntry.CachedPayloadMetadata.Retain(),\n\t\tCachedTransaction: stackEntry.CachedTransaction.Retain(),\n\t\tCachedTransactionMetadata: stackEntry.CachedTransactionMetadata.Retain(),\n\t}\n}", "func (transaction *TokenUpdateTransaction) GetFreezeKey() Key {\n\treturn transaction.freezeKey\n}", "func (outer outer) Free() bool {\r\n\treturn false\r\n}", "func (o *ParamsReg) Backup() {\n\tcopy(o.bkpTheta, o.theta)\n\to.bkpBias = o.bias\n\to.bkpLambda = o.lambda\n\to.bkpDegree = o.degree\n}", "func (ch *ClickHouse) FreezeTable(table Table) error {\n\tvar partitions []struct {\n\t\tPartitionID string `db:\"partition_id\"`\n\t}\n\tq := fmt.Sprintf(\"SELECT DISTINCT partition_id FROM system.parts WHERE database='%v' AND table='%v'\", table.Database, table.Name)\n\tif err := ch.conn.Select(&partitions, q); err != nil {\n\t\treturn fmt.Errorf(\"can't get partitions for \\\"%s.%s\\\" with %v\", table.Database, table.Name, err)\n\t}\n\tlog.Printf(\"Freeze '%v.%v'\", table.Database, table.Name)\n\tfor _, item := range partitions {\n\t\tif ch.DryRun {\n\t\t\tlog.Printf(\" partition '%v' ...skip becouse dry-run\", item.PartitionID)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Printf(\" partition '%v'\", item.PartitionID)\n\t\t// TODO: find out this magic\n\t\tif item.PartitionID == \"all\" {\n\t\t\titem.PartitionID = \"tuple()\"\n\t\t}\n\t\tif _, err := ch.conn.Exec(\n\t\t\tfmt.Sprintf(\n\t\t\t\t\"ALTER TABLE %v.%v FREEZE PARTITION %v;\",\n\t\t\t\ttable.Database,\n\t\t\t\ttable.Name,\n\t\t\t\titem.PartitionID,\n\t\t\t)); err == nil {\n\t\t\tcontinue\n\t\t}\n\t\tif _, err := ch.conn.Exec(\n\t\t\tfmt.Sprintf(\n\t\t\t\t\"ALTER TABLE %v.%v FREEZE PARTITION '%v';\",\n\t\t\t\ttable.Database,\n\t\t\t\ttable.Name,\n\t\t\t\titem.PartitionID,\n\t\t\t)); err != nil {\n\t\t\treturn fmt.Errorf(\"can't freze partiotion '%s' on '%s.%s' with: %v\", item.PartitionID, table.Database, table.Name, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (*noCopy) Lock() {}", "func Free() {\n\tflags = nil // Any future call to Get() will panic on a nil dereference.\n}", "func (x *Value) Free() {\n\tif x != nil && x.allocs23e8c9e3 != nil {\n\t\tx.allocs23e8c9e3.(*cgoAllocMap).Free()\n\t\tx.ref23e8c9e3 = nil\n\t}\n}", "func (f Fill) SetValue(value float64) Fill {\n\tf.value = value\n\treturn f\n}", "func (tb *tensorBase) Retain() {\n\tatomic.AddInt64(&tb.refCount, 1)\n}", "func SyncRuntimeDoSpin()", "func (_Storage *StorageCaller) AccountFrozen(opts *bind.CallOpts, addr common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Storage.contract.Call(opts, out, \"accountFrozen\", addr)\n\treturn *ret0, err\n}", "func (r RawValues) Retain(values ...string) RawValues {\n\ttoretain := make(map[string]bool)\n\tfor _, v := range values {\n\t\ttoretain[v] = true\n\t}\n\tfiltered := make([]RawValue, 0)\n\tfor _, rawValue := range r {\n\t\tif _, ok := toretain[rawValue.Value]; ok {\n\t\t\tfiltered = append(filtered, rawValue)\n\t\t}\n\t}\n\treturn filtered\n}", "func (v *VolumesServiceMock) Freeze(podUID string, name string) (vol *api.Volume, err error) {\n\targs := v.Called(podUID, name)\n\tx := args.Get(0)\n\tif x != nil {\n\t\tvol = x.(*api.Volume)\n\t}\n\terr = args.Error(1)\n\treturn\n}", "func FixFreezeLookupMigration(db *IndexerDb, state *MigrationState) error {\n\t// Technically with this query no transactions are needed, and the accounting state doesn't need to be locked.\n\tupdateQuery := \"INSERT INTO txn_participation (addr, round, intra) VALUES ($1, $2, $3) ON CONFLICT DO NOTHING\"\n\tquery := fmt.Sprintf(\"select decode(txn.txn->'txn'->>'fadd','base64'),round,intra from txn where typeenum = %d AND txn.txn->'txn'->'snd' != txn.txn->'txn'->'fadd'\", idb.TypeEnumAssetFreeze)\n\trows, err := db.db.Query(query)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to query transactions: %v\", err)\n\t}\n\tdefer rows.Close()\n\n\ttxprows := make([][]interface{}, 0)\n\n\t// Loop through all transactions and compute account data.\n\tdb.log.Print(\"loop through all freeze transactions\")\n\tfor rows.Next() {\n\t\tvar addr []byte\n\t\tvar round, intra uint64\n\t\terr = rows.Scan(&addr, &round, &intra)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error scanning row: %v\", err)\n\t\t}\n\n\t\ttxprows = append(txprows, []interface{}{addr, round, intra})\n\n\t\tif len(txprows) > 5000 {\n\t\t\terr = updateBatch(db, updateQuery, txprows)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"updating batch: %v\", err)\n\t\t\t}\n\t\t\ttxprows = txprows[:0]\n\t\t}\n\t}\n\n\tif rows.Err() != nil {\n\t\treturn fmt.Errorf(\"error while processing freeze transactions: %v\", rows.Err())\n\t}\n\n\t// Commit any leftovers\n\tif len(txprows) > 0 {\n\t\terr = updateBatch(db, updateQuery, txprows)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"updating batch: %v\", err)\n\t\t}\n\t}\n\n\t// Update migration state\n\treturn upsertMigrationState(db, state, true)\n}", "func (stateObj *stateObject) finalise(prefetch bool) {\n\tlog.Debugf(\"stateObject finalise. address:%x, prefetch:%v\", stateObj.address, prefetch)\n\n\tslotsToPrefetch := make([][]byte, 0, len(stateObj.dirtyStorage))\n\tfor key, value := range stateObj.dirtyStorage {\n\t\tstateObj.pendingStorage[key] = value\n\t\tif value != stateObj.originStorage[key] {\n\t\t\tslotsToPrefetch = append(slotsToPrefetch, common.CopyBytes(key[:])) // Copy needed for closure\n\t\t}\n\t}\n\tif stateObj.db.prefetcher != nil && prefetch && len(slotsToPrefetch) > 0 && stateObj.data.Root != emptyRoot {\n\t\tstateObj.db.prefetcher.prefetch(stateObj.addrHash, stateObj.data.Root, slotsToPrefetch)\n\t}\n\tif len(stateObj.dirtyStorage) > 0 {\n\t\tstateObj.dirtyStorage = make(Storage)\n\t}\n}", "func UnfreezeClock(t *testing.T) {\n\tif t == nil {\n\t\tpanic(\"nice try\")\n\t}\n\tc = &DefaultClock{}\n}", "func (this *Tidy) Bare(val bool) (bool, error) {\n\treturn this.optSetBool(C.TidyMakeBare, cBool(val))\n}", "func (o *FreezeParams) WithHTTPClient(client *http.Client) *FreezeParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (m *MemoryStateDB) ExecFrozen(tx *types.Transaction, addr string, amount int64) bool {\n\tif nil == tx {\n\t\tlog15.Error(\"ExecFrozen get nil tx\")\n\t\treturn false\n\t}\n\n\texecaddr := address.ExecAddress(string(tx.Execer))\n\tret, err := m.CoinsAccount.ExecFrozen(addr, execaddr, amount)\n\tif err != nil {\n\t\tlog15.Error(\"ExecFrozen error\", \"addr\", addr, \"execaddr\", execaddr, \"amount\", amount, \"err info\", err)\n\t\treturn false\n\t}\n\n\tm.addChange(balanceChange{\n\t\tamount: amount,\n\t\tdata: ret.KV,\n\t\tlogs: ret.Logs,\n\t})\n\n\treturn true\n}", "func (m *neighborEntryRWMutex) RLockBypass() {\n\tm.mu.RLock()\n}", "func (vals PendingValues) lockSet(inst, prop ident.Id, nilVal, want interface{}) (err error) {\n\tif reflect.TypeOf(nilVal) != reflect.TypeOf(want) {\n\t\terr = SetValueMismatch(inst, prop, nilVal, want)\n\t} else if curr, have := vals[prop]; have && curr != want {\n\t\terr = SetValueChanged(inst, prop, curr, want)\n\t} else {\n\t\tvals[prop] = want\n\t}\n\treturn err\n}", "func (s IPPortSecretArray) Retain(keep func(x IPPortSecret) bool) IPPortSecretArray {\n\tn := 0\n\tfor _, x := range s {\n\t\tif keep(x) {\n\t\t\ts[n] = x\n\t\t\tn++\n\t\t}\n\t}\n\ts = s[:n]\n\n\treturn s\n}", "func (f *IndexFile) Retain() { f.wg.Add(1) }", "func (_Token *TokenTransactor) FreezeTokens(opts *bind.TransactOpts, account common.Address, amount *big.Int) (*types.Transaction, error) {\n\treturn _Token.contract.Transact(opts, \"freezeTokens\", account, amount)\n}", "func (n *metadataNode) assertFrozen() {\n\tif n.acls != nil {\n\t\tpanic(\"not frozen yet\")\n\t}\n}", "func (b *Boolean) Reset() {\n\tb.Value = false\n\tb.Default = false\n\tb.Initialized = false\n\tBooleanPool.Put(b)\n}", "func (tester* FreezeTester) nonBpVote(t *testing.T, d *Dandelion) {\n\ta := assert.New(t)\n\ta.True(d.Contract(constants.COSSysAccount, frCrtName).CheckExist())\n\tfreezeAcct := tester.acc5\n\tsta := freezeAcct.GetFreeze()\n\tmemo := freezeAcct.GetFreezeMemo()\n\tnewSta := tester.mdFreezeStatus(sta)\n\ta.NotEqual(sta, newSta)\n\tmemoArray,nameArray := tester.getProposalMemoAndNameParams(d,[]*DandelionAccount{freezeAcct})\n\n\t//1.proposal\n\tApplyNoError(t, d, fmt.Sprintf(\"%s: %s.%s.proposalfreeze %s,%d,%s\", tester.acc0.Name, constants.COSSysAccount, frCrtName, nameArray, newSta, memoArray))\n\t//2.fetch proposal_id\n\tpropId,err := tester.getProposalId(d)\n\ta.NoError(err)\n\t//less than 2/3 bp vote to proposalId\n\ttester.voteById(t, d, propId, 0, tester.threshold-1)\n\t//non bp vote\n\tApplyError(t, d, fmt.Sprintf(\"%s: %s.%s.vote %v\", tester.acc4.Name, constants.COSSysAccount, frCrtName, propId))\n\t//final vote fail, set freeze fail\n\ta.Equal(sta, freezeAcct.GetFreeze())\n\ta.Equal(memo, freezeAcct.GetFreezeMemo())\n\n}", "func (s *Send) FreeVars() []Name {\n\tfv := []Name{}\n\tfor _, v := range s.Vals {\n\t\tfv = append(fv, v)\n\t}\n\tsort.Sort(byName(fv))\n\treturn RemDup(fv)\n}", "func (l *FixedLimiter) Reset() {\n\tl.mu.Lock()\n\tdefer l.mu.Unlock()\n\tl.value = 0\n}", "func (v Chunk) Retain() {\n\tv.buf.Retain()\n}", "func (p *PKGBUILD) RecomputeValues() {\n\tp.info.RecomputeValues()\n}", "func (m *neighborEntryRWMutex) RUnlockBypass() {\n\tm.mu.RUnlock()\n}", "func (this *FeedableBuffer) Minimize() {\n\tthis.Data = this.Data[:this.minByteCount]\n}", "func (lv *LazyValue) Reset() {\n\tlv.Lock()\n\tdefer lv.Unlock()\n\n\tlv.ready = false\n\tlv.value = values.None\n\tlv.err = nil\n}", "func (s *slotted) UnReserve() {\n\tif atomic.AddInt32((*int32)(s), -1) < 0 {\n\t\tatomic.StoreInt32((*int32)(s), 0)\n\t}\n}", "func (tl *TimeLockCondition) Fulfill(fulfillment UnlockFulfillment, ctx FulfillContext) error {\n\tif !tl.Fulfillable(FulfillableContext{BlockHeight: ctx.BlockHeight, BlockTime: ctx.BlockTime}) {\n\t\treturn errors.New(\"time lock has not yet been reached\")\n\t}\n\n\t// time lock hash been reached,\n\t// delegate the actual fulfillment to the given fulfillment, if supported\n\tswitch tf := fulfillment.(type) {\n\tcase *SingleSignatureFulfillment:\n\t\treturn tl.Condition.Fulfill(tf, ctx)\n\tcase *MultiSignatureFulfillment:\n\t\treturn tl.Condition.Fulfill(tf, ctx)\n\tdefault:\n\t\treturn ErrUnexpectedUnlockFulfillment\n\t}\n}", "func (f *Flag) Set() { atomic.CompareAndSwapUint32((*uint32)(unsafe.Pointer(f)), 0, 1) }", "func DeferLiveness() {\n\tvar x [10]int\n\tescape(&x)\n\tfn := func() {\n\t\tif x[0] != 42 {\n\t\t\tpanic(\"FAIL\")\n\t\t}\n\t}\n\tdefer fn()\n\n\tx[0] = 42\n\truntime.GC()\n\truntime.GC()\n\truntime.GC()\n}", "func (this *Hash) Shrink() {\n\tif this == nil {\n\t\treturn\n\t}\n\n\tif this.lock {\n\t\tthis.mu.Lock()\n\t\tdefer this.mu.Unlock()\n\t}\n\n\tthis.loose.shrink()\n\tthis.compact.shrink(this.loose.a)\n}", "func (cpu *CPU) writeHalfcarryFlag(val bool) {\n if val {\n cpu.f = cpu.f ^ ( 1 << 5 )\n }\n}", "func (n *Number) Reset() {\n\tn.Value = 0.0\n\tn.Initialized = false\n\tNumberPool.Put(n)\n}", "func (x *FzStrokeState) Free() {\n\tif x != nil {\n\t\tC.free(unsafe.Pointer(x))\n\t}\n}", "func (b *LeakyLimiter) Reset() time.Time {\n\tb.remaining = b.capacity\n\treturn b.reset\n}", "func (c *CycleState) Unlock() {\n\tc.mx.Unlock()\n}", "func Fill(value bool) *SimpleElement { return newSEBool(\"fill\", value) }", "func (c *Container) frozen() bool {\n\tif c == nil {\n\t\treturn true\n\t}\n\treturn (c.flags & flagFrozen) != 0\n}", "func (b *Buffer) Retain() {\n\tif b.mem != nil || b.parent != nil {\n\t\tatomic.AddInt64(&b.refCount, 1)\n\t}\n}", "func (m *Mutex) ForceRealse() {\n\tatomic.StoreUint32(&m.l, 0)\n}", "func Relax(out1 *LooseFieldElement, arg1 *TightFieldElement) {\n\tx1 := arg1[0]\n\tx2 := arg1[1]\n\tx3 := arg1[2]\n\tx4 := arg1[3]\n\tx5 := arg1[4]\n\tout1[0] = x1\n\tout1[1] = x2\n\tout1[2] = x3\n\tout1[3] = x4\n\tout1[4] = x5\n}" ]
[ "0.6292724", "0.6172568", "0.6090478", "0.60317403", "0.5963623", "0.58790493", "0.5820719", "0.57209927", "0.5712187", "0.56367314", "0.5491579", "0.5445859", "0.5395169", "0.53697157", "0.53642315", "0.5219047", "0.5203252", "0.52030545", "0.51646525", "0.5120046", "0.5095717", "0.5072036", "0.4992856", "0.49605715", "0.49496153", "0.49485943", "0.49303678", "0.4914931", "0.4905752", "0.49006045", "0.48866832", "0.48568454", "0.4823699", "0.48183236", "0.47987852", "0.47315022", "0.45978817", "0.45940936", "0.45935822", "0.4573915", "0.45477486", "0.45260292", "0.45235398", "0.4519089", "0.45092362", "0.45078444", "0.4478194", "0.44747028", "0.446194", "0.4442175", "0.44410563", "0.44038707", "0.438813", "0.43813273", "0.436612", "0.43581527", "0.43569055", "0.43386182", "0.4337029", "0.4318477", "0.43111184", "0.4294279", "0.42941493", "0.42703456", "0.4265715", "0.42557698", "0.4236136", "0.42307094", "0.4226536", "0.4221072", "0.42147005", "0.42088905", "0.42042136", "0.4176763", "0.4174508", "0.41704193", "0.4166596", "0.41631296", "0.41629037", "0.41622674", "0.41480255", "0.41469267", "0.41423237", "0.41345876", "0.41340083", "0.41299763", "0.4117786", "0.41175222", "0.41172984", "0.4116206", "0.41149288", "0.41127932", "0.41100523", "0.41078106", "0.40977663", "0.40956038", "0.4091769", "0.4089258", "0.4089158", "0.40816256" ]
0.5367814
14
Truth implements the starlark.Value.Truth() method.
func (p Path) Truth() starlark.Bool { return starlark.True }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func truth(r reflect.Value) bool {\nout:\n\tswitch r.Kind() {\n\tcase reflect.Array, reflect.Slice:\n\t\treturn r.Len() > 0\n\tcase reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\t\treturn r.Int() > 0\n\tcase reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64:\n\t\treturn r.Uint() > 0\n\tcase reflect.Float32, reflect.Float64:\n\t\treturn r.Float() > 0\n\tcase reflect.String:\n\t\treturn r.String() != \"\"\n\tcase reflect.Bool:\n\t\treturn r.Bool()\n\tcase reflect.Ptr, reflect.Interface:\n\t\tr = r.Elem()\n\t\tgoto out\n\tdefault:\n\t\treturn r.Interface() != nil\n\t}\n}", "func True() TermT {\n\treturn TermT(C.yices_true())\n}", "func True(actual interface{}) Truth {\n\tmustBeCleanStart()\n\treturn Truth{actual.(bool), fmt.Sprintf(\"%#v\", actual)}\n}", "func (v *Value) Bool() bool {\n return Util.ToBool(v.data)\n}", "func (t *Target) Truth() starlark.Bool { return starlark.Bool(t == nil) }", "func (m *kubePackage) Truth() starlark.Bool { return starlark.True }", "func True(t TestingT, v interface{}, extras ...interface{}) bool {\n\tval, ok := v.(bool)\n\tif !ok || val != true {\n\t\texps, acts := toString(true, v)\n\n\t\treturn Errorf(t, \"Expect to be true\", []labeledOutput{\n\t\t\t{\n\t\t\t\tlabel: labelMessages,\n\t\t\t\tcontent: formatExtras(extras...),\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"-expected\",\n\t\t\t\tcontent: exps,\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"+received\",\n\t\t\t\tcontent: acts,\n\t\t\t},\n\t\t})\n\t}\n\n\treturn true\n}", "func (n *Node) Bool() bool", "func (v Value) Bool() (bool, error) {\n\tif v.typ != Bool {\n\t\treturn false, v.newError(\"%s is not a bool\", v.Raw())\n\t}\n\treturn v.boo, nil\n}", "func (m *Value) Bool() bool { return m.BoolMock() }", "func (v Value) Bool() bool {\n\treturn v.Integer() != 0\n}", "func True(v bool) {\n\tassert(v, 1, \"assert true\")\n}", "func Bool(x bool) bool { return x }", "func (sp booleanSpace) BoolOutcome(value bool) Outcome {\n\tif value {\n\t\treturn 1\n\t} else {\n\t\treturn 0\n\t}\n}", "func (i *Instance) Truth() exprcore.Bool {\n\treturn exprcore.True\n}", "func (sp booleanSpace) BoolValue(outcome Outcome) bool {\n\treturn outcome != 0\n}", "func (v Value) Truthy() bool {\n\tswitch v.Type() {\n\tcase TypeUndefined, TypeNull:\n\t\treturn false\n\tcase TypeBoolean:\n\t\treturn v.Bool()\n\tcase TypeNumber:\n\t\treturn !v.IsNaN() && v.Float() != 0\n\tcase TypeString:\n\t\treturn v.String() != \"\"\n\tcase TypeSymbol, TypeFunction, TypeObject:\n\t\treturn true\n\tdefault:\n\t\tpanic(\"bad type\")\n\t}\n}", "func (r Record) Bool(key string) bool {\n\tif r.dropped {\n\t\tlog.Fatalf(\"Int called on dropped record\")\n\t}\n\n\tval, ok := r.values[key]\n\tif !ok {\n\t\treturn false\n\t}\n\n\tswitch val := val.(type) {\n\tcase nil:\n\t\treturn false\n\tcase string:\n\t\treturn val == \"true\"\n\tcase int64:\n\t\treturn val != 0\n\tcase float64:\n\t\treturn val != 0.0\n\tcase bool:\n\t\treturn val\n\t}\n\n\treturn false\n}", "func (o BoolOperand) Evaluate(cxt interface{}) (bool, error) {\n\treturn o.Value, nil\n}", "func (o Nil) Truthy() bool { return false }", "func (v Value) Truthy() bool {\n\tpanic(message)\n}", "func (tr Row) Bool(nn int) (val bool) {\n\tval, err := tr.BoolErr(nn)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn\n}", "func Bool(a bool) cell.I {\n\tif a {\n\t\treturn sym.True\n\t}\n\n\treturn pair.Null\n}", "func (b *Bool) Literal() {}", "func (tv *TypedBool) Bool() bool {\n\treturn tv.Bytes[0] == 1\n}", "func (v Value) Bool() bool {\n\treturn v.v.Bool()\n}", "func (b *Boolean) True() *Boolean {\n\treturn b.Equal(true)\n}", "func (BooleanLiteral) literalNode() {}", "func (r *TTNRandom) Bool() bool {\n\treturn r.Interface.Intn(2) == 0\n}", "func True(Right) bool {\n\treturn true\n}", "func (n Nil) True() bool { return false }", "func (v *Value) Bool() (bool, error) {\n\tif v.kind == kindBool {\n\t\treturn v.boolContent, nil\n\t}\n\treturn false, nil\n}", "func IsTruthy(val string) bool {\n\tswitch strings.ToLower(strings.TrimSpace(val)) {\n\tcase \"y\", \"yes\", \"true\", \"t\", \"on\", \"1\":\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func Truthy(t *testing.T, name string, v bool) {\n\tif !v {\n\t\tflux.FatalFailed(t, \"Expected truthy value for %s\", name)\n\t} else {\n\t\tflux.LogPassed(t, \"%s passed with truthy value\", name)\n\t}\n}", "func Bool(r interface{}, err error) (bool, error) {\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tswitch r := r.(type) {\n\tcase bool:\n\t\treturn r, err\n\t// Very common in redis to reply int64 with 0 for bool flag.\n\tcase int:\n\t\treturn r != 0, nil\n\tcase int64:\n\t\treturn r != 0, nil\n\tcase []byte:\n\t\treturn strconv.ParseBool(string(r))\n\tcase string:\n\t\treturn strconv.ParseBool(r)\n\tcase nil:\n\t\treturn false, simplesessions.ErrNil\n\t}\n\treturn false, simplesessions.ErrAssertType\n}", "func BoolVal(x Value) bool {\n\treturn constant.BoolVal(x)\n}", "func (o *FakeObject) Bool() bool { return o.Value.(bool) }", "func False() TermT {\n\treturn TermT(C.yices_false())\n}", "func True(t Testing, v interface{}, formatAndArgs ...interface{}) bool {\n\tvar tv bool\n\tswitch v.(type) {\n\tcase bool:\n\t\ttv = v.(bool)\n\t}\n\n\tif tv != true {\n\t\treturn Fail(t, pretty.Sprintf(\"Expected %# v to be true\", v), formatAndArgs...)\n\t}\n\n\treturn true\n}", "func BoolConstValue(t TermT, val *int32) int32 {\n\treturn int32(C.yices_bool_const_value(C.term_t(t), (*C.int32_t)(val)))\n}", "func (s *Smpval) Bool() bool {\n\treturn s.b\n}", "func Bool(v *Value, def bool) bool {\n\tb, err := v.Bool()\n\tif err != nil {\n\t\treturn def\n\t}\n\treturn b\n}", "func (n BoolWrapper) Value() (Value, error) {\n\tif !n.Valid {\n\t\treturn nil, nil\n\t}\n\treturn n.Bool, nil\n}", "func (s *Sub) Truth() starlark.Bool { return s == nil }", "func wrapWithIsTrue(ctx sessionctx.Context, keepNull bool, arg Expression) (Expression, error) {\n\tif arg.GetType().EvalType() == types.ETInt {\n\t\treturn arg, nil\n\t}\n\tfc := &isTrueOrFalseFunctionClass{baseFunctionClass{ast.IsTruth, 1, 1}, opcode.IsTruth, keepNull}\n\tf, err := fc.getFunction(ctx, []Expression{arg})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsf := &ScalarFunction{\n\t\tFuncName: model.NewCIStr(ast.IsTruth),\n\t\tFunction: f,\n\t\tRetType: f.getRetTp(),\n\t}\n\treturn FoldConstant(sf), nil\n}", "func True(t testing.TB, value bool, msgAndArgs ...interface{}) bool {\n\tif !value {\n\t\treturn failTest(t, 1, fmt.Sprintf(\"True: expected `true`, actual `%#v`\", value), msgAndArgs...)\n\t}\n\n\treturn true\n}", "func (e *Encoder) Bool(v bool) (int, error) {\n\tif v {\n\t\treturn e.Byte(0x1)\n\t}\n\treturn e.Byte(0x0)\n}", "func Bool(v bool) (p *bool) { return &v }", "func (n *NotLikeOp) IsTrue(left, right EvalResult) (bool, error) {\n\treturn false, nil\n}", "func (v Bool) Bool() bool {\n\treturn v.v\n}", "func BoolVal(b *bool) bool {\n\tif b == nil {\n\t\treturn false\n\t}\n\treturn *b\n}", "func Bool(val interface{}) (bool, error) {\n\tif val == nil {\n\t\treturn false, nil\n\t}\n\tswitch ret := val.(type) {\n\tcase bool:\n\t\treturn ret, nil\n\tcase int, int8, int16, int32, int64, float32, float64, uint, uint8, uint16, uint32, uint64:\n\t\treturn ret != 0, nil\n\tcase []byte:\n\t\treturn stringToBool(string(ret))\n\tcase string:\n\t\treturn stringToBool(ret)\n\tdefault:\n\t\treturn false, converError(val, \"bool\")\n\t}\n}", "func (b *Bool) Value() bool {\n\t// generate nil checks and faults early.\n\tref := &b.i\n\treturn atomic.LoadUint64(ref) == 1\n}", "func TestNewResult_bool(t *testing.T) {\n\tvar reading interface{} = true\n\treq := models.CommandRequest{\n\t\tDeviceResourceName: \"light\",\n\t\tType: common.ValueTypeBool,\n\t}\n\n\tcmdVal, err := newResult(req, reading)\n\tif err != nil {\n\t\tt.Fatalf(\"Fail to create new ReadCommand result, %v\", err)\n\t}\n\tval, err := cmdVal.BoolValue()\n\tif val != true || err != nil {\n\t\tt.Errorf(\"Convert new result(%v) failed, error: %v\", val, err)\n\t}\n}", "func IsTruthy(s string) bool {\n\treturn s == \"1\" || strings.EqualFold(s, \"true\")\n}", "func (n *SoupNode) Truth() starlark.Bool {\n\treturn n != nil\n}", "func (v *Value) Bool() bool {\n\treturn (bool)(C.value_get_bool(v.value))\n}", "func (h *Random) Bool() bool {\n\tboolList := []bool{true, false}\n\trandomIndex := rand.Intn(len(boolList))\n\treturn boolList[randomIndex]\n}", "func TestCheckBinaryExprBoolRhlBool(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `true >> true`, env,\n\t\t`invalid operation: true >> true (shift count type bool, must be unsigned integer)`,\n\t)\n\n}", "func Bool(value interface{}) bool {\n\tret, err := BoolE(value)\n\tif err != nil {\n\t\tmod.Error(err)\n\t}\n\treturn ret\n}", "func (nvp *NameValues) Bool(name string) (bool, bool) {\n\tvalue, _ := nvp.String(name)\n\treturn (value == \"true\" || value == \"yes\" || value == \"1\" || value == \"-1\" || value == \"on\"), true\n}", "func Boolean() Scalar {\n\treturn booleanTypeInstance\n}", "func TestCheckBinaryExprBoolNeqBool(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectConst(t, `true != true`, env, (true != true), ConstBool)\n}", "func (v Value) Bool() bool {\n\tpanic(message)\n}", "func (l *LikeOp) IsTrue(left, right EvalResult) (bool, error) {\n\treturn false, nil\n}", "func TestFetchTrueBool(t *testing.T) {\n\texpected := \"true\"\n\tinput := \"rue\"\n\treader := bytes.NewReader([]byte(input))\n\tlex := NewLexer(reader)\n\tif err := lex.fetchTrueBool(); err != nil {\n\t\tt.Error(err.Error())\n\t\treturn\n\t}\n\n\tif len(lex.tokens) != 1 {\n\t\tt.Error(\"expecting 1 token to be fetched\")\n\t\treturn\n\t}\n\n\ttoken := lex.tokens[0]\n\tif token.t != TokenBool {\n\t\tt.Errorf(\"unexpected token type %d (%s), expecting token type %d (%s)\", token.t, tokenTypeMap[token.t], TokenBool, tokenTypeMap[TokenBool])\n\t\treturn\n\t}\n\n\tif token.String() != expected {\n\t\tt.Errorf(\"unexpected %s, expecting %s\", token.String(), expected)\n\t}\n}", "func isTrue(val reflect.Value) (truth, ok bool) {\n\tif !val.IsValid() {\n\t\t// Something like var x interface{}, never set. It's a form of nil.\n\t\treturn false, true\n\t}\n\tswitch val.Kind() {\n\tcase reflect.Array, reflect.Map, reflect.Slice, reflect.String:\n\t\ttruth = val.Len() > 0\n\tcase reflect.Bool:\n\t\ttruth = val.Bool()\n\tcase reflect.Complex64, reflect.Complex128:\n\t\ttruth = val.Complex() != 0\n\tcase reflect.Chan, reflect.Func, reflect.Ptr, reflect.Interface:\n\t\ttruth = !val.IsNil()\n\tcase reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\t\ttruth = val.Int() != 0\n\tcase reflect.Float32, reflect.Float64:\n\t\ttruth = val.Float() != 0\n\tcase reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:\n\t\ttruth = val.Uint() != 0\n\tcase reflect.Struct:\n\t\ttruth = true // Struct values are always true.\n\tdefault:\n\t\treturn\n\t}\n\treturn truth, true\n}", "func (sr *StringResult) Bool() (bool, error) {\n\treturn redis.Bool(sr.val, nil)\n}", "func (r *Redis) Bool(reply interface{}, err error) (bool, error) {\n\treturn redigo.Bool(reply, err)\n}", "func TestBoolVal(t *testing.T) {\n\tConvey(\"Testing BoolVal\", t, func() {\n\t\ttrueValues := []string{\"yes\", \"ok\", \"true\", \"1\", \"enabled\", \"True\", \"TRUE\", \"YES\", \"Yes\"}\n\t\tfalseValues := []string{\"no\", \"0\", \"false\", \"False\", \"FALSE\", \"disabled\"}\n\t\tfor _, val := range trueValues {\n\t\t\tSo(BoolVal(val), ShouldBeTrue)\n\t\t}\n\t\tfor _, val := range falseValues {\n\t\t\tSo(BoolVal(val), ShouldBeFalse)\n\t\t}\n\t})\n}", "func TestCheckBinaryExprBoolEqlBool(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectConst(t, `true == true`, env, (true == true), ConstBool)\n}", "func TestCheckBinaryExprBoolGtrBool(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `true > true`, env,\n\t\t`invalid operation: true > true (operator > not defined on bool)`,\n\t)\n\n}", "func (opa *OPA) Bool(ctx context.Context, input interface{}, opts ...func(*rego.Rego)) (bool, error) {\n\n\tm := metrics.New()\n\tvar decisionID string\n\tvar revision string\n\tvar decision bool\n\n\terr := storage.Txn(ctx, opa.manager.Store, storage.TransactionParams{}, func(txn storage.Transaction) error {\n\n\t\tvar err error\n\n\t\trevision, err = getRevision(ctx, opa.manager.Store, txn)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdecisionID, err = uuid4()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\topts = append(opts,\n\t\t\trego.Metrics(m),\n\t\t\trego.Query(defaultDecision),\n\t\t\trego.Input(input),\n\t\t\trego.Compiler(opa.manager.GetCompiler()),\n\t\t\trego.Store(opa.manager.Store),\n\t\t\trego.Transaction(txn))\n\n\t\trs, err := rego.New(opts...).Eval(ctx)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t} else if len(rs) == 0 {\n\t\t\treturn fmt.Errorf(\"undefined decision\")\n\t\t} else if b, ok := rs[0].Expressions[0].Value.(bool); !ok || len(rs) > 1 {\n\t\t\treturn fmt.Errorf(\"non-boolean decision\")\n\t\t} else {\n\t\t\tdecision = b\n\t\t}\n\n\t\treturn nil\n\t})\n\n\tif opa.decisionLogsPlugin != nil {\n\t\trecord := &server.Info{\n\t\t\tRevision: revision,\n\t\t\tDecisionID: decisionID,\n\t\t\tTimestamp: time.Now(),\n\t\t\tQuery: defaultDecision,\n\t\t\tInput: input,\n\t\t\tError: err,\n\t\t\tMetrics: m,\n\t\t}\n\t\tif err == nil {\n\t\t\tvar x interface{} = decision\n\t\t\trecord.Results = &x\n\t\t}\n\t\topa.decisionLogsPlugin.Log(ctx, record)\n\t}\n\n\treturn decision, err\n}", "func False(v bool) {\n\tassert(!v, 1, \"assert false\")\n}", "func isTruthy(o object.Object) bool {\n\tswitch o {\n\tcase ConstFalse, ConstNil:\n\t\treturn false\n\tdefault:\n\t\t// special case: 0 or 0.0 is not truthy\n\t\tswitch o.Type() {\n\t\tcase object.IntType:\n\t\t\tif o.(*object.Integer).Value == 0 {\n\t\t\t\treturn false\n\t\t\t}\n\t\tcase object.FloatType:\n\t\t\tif o.(*object.Float).Value == 0.0 {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\n\t\treturn true\n\t}\n}", "func isBoolean(t Type) bool { return isBasic(t, IsBoolean) }", "func Bool(a bool, b bool) bool {\n\treturn a == b\n}", "func (r *Decoder) Bool() bool {\n\tr.Sync(SyncBool)\n\tx, err := r.Data.ReadByte()\n\tr.checkErr(err)\n\tassert(x < 2)\n\treturn x != 0\n}", "func Bool(v *bool) bool {\n\tif v != nil {\n\t\treturn *v\n\t}\n\treturn false\n}", "func (this *Not) Type() value.Type { return value.BOOLEAN }", "func False(t TestingT, v interface{}, extras ...interface{}) bool {\n\tval, ok := v.(bool)\n\tif !ok || val != false {\n\t\texps, acts := toString(false, v)\n\n\t\treturn Errorf(t, \"Expect to be false\", []labeledOutput{\n\t\t\t{\n\t\t\t\tlabel: labelMessages,\n\t\t\t\tcontent: formatExtras(extras...),\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"-expected\",\n\t\t\t\tcontent: exps,\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"+received\",\n\t\t\t\tcontent: acts,\n\t\t\t},\n\t\t})\n\t}\n\n\treturn true\n}", "func (data *Data) Bool(s ...string) bool {\n\treturn data.Interface(s...).(bool)\n}", "func (n *eeNum) bool() *bool { return (*bool)(unsafe.Pointer(&n.data)) }", "func (*Base) LiteralBoolean(p ASTPass, node *ast.LiteralBoolean, ctx Context) {\n}", "func (v Boolean) Bool() bool {\n\treturn v.v\n}", "func (b *Boolean) Raw() bool {\n\treturn b.value\n}", "func Bool(b bool) int {\n\tif b {\n\t\treturn 1\n\t}\n\treturn 0\n}", "func True(t testing.TB, ok bool, msgAndArgs ...interface{}) {\n\tif ok {\n\t\treturn\n\t}\n\tt.Helper()\n\tt.Fatal(formatMsgAndArgs(\"Expected expression to be true\", msgAndArgs...))\n}", "func BoolValue(t bool) Value {\n\tif t {\n\t\treturn Value{Typ: ':', IntegerV: 1}\n\t}\n\treturn Value{Typ: ':', IntegerV: 0}\n}", "func TestGetBooleanTrue(t *testing.T) {\n\tclient := newQueriesClient(t)\n\tresult, err := client.GetBooleanTrue(context.Background(), nil)\n\trequire.NoError(t, err)\n\trequire.Zero(t, result)\n}", "func Bool(b bool) Cell {\n\tif b {\n\t\treturn True\n\t}\n\treturn Nil\n}", "func (c *C) Bool() Type {\n\t// TODO: support custom bool types\n\treturn c.e.Go().Bool()\n}", "func (t *Typed) Bool(key string) bool {\n\treturn t.BoolOr(key, false)\n}", "func (w *Writer) Bool(b bool) error {\n\tif b {\n\t\treturn w.Bit(1)\n\t}\n\treturn w.Bit(0)\n}", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }" ]
[ "0.6300829", "0.60781026", "0.60108006", "0.58833617", "0.58359164", "0.5806103", "0.5771532", "0.57651114", "0.5737598", "0.57369924", "0.57280886", "0.57044286", "0.5695262", "0.5693679", "0.56920135", "0.568601", "0.5668572", "0.56666636", "0.563851", "0.5629889", "0.5616634", "0.5584309", "0.5529425", "0.5500306", "0.54801285", "0.5472126", "0.5467016", "0.5447669", "0.5446709", "0.5436271", "0.5424555", "0.5422969", "0.5417694", "0.5401511", "0.53957546", "0.5386414", "0.53824073", "0.5375091", "0.53748584", "0.53564847", "0.5336326", "0.5331273", "0.5313354", "0.53132534", "0.5312648", "0.53100294", "0.5308335", "0.53031814", "0.52836114", "0.52809644", "0.52798057", "0.52786", "0.52720594", "0.5271782", "0.52691036", "0.5267118", "0.5265091", "0.526321", "0.52605045", "0.5258576", "0.5251885", "0.5248594", "0.52274674", "0.52167714", "0.5213396", "0.5206933", "0.52049845", "0.5204642", "0.51976407", "0.51969045", "0.5194055", "0.51919717", "0.5185706", "0.5180901", "0.51807034", "0.51738477", "0.51703733", "0.5168347", "0.5154495", "0.515254", "0.51525223", "0.51307446", "0.5129085", "0.51255333", "0.51162857", "0.5104151", "0.5100952", "0.5096887", "0.50953585", "0.5090465", "0.50819314", "0.5081514", "0.5079985", "0.5057859", "0.50546056", "0.50546056", "0.50546056", "0.50546056", "0.50546056", "0.50546056" ]
0.5112665
85
Hash32 implements the Arg.Hash32() method.
func (p Path) Hash32(h hash.Hash32) { h.Write([]byte(p)) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *Target) Hash32(h hash.Hash32) {\n\th.Write([]byte(t.Name))\n\th.Write([]byte(t.Builder))\n\tfor _, arg := range t.Args {\n\t\targ.Hash32(h)\n\t}\n\tfor _, env := range t.Env {\n\t\th.Write([]byte(env))\n\t}\n}", "func CalcHash32(data []byte) Hash32 {\n\treturn hash.Sum(data)\n}", "func (s String) Hash32(h hash.Hash32) { h.Write([]byte(s)) }", "func HexToHash32(s string) Hash32 { return BytesToHash(util.FromHex(s)) }", "func (s *Sub) Hash32(h hash.Hash32) {\n\th.Write([]byte(s.Format))\n\tfor _, sub := range s.Substitutions {\n\t\th.Write([]byte(sub.Key))\n\t\tsub.Value.Hash32(h)\n\t}\n}", "func (ch *ConsistentHash) fnv32Hash(key string) uint32 {\n\tnew32Hash := fnv.New32()\n\tnew32Hash.Write([]byte(key))\n\treturn new32Hash.Sum32()\n}", "func hash(s string) uint32 {\n h := fnv.New32a()\n h.Write([]byte(s))\n return h.Sum32()\n}", "func hash(value string) uint32 {\n\th := fnv.New32a()\n\th.Write([]byte(value))\n\n\treturn h.Sum32()\n}", "func Hash(strings ...string) uint32 {\n\tdigester := fnv.New32()\n\tfor _, s := range strings {\n\t\t_, _ = io.WriteString(digester, s)\n\t}\n\treturn digester.Sum32()\n}", "func Hash(key string) uint32 {\n\treturn uint32(aeshashstr(noescape(unsafe.Pointer(&key)), 0))\n}", "func f32hash(p unsafe.Pointer, h uintptr) uintptr {\n\tf := *(*float32)(p)\n\tswitch {\n\tcase f == 0:\n\t\treturn c1 * (c0 ^ h) // +0, -0\n\tcase f != f:\n\t\treturn c1 * (c0 ^ h ^ uintptr(fastrand())) // any kind of NaN\n\tdefault:\n\t\treturn memhash(p, h, 4)\n\t}\n}", "func FNVHash32(value uint32) uint32 {\n\thash := FNVOffsetBasis32\n\tfor i := 0; i < 4; i++ {\n\t\toctet := value & 0x00FF\n\t\tvalue >>= 8\n\n\t\thash ^= octet\n\t\thash *= FNVPrime32\n\t}\n\treturn hash\n}", "func hash(x []byte) uint32 {\n\treturn crc32.ChecksumIEEE(x)\n}", "func Hash32(s []byte) uint32 {\n\tn := uint32(len(s))\n\tif n <= 24 {\n\t\tif n <= 12 {\n\t\t\tif n <= 4 {\n\t\t\t\treturn hash32Len0to4(s)\n\t\t\t}\n\t\t\treturn hash32Len5to12(s)\n\t\t}\n\t\treturn hash32Len13to24(s)\n\t}\n\n\t// n > 24\n\th := n\n\tg := c1 * n\n\tf := g\n\n\ta0 := ror32(fetch32(s[n-4:])*c1, 17) * c2\n\ta1 := ror32(fetch32(s[n-8:])*c1, 17) * c2\n\ta2 := ror32(fetch32(s[n-16:])*c1, 17) * c2\n\ta3 := ror32(fetch32(s[n-12:])*c1, 17) * c2\n\ta4 := ror32(fetch32(s[n-20:])*c1, 17) * c2\n\n\tconst magic = 0xe6546b64\n\th ^= a0\n\th = ror32(h, 19)\n\th = h*5 + magic\n\th ^= a2\n\th = ror32(h, 19)\n\th = h*5 + magic\n\tg ^= a1\n\tg = ror32(g, 19)\n\tg = g*5 + magic\n\tg ^= a3\n\tg = ror32(g, 19)\n\tg = g*5 + magic\n\tf += a4\n\tf = ror32(f, 19)\n\tf = f*5 + magic\n\tfor i := (n - 1) / 20; i != 0; i-- {\n\t\ta0 := ror32(fetch32(s)*c1, 17) * c2\n\t\ta1 := fetch32(s[4:])\n\t\ta2 := ror32(fetch32(s[8:])*c1, 17) * c2\n\t\ta3 := ror32(fetch32(s[12:])*c1, 17) * c2\n\t\ta4 := fetch32(s[16:])\n\t\th ^= a0\n\t\th = ror32(h, 18)\n\t\th = h*5 + magic\n\t\tf += a1\n\t\tf = ror32(f, 19)\n\t\tf = f * c1\n\t\tg += a2\n\t\tg = ror32(g, 18)\n\t\tg = g*5 + magic\n\t\th ^= a3 + a1\n\t\th = ror32(h, 19)\n\t\th = h*5 + magic\n\t\tg ^= a4\n\t\tg = bswap32(g) * 5\n\t\th += a4 * 5\n\t\th = bswap32(h)\n\t\tf += a0\n\t\tf, g, h = g, h, f // a.k.a. PERMUTE3\n\t\ts = s[20:]\n\t}\n\tg = ror32(g, 11) * c1\n\tg = ror32(g, 17) * c1\n\tf = ror32(f, 11) * c1\n\tf = ror32(f, 17) * c1\n\th = ror32(h+g, 19)\n\th = h*5 + magic\n\th = ror32(h, 17) * c1\n\th = ror32(h+f, 19)\n\th = h*5 + magic\n\th = ror32(h, 17) * c1\n\treturn h\n}", "func hash(s string) int {\n\th := fnv.New32a()\n\tif _, err := h.Write([]byte(s)); err != nil {\n\t\tpanic(err) // should never happen\n\t}\n\n\treturn int(h.Sum32() & 0x7FFFFFFF) // mask MSB of uint32 as this will be sign bit\n}", "func (gg GlobGroup) Hash32(h hash.Hash32) {\n\tfor _, p := range gg {\n\t\th.Write([]byte(p))\n\t}\n}", "func Hash32(s []byte) uint32 {\n\n\tslen := len(s)\n\n\tif slen <= 24 {\n\t\tif slen <= 12 {\n\t\t\tif slen <= 4 {\n\t\t\t\treturn hash32Len0to4(s, 0)\n\t\t\t}\n\t\t\treturn hash32Len5to12(s, 0)\n\t\t}\n\t\treturn hash32Len13to24Seed(s, 0)\n\t}\n\n\t// len > 24\n\th := uint32(slen)\n\tg := c1 * uint32(slen)\n\tf := g\n\ta0 := rotate32(fetch32(s, slen-4)*c1, 17) * c2\n\ta1 := rotate32(fetch32(s, slen-8)*c1, 17) * c2\n\ta2 := rotate32(fetch32(s, slen-16)*c1, 17) * c2\n\ta3 := rotate32(fetch32(s, slen-12)*c1, 17) * c2\n\ta4 := rotate32(fetch32(s, slen-20)*c1, 17) * c2\n\th ^= a0\n\th = rotate32(h, 19)\n\th = h*5 + 0xe6546b64\n\th ^= a2\n\th = rotate32(h, 19)\n\th = h*5 + 0xe6546b64\n\tg ^= a1\n\tg = rotate32(g, 19)\n\tg = g*5 + 0xe6546b64\n\tg ^= a3\n\tg = rotate32(g, 19)\n\tg = g*5 + 0xe6546b64\n\tf += a4\n\tf = rotate32(f, 19) + 113\n\titers := (slen - 1) / 20\n\tfor {\n\t\ta := fetch32(s, 0)\n\t\tb := fetch32(s, 4)\n\t\tc := fetch32(s, 8)\n\t\td := fetch32(s, 12)\n\t\te := fetch32(s, 16)\n\t\th += a\n\t\tg += b\n\t\tf += c\n\t\th = mur(d, h) + e\n\t\tg = mur(c, g) + a\n\t\tf = mur(b+e*c1, f) + d\n\t\tf += g\n\t\tg += f\n\t\ts = s[20:]\n\t\titers--\n\t\tif iters == 0 {\n\t\t\tbreak\n\t\t}\n\t}\n\tg = rotate32(g, 11) * c1\n\tg = rotate32(g, 17) * c1\n\tf = rotate32(f, 11) * c1\n\tf = rotate32(f, 17) * c1\n\th = rotate32(h+g, 19)\n\th = h*5 + 0xe6546b64\n\th = rotate32(h, 17) * c1\n\th = rotate32(h+f, 19)\n\th = h*5 + 0xe6546b64\n\th = rotate32(h, 17) * c1\n\treturn h\n}", "func (h Hash20) ToHash32() (h32 Hash32) {\n\tcopy(h32[:], h[:])\n\treturn\n}", "func TestExample(t *testing.T) {\n\tstr := \"hello world\"\n\tbytes := []byte(str)\n\thash := Hash32(bytes)\n\tfmt.Printf(\"Hash32(%s) is %x\\n\", str, hash)\n}", "func hash(data []byte) uint32 {\n\tvar h uint32 = binary.LittleEndian.Uint32(data) * kDictHashMul32\n\n\t/* The higher bits contain more mixture from the multiplication,\n\t so we take our results from there. */\n\treturn h >> uint(32-kDictNumBits)\n}", "func (t *hashReader) Sum32() uint32 {\n\treturn t.h.Sum32()\n}", "func (h *Hash) Sum32() (uint32, bool) {\n\th32, ok := h.Hash.(hash.Hash32)\n\tif !ok {\n\t\treturn 0, false\n\t}\n\n\treturn h32.Sum32(), true\n}", "func (d Data32) Hash() Hash {\n\treturn hash(d)\n}", "func hash3(u uint32, h uint8) uint32 {\n\treturn ((u << (32 - 24)) * prime3bytes) >> ((32 - h) & 31)\n}", "func hash(elements ...[32]byte) [32]byte {\n\tvar hash []byte\n\tfor i := range elements {\n\t\thash = append(hash, elements[i][:]...)\n\t}\n\treturn sha256.Sum256(hash)\n}", "func strhash(a unsafe.Pointer, h uintptr) uintptr", "func CalcObjectHash32(obj scale.Encodable) Hash32 {\n\tbytes, err := codec.Encode(obj)\n\tif err != nil {\n\t\tpanic(\"could not serialize object\")\n\t}\n\treturn CalcHash32(bytes)\n}", "func (h *Hash) IsHash32() bool {\n\t_, ok := h.Hash.(hash.Hash32)\n\treturn ok\n}", "func TestHash32(t *testing.T) {\n\tstdHash := crc32.New(crc32.IEEETable)\n\tif _, err := stdHash.Write([]byte(\"test\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\t// create a new hash with stdHash.Sum32() as initial crc\n\tcrcHash := New(stdHash.Sum32(), crc32.IEEETable)\n\n\tstdHashSize := stdHash.Size()\n\tcrcHashSize := crcHash.Size()\n\tif stdHashSize != crcHashSize {\n\t\tt.Fatalf(\"%d != %d\", stdHashSize, crcHashSize)\n\t}\n\n\tstdHashBlockSize := stdHash.BlockSize()\n\tcrcHashBlockSize := crcHash.BlockSize()\n\tif stdHashBlockSize != crcHashBlockSize {\n\t\tt.Fatalf(\"%d != %d\", stdHashBlockSize, crcHashBlockSize)\n\t}\n\n\tstdHashSum32 := stdHash.Sum32()\n\tcrcHashSum32 := crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n\n\tstdHashSum := stdHash.Sum(make([]byte, 32))\n\tcrcHashSum := crcHash.Sum(make([]byte, 32))\n\tif !reflect.DeepEqual(stdHashSum, crcHashSum) {\n\t\tt.Fatalf(\"sum = %v, want %v\", crcHashSum, stdHashSum)\n\t}\n\n\t// write something\n\tif _, err := stdHash.Write([]byte(\"hello\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif _, err := crcHash.Write([]byte(\"hello\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tstdHashSum32 = stdHash.Sum32()\n\tcrcHashSum32 = crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n\n\t// reset\n\tstdHash.Reset()\n\tcrcHash.Reset()\n\tstdHashSum32 = stdHash.Sum32()\n\tcrcHashSum32 = crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n}", "func Hash(length int, key string) int64 {\n\tif key == \"\" {\n\t\treturn 0\n\t}\n\thc := hashCode(key)\n\treturn (hc ^ (hc >> 16)) % int64(length)\n}", "func Sum32(key string) uint32 {\n\treturn Sum32Seed(key, 0)\n}", "func HashASM(k0, k1 uint64, p []byte) uint64", "func hash(key uint64) uint64 {\r\n\tkey ^= key >> 33\r\n\tkey *= 0xff51afd7ed558ccd\r\n\tkey ^= key >> 33\r\n\tkey *= 0xc4ceb9fe1a85ec53\r\n\tkey ^= key >> 33\r\n\treturn key\r\n}", "func strhash0(p unsafe.Pointer, h uintptr) uintptr", "func Hash(b []byte) uint32 {\n\tconst (\n\t\tseed = 0xbc9f1d34\n\t\tm = 0xc6a4a793\n\t)\n\th := uint32(seed) ^ uint32(len(b))*m\n\tfor ; len(b) >= 4; b = b[4:] {\n\t\th += uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24\n\t\th *= m\n\t\th ^= h >> 16\n\t}\n\tswitch len(b) {\n\tcase 3:\n\t\th += uint32(b[2]) << 16\n\t\tfallthrough\n\tcase 2:\n\t\th += uint32(b[1]) << 8\n\t\tfallthrough\n\tcase 1:\n\t\th += uint32(b[0])\n\t\th *= m\n\t\th ^= h >> 24\n\t}\n\treturn h\n}", "func HashFunction(buf []byte) uint32 {\n\tvar hash uint32 = 5381\n\tfor _, b := range buf {\n\t\thash = ((hash << 5) + hash) + uint32(b)\n\t}\n\treturn hash\n}", "func byteshash(p *[]byte, h uintptr) uintptr", "func Bytes32ToIpfsHash(value [32]byte) (string, error) {\n\tbyteArray := [34]byte{18, 32}\n\tcopy(byteArray[2:], value[:])\n\tif len(byteArray) != 34 {\n\t\treturn \"\", errors.New(\"invalid bytes32 value\")\n\t}\n\n\thash := base58.Encode(byteArray[:])\n\treturn hash, nil\n}", "func sumHash(c byte, h uint32) uint32 {\n\treturn (h * hashPrime) ^ uint32(c)\n}", "func (h Hash32) Hex() string { return util.Encode(h[:]) }", "func (this *Ring) Hash(key string) uint32 {\n\treturn crc32.ChecksumIEEE([]byte(key))\n}", "func hash(data []byte) [32]byte {\n\tvar hash [32]byte\n\n\th := sha256.New()\n\t// The hash interface never returns an error, for that reason\n\t// we are not handling the error below. For reference, it is\n\t// stated here https://golang.org/pkg/hash/#Hash\n\t// #nosec G104\n\th.Write(data)\n\th.Sum(hash[:0])\n\n\treturn hash\n}", "func FNV32(s string) uint32 {\n\treturn uint32Hasher(fnv.New32(), s)\n}", "func (h Hash32) Field() log.Field { return log.String(\"hash\", hex.EncodeToString(h[:])) }", "func hashInt(s string) uint32 {\n\tb := []byte(s)\n\th := crc32.ChecksumIEEE(b)\n\treturn h\n}", "func memhash(p unsafe.Pointer, h, s uintptr) uintptr", "func memhash(p unsafe.Pointer, h, s uintptr) uintptr", "func (_L1Block *L1BlockCaller) Hash(opts *bind.CallOpts) ([32]byte, error) {\n\tvar out []interface{}\n\terr := _L1Block.contract.Call(opts, &out, \"hash\")\n\n\tif err != nil {\n\t\treturn *new([32]byte), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new([32]byte)).(*[32]byte)\n\n\treturn out0, err\n\n}", "func TestHash32(t *testing.T) {\n\tt.Parallel()\n\n\tconst n = 400\n\n\tf := NewOptimized(Config{\n\t\tCapacity: n,\n\t\tFPRate: .01,\n\t})\n\n\tr := rand.New(rand.NewSource(32))\n\n\tfor i := 0; i < n; i++ {\n\t\tf.Add(uint64(r.Uint32()))\n\t}\n\n\tconst nrounds = 8\n\tfp := 0\n\tfor i := n; i < nrounds*n; i++ {\n\t\tif f.Has(uint64(r.Uint32())) {\n\t\t\tfp++\n\t\t}\n\t}\n\n\tfprate := float64(fp) / (nrounds * n)\n\tt.Logf(\"FP rate = %.2f%%\", 100*fprate)\n\tassert.LessOrEqual(t, fprate, .1)\n}", "func Hash32WithSeed(s []byte, seed uint32) uint32 {\n\tslen := len(s)\n\n\tif slen <= 24 {\n\t\tif slen >= 13 {\n\t\t\treturn hash32Len13to24Seed(s, seed*c1)\n\t\t}\n\t\tif slen >= 5 {\n\t\t\treturn hash32Len5to12(s, seed)\n\t\t}\n\t\treturn hash32Len0to4(s, seed)\n\t}\n\th := hash32Len13to24Seed(s[:24], seed^uint32(slen))\n\treturn mur(Hash32(s[24:])+seed, h)\n}", "func Hash(key []byte) uint64 {\n\treturn murmur3.Sum64(key)\n}", "func hash(s string) string {\n\thash := fnv.New32a()\n\thash.Write([]byte(s))\n\tintHash := hash.Sum32()\n\tresult := fmt.Sprintf(\"%08x\", intHash)\n\treturn result\n}", "func hash4(u uint32, h uint8) uint32 {\n\treturn (u * prime4bytes) >> ((32 - h) & 31)\n}", "func hash4x64(u uint64, h uint8) uint32 {\n\treturn (uint32(u) * prime4bytes) >> ((32 - h) & 31)\n}", "func sha3hash(t *testing.T, data ...[]byte) []byte {\n\tt.Helper()\n\th := sha3.NewLegacyKeccak256()\n\tr, err := doSum(h, nil, data...)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\treturn r\n}", "func (s *ShardMap) hash(v interface{}) int {\n\tswitch s.Type {\n\tcase \"string\":\n\t\tval, ok := v.(string)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\thash := fnv.New32()\n\t\thash.Write([]byte(val))\n\t\treturn int(hash.Sum32() % NumShards)\n\tcase \"int32\":\n\t\t// Values that come as numbers in JSON are of type float64.\n\t\tval, ok := v.(float64)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\treturn int(int32(val) % NumShards)\n\tdefault:\n\t\treturn -1\n\t}\n}", "func Checksum32(data []byte) uint32 {\n\treturn Checksum32Seed(data, 0)\n}", "func hash(s string) string {\n\th := fnv.New32a()\n\t_, err := h.Write([]byte(s))\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn fmt.Sprint(h.Sum32())\n}", "func Sha3256(bs []byte) ([]byte, error) {\n\treturn PerformHash(sha3.New256(), bs)\n}", "func addrHash(addr uint16) byte {\n\treturn (byte(addr) ^ byte(addr>>8)) & 0x7f\n}", "func (p Path) Hash() (uint32, error) {\n\treturn adler32.Checksum([]byte(p)), nil\n}", "func (r *RunCtx) Hash() (uint32, error) {\n\treturn 0, fmt.Errorf(\"not hashable\")\n}", "func hash(k Key) int {\n\tkey := fmt.Sprintf(\"%s\", k)\n\th := 0\n\tfor i := 0; i < len(key); i++ {\n\t\th = 31 * h + int(key[i])\n\t}\n\treturn h\n}", "func (t *Target) Hash() (uint32, error) {\n\th := adler32.New()\n\tt.Hash32(h)\n\treturn h.Sum32(), nil\n}", "func Sha256Hash(data []byte) [32]byte {\n\tsum := sha256.Sum256(data)\n\treturn sum\n}", "func strhash(p *string, h uintptr) uintptr", "func eb32(bits uint32, hi uint8, lo uint8) uint32 {\n\tm := uint32(((1 << (hi - lo)) - 1) << lo)\n\treturn (bits & m) >> lo\n}", "func Sum32Seed(key string, seed uint32) uint32 {\n\tvar nblocks = len(key) / 4\n\tvar nbytes = nblocks * 4\n\tvar h1 = seed\n\tconst c1 = 0xcc9e2d51\n\tconst c2 = 0x1b873593\n\tfor i := 0; i < nbytes; i += 4 {\n\t\tk1 := uint32(key[i+0]) | uint32(key[i+1])<<8 |\n\t\t\tuint32(key[i+2])<<16 | uint32(key[i+3])<<24\n\t\tk1 *= c1\n\t\tk1 = (k1 << 15) | (k1 >> 17)\n\t\tk1 *= c2\n\t\th1 ^= k1\n\t\th1 = (h1 << 13) | (h1 >> 19)\n\t\th1 = h1*5 + 0xe6546b64\n\t}\n\tvar k1 uint32\n\tswitch len(key) & 3 {\n\tcase 3:\n\t\tk1 ^= uint32(key[nbytes+2]) << 16\n\t\tfallthrough\n\tcase 2:\n\t\tk1 ^= uint32(key[nbytes+1]) << 8\n\t\tfallthrough\n\tcase 1:\n\t\tk1 ^= uint32(key[nbytes+0])\n\t\tk1 *= c1\n\t\tk1 = (k1 << 15) | (k1 >> 17)\n\t\tk1 *= c2\n\t\th1 ^= k1\n\t}\n\th1 ^= uint32(len(key))\n\th1 ^= h1 >> 16\n\th1 *= 0x85ebca6b\n\th1 ^= h1 >> 13\n\th1 *= 0xc2b2ae35\n\th1 ^= h1 >> 16\n\treturn h1\n}", "func Hash(data []byte) (string, int64) {\n\thasher := adler32.New()\n\tb, e := hasher.Write(data)\n\tif e != nil {\n\t\tlogs.WithFields(logs.Fields{\n\t\t\t\"Error\": e,\n\t\t}).Error(\"Unable to write chunk of data via hasher.Write\", e)\n\t}\n\treturn hex.EncodeToString(hasher.Sum(nil)), int64(b)\n}", "func HashXXH3_64(input []byte, seed uint64) (result uint64) {\n\treturn parser.HashXXH3_64(input, seed)\n}", "func CalcBlocksHash32(view []BlockID, additionalBytes []byte) Hash32 {\n\tsortedView := make([]BlockID, len(view))\n\tcopy(sortedView, view)\n\tSortBlockIDs(sortedView)\n\treturn CalcBlockHash32Presorted(sortedView, additionalBytes)\n}", "func NewHashFromArray(bytes [32]byte) Hash32 {\n\treturn NewHashFromBytes(bytes[:])\n}", "func (t *openAddressing) hash(key string, round int) uint32 {\n\tnum := uint(stringToInt(key))\n\tmax := uint(len(t.values) - 1)\n\treturn uint32((hashDivision(num, max) + uint(round)*hashDivision2(num, max)) % max)\n}", "func hash(addr mino.Address) *big.Int {\n\tsha := sha256.New()\n\tmarshalled, err := addr.MarshalText()\n\tif err != nil {\n\t\tmarshalled = []byte(addr.String())\n\t}\n\t// A hack to accommodate for minogrpc's design:\n\t// 1) the first byte is used to indicate if a node is orchestrator or not\n\t// 2) the only way to reach the orchestrator is to route a message to nil\n\t// from its server side, which has the same address but orchestrator byte\n\t// set to f.\n\t// We therefore have to ignore if a node is the orchestrator to be able to\n\t// route the message first to its server side, then from the server side to\n\t// the client side.\n\tsha.Write(marshalled[1:])\n\treturn byteArrayToBigInt(sha.Sum(nil))\n}", "func Hash(value int64) uint64 {\n\treturn FNVHash64(uint64(value))\n}", "func BytesToHash(b []byte) Hash32 {\n\tvar h Hash32\n\th.SetBytes(b)\n\treturn h\n}", "func hashFunction(key int, size int) int {\n\treturn key % size\n}", "func FNV32a(s string) uint32 {\n\treturn uint32Hasher(fnv.New32a(), s)\n}", "func Hash3Words(a, b, c, initval uint32) uint32 {\n\tconst iv = 0xdeadbeef + (3 << 2)\n\tinitval += iv\n\n\ta += initval\n\tb += initval\n\tc += initval\n\n\tc ^= b\n\tc -= rol32(b, 14)\n\ta ^= c\n\ta -= rol32(c, 11)\n\tb ^= a\n\tb -= rol32(a, 25)\n\tc ^= b\n\tc -= rol32(b, 16)\n\ta ^= c\n\ta -= rol32(c, 4)\n\tb ^= a\n\tb -= rol32(a, 14)\n\tc ^= b\n\tc -= rol32(b, 24)\n\n\treturn c\n}", "func (payload *ExtEthCompatPayload) Hash() B32 {\n\thash := crypto.Keccak256(payload.Value)\n\tvar phash B32\n\tcopy(phash[:], hash)\n\treturn phash\n}", "func SHA256(raw []byte) Hash {\n\treturn gosha256.Sum256(raw)\n}", "func HashKey(key int) int {\n\t/*\n\t\ttiedot should be compiled/run on x86-64 systems.\n\t\tIf you decide to compile tiedot on 32-bit systems, the following integer-smear algorithm will cause compilation failure\n\t\tdue to 32-bit interger overflow; therefore you must modify the algorithm.\n\t\tDo not remove the integer-smear process, and remember to run test cases to verify your mods.\n\t*/\n\t// ========== Integer-smear start =======\n\tkey = key ^ (key >> 4)\n\tkey = (key ^ 0xdeadbeef) + (key << 5)\n\tkey = key ^ (key >> 11)\n\t// ========== Integer-smear end =========\n\treturn key & ((1 << HASH_BITS) - 1) // Do not modify this line\n}", "func nilinterhash(a unsafe.Pointer, h uintptr) uintptr", "func fnv32a(s string) uint32 {\n\tconst (\n\t\tinitial = 2166136261\n\t\tprime = 16777619\n\t)\n\n\thash := uint32(initial)\n\tfor i := 0; i < len(s); i++ {\n\t\thash ^= uint32(s[i])\n\t\thash *= prime\n\t}\n\treturn hash\n}", "func CalculateHash(args []string) string {\n\tvar str = \"\"\n\tfor _,v := range args {\n\t\tstr += v\n\t}\n\thasher := sha256.New()\n\thasher.Write([]byte(str))\n\treturn hex.EncodeToString(hasher.Sum(nil))\n}", "func CalcBlockHash32Presorted(sortedView []BlockID, additionalBytes []byte) Hash32 {\n\thash := hash.New()\n\thash.Write(additionalBytes)\n\tfor _, id := range sortedView {\n\t\thash.Write(id.Bytes()) // this never returns an error: https://golang.org/pkg/hash/#Hash\n\t}\n\tvar res Hash32\n\thash.Sum(res[:0])\n\treturn res\n}", "func runtime_memhash(p unsafe.Pointer, seed, s uintptr) uintptr", "func hashCode(key string) int64 {\n\tv := int64(crc32.ChecksumIEEE([]byte(key)))\n\tif v >= 0 {\n\t\treturn v\n\t}\n\tif -v > 0 {\n\t\treturn -v\n\t}\n\t// v == MinInt\n\treturn 0\n}", "func (o *ObjectIndex) Hash() uint32 {\n\tvar h uint32 = 17\n\n\tvar str string\n\tstr += fmt.Sprintf(\"%08x\", o.machine)\n\tstr += fmt.Sprintf(\"%04x\", o.pid)\n\tstr += fmt.Sprintf(\"%08x\", o.id)\n\tstr += fmt.Sprintf(\"%08x\", o.Rand)\n\tfor _, v := range str {\n\t\th += h*23 + uint32(v)\n\t}\n\treturn h\n}", "func crc32Demo() {\n\t// hasher\n\th := crc32.NewIEEE()\n\tfmt.Println(reflect.TypeOf(h))\n\n\t// write a string converted to bytes\n\th.Write([]byte(\"test\"))\n\n\t// checksum\n\tv := h.Sum32()\n\tfmt.Println(reflect.TypeOf(v)) // uint32\n\tfmt.Println(v)\n}", "func (h Hash32) Bytes() []byte { return h[:] }", "func hashmapHash(data []byte) uint32 {\n\tvar result uint32 = 2166136261 // FNV offset basis\n\tfor _, c := range data {\n\t\tresult ^= uint32(c)\n\t\tresult *= 16777619 // FNV prime\n\t}\n\treturn result\n}", "func fletcher32(payload []byte) uint32 {\n\ts1 := uint16(0)\n\ts2 := uint16(0)\n\n\tsz := len(payload) & (^1)\n\tfor i := 0; i < sz; i += 2 {\n\t\ts1 += uint16(payload[i]) | (uint16(payload[i+1]) << 8)\n\t\ts2 += s1\n\t}\n\tif len(payload)&1 != 0 {\n\t\ts1 += uint16(payload[sz])\n\t\ts2 += s1\n\t}\n\treturn (uint32(s2) << 16) | uint32(s1)\n}", "func Hash(data []byte) [blake2b.Size]byte {\n\treturn blake2b.Sum512(data)\n}", "func (i *Instance) Hash() (uint32, error) {\n\th := fnv.New32()\n\tfmt.Fprintf(h, \"%s%s\", i.Name, i.Version)\n\n\tfn, err := i.Fn.Hash()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn fn ^ h.Sum32(), nil\n}", "func hash_func(x, y, n HashValue) (HashValue) {\n return (x*1640531513 ^ y*2654435789) % n\n}", "func HashBuildArgs(args interface{}) (string, error) {\n\tdata, err := json.Marshal(args)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\thash := sha256.Sum256(data)\n\treturn hex.EncodeToString(hash[:]), nil\n}", "func (gg GlobGroup) Hash() (uint32, error) {\n\th := adler32.New()\n\tgg.Hash32(h)\n\treturn h.Sum32(), nil\n}", "func (hasher *SHA256) HashLength() uint {\n\treturn 64\n}", "func (h *MemHash) Hash() uint32 {\n\tss := (*stringStruct)(unsafe.Pointer(&h.buf))\n\treturn uint32(memhash(ss.str, 0, uintptr(ss.len)))\n}" ]
[ "0.74894154", "0.72687817", "0.7188007", "0.69853705", "0.6892721", "0.67331547", "0.6720686", "0.67194086", "0.6649251", "0.66210663", "0.65797126", "0.6577248", "0.6553477", "0.6546589", "0.6525743", "0.64986825", "0.64651316", "0.6463679", "0.64490825", "0.63933766", "0.631653", "0.6316435", "0.6264623", "0.6239596", "0.6215961", "0.620281", "0.6120723", "0.61176646", "0.61121273", "0.60869133", "0.60768", "0.6070945", "0.6068143", "0.60301584", "0.6023994", "0.60172313", "0.60023165", "0.59999824", "0.5997599", "0.5994999", "0.59749955", "0.5944826", "0.59306836", "0.5927316", "0.5925298", "0.5898057", "0.5898057", "0.58708656", "0.5862113", "0.5855231", "0.5853075", "0.58529514", "0.58501244", "0.5840421", "0.58195925", "0.5813835", "0.58128905", "0.5808467", "0.58067775", "0.5790786", "0.57631826", "0.5756814", "0.5747848", "0.57439625", "0.57336444", "0.57286763", "0.5721987", "0.57191473", "0.5719119", "0.57175076", "0.5712127", "0.5683982", "0.56668335", "0.56535757", "0.5629039", "0.5622986", "0.5620013", "0.561975", "0.56082356", "0.56069034", "0.5603419", "0.5597878", "0.55931437", "0.55875325", "0.5586903", "0.558125", "0.5577949", "0.5559417", "0.55589855", "0.55566466", "0.5553403", "0.553855", "0.55382097", "0.5537555", "0.552272", "0.5500112", "0.5494353", "0.54865384", "0.54836756", "0.54825485" ]
0.7510771
0
Hash implements the starlark.Value.Hash() method.
func (p Path) Hash() (uint32, error) { return adler32.Checksum([]byte(p)), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Hash(value int64) uint64 {\n\treturn FNVHash64(uint64(value))\n}", "func (n *SoupNode) Hash() (uint32, error) {\n\treturn hashString(fmt.Sprintf(\"%v\", *n)), nil\n}", "func (o *ObjectIndex) Hash() uint32 {\n\tvar h uint32 = 17\n\n\tvar str string\n\tstr += fmt.Sprintf(\"%08x\", o.machine)\n\tstr += fmt.Sprintf(\"%04x\", o.pid)\n\tstr += fmt.Sprintf(\"%08x\", o.id)\n\tstr += fmt.Sprintf(\"%08x\", o.Rand)\n\tfor _, v := range str {\n\t\th += h*23 + uint32(v)\n\t}\n\treturn h\n}", "func hash(key, value string) int64 {\n\thash := siphash.New(sipConst)\n\thash.Write([]byte(key + \":::\" + value))\n\treturn int64(hash.Sum64())\n}", "func (term *Term) Hash() int {\n\treturn term.Value.Hash()\n}", "func (s *ShardMap) hash(v interface{}) int {\n\tswitch s.Type {\n\tcase \"string\":\n\t\tval, ok := v.(string)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\thash := fnv.New32()\n\t\thash.Write([]byte(val))\n\t\treturn int(hash.Sum32() % NumShards)\n\tcase \"int32\":\n\t\t// Values that come as numbers in JSON are of type float64.\n\t\tval, ok := v.(float64)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\treturn int(int32(val) % NumShards)\n\tdefault:\n\t\treturn -1\n\t}\n}", "func (this *Ring) Hash(key string) uint32 {\n\treturn crc32.ChecksumIEEE([]byte(key))\n}", "func Hash(length int, key string) int64 {\n\tif key == \"\" {\n\t\treturn 0\n\t}\n\thc := hashCode(key)\n\treturn (hc ^ (hc >> 16)) % int64(length)\n}", "func HashOf(v Value) []byte {\n\treturn quad.HashOf(v)\n}", "func hash(value string) uint32 {\n\th := fnv.New32a()\n\th.Write([]byte(value))\n\n\treturn h.Sum32()\n}", "func (ref Ref) Hash() int {\n\treturn termSliceHash(ref)\n}", "func (o *Object) Hash() string {\n\treturn Hash(o, true, false, true)\n}", "func (r Ref) Hash() int {\n\treturn termSliceHash(r)\n}", "func hash(obj interface{}) KHash {\n\tvar buffer bytes.Buffer\n\tencoder := json.NewEncoder(&buffer)\n\terr := encoder.Encode(obj)\n\tif err != nil {\n\t\tpanic(\"cannot encode object\")\n\t}\n\n\tdata := buffer.Bytes()\n\th := sha256.Sum256(data)\n\n\t// log.Printf(\"hashing %#v represented as %s with hash %X\", obj, data, h)\n\treturn h\n}", "func (i *Instance) Hash() (uint32, error) {\n\th := fnv.New32()\n\tfmt.Fprintf(h, \"%s%s\", i.Name, i.Version)\n\n\tfn, err := i.Fn.Hash()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn fn ^ h.Sum32(), nil\n}", "func (expr *Expr) Hash() int {\n\ts := expr.Index\n\tswitch ts := expr.Terms.(type) {\n\tcase []*Term:\n\t\tfor _, t := range ts {\n\t\t\ts += t.Value.Hash()\n\t\t}\n\tcase *Term:\n\t\ts += ts.Value.Hash()\n\t}\n\tif expr.Negated {\n\t\ts++\n\t}\n\treturn s\n}", "func (i *Index) Hash() (uint32, error) {\n\treturn 0, fmt.Errorf(\"unhashable: %s\", i.Type())\n}", "func (p *FiveTuple) Hash() uint64 {\n\treturn siphash.Hash(lookupKey, 0, p.data)\n}", "func (k Ktype) Hash() uint32 {\n\tif k == UnknownKtype {\n\t\treturn 0\n\t}\n\treturn hashers.FnvUint32([]byte(k.String()))\n}", "func (sc *SetComprehension) Hash() int {\n\treturn sc.Term.Hash() + sc.Body.Hash()\n}", "func Hash(data interface{}) string {\n\treturn hex.EncodeToString(RawHash(data))\n}", "func (obj *identifier) Hash() hash.Hash {\n\tif obj.IsElement() {\n\t\treturn obj.Element().Hash()\n\t}\n\n\treturn obj.Comparer().Hash()\n}", "func Hash(key []byte) uint64 {\n\treturn murmur3.Sum64(key)\n}", "func (o *Object) Hash(ht hash.Type) (string, error) {\n\terr := o.loadMetadataIfNotLoaded()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif ht&hash.MD5 == 0 {\n\t\treturn \"\", hash.ErrUnsupported\n\t}\n\treturn hex.EncodeToString(o.meta.Hash), nil\n}", "func (null Null) Hash() int {\n\treturn 0\n}", "func Hasher(value string) string {\n\th := fnv.New32a()\n\t_, _ = h.Write([]byte(value))\n\treturn fmt.Sprintf(\"%v\", h.Sum32())\n}", "func (p PropertyHashList) Hash() string {\n\tglobalSum := sha256.New()\n\tfor _, hash := range p {\n\t\t_, _ = globalSum.Write(hash.Hash)\n\t}\n\n\tsum := globalSum.Sum(nil)\n\n\treturn hex.EncodeToString(sum)\n}", "func (t *openAddressing) hash(key string, round int) uint32 {\n\tnum := uint(stringToInt(key))\n\tmax := uint(len(t.values) - 1)\n\treturn uint32((hashDivision(num, max) + uint(round)*hashDivision2(num, max)) % max)\n}", "func hash(t types.Type, x value) int {\n\tswitch x := x.(type) {\n\tcase bool:\n\t\tif x {\n\t\t\treturn 1\n\t\t}\n\t\treturn 0\n\tcase int:\n\t\treturn x\n\tcase int8:\n\t\treturn int(x)\n\tcase int16:\n\t\treturn int(x)\n\tcase int32:\n\t\treturn int(x)\n\tcase int64:\n\t\treturn int(x)\n\tcase uint:\n\t\treturn int(x)\n\tcase uint8:\n\t\treturn int(x)\n\tcase uint16:\n\t\treturn int(x)\n\tcase uint32:\n\t\treturn int(x)\n\tcase uint64:\n\t\treturn int(x)\n\tcase uintptr:\n\t\treturn int(x)\n\tcase float32:\n\t\treturn int(x)\n\tcase float64:\n\t\treturn int(x)\n\tcase complex64:\n\t\treturn int(real(x))\n\tcase complex128:\n\t\treturn int(real(x))\n\tcase string:\n\t\treturn hashString(x)\n\tcase *value:\n\t\treturn int(uintptr(unsafe.Pointer(x)))\n\tcase chan value:\n\t\treturn int(uintptr(reflect.ValueOf(x).Pointer()))\n\tcase structure:\n\t\treturn x.hash(t)\n\tcase array:\n\t\treturn x.hash(t)\n\tcase iface:\n\t\treturn x.hash(t)\n\tcase rtype:\n\t\treturn x.hash(t)\n\t}\n\tpanic(fmt.Sprintf(\"%T is unhashable\", x))\n}", "func (o *Object) Hash(ctx context.Context, r hash.Type) (string, error) {\n\treturn \"\", hash.ErrUnsupported\n}", "func Hash(s int, o Orientation) (int, error) {\n\n\tvar errVal int = 10\n\n\tif !(s >= 0 && s <= palletWidth*palletLength) {\n\t\treturn errVal, ErrSize\n\t}\n\tif o != HORIZONTAL && o != VERTICAL && o != SQUAREGRID {\n\t\treturn errVal, ErrOrient\n\t}\n\n\tvar hash int\n\n\tswitch s {\n\tcase 1, 2, 3, 6:\n\t\thash = s - 1\n\tcase 4:\n\t\tif o == SQUAREGRID {\n\t\t\thash = s\n\t\t} else {\n\t\t\thash = s - 1\n\t\t}\n\tcase 8:\n\t\thash = 6\n\tcase 9:\n\t\thash = 7\n\tcase 12:\n\t\thash = 8\n\tcase 16:\n\t\thash = 9\n\tdefault:\n\t\treturn errVal, ErrSize\n\t}\n\n\treturn hash, nil\n}", "func (oc *ObjectComprehension) Hash() int {\n\treturn oc.Key.Hash() + oc.Value.Hash() + oc.Body.Hash()\n}", "func (c *Cluster) Hash(v interface{}) (int, error) {\n\th, err := hashstructure.Hash(v, nil)\n\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\n\t// get cluster index\n\tci := int(h % uint64(c.metadata.NumShards))\n\n\treturn ci, nil\n}", "func (a Address) Hash() Hash { return BytesToHash(a[:]) }", "func (v Var) Hash() int {\n\th := xxhash.ChecksumString64S(string(v), hashSeed0)\n\treturn int(h)\n}", "func Hash(k0, k1 uint64, p []byte) uint64 {\n\tvar d digest\n\td.size = Size\n\td.k0 = k0\n\td.k1 = k1\n\td.Reset()\n\td.Write(p)\n\treturn d.Sum64()\n}", "func (m *Map) Hash() string {\n\treturn Hash(m, true, false, true)\n}", "func (b BlockChain) Hash() {\n\n}", "func (a *Array) Hash() string {\n\treturn Hash(a, true, false, true)\n}", "func Hash(strings ...string) uint32 {\n\tdigester := fnv.New32()\n\tfor _, s := range strings {\n\t\t_, _ = io.WriteString(digester, s)\n\t}\n\treturn digester.Sum32()\n}", "func (o *Object) Hash(ctx context.Context, t hash.Type) (string, error) {\n\treturn \"\", hash.ErrUnsupported\n}", "func (c Category) Hash() uint32 {\n\treturn hashers.FnvUint32([]byte(c))\n}", "func Hash(i interface{}) string {\n\tv := reflect.ValueOf(i)\n\tif v.Kind() != reflect.Ptr {\n\t\tif !v.CanAddr(){\n\t\t\treturn \"\"\n\t\t}\n\t\tv = v.Addr()\n\t}\n\n\tsize := unsafe.Sizeof(v.Interface())\n\tb := (*[1 << 10]uint8)(unsafe.Pointer(v.Pointer()))[:size:size]\n\n\th := md5.New()\n\treturn base64.StdEncoding.EncodeToString(h.Sum(b))\n}", "func (s *set) Hash() int {\n\treturn s.hash\n}", "func (obj *bucket) Hash() hash.Hash {\n\treturn obj.immutable.Hash()\n}", "func (s Sample) Hash() []byte {\n\tallVecs := make([]linalg.Vector, len(s.Inputs)+len(s.Outputs))\n\tcopy(allVecs, s.Inputs)\n\tcopy(allVecs[len(s.Inputs):], s.Outputs)\n\treturn sgd.HashVectors(allVecs...)\n}", "func hash(key uint64) uint64 {\r\n\tkey ^= key >> 33\r\n\tkey *= 0xff51afd7ed558ccd\r\n\tkey ^= key >> 33\r\n\tkey *= 0xc4ceb9fe1a85ec53\r\n\tkey ^= key >> 33\r\n\treturn key\r\n}", "func Hash(a interface{}) (string, error) {\n\th := sha1.New()\n\tif err := json.NewEncoder(h).Encode(a); err != nil {\n\t\treturn \"\", nil\n\t}\n\treturn base64.URLEncoding.EncodeToString(h.Sum(nil)), nil\n}", "func Hash(content string, size int) int {\n\tsequence := sliceutil.Atoi(content, \",\")\n\tcircleKnots := GetStringCircle(size)\n\tfor _, n := range sequence {\n\t\tcircleKnots.ComputeKnot(n)\n\t}\n\treturn circleKnots.GetHash()\n}", "func hash(key string) int{\n\tvar num = 0\n\t// get the lenght of the key\n\tvar length = len(key)\n\n\t// add the ascii character value to creat a sum \n\tfor i := 0; i < length; i++{\n\n\t\tnum += int(key[i])\n\t}\n\t\n\t// square in the middle hash method\n\tvar avg = num * int((math.Pow(5.0, 0.5) - 1)) / 2\n\tvar numeric = avg - int(math.Floor(float64(avg)))\n\n\n\t// hash value to place into the table slice between -1 and CAPACITY - 1\n\treturn int(math.Floor(float64(numeric * CAPACITY)))\n}", "func (bc *Blockchain) Hash() {\n\n}", "func (t Tuple1) Hash() uint32 {\n\tif t.E1 == nil {\n\t\treturn 0\n\t}\n\treturn t.E1.Hash()\n}", "func hash(s string) uint32 {\n h := fnv.New32a()\n h.Write([]byte(s))\n return h.Sum32()\n}", "func (obj *immutable) Hash() hash.Hash {\n\treturn obj.hash\n}", "func (o *Object) Hash(t fs.HashType) (string, error) {\n\treturn \"\", fs.ErrHashUnsupported\n}", "func Hash(key string) uint32 {\n\treturn uint32(aeshashstr(noescape(unsafe.Pointer(&key)), 0))\n}", "func (o *ObjectInfo) Hash(ht hash.Type) (string, error) {\n\tif o.meta == nil {\n\t\tmo, err := o.f.NewObject(generateMetadataName(o.Remote()))\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\to.meta = readMetadata(mo)\n\t}\n\tif ht&hash.MD5 == 0 {\n\t\treturn \"\", hash.ErrUnsupported\n\t}\n\treturn hex.EncodeToString(o.meta.Hash), nil\n}", "func (h *Header) Hash() [32]byte {\n\tvar f []string\n\tif h.Description.Value != \"\" {\n\t\tf = append(f, h.Description.Value)\n\t}\n\tf = append(f, fmt.Sprint(h.Required.Value))\n\tf = append(f, fmt.Sprint(h.Deprecated.Value))\n\tf = append(f, fmt.Sprint(h.AllowEmptyValue.Value))\n\tif h.Style.Value != \"\" {\n\t\tf = append(f, h.Style.Value)\n\t}\n\tf = append(f, fmt.Sprint(h.Explode.Value))\n\tf = append(f, fmt.Sprint(h.AllowReserved.Value))\n\tif h.Schema.Value != nil {\n\t\tf = append(f, low.GenerateHashString(h.Schema.Value))\n\t}\n\tif h.Example.Value != nil {\n\t\tf = append(f, fmt.Sprint(h.Example.Value))\n\t}\n\tif len(h.Examples.Value) > 0 {\n\t\tfor k := range h.Examples.Value {\n\t\t\tf = append(f, fmt.Sprintf(\"%s-%x\", k.Value, h.Examples.Value[k].Value.Hash()))\n\t\t}\n\t}\n\tif len(h.Content.Value) > 0 {\n\t\tfor k := range h.Content.Value {\n\t\t\tf = append(f, fmt.Sprintf(\"%s-%x\", k.Value, h.Content.Value[k].Value.Hash()))\n\t\t}\n\t}\n\tkeys := make([]string, len(h.Extensions))\n\tz := 0\n\tfor k := range h.Extensions {\n\t\tkeys[z] = fmt.Sprintf(\"%s-%x\", k.Value, sha256.Sum256([]byte(fmt.Sprint(h.Extensions[k].Value))))\n\t\tz++\n\t}\n\tsort.Strings(keys)\n\tf = append(f, keys...)\n\treturn sha256.Sum256([]byte(strings.Join(f, \"|\")))\n}", "func (l *LexerATNConfig) Hash() int {\n\tvar f int\n\tif l.passedThroughNonGreedyDecision {\n\t\tf = 1\n\t} else {\n\t\tf = 0\n\t}\n\th := murmurInit(7)\n\th = murmurUpdate(h, l.state.GetStateNumber())\n\th = murmurUpdate(h, l.alt)\n\th = murmurUpdate(h, l.context.Hash())\n\th = murmurUpdate(h, l.semanticContext.Hash())\n\th = murmurUpdate(h, f)\n\th = murmurUpdate(h, l.lexerActionExecutor.Hash())\n\th = murmurFinish(h, 6)\n\treturn h\n}", "func (bol Boolean) Hash() int {\n\tif bol {\n\t\treturn 1\n\t}\n\treturn 0\n}", "func (set *lalrSet) hash() (val uint32) {\n\t// Need hash to be order independent, so\n\t// just XOR everything.\n\tfor _, list := range(set.items) {\n\t\tfor _, item := range(list) {\n\t\t\tval = val ^ item.hash()\n\t\t}\n\t}\n\n\treturn\n}", "func (n *node) Hash() []byte {\n\treturn n.hash\n}", "func Hash(t *Token) (hash []byte) {\n var sum []byte\n\n // Compute the SHA1 sum of the Token\n {\n shasum := sha1.Sum([]byte(salt+string(*t)))\n copy(sum[:], shasum[:20])\n }\n\n // Encode the sum to hexadecimal\n hex.Encode(sum, sum)\n\n return\n}", "func (m MapEntry) Hash() uint32 {\n\treturn sequtil.Hash(m.key)\n}", "func hash(ls prometheus.Tags) uint64 {\n\tlbs := make(labels.Labels, 0, len(ls))\n\tfor k, v := range ls {\n\t\tlbs = append(lbs, labels.Label{\n\t\t\tName: k,\n\t\t\tValue: v,\n\t\t})\n\t}\n\n\tsort.Slice(lbs[:], func(i, j int) bool {\n\t\treturn lbs[i].Name < lbs[j].Name\n\t})\n\n\treturn lbs.Hash()\n}", "func (obj *object) Hash() int {\n\treturn obj.hash\n}", "func (c *ColumnValue) Hash() uint64 {\n\tif c == nil {\n\t\treturn cache.NewHash(FragmentType_ColumnValue, nil)\n\t}\n\treturn cache.NewHash(FragmentType_ColumnValue, c.Column, c.Operator, c.Value)\n}", "func (in *Instance) hash(x, y, mu *big.Int, T uint64) *big.Int {\n\tb := sha512.New()\n\tb.Write(x.Bytes())\n\tb.Write(y.Bytes())\n\tb.Write(mu.Bytes())\n\tbits := make([]byte, 8)\n\tbinary.LittleEndian.PutUint64(bits, T)\n\tb.Write(bits)\n\tres := new(big.Int).SetBytes(b.Sum(nil))\n\tres.Mod(res, in.rsaModulus)\n\treturn res\n}", "func (s *Sub) Hash() (uint32, error) {\n\th := adler32.New()\n\ts.Hash32(h)\n\treturn h.Sum32(), nil\n}", "func (s SampleList) Hash(i int) []byte {\n\tres := md5.Sum(s[i])\n\treturn res[:]\n}", "func (n Number) Hash() int {\n\tf, err := json.Number(n).Float64()\n\tif err != nil {\n\t\tbs := []byte(n)\n\t\th := xxhash.Checksum64(bs)\n\t\treturn int(h)\n\t}\n\treturn int(f)\n}", "func (source *Source) Hash() int {\n\tvar hash int\n\n\tif len(source.Prefix) > 0 {\n\t\tfor _, b := range source.Prefix {\n\t\t\thash = int(b*31) + hash\n\t\t}\n\t}\n\n\thash = int(source.PrefixLen*31) + hash\n\thash = int(source.RouterId*31) + hash\n\n\treturn hash\n}", "func (r *Restriction) hash() ([]byte, error) {\n\tj, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\treturn hashUtils.SHA512(j), nil\n}", "func (i *Instance) Hash(extraBytes []byte) (string, error) {\n\t//nolint:gosec // not being used for secure purposes\n\th := sha1.New()\n\n\t// copy by value to ignore ETag without affecting i\n\ti2 := *i\n\ti2.ETag = \"\"\n\n\tinstanceBytes, err := bson.Marshal(i2)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif _, err := h.Write(append(instanceBytes, extraBytes...)); err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn fmt.Sprintf(\"%x\", h.Sum(nil)), nil\n}", "func (h *kustHash) Hash(m ifc.Kunstructured) (string, error) {\n\tu := unstructured.Unstructured{\n\t\tObject: m.Map(),\n\t}\n\tkind := u.GetKind()\n\tswitch kind {\n\tcase \"ConfigMap\":\n\t\tcm, err := unstructuredToConfigmap(u)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn configMapHash(cm)\n\tcase \"Secret\":\n\t\tsec, err := unstructuredToSecret(u)\n\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn secretHash(sec)\n\tdefault:\n\t\treturn \"\", fmt.Errorf(\n\t\t\t\"type %s is not supported for hashing in %v\",\n\t\t\tkind, m.Map())\n\t}\n}", "func (p Primitive) Hash() string {\n\treturn p.Name()\n}", "func (d Data32) Hash() Hash {\n\treturn hash(d)\n}", "func (n *notifier) hash(other *memberlist.Node) uint64 {\n\treturn uint64(murmur.Murmur3([]byte(other.Name), murmur.M3Seed))\n}", "func encodeHash(x uint64, p, pPrime uint) (hashCode uint64) {\n\tif x&onesFromTo(64-pPrime, 63-p) == 0 {\n\t\tr := rho(extractShift(x, 0, 63-pPrime))\n\t\treturn concat([]concatInput{\n\t\t\t{x, 64 - pPrime, 63},\n\t\t\t{uint64(r), 0, 5},\n\t\t\t{1, 0, 0}, // this just adds a 1 bit at the end\n\t\t})\n\t} else {\n\t\treturn concat([]concatInput{\n\t\t\t{x, 64 - pPrime, 63},\n\t\t\t{0, 0, 0}, // this just adds a 0 bit at the end\n\t\t})\n\t}\n}", "func (obj *set) Hash() hash.Hash {\n\treturn obj.hash\n}", "func (h *MemHash) Hash() uint32 {\n\tss := (*stringStruct)(unsafe.Pointer(&h.buf))\n\treturn uint32(memhash(ss.str, 0, uintptr(ss.len)))\n}", "func (d Data256) Hash() Hash {\n\treturn hash(d)\n}", "func hash(values ...[]byte) ([]byte, error) {\n\th := swarm.NewHasher()\n\tfor _, v := range values {\n\t\t_, err := h.Write(v)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn h.Sum(nil), nil\n}", "func (c Call) Hash() int {\n\treturn termSliceHash(c)\n}", "func hash(stav Stav) uint64{\n\tstr := \"\"\n\n\tfor i := 0; i < len(stav.Auta); i++ {\n\t\tstr += stav.Auta[i].Farba\n\t\tstr += strconv.Itoa(int(stav.Auta[i].X))\n\t\tstr += strconv.Itoa(int(stav.Auta[i].Y))\n\t\tstr += strconv.FormatBool(stav.Auta[i].Smer)\n\t\tstr += strconv.Itoa(int(stav.Auta[i].Dlzka))\n\t}\n\n\th := fnv.New64a()\n\th.Write([]byte(str))\n\treturn h.Sum64()\n\n}", "func (gdt *Array) Hash() Int {\n\targ0 := gdt.getBase()\n\n\tret := C.go_godot_array_hash(GDNative.api, arg0)\n\n\treturn Int(ret)\n}", "func Hash(b []byte, seed uint64) uint64", "func (t *Table) hash(s string) int {\n\t// Good enough.\n\th := fnv.New32()\n\th.Write([]byte(s))\n\treturn int(h.Sum32()) % t.m\n}", "func (e EmptyNode) Hash() util.Uint256 {\n\tpanic(\"can't get hash of an EmptyNode\")\n}", "func (o *ExportData) Hash() string {\n\targs := make([]interface{}, 0)\n\targs = append(args, o.CustomerID)\n\targs = append(args, o.ID)\n\targs = append(args, o.IntegrationInstanceID)\n\targs = append(args, o.JobID)\n\targs = append(args, o.Objects)\n\targs = append(args, o.RefID)\n\targs = append(args, o.RefType)\n\to.Hashcode = hash.Values(args...)\n\treturn o.Hashcode\n}", "func hash(k Key) int {\n\tkey := fmt.Sprintf(\"%s\", k)\n\th := 0\n\tfor i := 0; i < len(key); i++ {\n\t\th = 31 * h + int(key[i])\n\t}\n\treturn h\n}", "func (t *Target) hash() uint64 {\n\th := fnv.New64a()\n\n\t//nolint: errcheck\n\th.Write([]byte(fmt.Sprintf(\"%016d\", t.labels.Hash())))\n\t//nolint: errcheck\n\th.Write([]byte(t.URL().String()))\n\n\treturn h.Sum64()\n}", "func hash(m datasource.Metric) uint64 {\n\thash := fnv.New64a()\n\tlabels := m.Labels\n\tsort.Slice(labels, func(i, j int) bool {\n\t\treturn labels[i].Name < labels[j].Name\n\t})\n\tfor _, l := range labels {\n\t\t// drop __name__ to be consistent with Prometheus alerting\n\t\tif l.Name == \"__name__\" {\n\t\t\tcontinue\n\t\t}\n\t\thash.Write([]byte(l.Name))\n\t\thash.Write([]byte(l.Value))\n\t\thash.Write([]byte(\"\\xff\"))\n\t}\n\treturn hash.Sum64()\n}", "func (t *smallFlatTable) Hash() hash.Hash { return t.hash }", "func (obj *chunk) Hash() hash.Hash {\n\treturn obj.immutable.Hash()\n}", "func (d Data) Hash() Hash {\n\treturn hash(d)\n}", "func (dtk *DcmTagKey) Hash() uint32 {\n\treturn ((uint32(int(dtk.group)<<16) & 0xffff0000) | (uint32(int(dtk.element) & 0xffff)))\n}", "func Hash(seed maphash.Seed, k Key) uint64 {\n\tvar buf [8]byte\n\tswitch v := k.(type) {\n\tcase mapKey:\n\t\treturn hashMapKey(seed, v)\n\tcase interfaceKey:\n\t\ts := v.Hash()\n\t\t// Mix up the hash to ensure it covers 64-bits\n\t\tbinary.LittleEndian.PutUint64(buf[:8], uint64(s))\n\t\treturn hashBytes(seed, buf[:8])\n\tcase strKey:\n\t\treturn hashString(seed, string(v))\n\tcase bytesKey:\n\t\treturn hashBytes(seed, []byte(v))\n\tcase int8Key:\n\t\tbuf[0] = byte(v)\n\t\treturn hashBytes(seed, buf[:1])\n\tcase int16Key:\n\t\tbinary.LittleEndian.PutUint16(buf[:2], uint16(v))\n\t\treturn hashBytes(seed, buf[:2])\n\tcase int32Key:\n\t\tbinary.LittleEndian.PutUint32(buf[:4], uint32(v))\n\t\treturn hashBytes(seed, buf[:4])\n\tcase int64Key:\n\t\tbinary.LittleEndian.PutUint64(buf[:8], uint64(v))\n\t\treturn hashBytes(seed, buf[:8])\n\tcase uint8Key:\n\t\tbuf[0] = byte(v)\n\t\treturn hashBytes(seed, buf[:1])\n\tcase uint16Key:\n\t\tbinary.LittleEndian.PutUint16(buf[:2], uint16(v))\n\t\treturn hashBytes(seed, buf[:2])\n\tcase uint32Key:\n\t\tbinary.LittleEndian.PutUint32(buf[:4], uint32(v))\n\t\treturn hashBytes(seed, buf[:4])\n\tcase uint64Key:\n\t\tbinary.LittleEndian.PutUint64(buf[:8], uint64(v))\n\t\treturn hashBytes(seed, buf[:8])\n\tcase float32Key:\n\t\tbinary.LittleEndian.PutUint32(buf[:4], math.Float32bits(float32(v)))\n\t\treturn hashBytes(seed, buf[:4])\n\tcase float64Key:\n\t\tbinary.LittleEndian.PutUint64(buf[:8], math.Float64bits(float64(v)))\n\t\treturn hashBytes(seed, buf[:8])\n\tcase boolKey:\n\t\tif v {\n\t\t\tbuf[0] = 1\n\t\t}\n\t\treturn hashBytes(seed, buf[:1])\n\tcase sliceKey:\n\t\treturn hashSliceKey(seed, v)\n\tcase pointerKey:\n\t\treturn hashSliceKey(seed, v.sliceKey)\n\tcase pathKey:\n\t\treturn hashSliceKey(seed, v.sliceKey)\n\tcase nilKey:\n\t\treturn hashBytes(seed, nil)\n\tcase Hashable:\n\t\t// Mix up the hash to ensure it covers 64-bits\n\t\tbinary.LittleEndian.PutUint64(buf[:8], v.Hash())\n\t\treturn hashBytes(seed, buf[:8])\n\tdefault:\n\t\ts := _nilinterhash(v.Key())\n\t\tbinary.LittleEndian.PutUint64(buf[:8], uint64(s))\n\t\treturn hashBytes(seed, buf[:8])\n\t}\n}", "func Hash(mem []byte) uint64 {\n\tvar hash uint64 = 5381\n\tfor _, b := range mem {\n\t\thash = (hash << 5) + hash + uint64(b)\n\t}\n\treturn hash\n}", "func (spec Spec) DeepHash() string {\n\thash := sha512.New512_224()\n\tspec.DefaultService.hash(hash)\n\tfor _, rule := range spec.Rules {\n\t\trule.hash(hash)\n\t}\n\tsvcs := make([]string, len(spec.AllServices))\n\ti := 0\n\tfor k := range spec.AllServices {\n\t\tsvcs[i] = k\n\t\ti++\n\t}\n\tsort.Strings(svcs)\n\tfor _, svc := range svcs {\n\t\thash.Write([]byte(svc))\n\t\tspec.AllServices[svc].hash(hash)\n\t}\n\tspec.ShardCluster.hash(hash)\n\thash.Write([]byte(spec.VCL))\n\tfor _, auth := range spec.Auths {\n\t\tauth.hash(hash)\n\t}\n\tfor _, acl := range spec.ACLs {\n\t\tacl.hash(hash)\n\t}\n\tfor _, rw := range spec.Rewrites {\n\t\trw.hash(hash)\n\t}\n\tfor _, reqDisp := range spec.Dispositions {\n\t\treqDisp.hash(hash)\n\t}\n\th := new(big.Int)\n\th.SetBytes(hash.Sum(nil))\n\treturn h.Text(62)\n}" ]
[ "0.7138284", "0.7119623", "0.7041903", "0.700204", "0.7000892", "0.69765645", "0.6946901", "0.69159335", "0.69111687", "0.6884624", "0.68813616", "0.6854573", "0.6839519", "0.68087643", "0.68017894", "0.6779525", "0.6749075", "0.67478395", "0.6744907", "0.6727617", "0.66851336", "0.6683878", "0.66792506", "0.6667716", "0.6653189", "0.66493165", "0.66328806", "0.6597665", "0.6597063", "0.6596404", "0.6592299", "0.6580989", "0.6574767", "0.65641445", "0.6557275", "0.6555275", "0.65539527", "0.6548095", "0.6542581", "0.6530705", "0.65281284", "0.65240616", "0.65198684", "0.6508872", "0.6507938", "0.6487107", "0.6486554", "0.6477679", "0.647389", "0.6473444", "0.64730334", "0.6470614", "0.64632285", "0.6462884", "0.6446518", "0.6442148", "0.64393836", "0.6434532", "0.6423657", "0.64192003", "0.6406657", "0.64000094", "0.6399469", "0.639899", "0.6398343", "0.6398244", "0.6397757", "0.63953525", "0.6391269", "0.6382421", "0.63788843", "0.63755345", "0.6375003", "0.63731766", "0.6359296", "0.6357537", "0.6354525", "0.634988", "0.6342756", "0.6339535", "0.6333529", "0.63264585", "0.6322571", "0.63204396", "0.63193166", "0.6316662", "0.6316102", "0.63160074", "0.63152075", "0.63105017", "0.6309715", "0.6307885", "0.63074785", "0.63059074", "0.630408", "0.63034046", "0.62977403", "0.62975895", "0.6296842", "0.62967694" ]
0.6400822
61
starlarkPath parses Starlark kw/args and returns a corresponding `Path`
func starlarkPath( args starlark.Tuple, kwargs []starlark.Tuple, ) (starlark.Value, error) { if len(args) != 1 { return nil, errors.Errorf( "Expected exactly 1 positional argument; found %d", len(args), ) } if len(kwargs) != 0 { return nil, errors.Errorf( "Expected exactly 0 positional arguments; found %d", len(kwargs), ) } if s, ok := args[0].(starlark.String); ok { return Path(s), nil } return nil, errors.Errorf( "TypeError: Expected a string argument; found %s", args[0].Type(), ) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (j *Js) Getpath(args ...string) *Js {\n\td := j\n\tfor i := range args {\n\t\tm := d.Getdata()\n\n\t\tif val, ok := m[args[i]]; ok {\n\t\t\td.data = val\n\t\t} else {\n\t\t\td.data = nil\n\t\t\treturn d\n\t\t}\n\t}\n\treturn d\n}", "func extractPath(c *gin.Context, tag string) (string, []string, error) {\n\tname, required, defaultVal, err := parseTagKey(tag)\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\tp := c.Param(name)\n\n\t// XXX: deprecated, use of \"default\" tag is preferred\n\tif p == \"\" && defaultVal != \"\" {\n\t\treturn name, []string{defaultVal}, nil\n\t}\n\t// XXX: deprecated, use of \"validate\" tag is preferred\n\tif p == \"\" && required {\n\t\treturn \"\", nil, fmt.Errorf(\"missing path parameter: %s\", name)\n\t}\n\n\treturn name, []string{p}, nil\n}", "func (r *BasicRequest) PathArgs() (map[string]string, error) {\n\treturn r.Path, nil\n}", "func path(stmt semantic.Node) snippets.Pathway {\n\tswitch expr := stmt.(type) {\n\tcase *semantic.Global:\n\t\treturn snippets.Variable(snippets.SymbolCategory_Global, expr.Name())\n\tcase *semantic.Parameter:\n\t\treturn snippets.Variable(snippets.SymbolCategory_Parameter, expr.Name())\n\tcase *semantic.Local:\n\t\treturn snippets.Variable(snippets.SymbolCategory_Local, expr.Name())\n\tcase *semantic.Member:\n\t\treturn snippets.Field(path(expr.Object), expr.Field.Name())\n\tcase *semantic.PointerRange:\n\t\treturn snippets.Range(path(expr.Pointer))\n\tcase *semantic.SliceRange:\n\t\t// Don't change the path, since range of slice is still a slice.\n\t\treturn path(expr.Slice)\n\tcase *semantic.ArrayIndex:\n\t\treturn snippets.Elem(path(expr.Array))\n\tcase *semantic.SliceIndex:\n\t\treturn snippets.Elem(path(expr.Slice))\n\tcase *semantic.MapIndex:\n\t\treturn snippets.Elem(path(expr.Map))\n\tcase *semantic.Observed:\n\t\treturn path(expr.Parameter)\n\tcase *semantic.Cast:\n\t\treturn path(expr.Object)\n\tdefault:\n\t\tpanic(fmt.Errorf(\"Unexpect path expression %T:%v\", stmt, stmt))\n\t}\n}", "func ParsePath(raw storj.Path) (path Path) {\n\t// A path may contain a bucket and an unencrypted path.\n\tparts := strings.SplitN(raw, \"/\", 2)\n\tpath.bucket = parts[0]\n\tif len(parts) > 1 {\n\t\tpath.unencPath = paths.NewUnencrypted(parts[1])\n\t}\n\tpath.raw = []byte(raw)\n\treturn path\n}", "func (q Query) Path(args ...string) string {\n\ttemp := []string{args[0], q.Key}\n\targs = append(temp, args[1:]...)\n\treturn fmt.Sprintf(\"custom/%s\", strings.Join(args, \"/\"))\n}", "func Parse(rawpath string) (Path, error) {\n\tif err := validateRawPath(rawpath); err != nil {\n\t\treturn Path{}, err\n\t}\n\trootName := root(rawpath)\n\t// package name cannot contain \"-\" so gracefully remove them\n\t// if they present.\n\tpackageName := stripNonAlphaNumeric(rootName)\n\tif err := validatePackageName(packageName); err != nil {\n\t\treturn Path{}, err\n\t}\n\tp := Path{\n\t\tRawPath: rawpath,\n\t\tRoot: rootName,\n\t\tPackage: packageName,\n\t}\n\n\treturn p, nil\n}", "func (o *OpenAPI3) ParsePath(ctx context.Context, swPath *openapi3.PathItem, opts *OpenAPI3Options) (*spec.Path, error) {\n\n\tpath := &spec.Path{\n\t\tDescription: swPath.Description,\n\t}\n\n\tvar ext OpenAPI3PathExtension\n\terr := o.GetExtension(opts.ExtensionName, swPath.Extensions, &ext)\n\tif err != nil && err != ErrExtNotFound {\n\t\treturn nil, err\n\t}\n\n\tif ext.Name != nil && *ext.Name != \"\" {\n\t\tpath.Name = *ext.Name\n\t}\n\n\t// Parse each operation individually\n\tfor method, op := range swPath.Operations() {\n\t\tspecOp, err := o.ParseOperation(ctx, op, opts)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tspecOp.Method = method\n\t\tpath.Operations = append(path.Operations, specOp)\n\t}\n\n\t// We also need to add the parameters defined\n\t// on the path to all the operations\n\tfor _, p := range swPath.Parameters {\n\t\tif p.Value == nil {\n\t\t\tcontinue\n\t\t}\n\t\tparams, err := o.ParseParameter(ctx, p, opts)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor _, op := range path.Operations {\n\t\t\top.Parameters = append(op.Parameters, params...)\n\t\t}\n\t}\n\n\treturn path, nil\n}", "func (hc *HealthCheckArgsOrString) Path() *string {\n\tif hc.IsBasic() {\n\t\treturn aws.String(hc.Basic)\n\t}\n\treturn hc.Advanced.Path\n}", "func (jm JSONMeta) Path() string {\n\tsb := make([]string, 0)\n\tvar c JSONMetaNode = jm\n\tfor c != nil {\n\t\tsb = append([]string{c.Key()}, sb...)\n\t\t// Prepend a \".\" for non-index segments.\n\t\tif _, ok := c.Parent().(JSONMetaContainerNode); ok {\n\t\t\tsb = append([]string{\".\"}, sb...)\n\t\t}\n\t\tc = c.Parent()\n\t}\n\n\treturn strings.TrimLeft(strings.Join(sb, \"\"), \".\")\n}", "func parseTwirpPath(path string) (string, string, string) {\n\tparts := strings.Split(path, \"/\")\n\tif len(parts) < 2 {\n\t\treturn \"\", \"\", \"\"\n\t}\n\tmethod := parts[len(parts)-1]\n\tpkgService := parts[len(parts)-2]\n\tprefix := strings.Join(parts[0:len(parts)-2], \"/\")\n\treturn prefix, pkgService, method\n}", "func parseTwirpPath(path string) (string, string, string) {\n\tparts := strings.Split(path, \"/\")\n\tif len(parts) < 2 {\n\t\treturn \"\", \"\", \"\"\n\t}\n\tmethod := parts[len(parts)-1]\n\tpkgService := parts[len(parts)-2]\n\tprefix := strings.Join(parts[0:len(parts)-2], \"/\")\n\treturn prefix, pkgService, method\n}", "func Path(path string) PathRetriever {\n\treturn func() (string, error) { return path, nil }\n}", "func parsePath(path string) (root string) {\n\troot = strings.Trim(path, \"/\")\n\treturn\n}", "func PathForKey(raw string) paths.Unencrypted {\n\treturn paths.NewUnencrypted(strings.TrimSuffix(raw, \"/\"))\n}", "func (s *Nap) Path(path string) *Nap {\n\tbaseURL, baseErr := url.Parse(s.rawURL)\n\tpathURL, pathErr := url.Parse(path)\n\tif baseErr == nil && pathErr == nil {\n\t\ts.rawURL = baseURL.ResolveReference(pathURL).String()\n\t\tif strings.HasSuffix(path, \"/\") && !strings.HasSuffix(s.rawURL, \"/\") {\n\t\t\ts.rawURL += \"/\"\n\t\t}\n\t\treturn s\n\t}\n\treturn s\n}", "func parsePattern(path string, stash Stash) string {\n\t// Standard placeholders\n\t// XXX: Add relaxed and wildcard placeholders\n\t// XXX: Add restricted placeholders\n\tpathPattern := \"\"\n\tlastIndex := 0\n\tfor _, v := range stdPlaceholder.FindAllStringSubmatchIndex(path, -1) {\n\t\t// v is an array of pairs of ints. Each pair is start and end\n\t\t// indexes of the match in the string. The first pair is the\n\t\t// entire match, and other pairs are the corresponding\n\t\t// submatches\n\t\tgap := path[lastIndex:v[0]]\n\t\tlastIndex = v[1]\n\n\t\tstart := path[v[2]:v[3]]\n\t\tif start != \"/\" {\n\t\t\tstart = \"\"\n\t\t}\n\n\t\t//placeType := path[v[4]:v[5]]\n\t\tplaceName := path[v[6]:v[7]]\n\t\t//end := path[v[8]:v[9]] // unused\n\n\t\tmatchType := \"+\" // required\n\t\tif _, ok := stash[placeName]; ok {\n\t\t\tmatchType = \"*\" // optional\n\t\t\tif start == \"/\" {\n\t\t\t\tstart += \"?\"\n\t\t\t}\n\t\t}\n\n\t\tpathPattern += gap + fmt.Sprintf(\"%s(?P<%s>[^/.]%s)\", start, placeName, matchType)\n\t}\n\n\t// If we never matched, there were no placeholders\n\tif pathPattern == \"\" {\n\t\treturn path\n\t}\n\n\treturn pathPattern\n}", "func ParsePath(slashSeparated string) Path {\n\treturn Path(strings.Split(slashSeparated, \"/\"))\n}", "func (r *Request) pathString() {\n\tr.RequestCode.PathString()\n}", "func New(path string) Path {\n\tinSegments := strings.Split(path, \"/\")\n\ttrailing := inSegments[len(inSegments)-1] == \"*\"\n\n\tvar outSegments []Segment\n\tif trailing {\n\t\toutSegments = make([]Segment, len(inSegments)-1)\n\t} else {\n\t\toutSegments = make([]Segment, len(inSegments))\n\t}\n\n\tfor i := 0; i < len(outSegments); i++ {\n\t\tif strings.HasPrefix(inSegments[i], \":\") {\n\t\t\tif strings.Contains(inSegments[i], \"?\") {\n\t\t\t\toutSegments[i] = Segment{\n\t\t\t\t\tIsParam: true,\n\t\t\t\t\tIsOptional: true,\n\t\t\t\t\tParam: inSegments[i][1 : len(inSegments[i])-1],\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\toutSegments[i] = Segment{IsParam: true, Param: inSegments[i][1:]}\n\t\t} else {\n\t\t\toutSegments[i] = Segment{IsParam: false, Const: inSegments[i]}\n\t\t}\n\t}\n\n\treturn Path{Segments: outSegments, Trailing: trailing}\n}", "func getPathForMeta(metaname string) string {\n\tfor _, mp := range metapaths {\n\t\tif strings.EqualFold(mp.name, metaname) {\n\t\t\treturn mp.path\n\t\t}\n\t}\n\n\t// Unknown, so use metaname\n\treturn metaname\n}", "func (r InboundRequest) PathString(key string) (string, bool) {\n s, err := r.PathParams.String(key)\n return s, err == nil\n}", "func (this *JsonDecode) Path(path string) *JsonDecode {\n\tthis.fields = strings.Split(path, \".\")\n\tthis.fieldLen = len(this.fields)\n\n\treturn this\n}", "func (r *Paths) Path(handler interface{}) []string {\n\t// перебираем статические пути\n\tfor url, h := range r.static {\n\t\tif h == handler {\n\t\t\treturn splitter(url)\n\t\t}\n\t}\n\t// перебираем все пути с параметрами\n\tfor _, records := range r.fields {\n\t\tfor _, record := range records {\n\t\t\t// сравниваем адреса методов\n\t\t\tif handler == record.handler {\n\t\t\t\treturn record.parts\n\t\t\t}\n\t\t}\n\t}\n\treturn nil // данный обработчик не зарегистрирован\n}", "func ParsePath(p string) Path {\n\t// TODO use delim() here\n\tsegs := strings.Split(strings.Trim(p, \" /\"), \"/\")\n\tis := make([]Value, len(segs))\n\tfor i := range segs {\n\t\tis[i] = []byte(segs[i])\n\t}\n\treturn is\n}", "func (j JSONRPCRequest) Path() string {\n\t// create a standard folder structure\n\tsegments := []string{j.Method}\n\n\tfor _, s := range j.Params {\n\t\tstr := fmt.Sprintf(\"%v\", s)\n\t\tsegments = append(segments, str)\n\t}\n\tfolder := filepath.Join(segments...)\n\treturn folder\n}", "func PathToAttributePath(p cty.Path) *tftypes.AttributePath {\n\tif p == nil || len(p) < 1 {\n\t\treturn nil\n\t}\n\tap := tftypes.NewAttributePath()\n\tfor _, step := range p {\n\t\tswitch selector := step.(type) {\n\t\tcase cty.GetAttrStep:\n\t\t\tap = ap.WithAttributeName(selector.Name)\n\n\t\tcase cty.IndexStep:\n\t\t\tkey := selector.Key\n\t\t\tswitch key.Type() {\n\t\t\tcase cty.String:\n\t\t\t\tap = ap.WithElementKeyString(key.AsString())\n\t\t\tcase cty.Number:\n\t\t\t\tv, _ := key.AsBigFloat().Int64()\n\t\t\t\tap = ap.WithElementKeyInt(int(v))\n\t\t\tdefault:\n\t\t\t\t// We'll bail early if we encounter anything else, and just\n\t\t\t\t// return the valid prefix.\n\t\t\t\treturn ap\n\t\t\t}\n\t\t}\n\t}\n\treturn ap\n}", "func (o CustomHttpPatternOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v CustomHttpPattern) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func NewPath(root, doc string) *restful.WebService {\n\tws := new(restful.WebService)\n\tws.Path(root).\n\t\tDoc(doc).\n\t\tConsumes(restful.MIME_JSON).\n\t\tProduces(restful.MIME_JSON)\n\n\treturn ws\n}", "func starlarkTarget(\n\targs starlark.Tuple,\n\tkwargs []starlark.Tuple,\n) (starlark.Value, error) {\n\t// For the sake of simpler parsing, we'll simply require that all args are\n\t// passed as kwargs (no positional args).\n\tif len(args) != 0 {\n\t\treturn nil, errors.Errorf(\n\t\t\t\"Expected 0 positional args; found %d\",\n\t\t\tlen(args),\n\t\t)\n\t}\n\n\t// Make sure we have exactly the right number of keyword arguments.\n\tif len(kwargs) != 4 {\n\t\tfound := make([]string, len(kwargs))\n\t\tfor i, kwarg := range kwargs {\n\t\t\tfound[i] = string(kwarg[0].(starlark.String))\n\t\t}\n\t\treturn nil, errors.Errorf(\n\t\t\t\"Expected kwargs {name, builder, args, env}; found {%s}\",\n\t\t\tstrings.Join(found, \", \"),\n\t\t)\n\t}\n\n\t// Iterate through the keyword arguments and grab the values for each\n\t// kwarg, putting them into the right `starlark.Value` variable. We'll\n\t// convert these to Go values for the `*Target` struct later.\n\tvar nameKwarg, builderKwarg, argsKwarg, envKwarg starlark.Value\n\tfor _, kwarg := range kwargs {\n\t\tswitch key := kwarg[0].(starlark.String); key {\n\t\tcase \"name\":\n\t\t\tif nameKwarg != nil {\n\t\t\t\treturn nil, errors.Errorf(\"Duplicate argument 'name' found\")\n\t\t\t}\n\t\t\tnameKwarg = kwarg[1]\n\t\tcase \"builder\":\n\t\t\tif builderKwarg != nil {\n\t\t\t\treturn nil, errors.Errorf(\"Duplicate argument 'builder' found\")\n\t\t\t}\n\t\t\tbuilderKwarg = kwarg[1]\n\t\tcase \"args\":\n\t\t\tif argsKwarg != nil {\n\t\t\t\treturn nil, errors.Errorf(\"Duplicate argument 'args' found\")\n\t\t\t}\n\t\t\targsKwarg = kwarg[1]\n\t\tcase \"env\":\n\t\t\tif envKwarg != nil {\n\t\t\t\treturn nil, errors.Errorf(\"Duplicate argument 'env' found\")\n\t\t\t}\n\t\t\tenvKwarg = kwarg[1]\n\t\tdefault:\n\t\t\treturn nil, errors.Errorf(\"Unexpected argument '%s' found\", key)\n\t\t}\n\t}\n\n\t// Ok, now we've made sure we have values for the required keyword args and\n\t// that no additional arguments were passed. Next, we'll convert these\n\t// `starlark.Value`-typed variables into Go values for the output `*Target`\n\t// struct.\n\n\t// Validate that the `name` kwarg was a string.\n\tname, ok := nameKwarg.(starlark.String)\n\tif !ok {\n\t\treturn nil, errors.Errorf(\n\t\t\t\"TypeError: argument 'name': expected str, got %s\",\n\t\t\tnameKwarg.Type(),\n\t\t)\n\t}\n\n\t// Validate that the `builder` kwarg was a string.\n\tbuilder, ok := builderKwarg.(starlark.String)\n\tif !ok {\n\t\treturn nil, errors.Errorf(\n\t\t\t\"TypeError: argument 'builder': expected str, got %s\",\n\t\t\tbuilderKwarg.Type(),\n\t\t)\n\t}\n\n\t// Validate that the `args` kwarg was a list of `Arg`s, and convert it\n\t// into a `[]Arg` for the `Target.Args` field.\n\targsSL, ok := argsKwarg.(*starlark.List)\n\tif !ok {\n\t\treturn nil, errors.Errorf(\n\t\t\t\"TypeError: argument 'args': expected list, got %s\",\n\t\t\targsKwarg.Type(),\n\t\t)\n\t}\n\targs_ := make([]Arg, argsSL.Len())\n\tfor i := range args_ {\n\t\targ, err := starlarkValueToArg(argsSL.Index(i))\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"Argument 'args[%d]'\", i)\n\t\t}\n\t\targs_[i] = arg\n\t}\n\n\t// Validate that the `env` kwarg was a list of strings, and convert it into\n\t// a `[]string` for the `Target.Env` field.\n\tenvSL, ok := envKwarg.(*starlark.List)\n\tif !ok {\n\t\treturn nil, errors.Errorf(\n\t\t\t\"TypeError: argument 'env': expected list, got %s\",\n\t\t\tenvKwarg.Type(),\n\t\t)\n\t}\n\tenv := make([]string, envSL.Len())\n\tfor i := range env {\n\t\tstr, ok := envSL.Index(i).(starlark.String)\n\t\tif !ok {\n\t\t\treturn nil, errors.Errorf(\n\t\t\t\t\"TypeError: argument 'env[%d]': expected string; found %s\",\n\t\t\t\ti,\n\t\t\t\tenvSL.Index(i).Type(),\n\t\t\t)\n\t\t}\n\t\tenv[i] = string(str)\n\t}\n\n\t// By now, all of the fields have been validated, so build and return the\n\t// final `*Target`.\n\treturn &Target{\n\t\tName: string(name),\n\t\tBuilder: string(builder),\n\t\tArgs: args_,\n\t\tEnv: env,\n\t}, nil\n}", "func (p *v1Provider) Path(elements ...string) string {\n\tparts := []string{\n\t\tstrings.TrimSuffix(p.Cluster.Config.CatalogURL, \"/\"),\n\t\t\"v1\",\n\t}\n\tparts = append(parts, elements...)\n\treturn strings.Join(parts, \"/\")\n}", "func (rule *Rule) Path(ns Ref) Ref {\n\treturn ns.Append(StringTerm(string(rule.Name)))\n}", "func (s *SVG) Path(str string, args map[string]interface{}) {\n\tpathStr := fmt.Sprintf(\"<path d='%s' %s />\", str, s.WriteArgs(args))\n\ts.svgString += pathStr\n}", "func (c classy) Path(path string) Classy {\n\tc.path = path\n\treturn c\n}", "func parsePath(path string, r *maps.SpeedLimitsRequest) {\n\tif path != \"\" {\n\t\tls := strings.Split(path, \"|\")\n\t\tfor _, l := range ls {\n\t\t\tll, err := maps.ParseLatLng(l)\n\t\t\tcheck(err)\n\t\t\tr.Path = append(r.Path, ll)\n\t\t}\n\t}\n}", "func ConstructAtomPath(basePath string, skip int, top int) string {\n\tvalues := url.Values{}\n\n\tif skip > 0 {\n\t\tvalues.Add(\"$skip\", fmt.Sprintf(\"%d\", skip))\n\t}\n\n\tif top > 0 {\n\t\tvalues.Add(\"$top\", fmt.Sprintf(\"%d\", top))\n\t}\n\n\tif len(values) == 0 {\n\t\treturn basePath\n\t}\n\n\treturn fmt.Sprintf(\"%s?%s\", basePath, values.Encode())\n}", "func GetFromPath(c *gin.Context) {\n\tname := c.Param(\"name\")\n\tage := c.Param(\"age\")\n\n\tc.JSON(200, gin.H{\n\t\t\"name\": name,\n\t\t\"age\": age,\n\t})\n}", "func identPath(ident string, opts ...Option) *PathOptions {\n\tpath := &PathOptions{Ident: ident}\n\tfor i := range opts {\n\t\topts[i](path)\n\t}\n\treturn path\n}", "func NewPath(path string) (*Path, error) {\n\tvar newPath = &Path{\n\t\tpath: path,\n\t\ttrailingSlash: strings.HasSuffix(path, \"/\"),\n\t}\n\n\tvar current part\n\tvar param string\n\tvar err error\n\n\tfor len(path) > 1 {\n\t\tif path[0] != '/' {\n\t\t\treturn nil, fmt.Errorf(\"path %q corrupted\", path)\n\t\t}\n\t\tpath = path[1:]\n\n\t\tswitch path[0] {\n\t\tcase ':':\n\t\t\tcurrent = new(partParam)\n\t\t\tpath, param, err = current.read(path)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tnewPath.parts = append(newPath.parts, current)\n\n\t\tcase '$':\n\t\t\tcurrent = new(partRegex)\n\t\t\tpath, param, err = current.read(path)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tnewPath.parts = append(newPath.parts, current)\n\n\t\tcase '*':\n\t\t\tcurrent = new(partWildcard)\n\t\t\tpath, param, err = current.read(path)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tnewPath.parts = append(newPath.parts, current)\n\n\t\tdefault:\n\t\t\tcurrent = new(partFixed)\n\t\t\tpath, param, err = current.read(path)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tnewPath.parts = append(newPath.parts, current)\n\t\t}\n\n\t\tif param != \"\" {\n\t\t\tnewPath.params = append(newPath.params, param)\n\t\t}\n\t}\n\n\tsort.Strings(newPath.params)\n\n\treturn newPath, nil\n}", "func Parse(s string) Path {\n\treturn ParseSeparator(s, defaultSep)\n}", "func parsePathKey(path []byte) (k string, rem []byte, err error) {\n\tvar parsed = make([]byte, 0, SHORT_STRING_OPTIMIZED_CAP)\n\tvar c byte\n\n\trem = path\n\n\tfor {\n\t\tswitch {\n\t\tcase rem[0] == '\\\\' && rem[1] == 'u':\n\t\t\tutf8str := make([]byte, 0, SHORT_STRING_OPTIMIZED_CAP)\n\t\t\tutf8str, rem, err = parseUnicode(rem)\n\t\t\tfor _, c := range utf8str {\n\t\t\t\tparsed = append(parsed, c)\n\t\t\t}\n\t\t\tcontinue\n\n\t\tcase rem[0] == '\\\\' && rem[1] == '.':\n\t\t\tparsed = append(parsed, '.')\n\t\t\trem = rem[2:]\n\t\t\tcontinue\n\n\t\tcase rem[0] == '\\\\' && rem[1] == '[':\n\t\t\tparsed = append(parsed, '[')\n\t\t\trem = rem[2:]\n\t\t\tcontinue\n\n\t\tcase rem[0] == '\\\\' && rem[1] == ']':\n\t\t\tparsed = append(parsed, ']')\n\t\t\trem = rem[2:]\n\t\t\tcontinue\n\n\t\tcase rem[0] == '\\\\' && rem[1] == ';':\n\t\t\tparsed = append(parsed, ';')\n\t\t\trem = rem[2:]\n\t\t\tcontinue\n\n\t\tcase rem[0] == '\\\\' && rem[1] != 'u':\n\t\t\tc, rem, err = parseEscaped(rem)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tparsed = append(parsed, c)\n\t\t\tcontinue\n\n\t\tcase rem[0] == '.' || rem[0] == '[' || rem[0] == ';':\n\t\t\tgoto End\n\n\t\tdefault:\n\t\t\tparsed = append(parsed, rem[0])\n\t\t\tpos.col += 1\n\t\t\trem = rem[1:]\n\t\t\tcontinue\n\t\t}\n\t}\nEnd:\n\treturn string(parsed), rem, nil\n}", "func getPathSpec(ni *util.NodeInfo) (*pathSpec, error) {\n\tfor _, a := range ni.Annotation {\n\t\tif p, ok := a.(*pathSpec); ok {\n\t\t\treturn p, nil\n\t\t}\n\t}\n\treturn nil, fmt.Errorf(\"could not find path specification annotation\")\n}", "func (node *SourceInfoNode) GetPath(path []string) (*SourceInfoNode, error) {\n\tif len(path) == 0 {\n\t\treturn node, nil\n\t}\n\n\tkey := path[0]\n\tswitch node.body.Kind {\n\tcase yaml.MappingNode:\n\t\tchild, err := node.GetKey(key)\n\t\tif err != nil {\n\t\t\treturn node, err\n\t\t} else {\n\t\t\treturn child.GetPath(path[1:])\n\t\t}\n\tcase yaml.SequenceNode:\n\t\tindex, err := strconv.Atoi(key)\n\t\tif err != nil {\n\t\t\treturn node, err\n\t\t} else {\n\t\t\tchild, err := node.GetIndex(index)\n\t\t\tif err != nil {\n\t\t\t\treturn node, err\n\t\t\t} else {\n\t\t\t\treturn child.GetPath(path[1:])\n\t\t\t}\n\t\t}\n\tdefault:\n\t\treturn node, fmt.Errorf(\"Expected %s or %s at key %s but got %s\", prettyKind(yaml.MappingNode), prettyKind(yaml.SequenceNode), key, prettyKind(node.body.Kind))\n\t}\n}", "func (this *Value) Path(path string) (*Value, error) {\n\t// aliases always have priority\n\n\tif this.alias != nil {\n\t\tresult, ok := this.alias[path]\n\t\tif ok {\n\t\t\treturn result, nil\n\t\t}\n\t}\n\t// next we already parsed, used that\n\tswitch parsedValue := this.parsedValue.(type) {\n\tcase map[string]*Value:\n\t\tresult, ok := parsedValue[path]\n\t\tif ok {\n\t\t\treturn result, nil\n\t\t}\n\t}\n\t// finally, consult the raw bytes\n\tif this.raw != nil {\n\t\tres, err := jsonpointer.Find(this.raw, \"/\"+path)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif res != nil {\n\t\t\treturn NewValueFromBytes(res), nil\n\t\t}\n\t}\n\n\treturn nil, &Undefined{path}\n}", "func kpath(key, keytail string) string {\n\tl, t := len(key), len(keytail)\n\treturn key[:l-t]\n}", "func pgParsePath(a []byte) (out [][2][]byte, closed bool) {\n\tclosed = (a[0] == '(')\n\ta = a[1 : len(a)-1] // drop the surrounding parentheses or square brackets\n\treturn pgParsePointList(a), closed\n}", "func Path(w time.Duration, typ StatType, prefix []string) []string {\n\treturn stat{window: w, typ: typ}.metaPath(prefix)\n}", "func parse(pattern string) ([]string, error) {\n\tconst format = \"[get, post, put, patch, delete, view]/{a-Z}+/{*}+/{**}\"\n\n\tif pattern = strings.TrimLeft(strings.TrimSpace(pattern), delimiter); pattern == \"\" {\n\t\treturn nil, errors.Errorf(\"pattern illegal, should in format of %s\", format)\n\t}\n\n\tpaths := strings.Split(pattern, delimiter)\n\tif len(paths) < 2 {\n\t\treturn nil, errors.Errorf(\"pattern illegal, should in format of %s\", format)\n\t}\n\n\tfor i := range paths {\n\t\tpaths[i] = strings.TrimSpace(paths[i])\n\t}\n\n\t// likes get/ get/* get/**\n\tif len(paths) == 2 && (paths[1] == empty || paths[1] == fuzzy || paths[1] == omitted) {\n\t\treturn nil, errors.New(\"illegal wildcard\")\n\t}\n\n\tswitch paths[0] = strings.ToUpper(paths[0]); paths[0] {\n\tcase http.MethodGet,\n\t\thttp.MethodPost,\n\t\thttp.MethodPut,\n\t\thttp.MethodPatch,\n\t\thttp.MethodDelete,\n\t\tmethodView:\n\tdefault:\n\t\treturn nil, errors.Errorf(\"only supports [%s %s %s %s %s %s]\",\n\t\t\thttp.MethodGet, http.MethodPost, http.MethodPut, http.MethodPatch, http.MethodDelete, methodView)\n\t}\n\n\tfor k := 1; k < len(paths); k++ {\n\t\tif paths[k] == empty && k+1 != len(paths) {\n\t\t\treturn nil, errors.New(\"pattern contains illegal empty path\")\n\t\t}\n\n\t\tif paths[k] == omitted && k+1 != len(paths) {\n\t\t\treturn nil, errors.New(\"pattern contains illegal omitted path\")\n\t\t}\n\t}\n\n\treturn paths, nil\n}", "func ParseEncodedPath(p string) (Path, error) {\n\t// TODO use delim() here\n\tsegs := strings.Split(strings.Trim(p, \" /\"), \"/\")\n\tis := make([]Value, len(segs))\n\tfor i := range segs {\n\t\ts, err := url.QueryUnescape(segs[i])\n\t\tif err != nil {\n\t\t\treturn Path{}, err\n\t\t}\n\t\tis[i] = []byte(s)\n\t}\n\n\treturn is, nil\n}", "func jsonizePath(mi *Model, path string) string {\n\texprs := strings.Split(path, ExprSep)\n\texprs = jsonizeExpr(mi, exprs)\n\treturn strings.Join(exprs, ExprSep)\n}", "func AttributePathToPath(ap *tftypes.AttributePath) cty.Path {\n\tvar p cty.Path\n\tif ap == nil {\n\t\treturn p\n\t}\n\tfor _, step := range ap.Steps() {\n\t\tswitch step.(type) {\n\t\tcase tftypes.AttributeName:\n\t\t\tp = p.GetAttr(string(step.(tftypes.AttributeName)))\n\t\tcase tftypes.ElementKeyString:\n\t\t\tp = p.Index(cty.StringVal(string(step.(tftypes.ElementKeyString))))\n\t\tcase tftypes.ElementKeyInt:\n\t\t\tp = p.Index(cty.NumberIntVal(int64(step.(tftypes.ElementKeyInt))))\n\t\t}\n\t}\n\treturn p\n}", "func jsonizePath(mi *modelInfo, path string) string {\n\texprs := strings.Split(path, ExprSep)\n\texprs = jsonizeExpr(mi, exprs)\n\treturn strings.Join(exprs, ExprSep)\n}", "func (a *Api) Path(r *http.Request) string {\n\tif a.Prefix != \"\" {\n\t\tif strings.HasSuffix(a.Prefix, \"/\") {\n\t\t\ta.Prefix = strings.TrimSuffix(a.Prefix, \"/\")\n\t\t}\n\n\t\treturn strings.TrimPrefix(r.URL.Path, a.Prefix)\n\t}\n\treturn r.URL.Path\n}", "func (o CustomHttpPatternResponseOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CustomHttpPatternResponse) string { return v.Path }).(pulumi.StringOutput)\n}", "func getMetaForPath(path string) string {\n\tfor _, mp := range metapaths {\n\t\tif mp.path == path {\n\t\t\treturn mp.name\n\t\t}\n\t}\n\n\t// Unknown, so use path\n\treturn path\n}", "func (*HttpSpSubcats) GetPath() string { return \"/api/objects/http/sp_subcat/\" }", "func (r *Route) Path(tpl string) *Route {\n\tr.err = r.addRegexpMatcher(tpl, regexpTypePath)\n\treturn r\n}", "func (f *File) parsePath(path string) {\n\tn := filepath.Base(path)\n\te := filepath.Ext(n)\n\td := filepath.Dir(path)\n\n\tf.Name = n\n\tf.Path = filepath.Join(\"/\", d, fmt.Sprintf(\"%s%s\", f.ID.Hex(), e))\n}", "func (o ArgoCDSpecGrafanaIngressOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecGrafanaIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (o *Object) Path(path string) *Value {\n\topChain := o.chain.enter(\"Path(%q)\", path)\n\tdefer opChain.leave()\n\n\treturn jsonPath(opChain, o.value, path)\n}", "func (f FinagleFmt) Path() string {\n\treturn BaseZnodePath(f.role, f.environment, f.service)\n}", "func (*ItfparamsSecondarys) GetPath() string { return \"/api/objects/itfparams/secondary/\" }", "func (blt Bolt) Path() string {\n\treturn blt.path\n}", "func (d *dataStream) parseDataPath() (string, string, error) {\n\tpathSlice := strings.Split(strings.Trim(d.url.EscapedPath(), \"/\"), \"/\")\n\tglog.Infof(\"DEBUG -- %v\", pathSlice)\n\treturn pathSlice[0], strings.Join(pathSlice[1:], \"/\"), nil\n}", "func (d *InfoOutput) Path() string {\n\tval := d.reply[\"path\"]\n\n\treturn val.(string)\n\n}", "func (o CustomHttpPatternPtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *CustomHttpPattern) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (d *Fs) getPath(key string) (fPath string) {\n\tfPath = d.basePath\n\trunes := []rune(key)\n\tif len(key) > 4 {\n\t\tfPath = filepath.Join(fPath, string(runes[0:2]), string(runes[2:4]))\n\t}\n\treturn\n}", "func (*ItfparamsPrimarys) GetPath() string { return \"/api/objects/itfparams/primary/\" }", "func NewPath(root string) Path {\n\treturn Path{\n\t\t{\n\t\t\tKind: levelCustom,\n\t\t\tContent: root,\n\t\t},\n\t}\n}", "func AttributePathToPath(ap *proto.AttributePath) cty.Path {\n\tvar p cty.Path\n\tfor _, step := range ap.Steps {\n\t\tswitch selector := step.Selector.(type) {\n\t\tcase *proto.AttributePath_Step_AttributeName:\n\t\t\tp = p.GetAttr(selector.AttributeName)\n\t\tcase *proto.AttributePath_Step_ElementKeyString:\n\t\t\tp = p.Index(cty.StringVal(selector.ElementKeyString))\n\t\tcase *proto.AttributePath_Step_ElementKeyInt:\n\t\t\tp = p.Index(cty.NumberIntVal(selector.ElementKeyInt))\n\t\t}\n\t}\n\treturn p\n}", "func (r *Request) PathPayload(t *testing.T, path string) interface{} {\n\tparts := strings.Split(path, \".\")\n\tv := reflect.ValueOf(r.Payload)\n\tfor _, part := range parts {\n\t\tif v.Kind() == reflect.Interface {\n\t\t\tv = v.Elem()\n\t\t}\n\t\ttyp := v.Type()\n\t\tif typ.Kind() != reflect.Map {\n\t\t\tt.Fatalf(\"expected to find path %#v, but part %#v is of type %s\", path, part, typ)\n\t\t}\n\t\tif typ.Key().Kind() != reflect.String {\n\t\t\tpanic(\"test: key of part \" + part + \" of path \" + path + \" is not of type string\")\n\t\t}\n\t\tv = v.MapIndex(reflect.ValueOf(part))\n\t\tif !v.IsValid() {\n\t\t\tt.Fatalf(\"expected to find path %#v, but missing map key %#v\", path, part)\n\t\t}\n\t}\n\n\treturn v.Interface()\n}", "func (el *Fill) Path() {}", "func PathToAttributePath(p cty.Path) *proto.AttributePath {\n\tap := &proto.AttributePath{}\n\tfor _, step := range p {\n\t\tswitch selector := step.(type) {\n\t\tcase cty.GetAttrStep:\n\t\t\tap.Steps = append(ap.Steps, &proto.AttributePath_Step{\n\t\t\t\tSelector: &proto.AttributePath_Step_AttributeName{\n\t\t\t\t\tAttributeName: selector.Name,\n\t\t\t\t},\n\t\t\t})\n\t\tcase cty.IndexStep:\n\t\t\tkey := selector.Key\n\t\t\tswitch key.Type() {\n\t\t\tcase cty.String:\n\t\t\t\tap.Steps = append(ap.Steps, &proto.AttributePath_Step{\n\t\t\t\t\tSelector: &proto.AttributePath_Step_ElementKeyString{\n\t\t\t\t\t\tElementKeyString: key.AsString(),\n\t\t\t\t\t},\n\t\t\t\t})\n\t\t\tcase cty.Number:\n\t\t\t\tv, _ := key.AsBigFloat().Int64()\n\t\t\t\tap.Steps = append(ap.Steps, &proto.AttributePath_Step{\n\t\t\t\t\tSelector: &proto.AttributePath_Step_ElementKeyInt{\n\t\t\t\t\t\tElementKeyInt: v,\n\t\t\t\t\t},\n\t\t\t\t})\n\t\t\tdefault:\n\t\t\t\t// We'll bail early if we encounter anything else, and just\n\t\t\t\t// return the valid prefix.\n\t\t\t\treturn ap\n\t\t\t}\n\t\t}\n\t}\n\treturn ap\n}", "func ShowLabelPath(spaceID uuid.UUID, labelID uuid.UUID) string {\n\tparam0 := spaceID.String()\n\tparam1 := labelID.String()\n\n\treturn fmt.Sprintf(\"/api/spaces/%s/labels/%s\", param0, param1)\n}", "func (o ArgoCDSpecServerIngressOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecServerIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (t *transpiler) VisitTargetPathHead(ctx *parser.TargetPathHeadContext) interface{} {\n\tif ctx.TOKEN() != nil && ctx.TOKEN().GetText() != \"\" {\n\t\treturn pathSpec{\n\t\t\targ: getTokenText(ctx.TOKEN()),\n\t\t}\n\t}\n\n\t// ROOT_INPUT is a special case path segment since normally they cannot contain $.\n\tif ctx.ROOT_INPUT() != nil && ctx.ROOT_INPUT().GetText() != \"\" {\n\t\treturn pathSpec{\n\t\t\targ: ctx.ROOT_INPUT().GetText(),\n\t\t}\n\t}\n\n\t// ROOT is a special case path segment since it is a keyword and does not get tokenized as a TOKEN\n\t// TODO(): Remove after sunset.\n\tif ctx.ROOT() != nil && ctx.ROOT().GetText() != \"\" {\n\t\treturn pathSpec{\n\t\t\targ: ctx.ROOT().GetText(),\n\t\t}\n\t}\n\n\tif ctx.Index() != nil && ctx.Index().GetText() != \"\" {\n\t\treturn pathSpec{\n\t\t\tindex: ctx.Index().GetText(),\n\t\t}\n\t}\n\n\tif ctx.WILDCARD() != nil && ctx.WILDCARD().GetText() != \"\" {\n\t\treturn pathSpec{\n\t\t\tindex: ctx.WILDCARD().GetText(),\n\t\t}\n\t}\n\n\tif ctx.ArrayMod() != nil && ctx.ArrayMod().GetText() != \"\" {\n\t\treturn pathSpec{\n\t\t\tindex: ctx.ArrayMod().GetText(),\n\t\t}\n\t}\n\n\tt.fail(ctx, fmt.Errorf(\"invalid target path head - no token, index, arraymod, or wildcard\"))\n\treturn nil\n}", "func starlarkSub(\n\targs starlark.Tuple,\n\tkwargs []starlark.Tuple,\n) (starlark.Value, error) {\n\t// Expect exactly one positional argument, which represents the format\n\t// string.\n\tif len(args) != 1 {\n\t\treturn nil, errors.Errorf(\n\t\t\t\"Expected 1 positional argument 'format'; found %d\",\n\t\t\tlen(args),\n\t\t)\n\t}\n\n\t// Validate that the positional argument is a string.\n\tformat, ok := args[0].(starlark.String)\n\tif !ok {\n\t\treturn nil, errors.Errorf(\n\t\t\t\"TypeError: Expected argument 'format' has type str; found %s\",\n\t\t\targs[0].Type(),\n\t\t)\n\t}\n\n\t// Treat the keyword arguments as substitutions, including parsing their\n\t// values into `Arg`s.\n\tsubstitutions := make([]Substitution, len(kwargs))\n\tfor i, kwarg := range kwargs {\n\t\tvalue, err := starlarkValueToArg(kwarg[1])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tsubstitutions[i] = Substitution{\n\t\t\tKey: string(kwarg[0].(starlark.String)),\n\t\t\tValue: value,\n\t\t}\n\t}\n\n\t// TODO: Error if there are substitution placeholders in the format string\n\t// (e.g., `${Foo}`) for which there are no corresponding substitutions.\n\t// This is particularly important since the placeholder syntax is valid\n\t// bash, for example, if the placeholder is `${PATH}`, it would resolve at\n\t// runtime to the PATH env var, which would be a different down-the-road\n\t// error if it errored at all.\n\n\t// Build and return the resulting `*Sub` structure.\n\treturn &Sub{Format: string(format), Substitutions: substitutions}, nil\n}", "func getSinglePath(v interface{}) (string, error) {\n\tif name, err := resourceName(v); err != nil {\n\t\treturn \"\", err\n\t} else if id, err := resourceId(v); err != nil {\n\t\treturn \"\", err\n\t} else {\n\t\treturn fmt.Sprintf(\"%s(%d)\", name, id), nil\n\t}\n}", "func getPath(svc addr.HostSVC, ps *seg.PathSegment,\n\ttopoProv topology.Provider) (*snet.SVCAddr, error) {\n\n\tp, err := legacyPath(ps)\n\tif err != nil {\n\t\treturn nil, serrors.WrapStr(\"constructing path from segment\", err)\n\t}\n\tif err := p.Reverse(); err != nil {\n\t\treturn nil, serrors.WrapStr(\"reversing path\", err)\n\t}\n\tif err := p.InitOffsets(); err != nil {\n\t\treturn nil, serrors.WrapStr(\"initializing offsets\", err)\n\t}\n\thopF, err := p.GetHopField(p.HopOff)\n\tif err != nil {\n\t\treturn nil, serrors.WrapStr(\"extracting first hop field\", err)\n\t}\n\ttopo := topoProv.Get()\n\tifID := hopF.ConsIngress\n\tUnderlayNextHop, ok := topo.UnderlayNextHop2(ifID)\n\tif !ok {\n\t\treturn nil, serrors.New(\"first-hop border router not found\", \"intf_id\", ifID)\n\t}\n\treturn &snet.SVCAddr{IA: ps.FirstIA(), Path: p, NextHop: UnderlayNextHop, SVC: svc}, nil\n}", "func (self *Map) JSONPath(query string, fallback ...interface{}) interface{} {\n\tif d, err := JSONPath(self.MapNative(), query); err == nil && d != nil {\n\t\treturn d\n\t}\n\n\treturn sliceutil.First(fallback)\n}", "func (o ArgoCDSpecPrometheusIngressOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecPrometheusIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func PathToRIDAction(path, query, prefix string) (string, string) {\n\tif len(path) == len(prefix) || !strings.HasPrefix(path, prefix) {\n\t\treturn \"\", \"\"\n\t}\n\n\tpath = path[len(prefix):]\n\n\t// Dot separator not allowed in path\n\tif strings.ContainsRune(path, '.') {\n\t\treturn \"\", \"\"\n\t}\n\n\tif path[0] == '/' {\n\t\tpath = path[1:]\n\t}\n\tparts := strings.Split(path, \"/\")\n\tif len(parts) < 2 {\n\t\treturn \"\", \"\"\n\t}\n\n\tfor i := len(parts) - 1; i >= 0; i-- {\n\t\tpart, err := url.PathUnescape(parts[i])\n\t\tif err != nil {\n\t\t\treturn \"\", \"\"\n\t\t}\n\t\tparts[i] = part\n\t}\n\n\trid := strings.Join(parts[:len(parts)-1], \".\")\n\tif query != \"\" {\n\t\trid += \"?\" + query\n\t}\n\n\treturn rid, parts[len(parts)-1]\n}", "func (attr *Attribute) Path() string {\n\treturn attr.path\n}", "func (r *Router) Path(method, url string, viewFn View) *Path {\n\tif method != strings.ToUpper(method) {\n\t\tpanicf(\"http method '%s' must be in uppercase\", method)\n\t}\n\n\tp := &Path{\n\t\trouter: r,\n\t\tmethod: method,\n\t\turl: url,\n\t\tfullURL: r.getGroupFullPath(url),\n\t\tview: viewFn,\n\t}\n\n\tr.handlePath(p)\n\n\tp.registered = true\n\n\treturn p\n}", "func AnnotationPath(resultID string, path string, op string) ([]byte, error) {\n\tpatches := []jsonpatch.JsonPatchOperation{{\n\t\tOperation: op,\n\t\tPath: path,\n\t\tValue: resultID,\n\t}}\n\treturn json.Marshal(patches)\n}", "func parseMetricPathParameter(c *gin.Context) *dataselect.MetricQuery {\n\tmetricNamesParam := c.Query(\"metricNames\")\n\tvar metricNames []string\n\tif metricNamesParam != \"\" {\n\t\tmetricNames = strings.Split(metricNamesParam, \",\")\n\t} else {\n\t\tmetricNames = nil\n\t}\n\taggregationsParam := c.Query(\"aggregations\")\n\tvar rawAggregations []string\n\tif aggregationsParam != \"\" {\n\t\trawAggregations = strings.Split(aggregationsParam, \",\")\n\t} else {\n\t\trawAggregations = nil\n\t}\n\taggregationModes := metricapi.AggregationModes{}\n\tfor _, e := range rawAggregations {\n\t\taggregationModes = append(aggregationModes, metricapi.AggregationMode(e))\n\t}\n\treturn dataselect.NewMetricQuery(metricNames, aggregationModes)\n\n}", "func annotationPath(val string, path string, op string) ([]byte, error) {\n\tpatches := []jsonpatch.JsonPatchOperation{{\n\t\tOperation: op,\n\t\tPath: path,\n\t\tValue: val,\n\t}}\n\treturn json.Marshal(patches)\n}", "func (o FaqOutput) S3Path() FaqS3PathOutput {\n\treturn o.ApplyT(func(v *Faq) FaqS3PathOutput { return v.S3Path }).(FaqS3PathOutput)\n}", "func MatchPath(subPaths ...string) MatcherFunc { return MatchPaths(subPaths) }", "func parseFuncPath(path string) (pkgPath, fnName string) {\n\tif len(path) < 1 {\n\t\treturn \"\", \"\"\n\t}\n\tswitch path[0] {\n\tcase '(':\n\t\tregex := regexp.MustCompile(`\\((?P<pkg>[^)]+)\\).(?P<fn>.+)`)\n\t\tsubmatches := regex.FindStringSubmatch(path)\n\t\tif len(submatches) >= 3 {\n\t\t\treturn submatches[1], submatches[2]\n\t\t}\n\tcase '\"':\n\t\tregex := regexp.MustCompile(`\"(?P<pkg>[^)]+)\".(?P<fn>.+)`)\n\t\tsubmatches := regex.FindStringSubmatch(path)\n\t\tif len(submatches) >= 3 {\n\t\t\treturn submatches[1], submatches[2]\n\t\t}\n\tdefault:\n\t\tparts := strings.Split(path, \".\")\n\t\tif len(parts) >= 2 {\n\t\t\treturn parts[0], parts[1]\n\t\t}\n\t}\n\treturn \"\", path\n}", "func Path(output string) string {\n\turl := \"https://api.ipify.org\"\n\tquery := \"/?format=json\"\n\n\tif output == strings.ToLower(\"json\") {\n\t\turl += query\n\t} else if output != \"text\" {\n\t\treturn \"\"\n\t}\n\n\treturn url\n}", "func PathUpdate(path string, m *interface{}, value interface{}) {\n\tmatches := re.FindAllStringSubmatch(path, -1)\n\tcomponents := make([]string, 0, len(matches))\n\tfor _, m := range matches {\n\t\tcomponents = append(components, m[1])\n\t}\n\tpathUpdate(components, m, value)\n}", "func (ne *NSEnter) LookPath(file string) (string, error) {\n\treturn \"\", fmt.Errorf(\"not implemented, error looking up : %s\", file)\n}", "func Path(p string) Option {\n\treturn func(cs *csrf) {\n\t\tcs.opts.Path = p\n\t}\n}", "func newSWANDataFromPath(\n\td *common.Domain,\n\tr *http.Request) ([]*swan.Pair, *swan.Error) {\n\tb := common.GetSWANDataFromRequest(r)\n\tif b == \"\" {\n\t\treturn nil, nil\n\t}\n\treturn newSWANData(d, b)\n}", "func (p *Path) Match(path string) *Match {\n\tvar match = &Match{\n\t\tValues: make(map[string]string),\n\t}\n\n\tfor _, part := range p.parts {\n\t\tif len(path) < 1 {\n\t\t\treturn nil\n\t\t}\n\n\t\tif path[0] != '/' {\n\t\t\treturn nil\n\t\t}\n\t\t// prefix /\n\t\tpath = path[1:]\n\n\t\tmatched, key, value, length := part.match(path)\n\n\t\t//log.Printf(\"%#v == %v (%d) %s\", part, matched, length, value)\n\n\t\tif !matched {\n\t\t\treturn nil\n\t\t}\n\n\t\tif key != \"\" {\n\t\t\tmatch.Values[key] = value\n\t\t}\n\t\tpath = path[length:]\n\t}\n\n\tif len(path) > 0 && path != \"/\" {\n\t\treturn nil\n\t}\n\n\treturn match\n}", "func JSONPath(fields ...string) string {\n\treturn strings.Join(fields, \".\")\n}", "func (r *Router) formatPath(path string) string {\n\tif r.autoPrefix {\n\t\tif strings.HasPrefix(path, pathPrefix) {\n\t\t\treturn path\n\t\t} else {\n\t\t\treturn pathPrefix + path\n\t\t}\n\t}\n\treturn path\n}", "func pkgpath(p *types.Package) string {\n\tpath := p.Path()\n\tname := p.Name()\n\tif path == \"\" || name == \"main\" {\n\t\tpath = p.Name()\n\t}\n\treturn path\n}", "func (d *DbBackendCouch) getPathSpecificStuff(params dragonfruit.QueryParams) ([][]string,\n\tcouchdbRow, string, interface{}, error) {\n\tvar v interface{}\n\n\tpathmap := dragonfruit.PathParamRe.FindAllStringSubmatch(params.Path, -1)\n\n\tdoc, id, err := d.getRootDocument(params)\n\n\tif err != nil {\n\t\treturn pathmap, doc, id, v, err\n\t}\n\n\t// unwrap the body of the document into an interface\n\tif len(params.Body) > 0 {\n\t\terr = json.Unmarshal(params.Body, &v)\n\t}\n\n\tif v == nil {\n\t\treturn pathmap, doc, id, v, err\n\t}\n\n\treturn pathmap, doc, id, v, nil\n}" ]
[ "0.5881082", "0.5746452", "0.5599261", "0.5507016", "0.5389365", "0.53872955", "0.53744584", "0.53303474", "0.5306618", "0.5272473", "0.5219563", "0.5219563", "0.5170733", "0.51603353", "0.515905", "0.5149575", "0.50718874", "0.5059588", "0.5042506", "0.50372833", "0.503696", "0.50342447", "0.50326836", "0.5022716", "0.5022075", "0.5013477", "0.50124687", "0.50071406", "0.5004446", "0.5001686", "0.4997385", "0.49876562", "0.4976975", "0.49769562", "0.49718538", "0.49672425", "0.49637184", "0.49620232", "0.4956009", "0.4954888", "0.49544948", "0.49369746", "0.4919361", "0.4914128", "0.4903059", "0.4898844", "0.489292", "0.4892411", "0.48914254", "0.48672286", "0.48661312", "0.48640612", "0.48505858", "0.48453867", "0.48410955", "0.48338446", "0.48328966", "0.48198286", "0.48181155", "0.4807674", "0.48068607", "0.48047668", "0.47962794", "0.47831666", "0.47756416", "0.47737476", "0.4766904", "0.47660503", "0.47658274", "0.47544196", "0.475296", "0.47518477", "0.47453266", "0.4744625", "0.47360653", "0.47311834", "0.47272262", "0.47257775", "0.47257644", "0.47257408", "0.47216454", "0.47207052", "0.47189453", "0.47006416", "0.46937564", "0.46890754", "0.46855336", "0.46855196", "0.46827778", "0.4680196", "0.4679884", "0.4679219", "0.46700248", "0.46690655", "0.46639004", "0.46612397", "0.46589315", "0.4649958", "0.4646593", "0.4637322" ]
0.80633986
0
GlobGroup Type implements the starlark.Value.Type() method.
func (gg GlobGroup) Type() string { return fmt.Sprintf("GlobGroup") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (NamespaceStar) Type() string { return TypeNamespaceStar }", "func (b *GroupsSearchBuilder) Type(v string) *GroupsSearchBuilder {\n\tb.Params[\"type\"] = v\n\treturn b\n}", "func (value *Value) Type() Type {\n\treturn value.valueType\n}", "func (b baseValue) Type() string {\n\treturn string(b.flagType)\n}", "func (o *AutoscalerResourceLimitsGPULimit) Type() string {\n\tif o != nil && o.bitmap_&2 != 0 {\n\t\treturn o.type_\n\t}\n\treturn \"\"\n}", "func Type(value r.Value) r.Type {\n\tif !value.IsValid() || value == None {\n\t\treturn nil\n\t}\n\treturn value.Type()\n}", "func (val Value) Type() Type {\n\treturn val.typ\n}", "func (e *Wildcard) Type() Type {\n\treturn UnknownType\n}", "func (flg Var) Type() string {\n\treturn \"@filename or JSON literal\"\n}", "func (p *Package) Type(name string) (t *Type) {\n\tt, _ = p.Members[name].(*Type)\n\treturn\n}", "func (v *Value) Type() kiwi.ValueType {\n\treturn Type\n}", "func (ts *TypeSet) Type(s string) Type {\n\tts.RLock()\n\tdefer ts.RUnlock()\n\treturn ts.types[s]\n}", "func (group *ContainerGroup_Spec_ARM) GetType() string {\n\treturn \"Microsoft.ContainerInstance/containerGroups\"\n}", "func (v Value) Type() Type {\n\treturn v.Typ\n}", "func (v Value) Type() Type {\n\treturn v.typ\n}", "func (it *Regex) Type() graph.Type {\n\treturn graph.Regex\n}", "func (d *Driver) Type() (t string) {\n\treturn \"go\"\n}", "func (g *GroupedAVP) Type() datatype.TypeID {\n\treturn GroupedAVPType\n}", "func (t TsTimestampAggregation) Type() TsAggregationType {\n\treturn t.kind\n}", "func (*MessageFileTypeGroup) TypeName() string {\n\treturn \"messageFileTypeGroup\"\n}", "func (msf *ModuleSetFlag) Type() string {\n\treturn \"Modules Set Flag\"\n}", "func (script *Script) Type(name string) (Type, bool) {\n\tgslangType, ok := script.types[name]\n\n\treturn gslangType, ok\n}", "func (o GetGroupIdentityOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetGroupIdentity) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o ReportGroupOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ReportGroup) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (this *Value) Type() int {\n\treturn this.parsedType\n}", "func Type(typ string) gomock.Matcher {\n\treturn &typeMatcher{typ: typ}\n}", "func (g *GistFile) GetType() string {\n\tif g == nil || g.Type == nil {\n\t\treturn \"\"\n\t}\n\treturn *g.Type\n}", "func (element *Element) Type(value string) *Element {\n\treturn element.Attr(\"type\", value)\n}", "func (o StorageSettingOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v StorageSetting) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (sv *SupernodesValue) Type() string {\n\treturn \"supernodes\"\n}", "func (o GroupIdentityOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GroupIdentity) string { return v.Type }).(pulumi.StringOutput)\n}", "func (m *RecurrencePattern) GetType()(*RecurrencePatternType) {\n return m.type_escaped\n}", "func (n *piName) Type() Type {\n\treturn n.t\n}", "func (s *Scope) Type() *RegisteredType {\n\treturn GoStructRegistry.Lookup(\"packageScope\")\n}", "func (l *DirName) Type() string {\n\treturn \"string\"\n}", "func (a ValueNode) GetType() string {\n\treturn \"ValueNode\"\n}", "func (v Value) Type() querypb.Type {\n\treturn v.typ\n}", "func (p *Package) Type(name string) (*Type, bool) {\n\tidx, ok := p.typIdx[name]\n\tif !ok {\n\t\treturn nil, false\n\t}\n\treturn p.Types[idx], true\n}", "func (data *Instance) Type() Value {\n\treturn data.TypeTag\n}", "func (o ReplicationRecoveryPlanRecoveryGroupOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ReplicationRecoveryPlanRecoveryGroup) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o *AutoscalerResourceLimitsGPULimit) GetType() (value string, ok bool) {\n\tok = o != nil && o.bitmap_&2 != 0\n\tif ok {\n\t\tvalue = o.type_\n\t}\n\treturn\n}", "func (this *Mod) Type() value.Type { return value.NUMBER }", "func (group *ResourceGroup) GetType() string {\n\treturn \"Microsoft.Resources/resourceGroups\"\n}", "func (group *ResourceGroup) GetType() string {\n\treturn \"Microsoft.Resources/resourceGroups\"\n}", "func (this *Element) Type() value.Type { return value.JSON }", "func (j *SearchHighlight) Type(v string) *SearchHighlight {\n\tj.Json.SetPath([]string{\"fields\", j.field, \"type\"}, v)\n\treturn j\n}", "func (v *Value) Type() *JSONType {\n\tt := C.zj_Type(v.V)\n\tif t == nil {\n\t\treturn nil\n\t}\n\tret := JSONType(*t)\n\treturn &ret\n}", "func (NamespaceNode) Type() string { return TypeNamespaceNode }", "func (f *FlagChoice) Type() string {\n\treturn choiceList(f.choices...)\n}", "func (m *GroupMutation) Type() string {\n\treturn m.typ\n}", "func (m *GroupMutation) Type() string {\n\treturn m.typ\n}", "func (o *MetricDefaultAggregation) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (o GetReplicationRecoveryPlanRecoveryGroupOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetReplicationRecoveryPlanRecoveryGroup) string { return v.Type }).(pulumi.StringOutput)\n}", "func (m *BarGroupMutation) Type() string {\n\treturn m.typ\n}", "func (myOperatingSystemType *OperatingSystemType) Type() (param string) {\n\treturn myOperatingSystemType.Typevar\n}", "func getRegistryType(fullType string) *registry.Type {\n\ttList := strings.Split(fullType, \":\")\n\tif len(tList) != 2 {\n\t\treturn nil\n\t}\n\n\treturn &registry.Type{\n\t\tName: tList[0],\n\t\tVersion: tList[1],\n\t}\n}", "func (s *StringEnum) Type() string { return \"string\" }", "func (L *LState) RegType(typename string, ins interface{}) {\r\n\texport.NewType(L.L(), typename, ins)\r\n}", "func (l *CacheMode) Type() string {\n\treturn \"string\"\n}", "func (s SetValue) Type(ctx context.Context) attr.Type {\n\treturn SetType{ElemType: s.ElementType(ctx)}\n}", "func (*BasicGroup) TypeName() string {\n\treturn \"basicGroup\"\n}", "func (o RegistryIdentityOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RegistryIdentity) string { return v.Type }).(pulumi.StringOutput)\n}", "func (i ID) Type() string {\n\tstuff := strings.Split(i.String(), \"/\")\n\treturn stuff[0]\n}", "func (this *Self) Type() value.Type { return value.JSON }", "func (m *varMap) Type() JSONType {\n\treturn ObjectType\n}", "func (r Resource) Type() string {\n\treturn r.typ\n}", "func (o BucketGrantOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketGrant) string { return v.Type }).(pulumi.StringOutput)\n}", "func (g *Generator) Type(typeName string) *toolbox.TypeInfo {\n\treturn g.fileInfo.Type(typeName)\n}", "func (t Type) Type() string {\n\treturn t.typeName\n}", "func (b *Builtin) Type() ObjectType {\n\treturn BUILTIN\n}", "func (o RegistryTaskBaseImageTriggerOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RegistryTaskBaseImageTrigger) string { return v.Type }).(pulumi.StringOutput)\n}", "func (e *Entry) Type() tree.Type {\n\tif e.notFound {\n\t\treturn tree.TYPE_NOTFOUND\n\t}\n\tswitch e.st.Mode() & os.ModeType {\n\tcase 0:\n\t\treturn tree.TYPE_REGULAR\n\tcase os.ModeDir:\n\t\treturn tree.TYPE_DIRECTORY\n\tcase os.ModeSymlink:\n\t\treturn tree.TYPE_SYMLINK\n\tdefault:\n\t\treturn tree.TYPE_UNKNOWN\n\t}\n}", "func (o AddressGroupOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *AddressGroup) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (inst *InstFMul) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (a Asset) Type() AssetType {\n\treturn a.kind\n}", "func (k *Key) Type() string {\n\treturn k.base.Type()\n}", "func (m *EnumLiteral) Type() Type {\n\treturn m.wtype\n}", "func (md *MetaData) Type(key ...string) string {\n\tif ki, ok := md.keyInfo[Key(key).String()]; ok {\n\t\treturn ki.tomlType.typeString()\n\t}\n\treturn \"\"\n}", "func (this *NowStr) Type() value.Type { return value.STRING }", "func (n *GlobalNotification) Type() string {\r\n\treturn \"global\"\r\n}", "func (t DataType) TypeName() string { return typeNames[t] }", "func (m *EventRSVPMutation) Type() string {\n\treturn m.typ\n}", "func (s *GDrive) Type() string {\n\treturn \"gdrive\"\n}", "func (css *Settings) Type() config.Type {\n\treturn css.TypeVal\n}", "func (o GroupIdentityPtrOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *GroupIdentity) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Type\n\t}).(pulumi.StringPtrOutput)\n}", "func (o *GroupReplaceRequest) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (p Path) Type(reference, value interface{}) Path { return p.with(Type, reference, value) }", "func (Storage) Type() string {\n\treturn Type\n}", "func (o *Object) Type() string {\n\treturn o.Get(\"og:type\")\n}", "func (pier *Pier) Type() string {\n\tif pier.config.Mode.Type != repo.UnionMode {\n\t\treturn pier.plugin.Type()\n\t}\n\treturn repo.UnionMode\n}", "func (s UserSet) Type() string {\n\tres, _ := s.RecordCollection.Get(models.NewFieldName(\"Type\", \"type\")).(string)\n\treturn res\n}", "func (t *TypeValue) TypeName() string {\n\treturn t.name\n}", "func (o *SecretValue) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func Type() *dataType {\n\treturn &dataType{str: field.StringType()}\n}", "func (v Value) Type() Type {\n\tpanic(message)\n}", "func (e REnv) Type() Type { return e.Value().Type() }", "func GroupType_Values() []string {\n\treturn []string{\n\t\tGroupTypeTabular,\n\t}\n}", "func (v projectsVal) Type() string {\n\tif v.acceptMultipleProjects {\n\t\treturn \"comma-separated list of GCE projects\"\n\t}\n\treturn \"GCE project name\"\n}", "func (o MonitoredResourceDescriptorOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v MonitoredResourceDescriptor) string { return v.Type }).(pulumi.StringOutput)\n}", "func (c *Chunk) Type() string {\n\tvar bf bytes.Buffer\n\tbf.Write(c.typ)\n\treturn bf.String()\n}" ]
[ "0.6424173", "0.63089347", "0.6207848", "0.6189664", "0.6150497", "0.61098784", "0.6105709", "0.61046934", "0.61017525", "0.6056659", "0.6042432", "0.60282564", "0.60250735", "0.6024554", "0.6006814", "0.60044366", "0.59927106", "0.5968349", "0.59500855", "0.59442514", "0.59287304", "0.59038484", "0.5890247", "0.58830804", "0.5874322", "0.58617616", "0.58606744", "0.58573604", "0.58223283", "0.581341", "0.5807084", "0.57987446", "0.5779363", "0.577919", "0.576531", "0.5762579", "0.576209", "0.5756967", "0.5751188", "0.5746017", "0.5737325", "0.5730239", "0.572001", "0.572001", "0.57178575", "0.57157844", "0.5710384", "0.57103336", "0.57043153", "0.57011133", "0.57011133", "0.56952786", "0.5680706", "0.56739503", "0.56694454", "0.5662262", "0.56563777", "0.5653081", "0.56419146", "0.5639539", "0.5634817", "0.5632432", "0.56314707", "0.56275386", "0.5622141", "0.5608667", "0.5593114", "0.5587086", "0.5586684", "0.5581472", "0.55790746", "0.55723816", "0.5568636", "0.5560425", "0.555104", "0.55496955", "0.5535622", "0.55307347", "0.55300915", "0.5526044", "0.55257374", "0.55228686", "0.55198914", "0.55194974", "0.55181384", "0.55095625", "0.55047745", "0.55026734", "0.55014807", "0.54983693", "0.5495792", "0.5493065", "0.5481277", "0.5480911", "0.5479515", "0.5475982", "0.5474581", "0.547335", "0.54712355", "0.5470682" ]
0.81681466
0
Freeze implements the starlark.Value.Freeze() method.
func (gg GlobGroup) Freeze() {}
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (i *Index) Freeze() {\n\ti.frozen = true\n}", "func (p *Poly) freeze() {\n\tfor i := 0; i < n; i++ {\n\t\tp[i] = freeze(p[i])\n\t}\n}", "func freeze(o *goja.Object) {\n\tfor _, key := range o.Keys() {\n\t\to.DefineDataProperty(key, o.Get(key), goja.FLAG_FALSE, goja.FLAG_FALSE, goja.FLAG_TRUE)\n\t}\n}", "func (df *DataFrame) Freeze() { df.frozen = true }", "func (i *Instance) Freeze() {\n}", "func (t *Target) Freeze() {}", "func (v *ReadCloserValue) Freeze() {}", "func Freeze(x int32) int32 {\n\tx -= 9829 * ((13*x) >> 17)\n\tx -= 9829 * ((427*x + 2097152) >> 22)\n\ty := x + 9829\n\tv := subtle.ConstantTimeLessOrEq(int(x), -1)\n\treturn int32(subtle.ConstantTimeSelect(v, int(y), int(x)))\n}", "func (f *chainFreezer) freeze(db database.KeyValueStore) {\n\tnfdb := &nofreezedb{KeyValueStore: db}\n\n\tvar (\n\t\tbackoff bool\n\t\ttriggered chan struct{} // Used in tests\n\t)\n\tfor {\n\t\tselect {\n\t\tcase <-f.quit:\n\t\t\tlog.Info(\"Freezer shutting down\")\n\t\t\treturn\n\t\tdefault:\n\t\t}\n\t\tif backoff {\n\t\t\t// If we were doing a manual trigger, notify it\n\t\t\tif triggered != nil {\n\t\t\t\ttriggered <- struct{}{}\n\t\t\t\ttriggered = nil\n\t\t\t}\n\t\t\tselect {\n\t\t\tcase <-time.NewTimer(freezerRecheckInterval).C:\n\t\t\t\tbackoff = false\n\t\t\tcase triggered = <-f.trigger:\n\t\t\t\tbackoff = false\n\t\t\tcase <-f.quit:\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\t// Retrieve the freezing threshold.\n\t\thash := ReadHeadBlockHash(nfdb)\n\t\tif hash == (common.Hash{}) {\n\t\t\tlog.Debug(\"Current full block hash unavailable\") // new chain, empty database\n\t\t\tbackoff = true\n\t\t\tcontinue\n\t\t}\n\t\tnumber := ReadHeaderNumber(nfdb, hash)\n\t\tthreshold := atomic.LoadUint64(&f.threshold)\n\t\tfrozen := atomic.LoadUint64(&f.frozen)\n\t\tswitch {\n\t\tcase number == nil:\n\t\t\tlog.Error(\"Current full block number unavailable\", \"hash\", hash)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\n\t\tcase *number < threshold:\n\t\t\tlog.Debug(\"Current full block not old enough\", \"number\", *number, \"hash\", hash, \"delay\", threshold)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\n\t\tcase *number-threshold <= frozen:\n\t\t\tlog.Debug(\"Ancient blocks frozen already\", \"number\", *number, \"hash\", hash, \"frozen\", frozen)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\t\t}\n\t\thead := ReadHeader(nfdb, hash, *number)\n\t\tif head == nil {\n\t\t\tlog.Error(\"Current full block unavailable\", \"number\", *number, \"hash\", hash)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\t\t}\n\n\t\t// Seems we have data ready to be frozen, process in usable batches\n\t\tvar (\n\t\t\tstart = time.Now()\n\t\t\tfirst, _ = f.Ancients()\n\t\t\tlimit = *number - threshold\n\t\t)\n\t\tif limit-first > freezerBatchLimit {\n\t\t\tlimit = first + freezerBatchLimit\n\t\t}\n\t\tancients, err := f.freezeRange(nfdb, first, limit)\n\t\tif err != nil {\n\t\t\tlog.Error(\"Error in block freeze operation\", \"err\", err)\n\t\t\tbackoff = true\n\t\t\tcontinue\n\t\t}\n\n\t\t// Batch of blocks have been frozen, flush them before wiping from leveldb\n\t\tif err := f.Sync(); err != nil {\n\t\t\tlog.Critical(\"Failed to flush frozen tables\", \"err\", err)\n\t\t}\n\n\t\t// Wipe out all data from the active database\n\t\tbatch := db.NewBatch()\n\t\tfor i := 0; i < len(ancients); i++ {\n\t\t\t// Always keep the genesis block in active database\n\t\t\tif first+uint64(i) != 0 {\n\t\t\t\tDeleteBlockWithoutNumber(batch, ancients[i], first+uint64(i))\n\t\t\t\tDeleteCanonicalHash(batch, first+uint64(i))\n\t\t\t}\n\t\t}\n\t\tif err := batch.Write(); err != nil {\n\t\t\tlog.Critical(\"Failed to delete frozen canonical blocks\", \"err\", err)\n\t\t}\n\t\tbatch.Reset()\n\n\t\t// Wipe out side chains also and track dangling side chains\n\t\tvar dangling []common.Hash\n\t\tfrozen = atomic.LoadUint64(&f.frozen) // Needs reload after during freezeRange\n\t\tfor number := first; number < frozen; number++ {\n\t\t\t// Always keep the genesis block in active database\n\t\t\tif number != 0 {\n\t\t\t\tdangling = ReadAllHashes(db, number)\n\t\t\t\tfor _, hash := range dangling {\n\t\t\t\t\tlog.Debug(\"Deleting side chain\", \"number\", number, \"hash\", hash)\n\t\t\t\t\tDeleteBlock(batch, hash, number)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif err := batch.Write(); err != nil {\n\t\t\tlog.Critical(\"Failed to delete frozen side blocks\", \"err\", err)\n\t\t}\n\t\tbatch.Reset()\n\n\t\t// Step into the future and delete and dangling side chains\n\t\tif frozen > 0 {\n\t\t\ttip := frozen\n\t\t\tfor len(dangling) > 0 {\n\t\t\t\tdrop := make(map[common.Hash]struct{})\n\t\t\t\tfor _, hash := range dangling {\n\t\t\t\t\tlog.Debug(\"Dangling parent from Freezer\", \"number\", tip-1, \"hash\", hash)\n\t\t\t\t\tdrop[hash] = struct{}{}\n\t\t\t\t}\n\t\t\t\tchildren := ReadAllHashes(db, tip)\n\t\t\t\tfor i := 0; i < len(children); i++ {\n\t\t\t\t\t// Dig up the child and ensure it's dangling\n\t\t\t\t\tchild := ReadHeader(nfdb, children[i], tip)\n\t\t\t\t\tif child == nil {\n\t\t\t\t\t\tlog.Error(\"Missing dangling header\", \"number\", tip, \"hash\", children[i])\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tif _, ok := drop[child.ParentHash]; !ok {\n\t\t\t\t\t\tchildren = append(children[:i], children[i+1:]...)\n\t\t\t\t\t\ti--\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\t// Delete all block data associated with the child\n\t\t\t\t\tlog.Debug(\"Deleting dangling block\", \"number\", tip, \"hash\", children[i], \"parent\", child.ParentHash)\n\t\t\t\t\tDeleteBlock(batch, children[i], tip)\n\t\t\t\t}\n\t\t\t\tdangling = children\n\t\t\t\ttip++\n\t\t\t}\n\t\t\tif err := batch.Write(); err != nil {\n\t\t\t\tlog.Critical(\"Failed to delete dangling side blocks\", \"err\", err)\n\t\t\t}\n\t\t}\n\n\t\t// Log something friendly for the user\n\t\tcontext := []interface{}{\n\t\t\t\"blocks\", frozen - first, \"elapsed\", common.PrettyDuration(time.Since(start)), \"number\", frozen - 1,\n\t\t}\n\t\tif n := len(ancients); n > 0 {\n\t\t\tcontext = append(context, []interface{}{\"hash\", ancients[n-1]}...)\n\t\t}\n\t\tlog.Infof(\"Deep froze chain segment: %+v\", context...)\n\n\t\t// Avoid database thrashing with tiny writes\n\t\tif frozen-first < freezerBatchLimit {\n\t\t\tbackoff = true\n\t\t}\n\t}\n}", "func (c *Container) Freeze() *Container {\n\tif c == nil {\n\t\treturn nil\n\t}\n\tif c.flags&flagDirty != 0 {\n\t\tif roaringParanoia {\n\t\t\tpanic(\"freezing dirty container\")\n\t\t}\n\t\t// c.Repair won't work if this is already frozen, but in\n\t\t// theory that can't happen?\n\t\tc.Repair()\n\t}\n\t// don't need to freeze\n\tif c.flags&flagFrozen != 0 {\n\t\treturn c\n\t}\n\tc.flags |= flagFrozen\n\treturn c\n}", "func (me TxsdAnimTimingAttrsFill) IsFreeze() bool { return me.String() == \"freeze\" }", "func (g *metadataGraph) freeze(ctx context.Context) {\n\tg.root.freeze(ctx)\n}", "func (d *Dam) Unlock() {\n\td.freeze.Unlock()\n}", "func (s *Sub) Freeze() {}", "func (p Path) Freeze() {}", "func main() {\n\tx := new(int)\n\t*x++ // ok\n\n\tx = freeze(x)\n\n\tfmt.Println(*x) // ok; prints 1\n\t//*x++ // not ok; panics!\n}", "func (recv *Object) FreezeNotify() {\n\tC.g_object_freeze_notify((*C.GObject)(recv.native))\n\n\treturn\n}", "func (sa *SuffixArray) Freeze() error { return sa.ba.Freeze() }", "func (s *inMemSpannerServer) Unfreeze() {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tclose(s.freezed)\n}", "func (iv *writeOnlyInterval) freeze(s *Schema) (*Interval, error) {\n\tif err := iv.closeCurrentSegment(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tiv.Segments = make([]*Segment, iv.NumSegments)\n\tfor i := 0; i < iv.NumSegments; i++ {\n\t\tif !iv.DiskBacked {\n\t\t\tiv.Segments[i] = &Segment{Bytes: iv.buffers[i].Bytes()}\n\t\t\tcontinue\n\t\t}\n\t\tfilename := iv.SegmentFilename(s, i)\n\t\tf, err := os.Open(filename)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tmapped, err := mmap.Map(f, mmap.RDONLY, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tiv.Segments[i] = &Segment{File: f, Bytes: mapped}\n\t}\n\treturn &iv.Interval, nil\n}", "func (d *Director) FreezeMonkey(rng *rand.Rand, intensity float64) {\n\tif intensity < 0.1 {\n\t\treturn\n\t}\n\ttarget := d.randomAgent(rng)\n\tduration := d.makeDuration(rng, 1000, intensity)\n\tlog.Printf(\"[monkey] Freezing %v for %v\", target, duration)\n\tgo target.Stop(duration)\n}", "func NewFreezeParams() *FreezeParams {\n\treturn &FreezeParams{\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (_OracleMgr *OracleMgrCaller) FreezePeriod(opts *bind.CallOpts) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _OracleMgr.contract.Call(opts, out, \"freezePeriod\")\n\treturn *ret0, err\n}", "func (*CapturedStacktrace) Freeze() {}", "func (s *Client) Freeze(username string) error {\n\tuser := s.Init(username)\n\tdata, err := json.Marshal(user)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\ttx, err := s.db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tstmt, err := tx.Prepare(`INSERT OR REPLACE INTO frozen_user (user_id, data)VALUES((SELECT id FROM users WHERE username = ?), ?);`)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\t_, err = stmt.Exec(username, string(data))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (n *SoupNode) Freeze() {}", "func (kb *KubeBackend) Unfreeze() error {\n\tlogrus.Infof(\"set deployment %s replica=1\", kb.ID())\n\tdeployment, err := kb.manager.di.Lister().Deployments(kb.manager.namespace).Get(kb.ID())\n\tif err != nil {\n\t\tif err := kb.manager.syncBackend(kb.ID()); err != nil {\n\t\t\tlogrus.Warnf(\"sycn app with error: %v\", err)\n\t\t}\n\t\treturn err\n\t}\n\tif *deployment.Spec.Replicas != 0 {\n\t\treturn nil\n\t}\n\tvar targetReplica int32 = 1\n\tdeployment.Spec.Replicas = &targetReplica\n\t_, err = kb.manager.client.AppsV1().Deployments(kb.manager.namespace).\n\t\tUpdate(context.Background(), deployment, metav1.UpdateOptions{})\n\tif err != nil {\n\t\treturn err\n\t}\n\tkb.Lock()\n\tkb.stateStarting = true\n\tkb.stateHealthy = false\n\tkb.Unlock()\n\treturn nil\n}", "func (n *metadataNode) freeze(ctx context.Context) {\n\tn.assertNonFrozen()\n\n\t// md may be already non-nil for the root, this is fine.\n\tif n.md == nil {\n\t\tn.md = mergeIntoPrefixMetadata(ctx, n.prefix, n.acls)\n\t}\n\tn.acls = nil // mark as frozen, release unnecessary memory\n\n\tfor _, child := range n.children {\n\t\tchild.freeze(ctx)\n\t}\n}", "func (o *FreezeParams) WithTimeout(timeout time.Duration) *FreezeParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (d *Dam) Lock() {\n\td.freeze.Lock()\n}", "func updateFrozenState(db *IndexerDb, assetID uint64, closedAt *uint64, creator, freeze, holder types.Address) error {\n\t// Semi-blocking migration.\n\t// Hold accountingLock for the duration of the Transaction search + account_asset update.\n\tdb.accountingLock.Lock()\n\tdefer db.accountingLock.Unlock()\n\n\tminRound := uint64(0)\n\tif closedAt != nil {\n\t\tminRound = *closedAt\n\t}\n\n\tholderb64 := encoding.Base64(holder[:])\n\trow := db.db.QueryRow(freezeTransactionsQuery, freeze[:], holderb64, assetID, minRound)\n\tvar found uint64\n\terr := row.Scan(&found)\n\tif err != nil && err != sql.ErrNoRows {\n\t\treturn err\n\t}\n\n\t// If there are any freeze transactions then the default no longer applies.\n\t// Exit early if the asset was frozen\n\tif found != 0 {\n\t\treturn nil\n\t}\n\n\t// If there were no freeze transactions, re-initialize the frozen value.\n\tfrozen := !bytes.Equal(creator[:], holder[:])\n\tdb.db.Exec(`UPDATE account_asset SET frozen = $1 WHERE assetid = $2 and addr = $3`, frozen, assetID, holder[:])\n\n\treturn nil\n}", "func (_OracleMgr *OracleMgrCallerSession) FreezePeriod() (*big.Int, error) {\n\treturn _OracleMgr.Contract.FreezePeriod(&_OracleMgr.CallOpts)\n}", "func (r *RunCtx) Freeze() {\n}", "func (o *FreezeParams) WithContext(ctx context.Context) *FreezeParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (r *AccountDIDRegistry) UnFreeze(did DID) error {\n\texist := r.HasAccountDID(did)\n\tif !exist {\n\t\treturn fmt.Errorf(\"did %s not existed\", did)\n\t}\n\treturn r.auditStatus(did, Normal)\n}", "func (transaction *TokenUpdateTransaction) SetFreezeKey(publicKey Key) *TokenUpdateTransaction {\n\ttransaction._RequireNotFrozen()\n\ttransaction.freezeKey = publicKey\n\treturn transaction\n}", "func (c *TickMap) Freeze() []string {\n\tc.mlock.Lock()\n\tdefer c.mlock.Unlock()\n\ts := make([]string, len(c.m))\n\ti := 0\n\tfor key, _ := range c.m {\n\t\ts[i] = key\n\t\ti++\n\t}\n\treturn s\n}", "func (_OracleMgr *OracleMgrSession) FreezePeriod() (*big.Int, error) {\n\treturn _OracleMgr.Contract.FreezePeriod(&_OracleMgr.CallOpts)\n}", "func (o *FreezeParams) WithDefaults() *FreezeParams {\n\to.SetDefaults()\n\treturn o\n}", "func (r *AccountDIDRegistry) Freeze(did DID) error {\n\texist := r.HasAccountDID(did)\n\tif !exist {\n\t\treturn fmt.Errorf(\"did %s not existed\", did)\n\t}\n\treturn r.auditStatus(did, Frozen)\n}", "func (m *ClusterService) clusterFreeze(ctx context.Context, args struct {\n\tStatus bool\n}) (*proto.GeneralResp, error) {\n\tif _, _, err := permissions(ctx, ADMIN); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := m.cluster.setDisableAutoAllocate(args.Status); err != nil {\n\t\treturn nil, err\n\t}\n\treturn proto.Success(\"success\"), nil\n}", "func freezetheworld() {\n\tatomic.Store(&freezing, 1)\n\t// stopwait and preemption requests can be lost\n\t// due to races with concurrently executing threads,\n\t// so try several times\n\tfor i := 0; i < 5; i++ {\n\t\t// this should tell the scheduler to not start any new goroutines\n\t\tsched.stopwait = freezeStopWait\n\t\tatomic.Store(&sched.gcwaiting, 1)\n\t\t// this should stop running goroutines\n\t\tif !preemptall() {\n\t\t\tbreak // no running goroutines\n\t\t}\n\t\tusleep(1000)\n\t}\n\t// to be sure\n\tusleep(1000)\n\tpreemptall()\n\tusleep(1000)\n}", "func (p *BailServiceClient) FreezeBail(dealerId int64, amount float64, orderId int64) (r *Bail, err error) {\n\tif err = p.sendFreezeBail(dealerId, amount, orderId); err != nil {\n\t\treturn\n\t}\n\treturn p.recvFreezeBail()\n}", "func (s *Service) Frozen(ctx context.Context, req *pb.FrozenRequest) (*pb.FrozenResponse, error) {\n\tif !strings.HasPrefix(strings.Title(req.Address), \"Mx\") {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"invalid address\")\n\t}\n\n\tcState := s.blockchain.CurrentState()\n\tcState.RLock()\n\tdefer cState.RUnlock()\n\n\tvar reqCoin *coins.Model\n\n\tif req.CoinId != nil {\n\t\tcoinID := types.CoinID(req.CoinId.GetValue())\n\t\treqCoin = cState.Coins().GetCoin(coinID)\n\t\tif reqCoin == nil {\n\t\t\treturn nil, s.createError(status.New(codes.NotFound, \"Coin not found\"), transaction.EncodeError(code.NewCoinNotExists(\"\", coinID.String())))\n\t\t}\n\t}\n\tvar frozen []*pb.FrozenResponse_Frozen\n\n\tcState.FrozenFunds().GetFrozenFunds(s.blockchain.Height())\n\n\tfor i := s.blockchain.Height(); i <= s.blockchain.Height()+candidates.UnbondPeriod; i++ {\n\n\t\tif timeoutStatus := s.checkTimeout(ctx); timeoutStatus != nil {\n\t\t\treturn nil, timeoutStatus.Err()\n\t\t}\n\n\t\tfunds := cState.FrozenFunds().GetFrozenFunds(i)\n\t\tif funds == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, fund := range funds.List {\n\t\t\tif fund.Address.String() != req.Address {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tcoin := reqCoin\n\t\t\tif coin == nil {\n\t\t\t\tcoin = cState.Coins().GetCoin(fund.Coin)\n\t\t\t} else {\n\t\t\t\tif coin.ID() != fund.Coin {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\tfrozen = append(frozen, &pb.FrozenResponse_Frozen{\n\t\t\t\tHeight: funds.Height(),\n\t\t\t\tAddress: fund.Address.String(),\n\t\t\t\tCandidateKey: fund.CandidateKey.String(),\n\t\t\t\tCoin: &pb.Coin{\n\t\t\t\t\tId: uint64(fund.Coin),\n\t\t\t\t\tSymbol: coin.GetFullSymbol(),\n\t\t\t\t},\n\t\t\t\tValue: fund.Value.String(),\n\t\t\t})\n\t\t}\n\t}\n\n\treturn &pb.FrozenResponse{Frozen: frozen}, nil\n}", "func (*NoCopy) Lock() {}", "func (px *Paxos) freeMemory() {\n // Assertion: px is already locked by the callee\n\n // reproduction of Min() without requesting a lock\n // Question: Can I do this without duplciating code?\n min := px.done[px.me]\n for i := 0; i < len(px.done); i++ {\n if px.done[i] < min {\n min = px.done[i]\n }\n }\n min += 1\n\n for i, _ := range px.Instances {\n if i < min {\n delete(px.Instances, i)\n }\n }\n}", "func (_Token *TokenFilterer) FilterFreeze(opts *bind.FilterOpts, from []common.Address) (*TokenFreezeIterator, error) {\n\n\tvar fromRule []interface{}\n\tfor _, fromItem := range from {\n\t\tfromRule = append(fromRule, fromItem)\n\t}\n\n\tlogs, sub, err := _Token.contract.FilterLogs(opts, \"Freeze\", fromRule)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &TokenFreezeIterator{contract: _Token.contract, event: \"Freeze\", logs: logs, sub: sub}, nil\n}", "func (_Token *TokenFilterer) WatchFreeze(opts *bind.WatchOpts, sink chan<- *TokenFreeze, from []common.Address) (event.Subscription, error) {\n\n\tvar fromRule []interface{}\n\tfor _, fromItem := range from {\n\t\tfromRule = append(fromRule, fromItem)\n\t}\n\n\tlogs, sub, err := _Token.contract.WatchLogs(opts, \"Freeze\", fromRule)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(TokenFreeze)\n\t\t\t\tif err := _Token.contract.UnpackLog(event, \"Freeze\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (s EncryptedChatWaitingArray) Retain(keep func(x EncryptedChatWaiting) bool) EncryptedChatWaitingArray {\n\tn := 0\n\tfor _, x := range s {\n\t\tif keep(x) {\n\t\t\ts[n] = x\n\t\t\tn++\n\t\t}\n\t}\n\ts = s[:n]\n\n\treturn s\n}", "func (stackEntry *valuePayloadPropagationStackEntry) Retain() *valuePayloadPropagationStackEntry {\n\treturn &valuePayloadPropagationStackEntry{\n\t\tCachedPayload: stackEntry.CachedPayload.Retain(),\n\t\tCachedPayloadMetadata: stackEntry.CachedPayloadMetadata.Retain(),\n\t\tCachedTransaction: stackEntry.CachedTransaction.Retain(),\n\t\tCachedTransactionMetadata: stackEntry.CachedTransactionMetadata.Retain(),\n\t}\n}", "func (p *BailServiceClient) UnfreezeBail(dealerId int64, amount float64, orderId int64) (r *Bail, err error) {\n\tif err = p.sendUnfreezeBail(dealerId, amount, orderId); err != nil {\n\t\treturn\n\t}\n\treturn p.recvUnfreezeBail()\n}", "func (transaction *TokenUpdateTransaction) GetFreezeKey() Key {\n\treturn transaction.freezeKey\n}", "func (outer outer) Free() bool {\r\n\treturn false\r\n}", "func (o *ParamsReg) Backup() {\n\tcopy(o.bkpTheta, o.theta)\n\to.bkpBias = o.bias\n\to.bkpLambda = o.lambda\n\to.bkpDegree = o.degree\n}", "func (ch *ClickHouse) FreezeTable(table Table) error {\n\tvar partitions []struct {\n\t\tPartitionID string `db:\"partition_id\"`\n\t}\n\tq := fmt.Sprintf(\"SELECT DISTINCT partition_id FROM system.parts WHERE database='%v' AND table='%v'\", table.Database, table.Name)\n\tif err := ch.conn.Select(&partitions, q); err != nil {\n\t\treturn fmt.Errorf(\"can't get partitions for \\\"%s.%s\\\" with %v\", table.Database, table.Name, err)\n\t}\n\tlog.Printf(\"Freeze '%v.%v'\", table.Database, table.Name)\n\tfor _, item := range partitions {\n\t\tif ch.DryRun {\n\t\t\tlog.Printf(\" partition '%v' ...skip becouse dry-run\", item.PartitionID)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Printf(\" partition '%v'\", item.PartitionID)\n\t\t// TODO: find out this magic\n\t\tif item.PartitionID == \"all\" {\n\t\t\titem.PartitionID = \"tuple()\"\n\t\t}\n\t\tif _, err := ch.conn.Exec(\n\t\t\tfmt.Sprintf(\n\t\t\t\t\"ALTER TABLE %v.%v FREEZE PARTITION %v;\",\n\t\t\t\ttable.Database,\n\t\t\t\ttable.Name,\n\t\t\t\titem.PartitionID,\n\t\t\t)); err == nil {\n\t\t\tcontinue\n\t\t}\n\t\tif _, err := ch.conn.Exec(\n\t\t\tfmt.Sprintf(\n\t\t\t\t\"ALTER TABLE %v.%v FREEZE PARTITION '%v';\",\n\t\t\t\ttable.Database,\n\t\t\t\ttable.Name,\n\t\t\t\titem.PartitionID,\n\t\t\t)); err != nil {\n\t\t\treturn fmt.Errorf(\"can't freze partiotion '%s' on '%s.%s' with: %v\", item.PartitionID, table.Database, table.Name, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (*noCopy) Lock() {}", "func Free() {\n\tflags = nil // Any future call to Get() will panic on a nil dereference.\n}", "func (f Fill) SetValue(value float64) Fill {\n\tf.value = value\n\treturn f\n}", "func (x *Value) Free() {\n\tif x != nil && x.allocs23e8c9e3 != nil {\n\t\tx.allocs23e8c9e3.(*cgoAllocMap).Free()\n\t\tx.ref23e8c9e3 = nil\n\t}\n}", "func (tb *tensorBase) Retain() {\n\tatomic.AddInt64(&tb.refCount, 1)\n}", "func SyncRuntimeDoSpin()", "func (r RawValues) Retain(values ...string) RawValues {\n\ttoretain := make(map[string]bool)\n\tfor _, v := range values {\n\t\ttoretain[v] = true\n\t}\n\tfiltered := make([]RawValue, 0)\n\tfor _, rawValue := range r {\n\t\tif _, ok := toretain[rawValue.Value]; ok {\n\t\t\tfiltered = append(filtered, rawValue)\n\t\t}\n\t}\n\treturn filtered\n}", "func (_Storage *StorageCaller) AccountFrozen(opts *bind.CallOpts, addr common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Storage.contract.Call(opts, out, \"accountFrozen\", addr)\n\treturn *ret0, err\n}", "func (v *VolumesServiceMock) Freeze(podUID string, name string) (vol *api.Volume, err error) {\n\targs := v.Called(podUID, name)\n\tx := args.Get(0)\n\tif x != nil {\n\t\tvol = x.(*api.Volume)\n\t}\n\terr = args.Error(1)\n\treturn\n}", "func FixFreezeLookupMigration(db *IndexerDb, state *MigrationState) error {\n\t// Technically with this query no transactions are needed, and the accounting state doesn't need to be locked.\n\tupdateQuery := \"INSERT INTO txn_participation (addr, round, intra) VALUES ($1, $2, $3) ON CONFLICT DO NOTHING\"\n\tquery := fmt.Sprintf(\"select decode(txn.txn->'txn'->>'fadd','base64'),round,intra from txn where typeenum = %d AND txn.txn->'txn'->'snd' != txn.txn->'txn'->'fadd'\", idb.TypeEnumAssetFreeze)\n\trows, err := db.db.Query(query)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to query transactions: %v\", err)\n\t}\n\tdefer rows.Close()\n\n\ttxprows := make([][]interface{}, 0)\n\n\t// Loop through all transactions and compute account data.\n\tdb.log.Print(\"loop through all freeze transactions\")\n\tfor rows.Next() {\n\t\tvar addr []byte\n\t\tvar round, intra uint64\n\t\terr = rows.Scan(&addr, &round, &intra)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error scanning row: %v\", err)\n\t\t}\n\n\t\ttxprows = append(txprows, []interface{}{addr, round, intra})\n\n\t\tif len(txprows) > 5000 {\n\t\t\terr = updateBatch(db, updateQuery, txprows)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"updating batch: %v\", err)\n\t\t\t}\n\t\t\ttxprows = txprows[:0]\n\t\t}\n\t}\n\n\tif rows.Err() != nil {\n\t\treturn fmt.Errorf(\"error while processing freeze transactions: %v\", rows.Err())\n\t}\n\n\t// Commit any leftovers\n\tif len(txprows) > 0 {\n\t\terr = updateBatch(db, updateQuery, txprows)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"updating batch: %v\", err)\n\t\t}\n\t}\n\n\t// Update migration state\n\treturn upsertMigrationState(db, state, true)\n}", "func (stateObj *stateObject) finalise(prefetch bool) {\n\tlog.Debugf(\"stateObject finalise. address:%x, prefetch:%v\", stateObj.address, prefetch)\n\n\tslotsToPrefetch := make([][]byte, 0, len(stateObj.dirtyStorage))\n\tfor key, value := range stateObj.dirtyStorage {\n\t\tstateObj.pendingStorage[key] = value\n\t\tif value != stateObj.originStorage[key] {\n\t\t\tslotsToPrefetch = append(slotsToPrefetch, common.CopyBytes(key[:])) // Copy needed for closure\n\t\t}\n\t}\n\tif stateObj.db.prefetcher != nil && prefetch && len(slotsToPrefetch) > 0 && stateObj.data.Root != emptyRoot {\n\t\tstateObj.db.prefetcher.prefetch(stateObj.addrHash, stateObj.data.Root, slotsToPrefetch)\n\t}\n\tif len(stateObj.dirtyStorage) > 0 {\n\t\tstateObj.dirtyStorage = make(Storage)\n\t}\n}", "func UnfreezeClock(t *testing.T) {\n\tif t == nil {\n\t\tpanic(\"nice try\")\n\t}\n\tc = &DefaultClock{}\n}", "func (this *Tidy) Bare(val bool) (bool, error) {\n\treturn this.optSetBool(C.TidyMakeBare, cBool(val))\n}", "func (o *FreezeParams) WithHTTPClient(client *http.Client) *FreezeParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (m *MemoryStateDB) ExecFrozen(tx *types.Transaction, addr string, amount int64) bool {\n\tif nil == tx {\n\t\tlog15.Error(\"ExecFrozen get nil tx\")\n\t\treturn false\n\t}\n\n\texecaddr := address.ExecAddress(string(tx.Execer))\n\tret, err := m.CoinsAccount.ExecFrozen(addr, execaddr, amount)\n\tif err != nil {\n\t\tlog15.Error(\"ExecFrozen error\", \"addr\", addr, \"execaddr\", execaddr, \"amount\", amount, \"err info\", err)\n\t\treturn false\n\t}\n\n\tm.addChange(balanceChange{\n\t\tamount: amount,\n\t\tdata: ret.KV,\n\t\tlogs: ret.Logs,\n\t})\n\n\treturn true\n}", "func (m *neighborEntryRWMutex) RLockBypass() {\n\tm.mu.RLock()\n}", "func (vals PendingValues) lockSet(inst, prop ident.Id, nilVal, want interface{}) (err error) {\n\tif reflect.TypeOf(nilVal) != reflect.TypeOf(want) {\n\t\terr = SetValueMismatch(inst, prop, nilVal, want)\n\t} else if curr, have := vals[prop]; have && curr != want {\n\t\terr = SetValueChanged(inst, prop, curr, want)\n\t} else {\n\t\tvals[prop] = want\n\t}\n\treturn err\n}", "func (s IPPortSecretArray) Retain(keep func(x IPPortSecret) bool) IPPortSecretArray {\n\tn := 0\n\tfor _, x := range s {\n\t\tif keep(x) {\n\t\t\ts[n] = x\n\t\t\tn++\n\t\t}\n\t}\n\ts = s[:n]\n\n\treturn s\n}", "func (f *IndexFile) Retain() { f.wg.Add(1) }", "func (_Token *TokenTransactor) FreezeTokens(opts *bind.TransactOpts, account common.Address, amount *big.Int) (*types.Transaction, error) {\n\treturn _Token.contract.Transact(opts, \"freezeTokens\", account, amount)\n}", "func (n *metadataNode) assertFrozen() {\n\tif n.acls != nil {\n\t\tpanic(\"not frozen yet\")\n\t}\n}", "func (b *Boolean) Reset() {\n\tb.Value = false\n\tb.Default = false\n\tb.Initialized = false\n\tBooleanPool.Put(b)\n}", "func (l *FixedLimiter) Reset() {\n\tl.mu.Lock()\n\tdefer l.mu.Unlock()\n\tl.value = 0\n}", "func (tester* FreezeTester) nonBpVote(t *testing.T, d *Dandelion) {\n\ta := assert.New(t)\n\ta.True(d.Contract(constants.COSSysAccount, frCrtName).CheckExist())\n\tfreezeAcct := tester.acc5\n\tsta := freezeAcct.GetFreeze()\n\tmemo := freezeAcct.GetFreezeMemo()\n\tnewSta := tester.mdFreezeStatus(sta)\n\ta.NotEqual(sta, newSta)\n\tmemoArray,nameArray := tester.getProposalMemoAndNameParams(d,[]*DandelionAccount{freezeAcct})\n\n\t//1.proposal\n\tApplyNoError(t, d, fmt.Sprintf(\"%s: %s.%s.proposalfreeze %s,%d,%s\", tester.acc0.Name, constants.COSSysAccount, frCrtName, nameArray, newSta, memoArray))\n\t//2.fetch proposal_id\n\tpropId,err := tester.getProposalId(d)\n\ta.NoError(err)\n\t//less than 2/3 bp vote to proposalId\n\ttester.voteById(t, d, propId, 0, tester.threshold-1)\n\t//non bp vote\n\tApplyError(t, d, fmt.Sprintf(\"%s: %s.%s.vote %v\", tester.acc4.Name, constants.COSSysAccount, frCrtName, propId))\n\t//final vote fail, set freeze fail\n\ta.Equal(sta, freezeAcct.GetFreeze())\n\ta.Equal(memo, freezeAcct.GetFreezeMemo())\n\n}", "func (s *Send) FreeVars() []Name {\n\tfv := []Name{}\n\tfor _, v := range s.Vals {\n\t\tfv = append(fv, v)\n\t}\n\tsort.Sort(byName(fv))\n\treturn RemDup(fv)\n}", "func (v Chunk) Retain() {\n\tv.buf.Retain()\n}", "func (p *PKGBUILD) RecomputeValues() {\n\tp.info.RecomputeValues()\n}", "func (m *neighborEntryRWMutex) RUnlockBypass() {\n\tm.mu.RUnlock()\n}", "func (lv *LazyValue) Reset() {\n\tlv.Lock()\n\tdefer lv.Unlock()\n\n\tlv.ready = false\n\tlv.value = values.None\n\tlv.err = nil\n}", "func (this *FeedableBuffer) Minimize() {\n\tthis.Data = this.Data[:this.minByteCount]\n}", "func (s *slotted) UnReserve() {\n\tif atomic.AddInt32((*int32)(s), -1) < 0 {\n\t\tatomic.StoreInt32((*int32)(s), 0)\n\t}\n}", "func (f *Flag) Set() { atomic.CompareAndSwapUint32((*uint32)(unsafe.Pointer(f)), 0, 1) }", "func (tl *TimeLockCondition) Fulfill(fulfillment UnlockFulfillment, ctx FulfillContext) error {\n\tif !tl.Fulfillable(FulfillableContext{BlockHeight: ctx.BlockHeight, BlockTime: ctx.BlockTime}) {\n\t\treturn errors.New(\"time lock has not yet been reached\")\n\t}\n\n\t// time lock hash been reached,\n\t// delegate the actual fulfillment to the given fulfillment, if supported\n\tswitch tf := fulfillment.(type) {\n\tcase *SingleSignatureFulfillment:\n\t\treturn tl.Condition.Fulfill(tf, ctx)\n\tcase *MultiSignatureFulfillment:\n\t\treturn tl.Condition.Fulfill(tf, ctx)\n\tdefault:\n\t\treturn ErrUnexpectedUnlockFulfillment\n\t}\n}", "func DeferLiveness() {\n\tvar x [10]int\n\tescape(&x)\n\tfn := func() {\n\t\tif x[0] != 42 {\n\t\t\tpanic(\"FAIL\")\n\t\t}\n\t}\n\tdefer fn()\n\n\tx[0] = 42\n\truntime.GC()\n\truntime.GC()\n\truntime.GC()\n}", "func (this *Hash) Shrink() {\n\tif this == nil {\n\t\treturn\n\t}\n\n\tif this.lock {\n\t\tthis.mu.Lock()\n\t\tdefer this.mu.Unlock()\n\t}\n\n\tthis.loose.shrink()\n\tthis.compact.shrink(this.loose.a)\n}", "func (cpu *CPU) writeHalfcarryFlag(val bool) {\n if val {\n cpu.f = cpu.f ^ ( 1 << 5 )\n }\n}", "func (n *Number) Reset() {\n\tn.Value = 0.0\n\tn.Initialized = false\n\tNumberPool.Put(n)\n}", "func (x *FzStrokeState) Free() {\n\tif x != nil {\n\t\tC.free(unsafe.Pointer(x))\n\t}\n}", "func (b *LeakyLimiter) Reset() time.Time {\n\tb.remaining = b.capacity\n\treturn b.reset\n}", "func (c *CycleState) Unlock() {\n\tc.mx.Unlock()\n}", "func Fill(value bool) *SimpleElement { return newSEBool(\"fill\", value) }", "func (c *Container) frozen() bool {\n\tif c == nil {\n\t\treturn true\n\t}\n\treturn (c.flags & flagFrozen) != 0\n}", "func (b *Buffer) Retain() {\n\tif b.mem != nil || b.parent != nil {\n\t\tatomic.AddInt64(&b.refCount, 1)\n\t}\n}", "func (m *Mutex) ForceRealse() {\n\tatomic.StoreUint32(&m.l, 0)\n}", "func Relax(out1 *LooseFieldElement, arg1 *TightFieldElement) {\n\tx1 := arg1[0]\n\tx2 := arg1[1]\n\tx3 := arg1[2]\n\tx4 := arg1[3]\n\tx5 := arg1[4]\n\tout1[0] = x1\n\tout1[1] = x2\n\tout1[2] = x3\n\tout1[3] = x4\n\tout1[4] = x5\n}" ]
[ "0.629121", "0.61709946", "0.6089411", "0.60298693", "0.59620976", "0.5877203", "0.58217835", "0.57199144", "0.5709729", "0.5635698", "0.5491367", "0.5444173", "0.5394565", "0.536854", "0.5366694", "0.53647643", "0.5217478", "0.5201497", "0.5199691", "0.5163645", "0.5118248", "0.5094646", "0.50705296", "0.49911997", "0.49585488", "0.49480417", "0.49474224", "0.4928673", "0.49135697", "0.4906327", "0.48996666", "0.4885424", "0.4855675", "0.48224056", "0.48167235", "0.4797175", "0.47290012", "0.45966762", "0.4593359", "0.45928454", "0.45717788", "0.45235473", "0.45216572", "0.4518003", "0.45089075", "0.45087507", "0.44762477", "0.44738653", "0.44621104", "0.4441628", "0.4440062", "0.44034517", "0.43876907", "0.43816802", "0.4362988", "0.43589383", "0.4355712", "0.43394676", "0.43389082", "0.4318574", "0.43118986", "0.4294725", "0.42940456", "0.4269193", "0.4262917", "0.42545116", "0.4234557", "0.42305636", "0.4225275", "0.4220057", "0.4214381", "0.42096722", "0.42051747", "0.4176754", "0.41728988", "0.4169668", "0.41670215", "0.41624624", "0.41623878", "0.41623512", "0.41481364", "0.41467074", "0.41422772", "0.41346204", "0.41344106", "0.41291258", "0.41181058", "0.41177505", "0.41171423", "0.41154474", "0.4115406", "0.4113516", "0.410938", "0.41078776", "0.40973213", "0.40971446", "0.40919363", "0.40890986", "0.40885434", "0.40812027" ]
0.45463032
41
Truth implements the starlark.Value.Truth() method.
func (gg GlobGroup) Truth() starlark.Bool { return starlark.True }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func truth(r reflect.Value) bool {\nout:\n\tswitch r.Kind() {\n\tcase reflect.Array, reflect.Slice:\n\t\treturn r.Len() > 0\n\tcase reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\t\treturn r.Int() > 0\n\tcase reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64:\n\t\treturn r.Uint() > 0\n\tcase reflect.Float32, reflect.Float64:\n\t\treturn r.Float() > 0\n\tcase reflect.String:\n\t\treturn r.String() != \"\"\n\tcase reflect.Bool:\n\t\treturn r.Bool()\n\tcase reflect.Ptr, reflect.Interface:\n\t\tr = r.Elem()\n\t\tgoto out\n\tdefault:\n\t\treturn r.Interface() != nil\n\t}\n}", "func True() TermT {\n\treturn TermT(C.yices_true())\n}", "func True(actual interface{}) Truth {\n\tmustBeCleanStart()\n\treturn Truth{actual.(bool), fmt.Sprintf(\"%#v\", actual)}\n}", "func (v *Value) Bool() bool {\n return Util.ToBool(v.data)\n}", "func (t *Target) Truth() starlark.Bool { return starlark.Bool(t == nil) }", "func (m *kubePackage) Truth() starlark.Bool { return starlark.True }", "func True(t TestingT, v interface{}, extras ...interface{}) bool {\n\tval, ok := v.(bool)\n\tif !ok || val != true {\n\t\texps, acts := toString(true, v)\n\n\t\treturn Errorf(t, \"Expect to be true\", []labeledOutput{\n\t\t\t{\n\t\t\t\tlabel: labelMessages,\n\t\t\t\tcontent: formatExtras(extras...),\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"-expected\",\n\t\t\t\tcontent: exps,\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"+received\",\n\t\t\t\tcontent: acts,\n\t\t\t},\n\t\t})\n\t}\n\n\treturn true\n}", "func (n *Node) Bool() bool", "func (v Value) Bool() (bool, error) {\n\tif v.typ != Bool {\n\t\treturn false, v.newError(\"%s is not a bool\", v.Raw())\n\t}\n\treturn v.boo, nil\n}", "func (m *Value) Bool() bool { return m.BoolMock() }", "func (v Value) Bool() bool {\n\treturn v.Integer() != 0\n}", "func True(v bool) {\n\tassert(v, 1, \"assert true\")\n}", "func Bool(x bool) bool { return x }", "func (sp booleanSpace) BoolOutcome(value bool) Outcome {\n\tif value {\n\t\treturn 1\n\t} else {\n\t\treturn 0\n\t}\n}", "func (i *Instance) Truth() exprcore.Bool {\n\treturn exprcore.True\n}", "func (sp booleanSpace) BoolValue(outcome Outcome) bool {\n\treturn outcome != 0\n}", "func (v Value) Truthy() bool {\n\tswitch v.Type() {\n\tcase TypeUndefined, TypeNull:\n\t\treturn false\n\tcase TypeBoolean:\n\t\treturn v.Bool()\n\tcase TypeNumber:\n\t\treturn !v.IsNaN() && v.Float() != 0\n\tcase TypeString:\n\t\treturn v.String() != \"\"\n\tcase TypeSymbol, TypeFunction, TypeObject:\n\t\treturn true\n\tdefault:\n\t\tpanic(\"bad type\")\n\t}\n}", "func (r Record) Bool(key string) bool {\n\tif r.dropped {\n\t\tlog.Fatalf(\"Int called on dropped record\")\n\t}\n\n\tval, ok := r.values[key]\n\tif !ok {\n\t\treturn false\n\t}\n\n\tswitch val := val.(type) {\n\tcase nil:\n\t\treturn false\n\tcase string:\n\t\treturn val == \"true\"\n\tcase int64:\n\t\treturn val != 0\n\tcase float64:\n\t\treturn val != 0.0\n\tcase bool:\n\t\treturn val\n\t}\n\n\treturn false\n}", "func (o BoolOperand) Evaluate(cxt interface{}) (bool, error) {\n\treturn o.Value, nil\n}", "func (o Nil) Truthy() bool { return false }", "func (v Value) Truthy() bool {\n\tpanic(message)\n}", "func (tr Row) Bool(nn int) (val bool) {\n\tval, err := tr.BoolErr(nn)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn\n}", "func Bool(a bool) cell.I {\n\tif a {\n\t\treturn sym.True\n\t}\n\n\treturn pair.Null\n}", "func (b *Bool) Literal() {}", "func (tv *TypedBool) Bool() bool {\n\treturn tv.Bytes[0] == 1\n}", "func (v Value) Bool() bool {\n\treturn v.v.Bool()\n}", "func (b *Boolean) True() *Boolean {\n\treturn b.Equal(true)\n}", "func (BooleanLiteral) literalNode() {}", "func (r *TTNRandom) Bool() bool {\n\treturn r.Interface.Intn(2) == 0\n}", "func True(Right) bool {\n\treturn true\n}", "func (n Nil) True() bool { return false }", "func (v *Value) Bool() (bool, error) {\n\tif v.kind == kindBool {\n\t\treturn v.boolContent, nil\n\t}\n\treturn false, nil\n}", "func IsTruthy(val string) bool {\n\tswitch strings.ToLower(strings.TrimSpace(val)) {\n\tcase \"y\", \"yes\", \"true\", \"t\", \"on\", \"1\":\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func Truthy(t *testing.T, name string, v bool) {\n\tif !v {\n\t\tflux.FatalFailed(t, \"Expected truthy value for %s\", name)\n\t} else {\n\t\tflux.LogPassed(t, \"%s passed with truthy value\", name)\n\t}\n}", "func Bool(r interface{}, err error) (bool, error) {\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tswitch r := r.(type) {\n\tcase bool:\n\t\treturn r, err\n\t// Very common in redis to reply int64 with 0 for bool flag.\n\tcase int:\n\t\treturn r != 0, nil\n\tcase int64:\n\t\treturn r != 0, nil\n\tcase []byte:\n\t\treturn strconv.ParseBool(string(r))\n\tcase string:\n\t\treturn strconv.ParseBool(r)\n\tcase nil:\n\t\treturn false, simplesessions.ErrNil\n\t}\n\treturn false, simplesessions.ErrAssertType\n}", "func BoolVal(x Value) bool {\n\treturn constant.BoolVal(x)\n}", "func (o *FakeObject) Bool() bool { return o.Value.(bool) }", "func False() TermT {\n\treturn TermT(C.yices_false())\n}", "func True(t Testing, v interface{}, formatAndArgs ...interface{}) bool {\n\tvar tv bool\n\tswitch v.(type) {\n\tcase bool:\n\t\ttv = v.(bool)\n\t}\n\n\tif tv != true {\n\t\treturn Fail(t, pretty.Sprintf(\"Expected %# v to be true\", v), formatAndArgs...)\n\t}\n\n\treturn true\n}", "func BoolConstValue(t TermT, val *int32) int32 {\n\treturn int32(C.yices_bool_const_value(C.term_t(t), (*C.int32_t)(val)))\n}", "func (s *Smpval) Bool() bool {\n\treturn s.b\n}", "func Bool(v *Value, def bool) bool {\n\tb, err := v.Bool()\n\tif err != nil {\n\t\treturn def\n\t}\n\treturn b\n}", "func (n BoolWrapper) Value() (Value, error) {\n\tif !n.Valid {\n\t\treturn nil, nil\n\t}\n\treturn n.Bool, nil\n}", "func (s *Sub) Truth() starlark.Bool { return s == nil }", "func wrapWithIsTrue(ctx sessionctx.Context, keepNull bool, arg Expression) (Expression, error) {\n\tif arg.GetType().EvalType() == types.ETInt {\n\t\treturn arg, nil\n\t}\n\tfc := &isTrueOrFalseFunctionClass{baseFunctionClass{ast.IsTruth, 1, 1}, opcode.IsTruth, keepNull}\n\tf, err := fc.getFunction(ctx, []Expression{arg})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsf := &ScalarFunction{\n\t\tFuncName: model.NewCIStr(ast.IsTruth),\n\t\tFunction: f,\n\t\tRetType: f.getRetTp(),\n\t}\n\treturn FoldConstant(sf), nil\n}", "func True(t testing.TB, value bool, msgAndArgs ...interface{}) bool {\n\tif !value {\n\t\treturn failTest(t, 1, fmt.Sprintf(\"True: expected `true`, actual `%#v`\", value), msgAndArgs...)\n\t}\n\n\treturn true\n}", "func (e *Encoder) Bool(v bool) (int, error) {\n\tif v {\n\t\treturn e.Byte(0x1)\n\t}\n\treturn e.Byte(0x0)\n}", "func Bool(v bool) (p *bool) { return &v }", "func (n *NotLikeOp) IsTrue(left, right EvalResult) (bool, error) {\n\treturn false, nil\n}", "func (v Bool) Bool() bool {\n\treturn v.v\n}", "func BoolVal(b *bool) bool {\n\tif b == nil {\n\t\treturn false\n\t}\n\treturn *b\n}", "func Bool(val interface{}) (bool, error) {\n\tif val == nil {\n\t\treturn false, nil\n\t}\n\tswitch ret := val.(type) {\n\tcase bool:\n\t\treturn ret, nil\n\tcase int, int8, int16, int32, int64, float32, float64, uint, uint8, uint16, uint32, uint64:\n\t\treturn ret != 0, nil\n\tcase []byte:\n\t\treturn stringToBool(string(ret))\n\tcase string:\n\t\treturn stringToBool(ret)\n\tdefault:\n\t\treturn false, converError(val, \"bool\")\n\t}\n}", "func (b *Bool) Value() bool {\n\t// generate nil checks and faults early.\n\tref := &b.i\n\treturn atomic.LoadUint64(ref) == 1\n}", "func TestNewResult_bool(t *testing.T) {\n\tvar reading interface{} = true\n\treq := models.CommandRequest{\n\t\tDeviceResourceName: \"light\",\n\t\tType: common.ValueTypeBool,\n\t}\n\n\tcmdVal, err := newResult(req, reading)\n\tif err != nil {\n\t\tt.Fatalf(\"Fail to create new ReadCommand result, %v\", err)\n\t}\n\tval, err := cmdVal.BoolValue()\n\tif val != true || err != nil {\n\t\tt.Errorf(\"Convert new result(%v) failed, error: %v\", val, err)\n\t}\n}", "func (n *SoupNode) Truth() starlark.Bool {\n\treturn n != nil\n}", "func (v *Value) Bool() bool {\n\treturn (bool)(C.value_get_bool(v.value))\n}", "func IsTruthy(s string) bool {\n\treturn s == \"1\" || strings.EqualFold(s, \"true\")\n}", "func (h *Random) Bool() bool {\n\tboolList := []bool{true, false}\n\trandomIndex := rand.Intn(len(boolList))\n\treturn boolList[randomIndex]\n}", "func TestCheckBinaryExprBoolRhlBool(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `true >> true`, env,\n\t\t`invalid operation: true >> true (shift count type bool, must be unsigned integer)`,\n\t)\n\n}", "func Bool(value interface{}) bool {\n\tret, err := BoolE(value)\n\tif err != nil {\n\t\tmod.Error(err)\n\t}\n\treturn ret\n}", "func (nvp *NameValues) Bool(name string) (bool, bool) {\n\tvalue, _ := nvp.String(name)\n\treturn (value == \"true\" || value == \"yes\" || value == \"1\" || value == \"-1\" || value == \"on\"), true\n}", "func Boolean() Scalar {\n\treturn booleanTypeInstance\n}", "func TestCheckBinaryExprBoolNeqBool(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectConst(t, `true != true`, env, (true != true), ConstBool)\n}", "func (v Value) Bool() bool {\n\tpanic(message)\n}", "func (l *LikeOp) IsTrue(left, right EvalResult) (bool, error) {\n\treturn false, nil\n}", "func TestFetchTrueBool(t *testing.T) {\n\texpected := \"true\"\n\tinput := \"rue\"\n\treader := bytes.NewReader([]byte(input))\n\tlex := NewLexer(reader)\n\tif err := lex.fetchTrueBool(); err != nil {\n\t\tt.Error(err.Error())\n\t\treturn\n\t}\n\n\tif len(lex.tokens) != 1 {\n\t\tt.Error(\"expecting 1 token to be fetched\")\n\t\treturn\n\t}\n\n\ttoken := lex.tokens[0]\n\tif token.t != TokenBool {\n\t\tt.Errorf(\"unexpected token type %d (%s), expecting token type %d (%s)\", token.t, tokenTypeMap[token.t], TokenBool, tokenTypeMap[TokenBool])\n\t\treturn\n\t}\n\n\tif token.String() != expected {\n\t\tt.Errorf(\"unexpected %s, expecting %s\", token.String(), expected)\n\t}\n}", "func isTrue(val reflect.Value) (truth, ok bool) {\n\tif !val.IsValid() {\n\t\t// Something like var x interface{}, never set. It's a form of nil.\n\t\treturn false, true\n\t}\n\tswitch val.Kind() {\n\tcase reflect.Array, reflect.Map, reflect.Slice, reflect.String:\n\t\ttruth = val.Len() > 0\n\tcase reflect.Bool:\n\t\ttruth = val.Bool()\n\tcase reflect.Complex64, reflect.Complex128:\n\t\ttruth = val.Complex() != 0\n\tcase reflect.Chan, reflect.Func, reflect.Ptr, reflect.Interface:\n\t\ttruth = !val.IsNil()\n\tcase reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\t\ttruth = val.Int() != 0\n\tcase reflect.Float32, reflect.Float64:\n\t\ttruth = val.Float() != 0\n\tcase reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:\n\t\ttruth = val.Uint() != 0\n\tcase reflect.Struct:\n\t\ttruth = true // Struct values are always true.\n\tdefault:\n\t\treturn\n\t}\n\treturn truth, true\n}", "func (sr *StringResult) Bool() (bool, error) {\n\treturn redis.Bool(sr.val, nil)\n}", "func (r *Redis) Bool(reply interface{}, err error) (bool, error) {\n\treturn redigo.Bool(reply, err)\n}", "func TestBoolVal(t *testing.T) {\n\tConvey(\"Testing BoolVal\", t, func() {\n\t\ttrueValues := []string{\"yes\", \"ok\", \"true\", \"1\", \"enabled\", \"True\", \"TRUE\", \"YES\", \"Yes\"}\n\t\tfalseValues := []string{\"no\", \"0\", \"false\", \"False\", \"FALSE\", \"disabled\"}\n\t\tfor _, val := range trueValues {\n\t\t\tSo(BoolVal(val), ShouldBeTrue)\n\t\t}\n\t\tfor _, val := range falseValues {\n\t\t\tSo(BoolVal(val), ShouldBeFalse)\n\t\t}\n\t})\n}", "func TestCheckBinaryExprBoolEqlBool(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectConst(t, `true == true`, env, (true == true), ConstBool)\n}", "func TestCheckBinaryExprBoolGtrBool(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `true > true`, env,\n\t\t`invalid operation: true > true (operator > not defined on bool)`,\n\t)\n\n}", "func (opa *OPA) Bool(ctx context.Context, input interface{}, opts ...func(*rego.Rego)) (bool, error) {\n\n\tm := metrics.New()\n\tvar decisionID string\n\tvar revision string\n\tvar decision bool\n\n\terr := storage.Txn(ctx, opa.manager.Store, storage.TransactionParams{}, func(txn storage.Transaction) error {\n\n\t\tvar err error\n\n\t\trevision, err = getRevision(ctx, opa.manager.Store, txn)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdecisionID, err = uuid4()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\topts = append(opts,\n\t\t\trego.Metrics(m),\n\t\t\trego.Query(defaultDecision),\n\t\t\trego.Input(input),\n\t\t\trego.Compiler(opa.manager.GetCompiler()),\n\t\t\trego.Store(opa.manager.Store),\n\t\t\trego.Transaction(txn))\n\n\t\trs, err := rego.New(opts...).Eval(ctx)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t} else if len(rs) == 0 {\n\t\t\treturn fmt.Errorf(\"undefined decision\")\n\t\t} else if b, ok := rs[0].Expressions[0].Value.(bool); !ok || len(rs) > 1 {\n\t\t\treturn fmt.Errorf(\"non-boolean decision\")\n\t\t} else {\n\t\t\tdecision = b\n\t\t}\n\n\t\treturn nil\n\t})\n\n\tif opa.decisionLogsPlugin != nil {\n\t\trecord := &server.Info{\n\t\t\tRevision: revision,\n\t\t\tDecisionID: decisionID,\n\t\t\tTimestamp: time.Now(),\n\t\t\tQuery: defaultDecision,\n\t\t\tInput: input,\n\t\t\tError: err,\n\t\t\tMetrics: m,\n\t\t}\n\t\tif err == nil {\n\t\t\tvar x interface{} = decision\n\t\t\trecord.Results = &x\n\t\t}\n\t\topa.decisionLogsPlugin.Log(ctx, record)\n\t}\n\n\treturn decision, err\n}", "func False(v bool) {\n\tassert(!v, 1, \"assert false\")\n}", "func isTruthy(o object.Object) bool {\n\tswitch o {\n\tcase ConstFalse, ConstNil:\n\t\treturn false\n\tdefault:\n\t\t// special case: 0 or 0.0 is not truthy\n\t\tswitch o.Type() {\n\t\tcase object.IntType:\n\t\t\tif o.(*object.Integer).Value == 0 {\n\t\t\t\treturn false\n\t\t\t}\n\t\tcase object.FloatType:\n\t\t\tif o.(*object.Float).Value == 0.0 {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\n\t\treturn true\n\t}\n}", "func isBoolean(t Type) bool { return isBasic(t, IsBoolean) }", "func Bool(a bool, b bool) bool {\n\treturn a == b\n}", "func (r *Decoder) Bool() bool {\n\tr.Sync(SyncBool)\n\tx, err := r.Data.ReadByte()\n\tr.checkErr(err)\n\tassert(x < 2)\n\treturn x != 0\n}", "func Bool(v *bool) bool {\n\tif v != nil {\n\t\treturn *v\n\t}\n\treturn false\n}", "func (this *Not) Type() value.Type { return value.BOOLEAN }", "func False(t TestingT, v interface{}, extras ...interface{}) bool {\n\tval, ok := v.(bool)\n\tif !ok || val != false {\n\t\texps, acts := toString(false, v)\n\n\t\treturn Errorf(t, \"Expect to be false\", []labeledOutput{\n\t\t\t{\n\t\t\t\tlabel: labelMessages,\n\t\t\t\tcontent: formatExtras(extras...),\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"-expected\",\n\t\t\t\tcontent: exps,\n\t\t\t},\n\t\t\t{\n\t\t\t\tlabel: \"+received\",\n\t\t\t\tcontent: acts,\n\t\t\t},\n\t\t})\n\t}\n\n\treturn true\n}", "func (n *eeNum) bool() *bool { return (*bool)(unsafe.Pointer(&n.data)) }", "func (data *Data) Bool(s ...string) bool {\n\treturn data.Interface(s...).(bool)\n}", "func (*Base) LiteralBoolean(p ASTPass, node *ast.LiteralBoolean, ctx Context) {\n}", "func (v Boolean) Bool() bool {\n\treturn v.v\n}", "func (p Path) Truth() starlark.Bool { return starlark.True }", "func (b *Boolean) Raw() bool {\n\treturn b.value\n}", "func Bool(b bool) int {\n\tif b {\n\t\treturn 1\n\t}\n\treturn 0\n}", "func BoolValue(t bool) Value {\n\tif t {\n\t\treturn Value{Typ: ':', IntegerV: 1}\n\t}\n\treturn Value{Typ: ':', IntegerV: 0}\n}", "func True(t testing.TB, ok bool, msgAndArgs ...interface{}) {\n\tif ok {\n\t\treturn\n\t}\n\tt.Helper()\n\tt.Fatal(formatMsgAndArgs(\"Expected expression to be true\", msgAndArgs...))\n}", "func TestGetBooleanTrue(t *testing.T) {\n\tclient := newQueriesClient(t)\n\tresult, err := client.GetBooleanTrue(context.Background(), nil)\n\trequire.NoError(t, err)\n\trequire.Zero(t, result)\n}", "func Bool(b bool) Cell {\n\tif b {\n\t\treturn True\n\t}\n\treturn Nil\n}", "func (c *C) Bool() Type {\n\t// TODO: support custom bool types\n\treturn c.e.Go().Bool()\n}", "func (t *Typed) Bool(key string) bool {\n\treturn t.BoolOr(key, false)\n}", "func (w *Writer) Bool(b bool) error {\n\tif b {\n\t\treturn w.Bit(1)\n\t}\n\treturn w.Bit(0)\n}", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }", "func Bool(v bool) *bool { return &v }" ]
[ "0.6302526", "0.6080101", "0.60129184", "0.58864635", "0.58375704", "0.5808124", "0.57730013", "0.5767958", "0.5739565", "0.57394236", "0.57306063", "0.5705854", "0.569673", "0.5695482", "0.56939113", "0.5689309", "0.56702256", "0.5668591", "0.5640195", "0.5632629", "0.561903", "0.5585214", "0.5531326", "0.5501409", "0.5480797", "0.54750484", "0.54680777", "0.54490614", "0.5448442", "0.5439179", "0.54279584", "0.5425224", "0.5417215", "0.5402517", "0.53970605", "0.5389342", "0.53853536", "0.53770196", "0.5376137", "0.5358787", "0.5339695", "0.53338957", "0.5316804", "0.53156185", "0.5313072", "0.531223", "0.53105205", "0.5305221", "0.5284437", "0.528295", "0.52826446", "0.5279908", "0.5275551", "0.5272672", "0.5268835", "0.5268199", "0.5267804", "0.5264897", "0.52616864", "0.52612823", "0.5253428", "0.5250413", "0.5227911", "0.52198505", "0.52143174", "0.5208065", "0.52067065", "0.52066696", "0.519961", "0.5198307", "0.5194877", "0.5192432", "0.51867753", "0.5182572", "0.51812017", "0.51743495", "0.5170889", "0.51696765", "0.5156138", "0.5154562", "0.5154103", "0.51323307", "0.5132321", "0.5126496", "0.5118254", "0.51149327", "0.51061475", "0.51024115", "0.5099301", "0.5098613", "0.5091731", "0.5083631", "0.50834954", "0.5081787", "0.5059703", "0.50565106", "0.50565106", "0.50565106", "0.50565106", "0.50565106", "0.50565106" ]
0.0
-1
Hash32 implements the Arg.Hash32() method.
func (gg GlobGroup) Hash32(h hash.Hash32) { for _, p := range gg { h.Write([]byte(p)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p Path) Hash32(h hash.Hash32) { h.Write([]byte(p)) }", "func (t *Target) Hash32(h hash.Hash32) {\n\th.Write([]byte(t.Name))\n\th.Write([]byte(t.Builder))\n\tfor _, arg := range t.Args {\n\t\targ.Hash32(h)\n\t}\n\tfor _, env := range t.Env {\n\t\th.Write([]byte(env))\n\t}\n}", "func CalcHash32(data []byte) Hash32 {\n\treturn hash.Sum(data)\n}", "func (s String) Hash32(h hash.Hash32) { h.Write([]byte(s)) }", "func HexToHash32(s string) Hash32 { return BytesToHash(util.FromHex(s)) }", "func (s *Sub) Hash32(h hash.Hash32) {\n\th.Write([]byte(s.Format))\n\tfor _, sub := range s.Substitutions {\n\t\th.Write([]byte(sub.Key))\n\t\tsub.Value.Hash32(h)\n\t}\n}", "func (ch *ConsistentHash) fnv32Hash(key string) uint32 {\n\tnew32Hash := fnv.New32()\n\tnew32Hash.Write([]byte(key))\n\treturn new32Hash.Sum32()\n}", "func hash(s string) uint32 {\n h := fnv.New32a()\n h.Write([]byte(s))\n return h.Sum32()\n}", "func hash(value string) uint32 {\n\th := fnv.New32a()\n\th.Write([]byte(value))\n\n\treturn h.Sum32()\n}", "func Hash(strings ...string) uint32 {\n\tdigester := fnv.New32()\n\tfor _, s := range strings {\n\t\t_, _ = io.WriteString(digester, s)\n\t}\n\treturn digester.Sum32()\n}", "func Hash(key string) uint32 {\n\treturn uint32(aeshashstr(noescape(unsafe.Pointer(&key)), 0))\n}", "func f32hash(p unsafe.Pointer, h uintptr) uintptr {\n\tf := *(*float32)(p)\n\tswitch {\n\tcase f == 0:\n\t\treturn c1 * (c0 ^ h) // +0, -0\n\tcase f != f:\n\t\treturn c1 * (c0 ^ h ^ uintptr(fastrand())) // any kind of NaN\n\tdefault:\n\t\treturn memhash(p, h, 4)\n\t}\n}", "func FNVHash32(value uint32) uint32 {\n\thash := FNVOffsetBasis32\n\tfor i := 0; i < 4; i++ {\n\t\toctet := value & 0x00FF\n\t\tvalue >>= 8\n\n\t\thash ^= octet\n\t\thash *= FNVPrime32\n\t}\n\treturn hash\n}", "func hash(x []byte) uint32 {\n\treturn crc32.ChecksumIEEE(x)\n}", "func Hash32(s []byte) uint32 {\n\tn := uint32(len(s))\n\tif n <= 24 {\n\t\tif n <= 12 {\n\t\t\tif n <= 4 {\n\t\t\t\treturn hash32Len0to4(s)\n\t\t\t}\n\t\t\treturn hash32Len5to12(s)\n\t\t}\n\t\treturn hash32Len13to24(s)\n\t}\n\n\t// n > 24\n\th := n\n\tg := c1 * n\n\tf := g\n\n\ta0 := ror32(fetch32(s[n-4:])*c1, 17) * c2\n\ta1 := ror32(fetch32(s[n-8:])*c1, 17) * c2\n\ta2 := ror32(fetch32(s[n-16:])*c1, 17) * c2\n\ta3 := ror32(fetch32(s[n-12:])*c1, 17) * c2\n\ta4 := ror32(fetch32(s[n-20:])*c1, 17) * c2\n\n\tconst magic = 0xe6546b64\n\th ^= a0\n\th = ror32(h, 19)\n\th = h*5 + magic\n\th ^= a2\n\th = ror32(h, 19)\n\th = h*5 + magic\n\tg ^= a1\n\tg = ror32(g, 19)\n\tg = g*5 + magic\n\tg ^= a3\n\tg = ror32(g, 19)\n\tg = g*5 + magic\n\tf += a4\n\tf = ror32(f, 19)\n\tf = f*5 + magic\n\tfor i := (n - 1) / 20; i != 0; i-- {\n\t\ta0 := ror32(fetch32(s)*c1, 17) * c2\n\t\ta1 := fetch32(s[4:])\n\t\ta2 := ror32(fetch32(s[8:])*c1, 17) * c2\n\t\ta3 := ror32(fetch32(s[12:])*c1, 17) * c2\n\t\ta4 := fetch32(s[16:])\n\t\th ^= a0\n\t\th = ror32(h, 18)\n\t\th = h*5 + magic\n\t\tf += a1\n\t\tf = ror32(f, 19)\n\t\tf = f * c1\n\t\tg += a2\n\t\tg = ror32(g, 18)\n\t\tg = g*5 + magic\n\t\th ^= a3 + a1\n\t\th = ror32(h, 19)\n\t\th = h*5 + magic\n\t\tg ^= a4\n\t\tg = bswap32(g) * 5\n\t\th += a4 * 5\n\t\th = bswap32(h)\n\t\tf += a0\n\t\tf, g, h = g, h, f // a.k.a. PERMUTE3\n\t\ts = s[20:]\n\t}\n\tg = ror32(g, 11) * c1\n\tg = ror32(g, 17) * c1\n\tf = ror32(f, 11) * c1\n\tf = ror32(f, 17) * c1\n\th = ror32(h+g, 19)\n\th = h*5 + magic\n\th = ror32(h, 17) * c1\n\th = ror32(h+f, 19)\n\th = h*5 + magic\n\th = ror32(h, 17) * c1\n\treturn h\n}", "func hash(s string) int {\n\th := fnv.New32a()\n\tif _, err := h.Write([]byte(s)); err != nil {\n\t\tpanic(err) // should never happen\n\t}\n\n\treturn int(h.Sum32() & 0x7FFFFFFF) // mask MSB of uint32 as this will be sign bit\n}", "func Hash32(s []byte) uint32 {\n\n\tslen := len(s)\n\n\tif slen <= 24 {\n\t\tif slen <= 12 {\n\t\t\tif slen <= 4 {\n\t\t\t\treturn hash32Len0to4(s, 0)\n\t\t\t}\n\t\t\treturn hash32Len5to12(s, 0)\n\t\t}\n\t\treturn hash32Len13to24Seed(s, 0)\n\t}\n\n\t// len > 24\n\th := uint32(slen)\n\tg := c1 * uint32(slen)\n\tf := g\n\ta0 := rotate32(fetch32(s, slen-4)*c1, 17) * c2\n\ta1 := rotate32(fetch32(s, slen-8)*c1, 17) * c2\n\ta2 := rotate32(fetch32(s, slen-16)*c1, 17) * c2\n\ta3 := rotate32(fetch32(s, slen-12)*c1, 17) * c2\n\ta4 := rotate32(fetch32(s, slen-20)*c1, 17) * c2\n\th ^= a0\n\th = rotate32(h, 19)\n\th = h*5 + 0xe6546b64\n\th ^= a2\n\th = rotate32(h, 19)\n\th = h*5 + 0xe6546b64\n\tg ^= a1\n\tg = rotate32(g, 19)\n\tg = g*5 + 0xe6546b64\n\tg ^= a3\n\tg = rotate32(g, 19)\n\tg = g*5 + 0xe6546b64\n\tf += a4\n\tf = rotate32(f, 19) + 113\n\titers := (slen - 1) / 20\n\tfor {\n\t\ta := fetch32(s, 0)\n\t\tb := fetch32(s, 4)\n\t\tc := fetch32(s, 8)\n\t\td := fetch32(s, 12)\n\t\te := fetch32(s, 16)\n\t\th += a\n\t\tg += b\n\t\tf += c\n\t\th = mur(d, h) + e\n\t\tg = mur(c, g) + a\n\t\tf = mur(b+e*c1, f) + d\n\t\tf += g\n\t\tg += f\n\t\ts = s[20:]\n\t\titers--\n\t\tif iters == 0 {\n\t\t\tbreak\n\t\t}\n\t}\n\tg = rotate32(g, 11) * c1\n\tg = rotate32(g, 17) * c1\n\tf = rotate32(f, 11) * c1\n\tf = rotate32(f, 17) * c1\n\th = rotate32(h+g, 19)\n\th = h*5 + 0xe6546b64\n\th = rotate32(h, 17) * c1\n\th = rotate32(h+f, 19)\n\th = h*5 + 0xe6546b64\n\th = rotate32(h, 17) * c1\n\treturn h\n}", "func (h Hash20) ToHash32() (h32 Hash32) {\n\tcopy(h32[:], h[:])\n\treturn\n}", "func TestExample(t *testing.T) {\n\tstr := \"hello world\"\n\tbytes := []byte(str)\n\thash := Hash32(bytes)\n\tfmt.Printf(\"Hash32(%s) is %x\\n\", str, hash)\n}", "func hash(data []byte) uint32 {\n\tvar h uint32 = binary.LittleEndian.Uint32(data) * kDictHashMul32\n\n\t/* The higher bits contain more mixture from the multiplication,\n\t so we take our results from there. */\n\treturn h >> uint(32-kDictNumBits)\n}", "func (h *Hash) Sum32() (uint32, bool) {\n\th32, ok := h.Hash.(hash.Hash32)\n\tif !ok {\n\t\treturn 0, false\n\t}\n\n\treturn h32.Sum32(), true\n}", "func (t *hashReader) Sum32() uint32 {\n\treturn t.h.Sum32()\n}", "func (d Data32) Hash() Hash {\n\treturn hash(d)\n}", "func hash3(u uint32, h uint8) uint32 {\n\treturn ((u << (32 - 24)) * prime3bytes) >> ((32 - h) & 31)\n}", "func hash(elements ...[32]byte) [32]byte {\n\tvar hash []byte\n\tfor i := range elements {\n\t\thash = append(hash, elements[i][:]...)\n\t}\n\treturn sha256.Sum256(hash)\n}", "func strhash(a unsafe.Pointer, h uintptr) uintptr", "func CalcObjectHash32(obj scale.Encodable) Hash32 {\n\tbytes, err := codec.Encode(obj)\n\tif err != nil {\n\t\tpanic(\"could not serialize object\")\n\t}\n\treturn CalcHash32(bytes)\n}", "func (h *Hash) IsHash32() bool {\n\t_, ok := h.Hash.(hash.Hash32)\n\treturn ok\n}", "func TestHash32(t *testing.T) {\n\tstdHash := crc32.New(crc32.IEEETable)\n\tif _, err := stdHash.Write([]byte(\"test\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\t// create a new hash with stdHash.Sum32() as initial crc\n\tcrcHash := New(stdHash.Sum32(), crc32.IEEETable)\n\n\tstdHashSize := stdHash.Size()\n\tcrcHashSize := crcHash.Size()\n\tif stdHashSize != crcHashSize {\n\t\tt.Fatalf(\"%d != %d\", stdHashSize, crcHashSize)\n\t}\n\n\tstdHashBlockSize := stdHash.BlockSize()\n\tcrcHashBlockSize := crcHash.BlockSize()\n\tif stdHashBlockSize != crcHashBlockSize {\n\t\tt.Fatalf(\"%d != %d\", stdHashBlockSize, crcHashBlockSize)\n\t}\n\n\tstdHashSum32 := stdHash.Sum32()\n\tcrcHashSum32 := crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n\n\tstdHashSum := stdHash.Sum(make([]byte, 32))\n\tcrcHashSum := crcHash.Sum(make([]byte, 32))\n\tif !reflect.DeepEqual(stdHashSum, crcHashSum) {\n\t\tt.Fatalf(\"sum = %v, want %v\", crcHashSum, stdHashSum)\n\t}\n\n\t// write something\n\tif _, err := stdHash.Write([]byte(\"hello\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif _, err := crcHash.Write([]byte(\"hello\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tstdHashSum32 = stdHash.Sum32()\n\tcrcHashSum32 = crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n\n\t// reset\n\tstdHash.Reset()\n\tcrcHash.Reset()\n\tstdHashSum32 = stdHash.Sum32()\n\tcrcHashSum32 = crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n}", "func Hash(length int, key string) int64 {\n\tif key == \"\" {\n\t\treturn 0\n\t}\n\thc := hashCode(key)\n\treturn (hc ^ (hc >> 16)) % int64(length)\n}", "func Sum32(key string) uint32 {\n\treturn Sum32Seed(key, 0)\n}", "func HashASM(k0, k1 uint64, p []byte) uint64", "func hash(key uint64) uint64 {\r\n\tkey ^= key >> 33\r\n\tkey *= 0xff51afd7ed558ccd\r\n\tkey ^= key >> 33\r\n\tkey *= 0xc4ceb9fe1a85ec53\r\n\tkey ^= key >> 33\r\n\treturn key\r\n}", "func strhash0(p unsafe.Pointer, h uintptr) uintptr", "func Hash(b []byte) uint32 {\n\tconst (\n\t\tseed = 0xbc9f1d34\n\t\tm = 0xc6a4a793\n\t)\n\th := uint32(seed) ^ uint32(len(b))*m\n\tfor ; len(b) >= 4; b = b[4:] {\n\t\th += uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24\n\t\th *= m\n\t\th ^= h >> 16\n\t}\n\tswitch len(b) {\n\tcase 3:\n\t\th += uint32(b[2]) << 16\n\t\tfallthrough\n\tcase 2:\n\t\th += uint32(b[1]) << 8\n\t\tfallthrough\n\tcase 1:\n\t\th += uint32(b[0])\n\t\th *= m\n\t\th ^= h >> 24\n\t}\n\treturn h\n}", "func HashFunction(buf []byte) uint32 {\n\tvar hash uint32 = 5381\n\tfor _, b := range buf {\n\t\thash = ((hash << 5) + hash) + uint32(b)\n\t}\n\treturn hash\n}", "func Bytes32ToIpfsHash(value [32]byte) (string, error) {\n\tbyteArray := [34]byte{18, 32}\n\tcopy(byteArray[2:], value[:])\n\tif len(byteArray) != 34 {\n\t\treturn \"\", errors.New(\"invalid bytes32 value\")\n\t}\n\n\thash := base58.Encode(byteArray[:])\n\treturn hash, nil\n}", "func byteshash(p *[]byte, h uintptr) uintptr", "func sumHash(c byte, h uint32) uint32 {\n\treturn (h * hashPrime) ^ uint32(c)\n}", "func (h Hash32) Hex() string { return util.Encode(h[:]) }", "func (this *Ring) Hash(key string) uint32 {\n\treturn crc32.ChecksumIEEE([]byte(key))\n}", "func hash(data []byte) [32]byte {\n\tvar hash [32]byte\n\n\th := sha256.New()\n\t// The hash interface never returns an error, for that reason\n\t// we are not handling the error below. For reference, it is\n\t// stated here https://golang.org/pkg/hash/#Hash\n\t// #nosec G104\n\th.Write(data)\n\th.Sum(hash[:0])\n\n\treturn hash\n}", "func FNV32(s string) uint32 {\n\treturn uint32Hasher(fnv.New32(), s)\n}", "func (h Hash32) Field() log.Field { return log.String(\"hash\", hex.EncodeToString(h[:])) }", "func hashInt(s string) uint32 {\n\tb := []byte(s)\n\th := crc32.ChecksumIEEE(b)\n\treturn h\n}", "func memhash(p unsafe.Pointer, h, s uintptr) uintptr", "func memhash(p unsafe.Pointer, h, s uintptr) uintptr", "func (_L1Block *L1BlockCaller) Hash(opts *bind.CallOpts) ([32]byte, error) {\n\tvar out []interface{}\n\terr := _L1Block.contract.Call(opts, &out, \"hash\")\n\n\tif err != nil {\n\t\treturn *new([32]byte), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new([32]byte)).(*[32]byte)\n\n\treturn out0, err\n\n}", "func TestHash32(t *testing.T) {\n\tt.Parallel()\n\n\tconst n = 400\n\n\tf := NewOptimized(Config{\n\t\tCapacity: n,\n\t\tFPRate: .01,\n\t})\n\n\tr := rand.New(rand.NewSource(32))\n\n\tfor i := 0; i < n; i++ {\n\t\tf.Add(uint64(r.Uint32()))\n\t}\n\n\tconst nrounds = 8\n\tfp := 0\n\tfor i := n; i < nrounds*n; i++ {\n\t\tif f.Has(uint64(r.Uint32())) {\n\t\t\tfp++\n\t\t}\n\t}\n\n\tfprate := float64(fp) / (nrounds * n)\n\tt.Logf(\"FP rate = %.2f%%\", 100*fprate)\n\tassert.LessOrEqual(t, fprate, .1)\n}", "func Hash32WithSeed(s []byte, seed uint32) uint32 {\n\tslen := len(s)\n\n\tif slen <= 24 {\n\t\tif slen >= 13 {\n\t\t\treturn hash32Len13to24Seed(s, seed*c1)\n\t\t}\n\t\tif slen >= 5 {\n\t\t\treturn hash32Len5to12(s, seed)\n\t\t}\n\t\treturn hash32Len0to4(s, seed)\n\t}\n\th := hash32Len13to24Seed(s[:24], seed^uint32(slen))\n\treturn mur(Hash32(s[24:])+seed, h)\n}", "func Hash(key []byte) uint64 {\n\treturn murmur3.Sum64(key)\n}", "func hash(s string) string {\n\thash := fnv.New32a()\n\thash.Write([]byte(s))\n\tintHash := hash.Sum32()\n\tresult := fmt.Sprintf(\"%08x\", intHash)\n\treturn result\n}", "func hash4(u uint32, h uint8) uint32 {\n\treturn (u * prime4bytes) >> ((32 - h) & 31)\n}", "func hash4x64(u uint64, h uint8) uint32 {\n\treturn (uint32(u) * prime4bytes) >> ((32 - h) & 31)\n}", "func sha3hash(t *testing.T, data ...[]byte) []byte {\n\tt.Helper()\n\th := sha3.NewLegacyKeccak256()\n\tr, err := doSum(h, nil, data...)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\treturn r\n}", "func Checksum32(data []byte) uint32 {\n\treturn Checksum32Seed(data, 0)\n}", "func (s *ShardMap) hash(v interface{}) int {\n\tswitch s.Type {\n\tcase \"string\":\n\t\tval, ok := v.(string)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\thash := fnv.New32()\n\t\thash.Write([]byte(val))\n\t\treturn int(hash.Sum32() % NumShards)\n\tcase \"int32\":\n\t\t// Values that come as numbers in JSON are of type float64.\n\t\tval, ok := v.(float64)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\treturn int(int32(val) % NumShards)\n\tdefault:\n\t\treturn -1\n\t}\n}", "func hash(s string) string {\n\th := fnv.New32a()\n\t_, err := h.Write([]byte(s))\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn fmt.Sprint(h.Sum32())\n}", "func Sha3256(bs []byte) ([]byte, error) {\n\treturn PerformHash(sha3.New256(), bs)\n}", "func addrHash(addr uint16) byte {\n\treturn (byte(addr) ^ byte(addr>>8)) & 0x7f\n}", "func (p Path) Hash() (uint32, error) {\n\treturn adler32.Checksum([]byte(p)), nil\n}", "func (r *RunCtx) Hash() (uint32, error) {\n\treturn 0, fmt.Errorf(\"not hashable\")\n}", "func hash(k Key) int {\n\tkey := fmt.Sprintf(\"%s\", k)\n\th := 0\n\tfor i := 0; i < len(key); i++ {\n\t\th = 31 * h + int(key[i])\n\t}\n\treturn h\n}", "func (t *Target) Hash() (uint32, error) {\n\th := adler32.New()\n\tt.Hash32(h)\n\treturn h.Sum32(), nil\n}", "func Sha256Hash(data []byte) [32]byte {\n\tsum := sha256.Sum256(data)\n\treturn sum\n}", "func strhash(p *string, h uintptr) uintptr", "func eb32(bits uint32, hi uint8, lo uint8) uint32 {\n\tm := uint32(((1 << (hi - lo)) - 1) << lo)\n\treturn (bits & m) >> lo\n}", "func Sum32Seed(key string, seed uint32) uint32 {\n\tvar nblocks = len(key) / 4\n\tvar nbytes = nblocks * 4\n\tvar h1 = seed\n\tconst c1 = 0xcc9e2d51\n\tconst c2 = 0x1b873593\n\tfor i := 0; i < nbytes; i += 4 {\n\t\tk1 := uint32(key[i+0]) | uint32(key[i+1])<<8 |\n\t\t\tuint32(key[i+2])<<16 | uint32(key[i+3])<<24\n\t\tk1 *= c1\n\t\tk1 = (k1 << 15) | (k1 >> 17)\n\t\tk1 *= c2\n\t\th1 ^= k1\n\t\th1 = (h1 << 13) | (h1 >> 19)\n\t\th1 = h1*5 + 0xe6546b64\n\t}\n\tvar k1 uint32\n\tswitch len(key) & 3 {\n\tcase 3:\n\t\tk1 ^= uint32(key[nbytes+2]) << 16\n\t\tfallthrough\n\tcase 2:\n\t\tk1 ^= uint32(key[nbytes+1]) << 8\n\t\tfallthrough\n\tcase 1:\n\t\tk1 ^= uint32(key[nbytes+0])\n\t\tk1 *= c1\n\t\tk1 = (k1 << 15) | (k1 >> 17)\n\t\tk1 *= c2\n\t\th1 ^= k1\n\t}\n\th1 ^= uint32(len(key))\n\th1 ^= h1 >> 16\n\th1 *= 0x85ebca6b\n\th1 ^= h1 >> 13\n\th1 *= 0xc2b2ae35\n\th1 ^= h1 >> 16\n\treturn h1\n}", "func Hash(data []byte) (string, int64) {\n\thasher := adler32.New()\n\tb, e := hasher.Write(data)\n\tif e != nil {\n\t\tlogs.WithFields(logs.Fields{\n\t\t\t\"Error\": e,\n\t\t}).Error(\"Unable to write chunk of data via hasher.Write\", e)\n\t}\n\treturn hex.EncodeToString(hasher.Sum(nil)), int64(b)\n}", "func HashXXH3_64(input []byte, seed uint64) (result uint64) {\n\treturn parser.HashXXH3_64(input, seed)\n}", "func CalcBlocksHash32(view []BlockID, additionalBytes []byte) Hash32 {\n\tsortedView := make([]BlockID, len(view))\n\tcopy(sortedView, view)\n\tSortBlockIDs(sortedView)\n\treturn CalcBlockHash32Presorted(sortedView, additionalBytes)\n}", "func NewHashFromArray(bytes [32]byte) Hash32 {\n\treturn NewHashFromBytes(bytes[:])\n}", "func (t *openAddressing) hash(key string, round int) uint32 {\n\tnum := uint(stringToInt(key))\n\tmax := uint(len(t.values) - 1)\n\treturn uint32((hashDivision(num, max) + uint(round)*hashDivision2(num, max)) % max)\n}", "func hash(addr mino.Address) *big.Int {\n\tsha := sha256.New()\n\tmarshalled, err := addr.MarshalText()\n\tif err != nil {\n\t\tmarshalled = []byte(addr.String())\n\t}\n\t// A hack to accommodate for minogrpc's design:\n\t// 1) the first byte is used to indicate if a node is orchestrator or not\n\t// 2) the only way to reach the orchestrator is to route a message to nil\n\t// from its server side, which has the same address but orchestrator byte\n\t// set to f.\n\t// We therefore have to ignore if a node is the orchestrator to be able to\n\t// route the message first to its server side, then from the server side to\n\t// the client side.\n\tsha.Write(marshalled[1:])\n\treturn byteArrayToBigInt(sha.Sum(nil))\n}", "func Hash(value int64) uint64 {\n\treturn FNVHash64(uint64(value))\n}", "func BytesToHash(b []byte) Hash32 {\n\tvar h Hash32\n\th.SetBytes(b)\n\treturn h\n}", "func hashFunction(key int, size int) int {\n\treturn key % size\n}", "func FNV32a(s string) uint32 {\n\treturn uint32Hasher(fnv.New32a(), s)\n}", "func Hash3Words(a, b, c, initval uint32) uint32 {\n\tconst iv = 0xdeadbeef + (3 << 2)\n\tinitval += iv\n\n\ta += initval\n\tb += initval\n\tc += initval\n\n\tc ^= b\n\tc -= rol32(b, 14)\n\ta ^= c\n\ta -= rol32(c, 11)\n\tb ^= a\n\tb -= rol32(a, 25)\n\tc ^= b\n\tc -= rol32(b, 16)\n\ta ^= c\n\ta -= rol32(c, 4)\n\tb ^= a\n\tb -= rol32(a, 14)\n\tc ^= b\n\tc -= rol32(b, 24)\n\n\treturn c\n}", "func (payload *ExtEthCompatPayload) Hash() B32 {\n\thash := crypto.Keccak256(payload.Value)\n\tvar phash B32\n\tcopy(phash[:], hash)\n\treturn phash\n}", "func SHA256(raw []byte) Hash {\n\treturn gosha256.Sum256(raw)\n}", "func HashKey(key int) int {\n\t/*\n\t\ttiedot should be compiled/run on x86-64 systems.\n\t\tIf you decide to compile tiedot on 32-bit systems, the following integer-smear algorithm will cause compilation failure\n\t\tdue to 32-bit interger overflow; therefore you must modify the algorithm.\n\t\tDo not remove the integer-smear process, and remember to run test cases to verify your mods.\n\t*/\n\t// ========== Integer-smear start =======\n\tkey = key ^ (key >> 4)\n\tkey = (key ^ 0xdeadbeef) + (key << 5)\n\tkey = key ^ (key >> 11)\n\t// ========== Integer-smear end =========\n\treturn key & ((1 << HASH_BITS) - 1) // Do not modify this line\n}", "func nilinterhash(a unsafe.Pointer, h uintptr) uintptr", "func fnv32a(s string) uint32 {\n\tconst (\n\t\tinitial = 2166136261\n\t\tprime = 16777619\n\t)\n\n\thash := uint32(initial)\n\tfor i := 0; i < len(s); i++ {\n\t\thash ^= uint32(s[i])\n\t\thash *= prime\n\t}\n\treturn hash\n}", "func CalculateHash(args []string) string {\n\tvar str = \"\"\n\tfor _,v := range args {\n\t\tstr += v\n\t}\n\thasher := sha256.New()\n\thasher.Write([]byte(str))\n\treturn hex.EncodeToString(hasher.Sum(nil))\n}", "func CalcBlockHash32Presorted(sortedView []BlockID, additionalBytes []byte) Hash32 {\n\thash := hash.New()\n\thash.Write(additionalBytes)\n\tfor _, id := range sortedView {\n\t\thash.Write(id.Bytes()) // this never returns an error: https://golang.org/pkg/hash/#Hash\n\t}\n\tvar res Hash32\n\thash.Sum(res[:0])\n\treturn res\n}", "func runtime_memhash(p unsafe.Pointer, seed, s uintptr) uintptr", "func hashCode(key string) int64 {\n\tv := int64(crc32.ChecksumIEEE([]byte(key)))\n\tif v >= 0 {\n\t\treturn v\n\t}\n\tif -v > 0 {\n\t\treturn -v\n\t}\n\t// v == MinInt\n\treturn 0\n}", "func (o *ObjectIndex) Hash() uint32 {\n\tvar h uint32 = 17\n\n\tvar str string\n\tstr += fmt.Sprintf(\"%08x\", o.machine)\n\tstr += fmt.Sprintf(\"%04x\", o.pid)\n\tstr += fmt.Sprintf(\"%08x\", o.id)\n\tstr += fmt.Sprintf(\"%08x\", o.Rand)\n\tfor _, v := range str {\n\t\th += h*23 + uint32(v)\n\t}\n\treturn h\n}", "func crc32Demo() {\n\t// hasher\n\th := crc32.NewIEEE()\n\tfmt.Println(reflect.TypeOf(h))\n\n\t// write a string converted to bytes\n\th.Write([]byte(\"test\"))\n\n\t// checksum\n\tv := h.Sum32()\n\tfmt.Println(reflect.TypeOf(v)) // uint32\n\tfmt.Println(v)\n}", "func (h Hash32) Bytes() []byte { return h[:] }", "func hashmapHash(data []byte) uint32 {\n\tvar result uint32 = 2166136261 // FNV offset basis\n\tfor _, c := range data {\n\t\tresult ^= uint32(c)\n\t\tresult *= 16777619 // FNV prime\n\t}\n\treturn result\n}", "func Hash(data []byte) [blake2b.Size]byte {\n\treturn blake2b.Sum512(data)\n}", "func fletcher32(payload []byte) uint32 {\n\ts1 := uint16(0)\n\ts2 := uint16(0)\n\n\tsz := len(payload) & (^1)\n\tfor i := 0; i < sz; i += 2 {\n\t\ts1 += uint16(payload[i]) | (uint16(payload[i+1]) << 8)\n\t\ts2 += s1\n\t}\n\tif len(payload)&1 != 0 {\n\t\ts1 += uint16(payload[sz])\n\t\ts2 += s1\n\t}\n\treturn (uint32(s2) << 16) | uint32(s1)\n}", "func (i *Instance) Hash() (uint32, error) {\n\th := fnv.New32()\n\tfmt.Fprintf(h, \"%s%s\", i.Name, i.Version)\n\n\tfn, err := i.Fn.Hash()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn fn ^ h.Sum32(), nil\n}", "func hash_func(x, y, n HashValue) (HashValue) {\n return (x*1640531513 ^ y*2654435789) % n\n}", "func HashBuildArgs(args interface{}) (string, error) {\n\tdata, err := json.Marshal(args)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\thash := sha256.Sum256(data)\n\treturn hex.EncodeToString(hash[:]), nil\n}", "func (gg GlobGroup) Hash() (uint32, error) {\n\th := adler32.New()\n\tgg.Hash32(h)\n\treturn h.Sum32(), nil\n}", "func (hasher *SHA256) HashLength() uint {\n\treturn 64\n}", "func (h *MemHash) Hash() uint32 {\n\tss := (*stringStruct)(unsafe.Pointer(&h.buf))\n\treturn uint32(memhash(ss.str, 0, uintptr(ss.len)))\n}" ]
[ "0.7513112", "0.7491727", "0.72715765", "0.7190082", "0.6986059", "0.68957806", "0.67351377", "0.67205036", "0.67193633", "0.6649574", "0.6621809", "0.65809476", "0.6579308", "0.65534127", "0.6548068", "0.65251803", "0.64664733", "0.6466416", "0.64491963", "0.6394227", "0.63191426", "0.63190085", "0.62662023", "0.62407404", "0.62161326", "0.62012714", "0.61236066", "0.61206007", "0.6113711", "0.60865384", "0.6079078", "0.6070839", "0.60679626", "0.6028909", "0.60234565", "0.60172623", "0.6001384", "0.60012794", "0.5998455", "0.5994997", "0.5975615", "0.594546", "0.5931275", "0.5926682", "0.5925481", "0.5897024", "0.5897024", "0.58704674", "0.586417", "0.58550256", "0.5853209", "0.5851851", "0.585127", "0.5840641", "0.58194196", "0.5815427", "0.5813978", "0.5808147", "0.58071226", "0.57908183", "0.57635707", "0.57561874", "0.57481956", "0.5744478", "0.57343984", "0.57275856", "0.57239074", "0.57196003", "0.57193524", "0.57163304", "0.5714951", "0.56846184", "0.5666647", "0.5652715", "0.56283873", "0.56237423", "0.5620443", "0.56199783", "0.5608764", "0.56069165", "0.56030476", "0.55983585", "0.55925786", "0.5587579", "0.55864966", "0.558254", "0.5576032", "0.55597", "0.5559594", "0.5557764", "0.55558187", "0.5539344", "0.5538027", "0.55377007", "0.552354", "0.5500545", "0.5492596", "0.54867786", "0.548338", "0.5482333" ]
0.6501993
16
Hash implements the starlark.Value.Hash() method.
func (gg GlobGroup) Hash() (uint32, error) { h := adler32.New() gg.Hash32(h) return h.Sum32(), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Hash(value int64) uint64 {\n\treturn FNVHash64(uint64(value))\n}", "func (n *SoupNode) Hash() (uint32, error) {\n\treturn hashString(fmt.Sprintf(\"%v\", *n)), nil\n}", "func (o *ObjectIndex) Hash() uint32 {\n\tvar h uint32 = 17\n\n\tvar str string\n\tstr += fmt.Sprintf(\"%08x\", o.machine)\n\tstr += fmt.Sprintf(\"%04x\", o.pid)\n\tstr += fmt.Sprintf(\"%08x\", o.id)\n\tstr += fmt.Sprintf(\"%08x\", o.Rand)\n\tfor _, v := range str {\n\t\th += h*23 + uint32(v)\n\t}\n\treturn h\n}", "func hash(key, value string) int64 {\n\thash := siphash.New(sipConst)\n\thash.Write([]byte(key + \":::\" + value))\n\treturn int64(hash.Sum64())\n}", "func (term *Term) Hash() int {\n\treturn term.Value.Hash()\n}", "func (s *ShardMap) hash(v interface{}) int {\n\tswitch s.Type {\n\tcase \"string\":\n\t\tval, ok := v.(string)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\thash := fnv.New32()\n\t\thash.Write([]byte(val))\n\t\treturn int(hash.Sum32() % NumShards)\n\tcase \"int32\":\n\t\t// Values that come as numbers in JSON are of type float64.\n\t\tval, ok := v.(float64)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\treturn int(int32(val) % NumShards)\n\tdefault:\n\t\treturn -1\n\t}\n}", "func (this *Ring) Hash(key string) uint32 {\n\treturn crc32.ChecksumIEEE([]byte(key))\n}", "func Hash(length int, key string) int64 {\n\tif key == \"\" {\n\t\treturn 0\n\t}\n\thc := hashCode(key)\n\treturn (hc ^ (hc >> 16)) % int64(length)\n}", "func HashOf(v Value) []byte {\n\treturn quad.HashOf(v)\n}", "func hash(value string) uint32 {\n\th := fnv.New32a()\n\th.Write([]byte(value))\n\n\treturn h.Sum32()\n}", "func (ref Ref) Hash() int {\n\treturn termSliceHash(ref)\n}", "func (o *Object) Hash() string {\n\treturn Hash(o, true, false, true)\n}", "func (r Ref) Hash() int {\n\treturn termSliceHash(r)\n}", "func hash(obj interface{}) KHash {\n\tvar buffer bytes.Buffer\n\tencoder := json.NewEncoder(&buffer)\n\terr := encoder.Encode(obj)\n\tif err != nil {\n\t\tpanic(\"cannot encode object\")\n\t}\n\n\tdata := buffer.Bytes()\n\th := sha256.Sum256(data)\n\n\t// log.Printf(\"hashing %#v represented as %s with hash %X\", obj, data, h)\n\treturn h\n}", "func (i *Instance) Hash() (uint32, error) {\n\th := fnv.New32()\n\tfmt.Fprintf(h, \"%s%s\", i.Name, i.Version)\n\n\tfn, err := i.Fn.Hash()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn fn ^ h.Sum32(), nil\n}", "func (expr *Expr) Hash() int {\n\ts := expr.Index\n\tswitch ts := expr.Terms.(type) {\n\tcase []*Term:\n\t\tfor _, t := range ts {\n\t\t\ts += t.Value.Hash()\n\t\t}\n\tcase *Term:\n\t\ts += ts.Value.Hash()\n\t}\n\tif expr.Negated {\n\t\ts++\n\t}\n\treturn s\n}", "func (i *Index) Hash() (uint32, error) {\n\treturn 0, fmt.Errorf(\"unhashable: %s\", i.Type())\n}", "func (p *FiveTuple) Hash() uint64 {\n\treturn siphash.Hash(lookupKey, 0, p.data)\n}", "func (k Ktype) Hash() uint32 {\n\tif k == UnknownKtype {\n\t\treturn 0\n\t}\n\treturn hashers.FnvUint32([]byte(k.String()))\n}", "func (sc *SetComprehension) Hash() int {\n\treturn sc.Term.Hash() + sc.Body.Hash()\n}", "func Hash(data interface{}) string {\n\treturn hex.EncodeToString(RawHash(data))\n}", "func (obj *identifier) Hash() hash.Hash {\n\tif obj.IsElement() {\n\t\treturn obj.Element().Hash()\n\t}\n\n\treturn obj.Comparer().Hash()\n}", "func Hash(key []byte) uint64 {\n\treturn murmur3.Sum64(key)\n}", "func (o *Object) Hash(ht hash.Type) (string, error) {\n\terr := o.loadMetadataIfNotLoaded()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif ht&hash.MD5 == 0 {\n\t\treturn \"\", hash.ErrUnsupported\n\t}\n\treturn hex.EncodeToString(o.meta.Hash), nil\n}", "func (null Null) Hash() int {\n\treturn 0\n}", "func Hasher(value string) string {\n\th := fnv.New32a()\n\t_, _ = h.Write([]byte(value))\n\treturn fmt.Sprintf(\"%v\", h.Sum32())\n}", "func (p PropertyHashList) Hash() string {\n\tglobalSum := sha256.New()\n\tfor _, hash := range p {\n\t\t_, _ = globalSum.Write(hash.Hash)\n\t}\n\n\tsum := globalSum.Sum(nil)\n\n\treturn hex.EncodeToString(sum)\n}", "func (t *openAddressing) hash(key string, round int) uint32 {\n\tnum := uint(stringToInt(key))\n\tmax := uint(len(t.values) - 1)\n\treturn uint32((hashDivision(num, max) + uint(round)*hashDivision2(num, max)) % max)\n}", "func hash(t types.Type, x value) int {\n\tswitch x := x.(type) {\n\tcase bool:\n\t\tif x {\n\t\t\treturn 1\n\t\t}\n\t\treturn 0\n\tcase int:\n\t\treturn x\n\tcase int8:\n\t\treturn int(x)\n\tcase int16:\n\t\treturn int(x)\n\tcase int32:\n\t\treturn int(x)\n\tcase int64:\n\t\treturn int(x)\n\tcase uint:\n\t\treturn int(x)\n\tcase uint8:\n\t\treturn int(x)\n\tcase uint16:\n\t\treturn int(x)\n\tcase uint32:\n\t\treturn int(x)\n\tcase uint64:\n\t\treturn int(x)\n\tcase uintptr:\n\t\treturn int(x)\n\tcase float32:\n\t\treturn int(x)\n\tcase float64:\n\t\treturn int(x)\n\tcase complex64:\n\t\treturn int(real(x))\n\tcase complex128:\n\t\treturn int(real(x))\n\tcase string:\n\t\treturn hashString(x)\n\tcase *value:\n\t\treturn int(uintptr(unsafe.Pointer(x)))\n\tcase chan value:\n\t\treturn int(uintptr(reflect.ValueOf(x).Pointer()))\n\tcase structure:\n\t\treturn x.hash(t)\n\tcase array:\n\t\treturn x.hash(t)\n\tcase iface:\n\t\treturn x.hash(t)\n\tcase rtype:\n\t\treturn x.hash(t)\n\t}\n\tpanic(fmt.Sprintf(\"%T is unhashable\", x))\n}", "func (o *Object) Hash(ctx context.Context, r hash.Type) (string, error) {\n\treturn \"\", hash.ErrUnsupported\n}", "func Hash(s int, o Orientation) (int, error) {\n\n\tvar errVal int = 10\n\n\tif !(s >= 0 && s <= palletWidth*palletLength) {\n\t\treturn errVal, ErrSize\n\t}\n\tif o != HORIZONTAL && o != VERTICAL && o != SQUAREGRID {\n\t\treturn errVal, ErrOrient\n\t}\n\n\tvar hash int\n\n\tswitch s {\n\tcase 1, 2, 3, 6:\n\t\thash = s - 1\n\tcase 4:\n\t\tif o == SQUAREGRID {\n\t\t\thash = s\n\t\t} else {\n\t\t\thash = s - 1\n\t\t}\n\tcase 8:\n\t\thash = 6\n\tcase 9:\n\t\thash = 7\n\tcase 12:\n\t\thash = 8\n\tcase 16:\n\t\thash = 9\n\tdefault:\n\t\treturn errVal, ErrSize\n\t}\n\n\treturn hash, nil\n}", "func (oc *ObjectComprehension) Hash() int {\n\treturn oc.Key.Hash() + oc.Value.Hash() + oc.Body.Hash()\n}", "func (c *Cluster) Hash(v interface{}) (int, error) {\n\th, err := hashstructure.Hash(v, nil)\n\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\n\t// get cluster index\n\tci := int(h % uint64(c.metadata.NumShards))\n\n\treturn ci, nil\n}", "func (a Address) Hash() Hash { return BytesToHash(a[:]) }", "func (v Var) Hash() int {\n\th := xxhash.ChecksumString64S(string(v), hashSeed0)\n\treturn int(h)\n}", "func Hash(k0, k1 uint64, p []byte) uint64 {\n\tvar d digest\n\td.size = Size\n\td.k0 = k0\n\td.k1 = k1\n\td.Reset()\n\td.Write(p)\n\treturn d.Sum64()\n}", "func (m *Map) Hash() string {\n\treturn Hash(m, true, false, true)\n}", "func (b BlockChain) Hash() {\n\n}", "func (a *Array) Hash() string {\n\treturn Hash(a, true, false, true)\n}", "func Hash(strings ...string) uint32 {\n\tdigester := fnv.New32()\n\tfor _, s := range strings {\n\t\t_, _ = io.WriteString(digester, s)\n\t}\n\treturn digester.Sum32()\n}", "func (o *Object) Hash(ctx context.Context, t hash.Type) (string, error) {\n\treturn \"\", hash.ErrUnsupported\n}", "func (c Category) Hash() uint32 {\n\treturn hashers.FnvUint32([]byte(c))\n}", "func Hash(i interface{}) string {\n\tv := reflect.ValueOf(i)\n\tif v.Kind() != reflect.Ptr {\n\t\tif !v.CanAddr(){\n\t\t\treturn \"\"\n\t\t}\n\t\tv = v.Addr()\n\t}\n\n\tsize := unsafe.Sizeof(v.Interface())\n\tb := (*[1 << 10]uint8)(unsafe.Pointer(v.Pointer()))[:size:size]\n\n\th := md5.New()\n\treturn base64.StdEncoding.EncodeToString(h.Sum(b))\n}", "func (s *set) Hash() int {\n\treturn s.hash\n}", "func (obj *bucket) Hash() hash.Hash {\n\treturn obj.immutable.Hash()\n}", "func hash(key uint64) uint64 {\r\n\tkey ^= key >> 33\r\n\tkey *= 0xff51afd7ed558ccd\r\n\tkey ^= key >> 33\r\n\tkey *= 0xc4ceb9fe1a85ec53\r\n\tkey ^= key >> 33\r\n\treturn key\r\n}", "func (s Sample) Hash() []byte {\n\tallVecs := make([]linalg.Vector, len(s.Inputs)+len(s.Outputs))\n\tcopy(allVecs, s.Inputs)\n\tcopy(allVecs[len(s.Inputs):], s.Outputs)\n\treturn sgd.HashVectors(allVecs...)\n}", "func Hash(a interface{}) (string, error) {\n\th := sha1.New()\n\tif err := json.NewEncoder(h).Encode(a); err != nil {\n\t\treturn \"\", nil\n\t}\n\treturn base64.URLEncoding.EncodeToString(h.Sum(nil)), nil\n}", "func hash(key string) int{\n\tvar num = 0\n\t// get the lenght of the key\n\tvar length = len(key)\n\n\t// add the ascii character value to creat a sum \n\tfor i := 0; i < length; i++{\n\n\t\tnum += int(key[i])\n\t}\n\t\n\t// square in the middle hash method\n\tvar avg = num * int((math.Pow(5.0, 0.5) - 1)) / 2\n\tvar numeric = avg - int(math.Floor(float64(avg)))\n\n\n\t// hash value to place into the table slice between -1 and CAPACITY - 1\n\treturn int(math.Floor(float64(numeric * CAPACITY)))\n}", "func Hash(content string, size int) int {\n\tsequence := sliceutil.Atoi(content, \",\")\n\tcircleKnots := GetStringCircle(size)\n\tfor _, n := range sequence {\n\t\tcircleKnots.ComputeKnot(n)\n\t}\n\treturn circleKnots.GetHash()\n}", "func (bc *Blockchain) Hash() {\n\n}", "func (t Tuple1) Hash() uint32 {\n\tif t.E1 == nil {\n\t\treturn 0\n\t}\n\treturn t.E1.Hash()\n}", "func hash(s string) uint32 {\n h := fnv.New32a()\n h.Write([]byte(s))\n return h.Sum32()\n}", "func (obj *immutable) Hash() hash.Hash {\n\treturn obj.hash\n}", "func (o *Object) Hash(t fs.HashType) (string, error) {\n\treturn \"\", fs.ErrHashUnsupported\n}", "func Hash(key string) uint32 {\n\treturn uint32(aeshashstr(noescape(unsafe.Pointer(&key)), 0))\n}", "func (o *ObjectInfo) Hash(ht hash.Type) (string, error) {\n\tif o.meta == nil {\n\t\tmo, err := o.f.NewObject(generateMetadataName(o.Remote()))\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\to.meta = readMetadata(mo)\n\t}\n\tif ht&hash.MD5 == 0 {\n\t\treturn \"\", hash.ErrUnsupported\n\t}\n\treturn hex.EncodeToString(o.meta.Hash), nil\n}", "func (h *Header) Hash() [32]byte {\n\tvar f []string\n\tif h.Description.Value != \"\" {\n\t\tf = append(f, h.Description.Value)\n\t}\n\tf = append(f, fmt.Sprint(h.Required.Value))\n\tf = append(f, fmt.Sprint(h.Deprecated.Value))\n\tf = append(f, fmt.Sprint(h.AllowEmptyValue.Value))\n\tif h.Style.Value != \"\" {\n\t\tf = append(f, h.Style.Value)\n\t}\n\tf = append(f, fmt.Sprint(h.Explode.Value))\n\tf = append(f, fmt.Sprint(h.AllowReserved.Value))\n\tif h.Schema.Value != nil {\n\t\tf = append(f, low.GenerateHashString(h.Schema.Value))\n\t}\n\tif h.Example.Value != nil {\n\t\tf = append(f, fmt.Sprint(h.Example.Value))\n\t}\n\tif len(h.Examples.Value) > 0 {\n\t\tfor k := range h.Examples.Value {\n\t\t\tf = append(f, fmt.Sprintf(\"%s-%x\", k.Value, h.Examples.Value[k].Value.Hash()))\n\t\t}\n\t}\n\tif len(h.Content.Value) > 0 {\n\t\tfor k := range h.Content.Value {\n\t\t\tf = append(f, fmt.Sprintf(\"%s-%x\", k.Value, h.Content.Value[k].Value.Hash()))\n\t\t}\n\t}\n\tkeys := make([]string, len(h.Extensions))\n\tz := 0\n\tfor k := range h.Extensions {\n\t\tkeys[z] = fmt.Sprintf(\"%s-%x\", k.Value, sha256.Sum256([]byte(fmt.Sprint(h.Extensions[k].Value))))\n\t\tz++\n\t}\n\tsort.Strings(keys)\n\tf = append(f, keys...)\n\treturn sha256.Sum256([]byte(strings.Join(f, \"|\")))\n}", "func (l *LexerATNConfig) Hash() int {\n\tvar f int\n\tif l.passedThroughNonGreedyDecision {\n\t\tf = 1\n\t} else {\n\t\tf = 0\n\t}\n\th := murmurInit(7)\n\th = murmurUpdate(h, l.state.GetStateNumber())\n\th = murmurUpdate(h, l.alt)\n\th = murmurUpdate(h, l.context.Hash())\n\th = murmurUpdate(h, l.semanticContext.Hash())\n\th = murmurUpdate(h, f)\n\th = murmurUpdate(h, l.lexerActionExecutor.Hash())\n\th = murmurFinish(h, 6)\n\treturn h\n}", "func (bol Boolean) Hash() int {\n\tif bol {\n\t\treturn 1\n\t}\n\treturn 0\n}", "func (set *lalrSet) hash() (val uint32) {\n\t// Need hash to be order independent, so\n\t// just XOR everything.\n\tfor _, list := range(set.items) {\n\t\tfor _, item := range(list) {\n\t\t\tval = val ^ item.hash()\n\t\t}\n\t}\n\n\treturn\n}", "func (p Path) Hash() (uint32, error) {\n\treturn adler32.Checksum([]byte(p)), nil\n}", "func Hash(t *Token) (hash []byte) {\n var sum []byte\n\n // Compute the SHA1 sum of the Token\n {\n shasum := sha1.Sum([]byte(salt+string(*t)))\n copy(sum[:], shasum[:20])\n }\n\n // Encode the sum to hexadecimal\n hex.Encode(sum, sum)\n\n return\n}", "func (m MapEntry) Hash() uint32 {\n\treturn sequtil.Hash(m.key)\n}", "func hash(ls prometheus.Tags) uint64 {\n\tlbs := make(labels.Labels, 0, len(ls))\n\tfor k, v := range ls {\n\t\tlbs = append(lbs, labels.Label{\n\t\t\tName: k,\n\t\t\tValue: v,\n\t\t})\n\t}\n\n\tsort.Slice(lbs[:], func(i, j int) bool {\n\t\treturn lbs[i].Name < lbs[j].Name\n\t})\n\n\treturn lbs.Hash()\n}", "func (n *node) Hash() []byte {\n\treturn n.hash\n}", "func (obj *object) Hash() int {\n\treturn obj.hash\n}", "func (c *ColumnValue) Hash() uint64 {\n\tif c == nil {\n\t\treturn cache.NewHash(FragmentType_ColumnValue, nil)\n\t}\n\treturn cache.NewHash(FragmentType_ColumnValue, c.Column, c.Operator, c.Value)\n}", "func (in *Instance) hash(x, y, mu *big.Int, T uint64) *big.Int {\n\tb := sha512.New()\n\tb.Write(x.Bytes())\n\tb.Write(y.Bytes())\n\tb.Write(mu.Bytes())\n\tbits := make([]byte, 8)\n\tbinary.LittleEndian.PutUint64(bits, T)\n\tb.Write(bits)\n\tres := new(big.Int).SetBytes(b.Sum(nil))\n\tres.Mod(res, in.rsaModulus)\n\treturn res\n}", "func (s *Sub) Hash() (uint32, error) {\n\th := adler32.New()\n\ts.Hash32(h)\n\treturn h.Sum32(), nil\n}", "func (s SampleList) Hash(i int) []byte {\n\tres := md5.Sum(s[i])\n\treturn res[:]\n}", "func (n Number) Hash() int {\n\tf, err := json.Number(n).Float64()\n\tif err != nil {\n\t\tbs := []byte(n)\n\t\th := xxhash.Checksum64(bs)\n\t\treturn int(h)\n\t}\n\treturn int(f)\n}", "func (source *Source) Hash() int {\n\tvar hash int\n\n\tif len(source.Prefix) > 0 {\n\t\tfor _, b := range source.Prefix {\n\t\t\thash = int(b*31) + hash\n\t\t}\n\t}\n\n\thash = int(source.PrefixLen*31) + hash\n\thash = int(source.RouterId*31) + hash\n\n\treturn hash\n}", "func (r *Restriction) hash() ([]byte, error) {\n\tj, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\treturn hashUtils.SHA512(j), nil\n}", "func (i *Instance) Hash(extraBytes []byte) (string, error) {\n\t//nolint:gosec // not being used for secure purposes\n\th := sha1.New()\n\n\t// copy by value to ignore ETag without affecting i\n\ti2 := *i\n\ti2.ETag = \"\"\n\n\tinstanceBytes, err := bson.Marshal(i2)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif _, err := h.Write(append(instanceBytes, extraBytes...)); err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn fmt.Sprintf(\"%x\", h.Sum(nil)), nil\n}", "func (h *kustHash) Hash(m ifc.Kunstructured) (string, error) {\n\tu := unstructured.Unstructured{\n\t\tObject: m.Map(),\n\t}\n\tkind := u.GetKind()\n\tswitch kind {\n\tcase \"ConfigMap\":\n\t\tcm, err := unstructuredToConfigmap(u)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn configMapHash(cm)\n\tcase \"Secret\":\n\t\tsec, err := unstructuredToSecret(u)\n\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn secretHash(sec)\n\tdefault:\n\t\treturn \"\", fmt.Errorf(\n\t\t\t\"type %s is not supported for hashing in %v\",\n\t\t\tkind, m.Map())\n\t}\n}", "func (p Primitive) Hash() string {\n\treturn p.Name()\n}", "func (d Data32) Hash() Hash {\n\treturn hash(d)\n}", "func (n *notifier) hash(other *memberlist.Node) uint64 {\n\treturn uint64(murmur.Murmur3([]byte(other.Name), murmur.M3Seed))\n}", "func encodeHash(x uint64, p, pPrime uint) (hashCode uint64) {\n\tif x&onesFromTo(64-pPrime, 63-p) == 0 {\n\t\tr := rho(extractShift(x, 0, 63-pPrime))\n\t\treturn concat([]concatInput{\n\t\t\t{x, 64 - pPrime, 63},\n\t\t\t{uint64(r), 0, 5},\n\t\t\t{1, 0, 0}, // this just adds a 1 bit at the end\n\t\t})\n\t} else {\n\t\treturn concat([]concatInput{\n\t\t\t{x, 64 - pPrime, 63},\n\t\t\t{0, 0, 0}, // this just adds a 0 bit at the end\n\t\t})\n\t}\n}", "func (obj *set) Hash() hash.Hash {\n\treturn obj.hash\n}", "func (h *MemHash) Hash() uint32 {\n\tss := (*stringStruct)(unsafe.Pointer(&h.buf))\n\treturn uint32(memhash(ss.str, 0, uintptr(ss.len)))\n}", "func (d Data256) Hash() Hash {\n\treturn hash(d)\n}", "func hash(values ...[]byte) ([]byte, error) {\n\th := swarm.NewHasher()\n\tfor _, v := range values {\n\t\t_, err := h.Write(v)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn h.Sum(nil), nil\n}", "func (c Call) Hash() int {\n\treturn termSliceHash(c)\n}", "func hash(stav Stav) uint64{\n\tstr := \"\"\n\n\tfor i := 0; i < len(stav.Auta); i++ {\n\t\tstr += stav.Auta[i].Farba\n\t\tstr += strconv.Itoa(int(stav.Auta[i].X))\n\t\tstr += strconv.Itoa(int(stav.Auta[i].Y))\n\t\tstr += strconv.FormatBool(stav.Auta[i].Smer)\n\t\tstr += strconv.Itoa(int(stav.Auta[i].Dlzka))\n\t}\n\n\th := fnv.New64a()\n\th.Write([]byte(str))\n\treturn h.Sum64()\n\n}", "func (t *Table) hash(s string) int {\n\t// Good enough.\n\th := fnv.New32()\n\th.Write([]byte(s))\n\treturn int(h.Sum32()) % t.m\n}", "func Hash(b []byte, seed uint64) uint64", "func (gdt *Array) Hash() Int {\n\targ0 := gdt.getBase()\n\n\tret := C.go_godot_array_hash(GDNative.api, arg0)\n\n\treturn Int(ret)\n}", "func (e EmptyNode) Hash() util.Uint256 {\n\tpanic(\"can't get hash of an EmptyNode\")\n}", "func hash(k Key) int {\n\tkey := fmt.Sprintf(\"%s\", k)\n\th := 0\n\tfor i := 0; i < len(key); i++ {\n\t\th = 31 * h + int(key[i])\n\t}\n\treturn h\n}", "func (o *ExportData) Hash() string {\n\targs := make([]interface{}, 0)\n\targs = append(args, o.CustomerID)\n\targs = append(args, o.ID)\n\targs = append(args, o.IntegrationInstanceID)\n\targs = append(args, o.JobID)\n\targs = append(args, o.Objects)\n\targs = append(args, o.RefID)\n\targs = append(args, o.RefType)\n\to.Hashcode = hash.Values(args...)\n\treturn o.Hashcode\n}", "func hash(m datasource.Metric) uint64 {\n\thash := fnv.New64a()\n\tlabels := m.Labels\n\tsort.Slice(labels, func(i, j int) bool {\n\t\treturn labels[i].Name < labels[j].Name\n\t})\n\tfor _, l := range labels {\n\t\t// drop __name__ to be consistent with Prometheus alerting\n\t\tif l.Name == \"__name__\" {\n\t\t\tcontinue\n\t\t}\n\t\thash.Write([]byte(l.Name))\n\t\thash.Write([]byte(l.Value))\n\t\thash.Write([]byte(\"\\xff\"))\n\t}\n\treturn hash.Sum64()\n}", "func (t *Target) hash() uint64 {\n\th := fnv.New64a()\n\n\t//nolint: errcheck\n\th.Write([]byte(fmt.Sprintf(\"%016d\", t.labels.Hash())))\n\t//nolint: errcheck\n\th.Write([]byte(t.URL().String()))\n\n\treturn h.Sum64()\n}", "func (t *smallFlatTable) Hash() hash.Hash { return t.hash }", "func (obj *chunk) Hash() hash.Hash {\n\treturn obj.immutable.Hash()\n}", "func (d Data) Hash() Hash {\n\treturn hash(d)\n}", "func Hash(mem []byte) uint64 {\n\tvar hash uint64 = 5381\n\tfor _, b := range mem {\n\t\thash = (hash << 5) + hash + uint64(b)\n\t}\n\treturn hash\n}", "func Hash(seed maphash.Seed, k Key) uint64 {\n\tvar buf [8]byte\n\tswitch v := k.(type) {\n\tcase mapKey:\n\t\treturn hashMapKey(seed, v)\n\tcase interfaceKey:\n\t\ts := v.Hash()\n\t\t// Mix up the hash to ensure it covers 64-bits\n\t\tbinary.LittleEndian.PutUint64(buf[:8], uint64(s))\n\t\treturn hashBytes(seed, buf[:8])\n\tcase strKey:\n\t\treturn hashString(seed, string(v))\n\tcase bytesKey:\n\t\treturn hashBytes(seed, []byte(v))\n\tcase int8Key:\n\t\tbuf[0] = byte(v)\n\t\treturn hashBytes(seed, buf[:1])\n\tcase int16Key:\n\t\tbinary.LittleEndian.PutUint16(buf[:2], uint16(v))\n\t\treturn hashBytes(seed, buf[:2])\n\tcase int32Key:\n\t\tbinary.LittleEndian.PutUint32(buf[:4], uint32(v))\n\t\treturn hashBytes(seed, buf[:4])\n\tcase int64Key:\n\t\tbinary.LittleEndian.PutUint64(buf[:8], uint64(v))\n\t\treturn hashBytes(seed, buf[:8])\n\tcase uint8Key:\n\t\tbuf[0] = byte(v)\n\t\treturn hashBytes(seed, buf[:1])\n\tcase uint16Key:\n\t\tbinary.LittleEndian.PutUint16(buf[:2], uint16(v))\n\t\treturn hashBytes(seed, buf[:2])\n\tcase uint32Key:\n\t\tbinary.LittleEndian.PutUint32(buf[:4], uint32(v))\n\t\treturn hashBytes(seed, buf[:4])\n\tcase uint64Key:\n\t\tbinary.LittleEndian.PutUint64(buf[:8], uint64(v))\n\t\treturn hashBytes(seed, buf[:8])\n\tcase float32Key:\n\t\tbinary.LittleEndian.PutUint32(buf[:4], math.Float32bits(float32(v)))\n\t\treturn hashBytes(seed, buf[:4])\n\tcase float64Key:\n\t\tbinary.LittleEndian.PutUint64(buf[:8], math.Float64bits(float64(v)))\n\t\treturn hashBytes(seed, buf[:8])\n\tcase boolKey:\n\t\tif v {\n\t\t\tbuf[0] = 1\n\t\t}\n\t\treturn hashBytes(seed, buf[:1])\n\tcase sliceKey:\n\t\treturn hashSliceKey(seed, v)\n\tcase pointerKey:\n\t\treturn hashSliceKey(seed, v.sliceKey)\n\tcase pathKey:\n\t\treturn hashSliceKey(seed, v.sliceKey)\n\tcase nilKey:\n\t\treturn hashBytes(seed, nil)\n\tcase Hashable:\n\t\t// Mix up the hash to ensure it covers 64-bits\n\t\tbinary.LittleEndian.PutUint64(buf[:8], v.Hash())\n\t\treturn hashBytes(seed, buf[:8])\n\tdefault:\n\t\ts := _nilinterhash(v.Key())\n\t\tbinary.LittleEndian.PutUint64(buf[:8], uint64(s))\n\t\treturn hashBytes(seed, buf[:8])\n\t}\n}", "func (dtk *DcmTagKey) Hash() uint32 {\n\treturn ((uint32(int(dtk.group)<<16) & 0xffff0000) | (uint32(int(dtk.element) & 0xffff)))\n}", "func (spec Spec) DeepHash() string {\n\thash := sha512.New512_224()\n\tspec.DefaultService.hash(hash)\n\tfor _, rule := range spec.Rules {\n\t\trule.hash(hash)\n\t}\n\tsvcs := make([]string, len(spec.AllServices))\n\ti := 0\n\tfor k := range spec.AllServices {\n\t\tsvcs[i] = k\n\t\ti++\n\t}\n\tsort.Strings(svcs)\n\tfor _, svc := range svcs {\n\t\thash.Write([]byte(svc))\n\t\tspec.AllServices[svc].hash(hash)\n\t}\n\tspec.ShardCluster.hash(hash)\n\thash.Write([]byte(spec.VCL))\n\tfor _, auth := range spec.Auths {\n\t\tauth.hash(hash)\n\t}\n\tfor _, acl := range spec.ACLs {\n\t\tacl.hash(hash)\n\t}\n\tfor _, rw := range spec.Rewrites {\n\t\trw.hash(hash)\n\t}\n\tfor _, reqDisp := range spec.Dispositions {\n\t\treqDisp.hash(hash)\n\t}\n\th := new(big.Int)\n\th.SetBytes(hash.Sum(nil))\n\treturn h.Text(62)\n}" ]
[ "0.71392304", "0.7118988", "0.70433563", "0.70033073", "0.7000457", "0.69771796", "0.6946999", "0.69167256", "0.6911272", "0.6885437", "0.6881275", "0.6854698", "0.6839323", "0.6808889", "0.6802325", "0.67797977", "0.6749414", "0.6748137", "0.6745063", "0.67278856", "0.66853213", "0.6683674", "0.6680236", "0.6667324", "0.6652535", "0.6650117", "0.6633402", "0.65988237", "0.6597991", "0.65964407", "0.65933704", "0.6581115", "0.65747106", "0.6563886", "0.6557271", "0.6556366", "0.6554731", "0.65483683", "0.654223", "0.65323585", "0.6528166", "0.65242773", "0.651947", "0.65086025", "0.6507619", "0.6488133", "0.64871293", "0.6477431", "0.64751476", "0.647411", "0.6473493", "0.6471136", "0.64645445", "0.6462771", "0.64463836", "0.64432764", "0.64392227", "0.6434115", "0.6424212", "0.6419294", "0.6406749", "0.64006823", "0.64000505", "0.63998294", "0.63993245", "0.6399127", "0.6398219", "0.6397337", "0.63966537", "0.6391968", "0.63831466", "0.63792545", "0.6376018", "0.6374794", "0.63734704", "0.63600224", "0.6357834", "0.6355114", "0.63505995", "0.63443583", "0.63391376", "0.633428", "0.63270235", "0.63235474", "0.6320736", "0.6319919", "0.6317253", "0.6316801", "0.6316565", "0.63146734", "0.6311661", "0.63106495", "0.6308854", "0.6308431", "0.6305678", "0.63046443", "0.6303732", "0.62989837", "0.62983876", "0.6298272", "0.6296924" ]
0.0
-1
String Hash32 implements the Arg.Hash32() method.
func (s String) Hash32(h hash.Hash32) { h.Write([]byte(s)) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func HexToHash32(s string) Hash32 { return BytesToHash(util.FromHex(s)) }", "func (p Path) Hash32(h hash.Hash32) { h.Write([]byte(p)) }", "func hash(s string) uint32 {\n h := fnv.New32a()\n h.Write([]byte(s))\n return h.Sum32()\n}", "func (t *Target) Hash32(h hash.Hash32) {\n\th.Write([]byte(t.Name))\n\th.Write([]byte(t.Builder))\n\tfor _, arg := range t.Args {\n\t\targ.Hash32(h)\n\t}\n\tfor _, env := range t.Env {\n\t\th.Write([]byte(env))\n\t}\n}", "func hash(value string) uint32 {\n\th := fnv.New32a()\n\th.Write([]byte(value))\n\n\treturn h.Sum32()\n}", "func Hash(strings ...string) uint32 {\n\tdigester := fnv.New32()\n\tfor _, s := range strings {\n\t\t_, _ = io.WriteString(digester, s)\n\t}\n\treturn digester.Sum32()\n}", "func Hash(key string) uint32 {\n\treturn uint32(aeshashstr(noescape(unsafe.Pointer(&key)), 0))\n}", "func hash(s string) int {\n\th := fnv.New32a()\n\tif _, err := h.Write([]byte(s)); err != nil {\n\t\tpanic(err) // should never happen\n\t}\n\n\treturn int(h.Sum32() & 0x7FFFFFFF) // mask MSB of uint32 as this will be sign bit\n}", "func TestExample(t *testing.T) {\n\tstr := \"hello world\"\n\tbytes := []byte(str)\n\thash := Hash32(bytes)\n\tfmt.Printf(\"Hash32(%s) is %x\\n\", str, hash)\n}", "func CalcHash32(data []byte) Hash32 {\n\treturn hash.Sum(data)\n}", "func hash(s string) string {\n\th := fnv.New32a()\n\t_, err := h.Write([]byte(s))\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn fmt.Sprint(h.Sum32())\n}", "func hash(s string) string {\n\thash := fnv.New32a()\n\thash.Write([]byte(s))\n\tintHash := hash.Sum32()\n\tresult := fmt.Sprintf(\"%08x\", intHash)\n\treturn result\n}", "func (s *Sub) Hash32(h hash.Hash32) {\n\th.Write([]byte(s.Format))\n\tfor _, sub := range s.Substitutions {\n\t\th.Write([]byte(sub.Key))\n\t\tsub.Value.Hash32(h)\n\t}\n}", "func strhash(a unsafe.Pointer, h uintptr) uintptr", "func (h Hash32) Hex() string { return util.Encode(h[:]) }", "func Hash32(s []byte) uint32 {\n\tn := uint32(len(s))\n\tif n <= 24 {\n\t\tif n <= 12 {\n\t\t\tif n <= 4 {\n\t\t\t\treturn hash32Len0to4(s)\n\t\t\t}\n\t\t\treturn hash32Len5to12(s)\n\t\t}\n\t\treturn hash32Len13to24(s)\n\t}\n\n\t// n > 24\n\th := n\n\tg := c1 * n\n\tf := g\n\n\ta0 := ror32(fetch32(s[n-4:])*c1, 17) * c2\n\ta1 := ror32(fetch32(s[n-8:])*c1, 17) * c2\n\ta2 := ror32(fetch32(s[n-16:])*c1, 17) * c2\n\ta3 := ror32(fetch32(s[n-12:])*c1, 17) * c2\n\ta4 := ror32(fetch32(s[n-20:])*c1, 17) * c2\n\n\tconst magic = 0xe6546b64\n\th ^= a0\n\th = ror32(h, 19)\n\th = h*5 + magic\n\th ^= a2\n\th = ror32(h, 19)\n\th = h*5 + magic\n\tg ^= a1\n\tg = ror32(g, 19)\n\tg = g*5 + magic\n\tg ^= a3\n\tg = ror32(g, 19)\n\tg = g*5 + magic\n\tf += a4\n\tf = ror32(f, 19)\n\tf = f*5 + magic\n\tfor i := (n - 1) / 20; i != 0; i-- {\n\t\ta0 := ror32(fetch32(s)*c1, 17) * c2\n\t\ta1 := fetch32(s[4:])\n\t\ta2 := ror32(fetch32(s[8:])*c1, 17) * c2\n\t\ta3 := ror32(fetch32(s[12:])*c1, 17) * c2\n\t\ta4 := fetch32(s[16:])\n\t\th ^= a0\n\t\th = ror32(h, 18)\n\t\th = h*5 + magic\n\t\tf += a1\n\t\tf = ror32(f, 19)\n\t\tf = f * c1\n\t\tg += a2\n\t\tg = ror32(g, 18)\n\t\tg = g*5 + magic\n\t\th ^= a3 + a1\n\t\th = ror32(h, 19)\n\t\th = h*5 + magic\n\t\tg ^= a4\n\t\tg = bswap32(g) * 5\n\t\th += a4 * 5\n\t\th = bswap32(h)\n\t\tf += a0\n\t\tf, g, h = g, h, f // a.k.a. PERMUTE3\n\t\ts = s[20:]\n\t}\n\tg = ror32(g, 11) * c1\n\tg = ror32(g, 17) * c1\n\tf = ror32(f, 11) * c1\n\tf = ror32(f, 17) * c1\n\th = ror32(h+g, 19)\n\th = h*5 + magic\n\th = ror32(h, 17) * c1\n\th = ror32(h+f, 19)\n\th = h*5 + magic\n\th = ror32(h, 17) * c1\n\treturn h\n}", "func Hash32(s []byte) uint32 {\n\n\tslen := len(s)\n\n\tif slen <= 24 {\n\t\tif slen <= 12 {\n\t\t\tif slen <= 4 {\n\t\t\t\treturn hash32Len0to4(s, 0)\n\t\t\t}\n\t\t\treturn hash32Len5to12(s, 0)\n\t\t}\n\t\treturn hash32Len13to24Seed(s, 0)\n\t}\n\n\t// len > 24\n\th := uint32(slen)\n\tg := c1 * uint32(slen)\n\tf := g\n\ta0 := rotate32(fetch32(s, slen-4)*c1, 17) * c2\n\ta1 := rotate32(fetch32(s, slen-8)*c1, 17) * c2\n\ta2 := rotate32(fetch32(s, slen-16)*c1, 17) * c2\n\ta3 := rotate32(fetch32(s, slen-12)*c1, 17) * c2\n\ta4 := rotate32(fetch32(s, slen-20)*c1, 17) * c2\n\th ^= a0\n\th = rotate32(h, 19)\n\th = h*5 + 0xe6546b64\n\th ^= a2\n\th = rotate32(h, 19)\n\th = h*5 + 0xe6546b64\n\tg ^= a1\n\tg = rotate32(g, 19)\n\tg = g*5 + 0xe6546b64\n\tg ^= a3\n\tg = rotate32(g, 19)\n\tg = g*5 + 0xe6546b64\n\tf += a4\n\tf = rotate32(f, 19) + 113\n\titers := (slen - 1) / 20\n\tfor {\n\t\ta := fetch32(s, 0)\n\t\tb := fetch32(s, 4)\n\t\tc := fetch32(s, 8)\n\t\td := fetch32(s, 12)\n\t\te := fetch32(s, 16)\n\t\th += a\n\t\tg += b\n\t\tf += c\n\t\th = mur(d, h) + e\n\t\tg = mur(c, g) + a\n\t\tf = mur(b+e*c1, f) + d\n\t\tf += g\n\t\tg += f\n\t\ts = s[20:]\n\t\titers--\n\t\tif iters == 0 {\n\t\t\tbreak\n\t\t}\n\t}\n\tg = rotate32(g, 11) * c1\n\tg = rotate32(g, 17) * c1\n\tf = rotate32(f, 11) * c1\n\tf = rotate32(f, 17) * c1\n\th = rotate32(h+g, 19)\n\th = h*5 + 0xe6546b64\n\th = rotate32(h, 17) * c1\n\th = rotate32(h+f, 19)\n\th = h*5 + 0xe6546b64\n\th = rotate32(h, 17) * c1\n\treturn h\n}", "func strhash0(p unsafe.Pointer, h uintptr) uintptr", "func (ch *ConsistentHash) fnv32Hash(key string) uint32 {\n\tnew32Hash := fnv.New32()\n\tnew32Hash.Write([]byte(key))\n\treturn new32Hash.Sum32()\n}", "func strhash(p *string, h uintptr) uintptr", "func (h Hash32) Field() log.Field { return log.String(\"hash\", hex.EncodeToString(h[:])) }", "func stringToHash(s string) uint32 {\n\th := fnv.New32a()\n\th.Write([]byte(s))\n\treturn h.Sum32()\n}", "func Hash(length int, key string) int64 {\n\tif key == \"\" {\n\t\treturn 0\n\t}\n\thc := hashCode(key)\n\treturn (hc ^ (hc >> 16)) % int64(length)\n}", "func (h Hash32) String() string {\n\treturn h.Hex()\n}", "func hashInt(s string) uint32 {\n\tb := []byte(s)\n\th := crc32.ChecksumIEEE(b)\n\treturn h\n}", "func hash(x []byte) uint32 {\n\treturn crc32.ChecksumIEEE(x)\n}", "func hashstr(s string) HashType {\n\tvar result uint32\n\tresult = 0\n\tbytes := []byte(s)\n\n\tfor n, b := range bytes {\n\t\tresult ^= uint32(b) ^ (uint32(n*int(b)) << 8)\n\t\ttmp := result >> 27\n\t\tresult = result << 5\n\t\tresult = result | tmp\n\t}\n\n\treturn HashType(result)\n\n}", "func (gg GlobGroup) Hash32(h hash.Hash32) {\n\tfor _, p := range gg {\n\t\th.Write([]byte(p))\n\t}\n}", "func FNV32(s string) uint32 {\n\treturn uint32Hasher(fnv.New32(), s)\n}", "func (h Hash20) ToHash32() (h32 Hash32) {\n\tcopy(h32[:], h[:])\n\treturn\n}", "func stringHash(s string, seed uintptr) uintptr", "func (d Data32) Hash() Hash {\n\treturn hash(d)\n}", "func stringHash(s string) uint32 {\n\tbytes := []byte(s)\n\tn := len(bytes)\n\thash := uint32(0)\n\tfor index, b := range bytes {\n\t\thash += uint32(b) * uint32(math.Pow(31, float64(n-index)))\n\t}\n\treturn hash\n}", "func htkey_hash_str(k1 voidptr, len int) usize {\n\tvar k1p byteptr = (byteptr)(k1)\n\tvar hash usize\n\n\thash = 0 + 5381 + len + 1\n\tfor i := 0; i < len; i++ {\n\t\tc := k1p[i]\n\t\thash = ((hash << 5) + hash) ^ usize(c)\n\t}\n\n\treturn hash\n}", "func fnv32a(s string) uint32 {\n\tconst (\n\t\tinitial = 2166136261\n\t\tprime = 16777619\n\t)\n\n\thash := uint32(initial)\n\tfor i := 0; i < len(s); i++ {\n\t\thash ^= uint32(s[i])\n\t\thash *= prime\n\t}\n\treturn hash\n}", "func (this *Ring) Hash(key string) uint32 {\n\treturn crc32.ChecksumIEEE([]byte(key))\n}", "func f32hash(p unsafe.Pointer, h uintptr) uintptr {\n\tf := *(*float32)(p)\n\tswitch {\n\tcase f == 0:\n\t\treturn c1 * (c0 ^ h) // +0, -0\n\tcase f != f:\n\t\treturn c1 * (c0 ^ h ^ uintptr(fastrand())) // any kind of NaN\n\tdefault:\n\t\treturn memhash(p, h, 4)\n\t}\n}", "func Bytes32ToIpfsHash(value [32]byte) (string, error) {\n\tbyteArray := [34]byte{18, 32}\n\tcopy(byteArray[2:], value[:])\n\tif len(byteArray) != 34 {\n\t\treturn \"\", errors.New(\"invalid bytes32 value\")\n\t}\n\n\thash := base58.Encode(byteArray[:])\n\treturn hash, nil\n}", "func hash3(u uint32, h uint8) uint32 {\n\treturn ((u << (32 - 24)) * prime3bytes) >> ((32 - h) & 31)\n}", "func Sha3256(bs []byte) ([]byte, error) {\n\treturn PerformHash(sha3.New256(), bs)\n}", "func Hash(str string) string {\n\thasher := fnv.New64a()\n\thasher.Write([]byte(str))\n\treturn hex.EncodeToString(hasher.Sum(nil))\n}", "func FNVHash32(value uint32) uint32 {\n\thash := FNVOffsetBasis32\n\tfor i := 0; i < 4; i++ {\n\t\toctet := value & 0x00FF\n\t\tvalue >>= 8\n\n\t\thash ^= octet\n\t\thash *= FNVPrime32\n\t}\n\treturn hash\n}", "func HashASM(k0, k1 uint64, p []byte) uint64", "func CalculateHash(args []string) string {\n\tvar str = \"\"\n\tfor _,v := range args {\n\t\tstr += v\n\t}\n\thasher := sha256.New()\n\thasher.Write([]byte(str))\n\treturn hex.EncodeToString(hasher.Sum(nil))\n}", "func Sum32(key string) uint32 {\n\treturn Sum32Seed(key, 0)\n}", "func hexHash(input string) string {\n return idToString(hash([]byte(input)))\n}", "func Hasher(value string) string {\n\th := fnv.New32a()\n\t_, _ = h.Write([]byte(value))\n\treturn fmt.Sprintf(\"%v\", h.Sum32())\n}", "func Hash32WithSeed(s []byte, seed uint32) uint32 {\n\tslen := len(s)\n\n\tif slen <= 24 {\n\t\tif slen >= 13 {\n\t\t\treturn hash32Len13to24Seed(s, seed*c1)\n\t\t}\n\t\tif slen >= 5 {\n\t\t\treturn hash32Len5to12(s, seed)\n\t\t}\n\t\treturn hash32Len0to4(s, seed)\n\t}\n\th := hash32Len13to24Seed(s[:24], seed^uint32(slen))\n\treturn mur(Hash32(s[24:])+seed, h)\n}", "func HashFunction(buf []byte) uint32 {\n\tvar hash uint32 = 5381\n\tfor _, b := range buf {\n\t\thash = ((hash << 5) + hash) + uint32(b)\n\t}\n\treturn hash\n}", "func StrToBytes32(s string) Bytes32 {\n\treturn BytesToBytes32([]byte(s))\n}", "func hash(elements ...[32]byte) [32]byte {\n\tvar hash []byte\n\tfor i := range elements {\n\t\thash = append(hash, elements[i][:]...)\n\t}\n\treturn sha256.Sum256(hash)\n}", "func getHash(s string) uint32 {\n\ttbl := crc32.MakeTable(crc32.IEEE)\n\treturn crc32.Checksum([]byte(s), tbl)\n}", "func hash(data []byte) uint32 {\n\tvar h uint32 = binary.LittleEndian.Uint32(data) * kDictHashMul32\n\n\t/* The higher bits contain more mixture from the multiplication,\n\t so we take our results from there. */\n\treturn h >> uint(32-kDictNumBits)\n}", "func FNV32a(s string) uint32 {\n\treturn uint32Hasher(fnv.New32a(), s)\n}", "func hashString(s string) int {\n\tvar h uint32\n\tfor i := 0; i < len(s); i++ {\n\t\th ^= uint32(s[i])\n\t\th *= 16777619\n\t}\n\treturn int(h)\n}", "func Hash(p string) (string, string) {\n\ts := GenString(32)\n\treturn ComputeHash(p, s), s\n}", "func (t *hashReader) Sum32() uint32 {\n\treturn t.h.Sum32()\n}", "func hashSHA256(str string) string {\n\ts := sha256.New()\n\ts.Write([]byte(str))\n\treturn base64.StdEncoding.EncodeToString(s.Sum(nil))\n}", "func hash(data []byte) [32]byte {\n\tvar hash [32]byte\n\n\th := sha256.New()\n\t// The hash interface never returns an error, for that reason\n\t// we are not handling the error below. For reference, it is\n\t// stated here https://golang.org/pkg/hash/#Hash\n\t// #nosec G104\n\th.Write(data)\n\th.Sum(hash[:0])\n\n\treturn hash\n}", "func hash(key uint64) uint64 {\r\n\tkey ^= key >> 33\r\n\tkey *= 0xff51afd7ed558ccd\r\n\tkey ^= key >> 33\r\n\tkey *= 0xc4ceb9fe1a85ec53\r\n\tkey ^= key >> 33\r\n\treturn key\r\n}", "func hashSHA256(input string) string {\n\th := sha1.New()\n\th.Write([]byte(input))\n\treturn hex.EncodeToString(h.Sum(nil))\n}", "func StringToHash(s string) int {\n\tv := int(crc32.ChecksumIEEE([]byte(s)))\n\tif v >= 0 {\n\t\treturn v\n\t}\n\tif -v >= 0 {\n\t\treturn -v\n\t}\n\t// v == MinInt\n\treturn 0\n}", "func Hash(b []byte) uint32 {\n\tconst (\n\t\tseed = 0xbc9f1d34\n\t\tm = 0xc6a4a793\n\t)\n\th := uint32(seed) ^ uint32(len(b))*m\n\tfor ; len(b) >= 4; b = b[4:] {\n\t\th += uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24\n\t\th *= m\n\t\th ^= h >> 16\n\t}\n\tswitch len(b) {\n\tcase 3:\n\t\th += uint32(b[2]) << 16\n\t\tfallthrough\n\tcase 2:\n\t\th += uint32(b[1]) << 8\n\t\tfallthrough\n\tcase 1:\n\t\th += uint32(b[0])\n\t\th *= m\n\t\th ^= h >> 24\n\t}\n\treturn h\n}", "func Crc32(str string) uint32 {\n\treturn crc32.ChecksumIEEE([]byte(str))\n}", "func hash(k Key) int {\n\tkey := fmt.Sprintf(\"%s\", k)\n\th := 0\n\tfor i := 0; i < len(key); i++ {\n\t\th = 31 * h + int(key[i])\n\t}\n\treturn h\n}", "func CalcObjectHash32(obj scale.Encodable) Hash32 {\n\tbytes, err := codec.Encode(obj)\n\tif err != nil {\n\t\tpanic(\"could not serialize object\")\n\t}\n\treturn CalcHash32(bytes)\n}", "func (h *Hash) IsHash32() bool {\n\t_, ok := h.Hash.(hash.Hash32)\n\treturn ok\n}", "func sha3hash(t *testing.T, data ...[]byte) []byte {\n\tt.Helper()\n\th := sha3.NewLegacyKeccak256()\n\tr, err := doSum(h, nil, data...)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\treturn r\n}", "func Hash(data []byte) (string, int64) {\n\thasher := adler32.New()\n\tb, e := hasher.Write(data)\n\tif e != nil {\n\t\tlogs.WithFields(logs.Fields{\n\t\t\t\"Error\": e,\n\t\t}).Error(\"Unable to write chunk of data via hasher.Write\", e)\n\t}\n\treturn hex.EncodeToString(hasher.Sum(nil)), int64(b)\n}", "func (h *Hash) Sum32() (uint32, bool) {\n\th32, ok := h.Hash.(hash.Hash32)\n\tif !ok {\n\t\treturn 0, false\n\t}\n\n\treturn h32.Sum32(), true\n}", "func hash(ba string) string {\n\th := sha256.New()\n\th.Write([]byte(ba))\n\n\treturn fmt.Sprintf(\"%x\", h.Sum(nil))\n}", "func (h *MemHash) Hash() uint32 {\n\tss := (*stringStruct)(unsafe.Pointer(&h.buf))\n\treturn uint32(memhash(ss.str, 0, uintptr(ss.len)))\n}", "func hashcode(s string) int {\n\treturn tpgresource.Hashcode(s)\n}", "func HashBuildArgs(args interface{}) (string, error) {\n\tdata, err := json.Marshal(args)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\thash := sha256.Sum256(data)\n\treturn hex.EncodeToString(hash[:]), nil\n}", "func Hash(data interface{}) string {\n\treturn hex.EncodeToString(RawHash(data))\n}", "func TestHashString(t *testing.T) {\n\t// Block 100000 hash.\n\twantStr := \"000000000003ba27aa200b1cecaad478d2b00432346c3f1f3986da1afd33e506\"\n\thash := Hash([HashSize]byte{ // Make go vet happy.\n\t\t0x06, 0xe5, 0x33, 0xfd, 0x1a, 0xda, 0x86, 0x39,\n\t\t0x1f, 0x3f, 0x6c, 0x34, 0x32, 0x04, 0xb0, 0xd2,\n\t\t0x78, 0xd4, 0xaa, 0xec, 0x1c, 0x0b, 0x20, 0xaa,\n\t\t0x27, 0xba, 0x03, 0x00, 0x00, 0x00, 0x00, 0x00,\n\t})\n\n\thashStr := hash.String()\n\tassert.Equal(t, wantStr, hashStr)\n}", "func Hash(input []byte) string {\n\treturn fmt.Sprintf(\"%x\", sha256.Sum256(input))\n}", "func hash(key, value string) int64 {\n\thash := siphash.New(sipConst)\n\thash.Write([]byte(key + \":::\" + value))\n\treturn int64(hash.Sum64())\n}", "func Sum32Seed(key string, seed uint32) uint32 {\n\tvar nblocks = len(key) / 4\n\tvar nbytes = nblocks * 4\n\tvar h1 = seed\n\tconst c1 = 0xcc9e2d51\n\tconst c2 = 0x1b873593\n\tfor i := 0; i < nbytes; i += 4 {\n\t\tk1 := uint32(key[i+0]) | uint32(key[i+1])<<8 |\n\t\t\tuint32(key[i+2])<<16 | uint32(key[i+3])<<24\n\t\tk1 *= c1\n\t\tk1 = (k1 << 15) | (k1 >> 17)\n\t\tk1 *= c2\n\t\th1 ^= k1\n\t\th1 = (h1 << 13) | (h1 >> 19)\n\t\th1 = h1*5 + 0xe6546b64\n\t}\n\tvar k1 uint32\n\tswitch len(key) & 3 {\n\tcase 3:\n\t\tk1 ^= uint32(key[nbytes+2]) << 16\n\t\tfallthrough\n\tcase 2:\n\t\tk1 ^= uint32(key[nbytes+1]) << 8\n\t\tfallthrough\n\tcase 1:\n\t\tk1 ^= uint32(key[nbytes+0])\n\t\tk1 *= c1\n\t\tk1 = (k1 << 15) | (k1 >> 17)\n\t\tk1 *= c2\n\t\th1 ^= k1\n\t}\n\th1 ^= uint32(len(key))\n\th1 ^= h1 >> 16\n\th1 *= 0x85ebca6b\n\th1 ^= h1 >> 13\n\th1 *= 0xc2b2ae35\n\th1 ^= h1 >> 16\n\treturn h1\n}", "func byteshash(p *[]byte, h uintptr) uintptr", "func shortHash(s string, l int) string {\n\tsum := md5.Sum([]byte(s)) //nolint:gosec\n\tvar b = sum[:]\n\th := base32.StdEncoding.EncodeToString(b)\n\n\treturn strings.ToLower(h)[:l]\n}", "func TestHash32(t *testing.T) {\n\tstdHash := crc32.New(crc32.IEEETable)\n\tif _, err := stdHash.Write([]byte(\"test\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\t// create a new hash with stdHash.Sum32() as initial crc\n\tcrcHash := New(stdHash.Sum32(), crc32.IEEETable)\n\n\tstdHashSize := stdHash.Size()\n\tcrcHashSize := crcHash.Size()\n\tif stdHashSize != crcHashSize {\n\t\tt.Fatalf(\"%d != %d\", stdHashSize, crcHashSize)\n\t}\n\n\tstdHashBlockSize := stdHash.BlockSize()\n\tcrcHashBlockSize := crcHash.BlockSize()\n\tif stdHashBlockSize != crcHashBlockSize {\n\t\tt.Fatalf(\"%d != %d\", stdHashBlockSize, crcHashBlockSize)\n\t}\n\n\tstdHashSum32 := stdHash.Sum32()\n\tcrcHashSum32 := crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n\n\tstdHashSum := stdHash.Sum(make([]byte, 32))\n\tcrcHashSum := crcHash.Sum(make([]byte, 32))\n\tif !reflect.DeepEqual(stdHashSum, crcHashSum) {\n\t\tt.Fatalf(\"sum = %v, want %v\", crcHashSum, stdHashSum)\n\t}\n\n\t// write something\n\tif _, err := stdHash.Write([]byte(\"hello\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif _, err := crcHash.Write([]byte(\"hello\")); err != nil {\n\t\tt.Fatal(err)\n\t}\n\tstdHashSum32 = stdHash.Sum32()\n\tcrcHashSum32 = crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n\n\t// reset\n\tstdHash.Reset()\n\tcrcHash.Reset()\n\tstdHashSum32 = stdHash.Sum32()\n\tcrcHashSum32 = crcHash.Sum32()\n\tif stdHashSum32 != crcHashSum32 {\n\t\tt.Fatalf(\"%d != %d\", stdHashSum32, crcHashSum32)\n\t}\n}", "func hashname(l int, s ...string) string {\n\tconst shorthash = 6 // the length of the shorthash\n\n\tr := strings.Join(s, \"/\")\n\tif l > len(r) {\n\t\t// we're under the limit, nothing to do\n\t\treturn r\n\t}\n\thash := fmt.Sprintf(\"%x\", sha256.Sum256([]byte(r)))\n\tfor n := len(s) - 1; n >= 0; n-- {\n\t\ts[n] = truncate(l/len(s), s[n], hash[:shorthash])\n\t\tr = strings.Join(s, \"/\")\n\t\tif l > len(r) {\n\t\t\treturn r\n\t\t}\n\t}\n\t// truncated everything, but we're still too long\n\t// just return the hash truncated to l.\n\treturn hash[:min(len(hash), l)]\n}", "func addrHash(addr uint16) byte {\n\treturn (byte(addr) ^ byte(addr>>8)) & 0x7f\n}", "func (_Contract *ContractCallerSession) Bytes32ToString(_bytes32 [32]byte) (string, error) {\n\treturn _Contract.Contract.Bytes32ToString(&_Contract.CallOpts, _bytes32)\n}", "func hash(id, app string) string {\n\treturn id + \"|\" + app\n}", "func (p PlainTextHasher) Hash(str string) string { return str }", "func (s *ShardMap) hash(v interface{}) int {\n\tswitch s.Type {\n\tcase \"string\":\n\t\tval, ok := v.(string)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\thash := fnv.New32()\n\t\thash.Write([]byte(val))\n\t\treturn int(hash.Sum32() % NumShards)\n\tcase \"int32\":\n\t\t// Values that come as numbers in JSON are of type float64.\n\t\tval, ok := v.(float64)\n\t\tif !ok {\n\t\t\treturn -1\n\t\t}\n\n\t\treturn int(int32(val) % NumShards)\n\tdefault:\n\t\treturn -1\n\t}\n}", "func TestHash(t *testing.T) {\n\tdata := \"bce-auth-v1/aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa/2015-04-27T08:23:49Z/1800\"\n\tkey := \"bbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbb\"\n\tresult := \"1d5ce5f464064cbee060330d973218821825ac6952368a482a592e6615aef479\"\n\ttestResult := hash(data, key)\n\tif result == testResult {\n\t\tt.Log(\"hash test success\")\n\t} else {\n\t\tt.Error(\"hash test fail\")\n\t}\n}", "func SHA256(text string) string {\n\talgorithm := sha256.New()\n\treturn stringHasher(algorithm, text)\n}", "func (str String) Hash() int {\n\th := xxhash.ChecksumString64S(string(str), hashSeed0)\n\treturn int(h)\n}", "func sumHash(c byte, h uint32) uint32 {\n\treturn (h * hashPrime) ^ uint32(c)\n}", "func crc32Demo() {\n\t// hasher\n\th := crc32.NewIEEE()\n\tfmt.Println(reflect.TypeOf(h))\n\n\t// write a string converted to bytes\n\th.Write([]byte(\"test\"))\n\n\t// checksum\n\tv := h.Sum32()\n\tfmt.Println(reflect.TypeOf(v)) // uint32\n\tfmt.Println(v)\n}", "func Hash(text string) (string, error) {\n\treturn hashText(fnv.New64a(), text)\n}", "func Hash(s string) string {\n\tfmt.Println(\"Hashing string:\", s)\n\thash := sha256.Sum256([]byte(s))\n\tniceHash := fmt.Sprintf(\"%x\", hash)\n\tfmt.Println(\"Created hash:\", hash)\n\treturn niceHash\n}", "func (p Path) Hash() (uint32, error) {\n\treturn adler32.Checksum([]byte(p)), nil\n}", "func (_Contract *ContractCaller) Bytes32ToString(opts *bind.CallOpts, _bytes32 [32]byte) (string, error) {\n\tvar out []interface{}\n\terr := _Contract.contract.Call(opts, &out, \"bytes32ToString\", _bytes32)\n\n\tif err != nil {\n\t\treturn *new(string), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(string)).(*string)\n\n\treturn out0, err\n\n}", "func hash4(u uint32, h uint8) uint32 {\n\treturn (u * prime4bytes) >> ((32 - h) & 31)\n}", "func (s sha256hashing) HashString(hashme string, digest []byte) error {\n\thashAlgo := sha256.New()\n\thashAlgo.Write([]byte(hashme))\n\n\tcopySHA256(digest, hashAlgo.Sum(nil))\n\n\treturn nil\n}", "func (h *ArgonHasher) Hash(password string) (string, error) {\n\tparams := h.ArgonParams\n\n\tsalt, err := randomBytes(params.SaltLength)\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\thash := argon2.IDKey([]byte(password), salt, params.Iterations, params.Memory, params.Threads, params.KeyLength)\n\tb64Salt := base64.RawStdEncoding.EncodeToString(salt)\n\tb64Hash := base64.RawStdEncoding.EncodeToString(hash)\n\tencoded := fmt.Sprintf(\"$argon2id$v=%d$m=%d,t=%d,p=%d$%s$%s\", argon2.Version, params.Memory, params.Iterations, params.Threads, b64Salt, b64Hash)\n\n\treturn encoded, nil\n}" ]
[ "0.7575041", "0.7430379", "0.7387939", "0.7271739", "0.7122693", "0.6981774", "0.6963979", "0.69053805", "0.68457013", "0.68161345", "0.6785067", "0.67835623", "0.67606795", "0.6733655", "0.67066145", "0.670613", "0.66108847", "0.6555351", "0.6541611", "0.6475872", "0.64596903", "0.6396756", "0.6336968", "0.6336079", "0.62828106", "0.6255263", "0.62541646", "0.62458867", "0.61925715", "0.6174756", "0.6124632", "0.61207646", "0.61198455", "0.6098404", "0.6091043", "0.6080642", "0.60675246", "0.60571253", "0.60543084", "0.6044169", "0.6038852", "0.60284233", "0.60173947", "0.601395", "0.6002728", "0.59844106", "0.5962033", "0.5957986", "0.5955375", "0.5949933", "0.5925127", "0.5909665", "0.5889476", "0.58748716", "0.5872962", "0.5855659", "0.58504504", "0.5848136", "0.5846018", "0.5828848", "0.58048725", "0.57895166", "0.57856345", "0.57812357", "0.5753921", "0.57537556", "0.5747864", "0.5737445", "0.57367676", "0.5733663", "0.57300705", "0.57287693", "0.5727093", "0.57261", "0.57244307", "0.57234484", "0.571149", "0.5700931", "0.5699886", "0.5698005", "0.5693854", "0.5671225", "0.5662596", "0.56500834", "0.56362814", "0.56292295", "0.56268483", "0.56262654", "0.56215656", "0.5619436", "0.56173664", "0.56170297", "0.5612858", "0.5611557", "0.5610468", "0.560847", "0.56013924", "0.5600617", "0.5597706", "0.55965084" ]
0.80541074
0
Sub Type implements the starlark.Value.Type() method.
func (s *Sub) Type() string { return "Sub" }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (value *Value) Type() Type {\n\treturn value.valueType\n}", "func (v Value) Type() Type {\n\treturn v.Typ\n}", "func (val Value) Type() Type {\n\treturn val.typ\n}", "func (v Value) Type() Type {\n\treturn v.typ\n}", "func (v Value) Type() ValueType {\n\tif v.iface == nil {\n\t\treturn NilType\n\t}\n\tswitch v.iface.(type) {\n\tcase int64:\n\t\treturn IntType\n\tcase float64:\n\t\treturn FloatType\n\tcase bool:\n\t\treturn BoolType\n\tcase string:\n\t\treturn StringType\n\tcase *Table:\n\t\treturn TableType\n\tcase *Code:\n\t\treturn CodeType\n\tcase Callable:\n\t\treturn FunctionType\n\tcase *Thread:\n\t\treturn ThreadType\n\tcase *UserData:\n\t\treturn UserDataType\n\tdefault:\n\t\treturn UnknownType\n\t}\n}", "func (v Value) Type() Type {\n\tif !v.v.IsValid() {\n\t\treturn TypeUndefined\n\t}\n\n\tif v.v.CanInterface() {\n\t\ti := v.v.Interface()\n\t\tswitch i.(type) {\n\t\tcase Function:\n\t\t\treturn TypeFunction\n\t\tcase Object:\n\t\t\tif _, ok := i.(stringObject); ok {\n\t\t\t\treturn TypeString\n\t\t\t}\n\t\t\treturn TypeObject\n\t\t}\n\t}\n\n\tswitch v.v.Kind() {\n\tcase reflect.Ptr:\n\t\treturn TypeNull\n\tcase reflect.Bool:\n\t\treturn TypeBoolean\n\tcase reflect.Float64:\n\t\treturn TypeNumber\n\tdefault:\n\t\treturn TypeUndefined\n\t}\n\n}", "func (this *Value) Type() int {\n\treturn this.parsedType\n}", "func (v *Value) Type() kiwi.ValueType {\n\treturn Type\n}", "func (v Value) Type() querypb.Type {\n\treturn v.typ\n}", "func (inst *InstFSub) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (v Value) Type() Type {\n\tpanic(message)\n}", "func (inst *InstSub) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (this *Self) Type() value.Type { return value.JSON }", "func (data *Instance) Type() Value {\n\treturn data.TypeTag\n}", "func (a ValueNode) GetType() string {\n\treturn \"ValueNode\"\n}", "func Type(value r.Value) r.Type {\n\tif !value.IsValid() || value == None {\n\t\treturn nil\n\t}\n\treturn value.Type()\n}", "func (object Object) Type(value interface{}) Object {\n\treturn object.Property(as.PropertyType, value)\n}", "func (s S) Type() Type {\n\treturn s.typ\n}", "func (a AttributeValue) Type() AttributeValueType {\n\tif a.orig.Value == nil {\n\t\treturn AttributeValueNULL\n\t}\n\tswitch a.orig.Value.(type) {\n\tcase *otlpcommon.AnyValue_StringValue:\n\t\treturn AttributeValueSTRING\n\tcase *otlpcommon.AnyValue_BoolValue:\n\t\treturn AttributeValueBOOL\n\tcase *otlpcommon.AnyValue_IntValue:\n\t\treturn AttributeValueINT\n\tcase *otlpcommon.AnyValue_DoubleValue:\n\t\treturn AttributeValueDOUBLE\n\tcase *otlpcommon.AnyValue_KvlistValue:\n\t\treturn AttributeValueMAP\n\tcase *otlpcommon.AnyValue_ArrayValue:\n\t\treturn AttributeValueARRAY\n\t}\n\treturn AttributeValueNULL\n}", "func (node *GoValueNode) GetType() (reflect.Type, error) {\n\n\treturn node.thisValue.Type(), nil\n}", "func (sv *SupernodesValue) Type() string {\n\treturn \"supernodes\"\n}", "func (b baseValue) Type() string {\n\treturn string(b.flagType)\n}", "func (this *ObjectUnwrap) Type() value.Type {\n\n\t// this is the succinct version of the above...\n\treturn this.Operand().Type()\n}", "func (this *Element) Type() value.Type { return value.JSON }", "func (v *Value) Type() *JSONType {\n\tt := C.zj_Type(v.V)\n\tif t == nil {\n\t\treturn nil\n\t}\n\tret := JSONType(*t)\n\treturn &ret\n}", "func (record Packed) Type() TagType {\n\trecordType, n := util.FromVarint64(record)\n\tif 0 == n {\n\t\treturn NullTag\n\t}\n\treturn TagType(recordType)\n}", "func (j *JSONData) Type() JSONType {\n\tif j == nil || j.value == nil { // no data\n\t\treturn JSONnil\n\t}\n\tvalue := *j.value\n\tif value == nil {\n\t\treturn JSONnull\n\t}\n\tif _, ok := value.(bool); ok {\n\t\treturn JSONboolean\n\t}\n\tif _, ok := value.(float64); ok {\n\t\treturn JSONnumber\n\t}\n\tif _, ok := value.(string); ok {\n\t\treturn JSONstring\n\t}\n\tif _, ok := value.([]interface{}); ok {\n\t\treturn JSONarray\n\t}\n\tif _, ok := value.(map[string]interface{}); ok {\n\t\treturn JSONobject\n\t}\n\tpanic(errors.New(\"JSONData corrupt\"))\n}", "func (obj GoObject) Type() Type {\n\treturn GetGoType(reflect.TypeOf(obj.val))\n}", "func (e *Wildcard) Type() Type {\n\treturn UnknownType\n}", "func (m *BinaryOperatorSub) Type() Type {\n\treturn IntType{}\n}", "func (v *Variant) Type() *VariantType {\n\t// The return value is valid for the lifetime of value and must not be freed.\n\treturn newVariantType(C.g_variant_get_type(v.native()))\n}", "func (s Schemas) Type() Type {\n\tif s.TF != nil {\n\t\tswitch s.TF.Type {\n\t\tcase schema.TypeBool:\n\t\t\treturn TypeBool\n\t\tcase schema.TypeInt, schema.TypeFloat:\n\t\t\treturn TypeNumber\n\t\tcase schema.TypeString:\n\t\t\treturn TypeString\n\t\tcase schema.TypeList, schema.TypeSet:\n\t\t\treturn s.ElemSchemas().Type().ListOf()\n\t\tcase schema.TypeMap:\n\t\t\treturn TypeMap\n\t\tdefault:\n\t\t\treturn TypeUnknown\n\t\t}\n\t}\n\n\treturn TypeUnknown\n}", "func (this *ObjectLength) Type() value.Type { return value.NUMBER }", "func (this *ObjectLength) Type() value.Type { return value.NUMBER }", "func (s *Smpval) Type() reflect.Type {\n\treturn s.val.Type()\n}", "func (v *VInteger) Type() string {\n\treturn \"integer\"\n}", "func (d *Decoder) Type() (Type, error) {\n\n\t// start with 1 byte and append to it until we get a clean varint\n\tvar (\n\t\ttag uint64\n\t\ttagBytes []byte\n\t)\n\nreadTagByte:\n\tfor {\n\t\tvar singleByte = make([]byte, 1)\n\t\t_, err := io.ReadFull(d.input, singleByte)\n\t\tif err != nil {\n\t\t\treturn typeUninited, err\n\t\t}\n\t\ttagBytes = append(tagBytes, singleByte[0])\n\n\t\tvar byteCount int\n\t\ttag, byteCount = varint.ConsumeVarint(tagBytes)\n\t\tswitch {\n\t\tcase byteCount == varint.ErrCodeTruncated:\n\t\t\tcontinue readTagByte\n\t\tcase byteCount > 0:\n\t\t\tfmt.Fprintln(dbg, \"\\tvarint byteCount:\", byteCount)\n\t\t\tbreak readTagByte // we got a varint!\n\t\tdefault:\n\t\t\treturn typeUninited, fmt.Errorf(\"bipf: broken varint tag field\")\n\t\t}\n\t}\n\n\tfmt.Fprintf(dbg, \"\\tdecoded %x to tag: %d\\n\", tagBytes, tag)\n\n\t// apply mask to get type\n\td.currentType = Type(tag & tagMask)\n\tif d.currentType >= TypeReserved {\n\t\treturn 0, fmt.Errorf(\"bipf: invalid type: %s\", d.currentType)\n\t}\n\n\t// shift right to get length\n\td.currentLen = uint64(tag >> tagSize)\n\n\t// drop some debugging info\n\tfmt.Fprintln(dbg, \"\\tvalue type:\", d.currentType)\n\tfmt.Fprintln(dbg, \"\\tvalue length:\", d.currentLen)\n\tfmt.Fprintln(dbg)\n\tdbg.Sync()\n\n\treturn d.currentType, nil\n}", "func (t *Type) Val() *Type", "func (def TypeDefinition) Type() Type {\n\treturn def.theType\n}", "func (a AttributeValue) Type() AttributeValueType {\n\treturn AttributeValueType(a.orig.Type)\n}", "func (l *NullLiteral) Type() Type {\n\treturn UnknownType\n}", "func (s *Struct) Type() Type {\n\treturn s.ty\n}", "func (m *EnumLiteral) Type() Type {\n\treturn m.wtype\n}", "func (e REnv) Type() Type { return e.Value().Type() }", "func (l *UnknownLiteral) Type() Type {\n\treturn UnknownType\n}", "func (m *PairLiteral) Type() Type {\n\treturn PairType{first: m.fst.Type(), second: m.snd.Type()}\n}", "func (p RProc) Type() Type { return p.Value().Type() }", "func (s *Swift) Type() Type {\n\tif hasBranchCode(s.value) {\n\t\treturn Type11\n\t}\n\treturn Type8\n}", "func (ts *TypeSet) Type(s string) Type {\n\tts.RLock()\n\tdefer ts.RUnlock()\n\treturn ts.types[s]\n}", "func (t *TypeValue) TypeName() string {\n\treturn t.name\n}", "func (t *jsonDataType) Type() interface{} {\n\treturn \"\"\n}", "func (inst *InstFAdd) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (s SetValue) Type(ctx context.Context) attr.Type {\n\treturn SetType{ElemType: s.ElementType(ctx)}\n}", "func (v Var) Type() (t Type, err error) {\n\tvar typ C.nc_type\n\terr = newError(C.nc_inq_vartype(C.int(v.ds), C.int(v.id), &typ))\n\tt = Type(typ)\n\treturn\n}", "func (this *Mod) Type() value.Type { return value.NUMBER }", "func (l *List) Type() Type {\n\treturn &ListType{\n\t\tElemType: l.valType,\n\t}\n}", "func Type() *dataType {\n\treturn &dataType{str: field.StringType()}\n}", "func (s *StringPointerValue) Type() string {\n\treturn \"string\"\n}", "func (o *PairAnyValueAnyValue) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (b *BaseType) Type() string {\n\treturn b.baseType\n}", "func (t FieldType) Type() FieldType {\n\treturn t\n}", "func (this *NowStr) Type() value.Type { return value.STRING }", "func (Integer) Type() types.Type {\n\treturn types.Number\n}", "func (inst *InstAdd) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (myOperatingSystemType *OperatingSystemType) Type() (param string) {\n\treturn myOperatingSystemType.Typevar\n}", "func (u UseSRTP) TypeValue() TypeValue {\n\treturn UseSRTPTypeValue\n}", "func (rv *ReturnValue) Type() ObjectType {\n\treturn RETURN_VALUE\n}", "func (n *piName) Type() Type {\n\treturn n.t\n}", "func (rv *ReturnValue) Type() ObjectType { return RETURN_VALUE_OBJ }", "func (o JavaScriptFunctionBindingOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v JavaScriptFunctionBinding) string { return v.Type }).(pulumi.StringOutput)\n}", "func (s Series) Type() Type {\n\treturn s.t\n}", "func (n NodeType) Type() NodeType {\n\treturn n\n}", "func (t *IntDataType) Type() interface{} {\n\treturn 0\n}", "func (inst *InstFMul) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (tv *TypedLeafListUint) ValueType() ValueType {\n\treturn tv.Type\n}", "func (v *ClassValue) Type() semantic.Type {\n\treturn v.Class\n}", "func (t *StringDataType) Type() interface{} {\n\treturn \"\"\n}", "func (inst *InstSDiv) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (i *Integer) Type() ObjectType { return IntegerObject }", "func (this *ObjectValues) Type() value.Type { return value.ARRAY }", "func (this *ObjectValues) Type() value.Type { return value.ARRAY }", "func (c ValueColumn) Type() string {\n\treturn c.ctype\n}", "func (o CustomClrSerializationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CustomClrSerialization) string { return v.Type }).(pulumi.StringOutput)\n}", "func (i *Integer) Type() ObjectType {\n\treturn INTEGER\n}", "func (r *VarRef) Type() Type {\n\treturn r.ExprType\n}", "func (g *Grammar) Type(tokenType int) (out Term) {\n\treturn TypedTerm(tokenType)\n}", "func (o FieldOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Field) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (l *BooleanLiteral) Type() Type {\n\treturn Boolean\n}", "func (tv *TypedLeafListBytes) ValueType() ValueType {\n\treturn tv.Type\n}", "func (t Type) Type() string {\n\treturn t.typeName\n}", "func (inst *InstURem) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (inst *InstSRem) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (this *ObjectInnerValues) Type() value.Type { return value.ARRAY }", "func (inst *InstFRem) Type() types.Type {\n\t// Cache type if not present.\n\tif inst.Typ == nil {\n\t\tinst.Typ = inst.X.Type()\n\t}\n\treturn inst.Typ\n}", "func (element *Element) Type(value string) *Element {\n\treturn element.Attr(\"type\", value)\n}", "func TypeOf(data []byte) Type {\n\tt := data[0]\n\n\t// FIXME: add additional validation\n\n\tswitch {\n\tdefault:\n\t\treturn invalid\n\tcase t == 'i':\n\t\treturn integer\n\tcase t >= '0' && t <= '9':\n\t\treturn str\n\tcase t == 'l':\n\t\treturn list\n\tcase t == 'd':\n\t\treturn dictionary\n\t}\n}", "func (l *UUIDLiteral) Type() Type {\n\treturn UUID\n}", "func (t Token) Type() TokenType {\n\tif len(t) == 0 {\n\t\treturn Empty\n\t}\n\ttyp := Empty\n\tfor _, r := range t {\n\t\ttmp := runeType(r)\n\t\tif typ != Empty && typ != tmp {\n\t\t\treturn Mixed\n\t\t}\n\t\ttyp = tmp\n\t}\n\treturn typ\n}", "func (v Struct) Type() Type {\n\tt := make(typeStruct, 0, len(v))\n\tfor _, field := range v {\n\t\tt = append(t, typeStructField{Name: field.Name, Type: field.Value.Type()})\n\t}\n\treturn t\n}", "func (o ScalarFunctionPropertiesOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ScalarFunctionProperties) string { return v.Type }).(pulumi.StringOutput)\n}" ]
[ "0.755581", "0.7475477", "0.74640226", "0.74598986", "0.7345107", "0.7240795", "0.7237268", "0.71322525", "0.70109874", "0.6932564", "0.6901902", "0.68951213", "0.68916506", "0.68422705", "0.67929244", "0.67622596", "0.6681911", "0.6673024", "0.6662416", "0.6656612", "0.6624057", "0.6623154", "0.6607991", "0.65920126", "0.6565786", "0.6532572", "0.65209764", "0.651465", "0.65109885", "0.6507539", "0.6499952", "0.6477376", "0.64719903", "0.64719903", "0.6432472", "0.6419071", "0.64139426", "0.6405478", "0.6383998", "0.63761777", "0.6363199", "0.6274093", "0.62612826", "0.62515354", "0.6240828", "0.62323135", "0.6221681", "0.62113494", "0.6199548", "0.618952", "0.6189396", "0.6174667", "0.61568487", "0.61533993", "0.61502624", "0.6146661", "0.61429155", "0.61280173", "0.61107296", "0.6108771", "0.61051077", "0.60961777", "0.6095813", "0.60926217", "0.60914415", "0.6081123", "0.60673606", "0.60647655", "0.6064445", "0.60597", "0.6057894", "0.60498965", "0.60489756", "0.6043783", "0.6036228", "0.602673", "0.6023577", "0.6020027", "0.60115856", "0.60115653", "0.60115653", "0.6007855", "0.6005541", "0.60032046", "0.5998557", "0.5998079", "0.59927315", "0.5992368", "0.5992336", "0.59845793", "0.5983643", "0.5981723", "0.59775484", "0.59686846", "0.5960101", "0.59569174", "0.59537166", "0.5948143", "0.5944768", "0.5942745" ]
0.61994
49