query
stringlengths
9
9.05k
document
stringlengths
10
222k
metadata
dict
negatives
listlengths
30
30
negative_scores
listlengths
30
30
document_score
stringlengths
4
10
document_rank
stringclasses
2 values
Implementing set start date functionality
def set_start_date(self, start_date): self.set_value_into_input_field(self.start_date_inputbox_locator, start_date)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_start_date(self, date):\n pass", "def start_date(self, start_date):\n self._start_date = start_date", "def start_date(self, start_date):\n \n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def start_date(self, start_date):\n\n self._start_date = start_date", "def date_start(self, date_start):\n\n self._date_start = date_start", "def set_begin_date(self, begin_date):\n self.set_value_into_input_field(self.begin_date_inputbox_locator, begin_date)", "def start_date(self, start_date):\n if start_date is None:\n raise ValueError(\"Invalid value for `start_date`, must not be `None`\")\n\n self._start_date = start_date", "def start_date(self, start_date):\n if start_date is None:\n start_date = datetime.now() - timedelta(days=365)\n\n self._start_date = dt_utils.parse_date(start_date)", "def model_start_date(self, model_start_date):\n\n self._model_start_date = model_start_date", "def onchange_start_date(self, start_date=False):\n if not start_date:\n return {}\n result = {'value': {'last_renovation_date': start_date}}\n return result", "def set_statement_begin_date(self, begin_date):\n begin_date_to_set = None\n if begin_date != \"\":\n begin_date_to_set = begin_date\n else:\n self.first_day_of_previous_month = self.get_date(first_day_of_last_month=True)\n begin_date_to_set = self.first_day_of_previous_month\n self.set_value_into_input_field(self.statement_begin_date_locator, begin_date_to_set)\n return begin_date_to_set", "def set_begin_date_for_search(self, begin_date):\n self.set_value_into_input_field(self.begin_date_locator, begin_date)", "def _get_start_date(self):\n today = datetimedate.today()\n if self.start == 'week':\n start_date = today - timedelta(days=today.weekday())\n elif self.start == 'month':\n start_date = today.replace(day=1)\n elif self.start == 'quarter':\n quarter = math.ceil(today.month / 3)\n start_date = datetimedate(\n today.year,\n ((quarter - 1) * 3) + 1,\n 1\n )\n elif self.start == 'year':\n start_date = datetimedate(today.year, 1, 1)\n elif self.start == 'all':\n start_date = datetimedate(2010, 1, 1)\n else:\n try:\n start_date = datetime.strptime(self.start, \"%Y-%m-%d\").date()\n except Exception as e:\n raise ParseError(\"start argument not valid\")\n\n self.start_date = start_date", "def begin_date(self, value):\n\n if not isinstance(value, datetime):\n raise TypeError(_pretty_message(\n '''\n begin_date must be an instance of datetime.datetime, not %s\n ''',\n _type_name(value)\n ))\n\n self._begin_date = value", "def setDate(self, p_int, p_int_1, p_int_2): # real signature unknown; restored from __doc__\r\n return False", "def test_build__set_start_date(self, valid_service: fixture) -> None:\n service: Service = valid_service\n\n assert date.isoformat(date.today() - timedelta(days=1)) in service.start_date", "def set_billing_cycle_begin_date(self, begin_date):\n if begin_date == \"\":\n current_date = datetime.date.today()\n begin_date = current_date.replace(day=1)\n begin_date = begin_date.strftime(\"%m/%d/%Y\")\n self.set_value_into_input_field(self.billing_cycle_begin_date_inputbox_locator, begin_date)", "def set_bulk_add_begin_date(self, begin_date):\n if begin_date == \"\":\n begin_date = self.get_date(current_date=True)\n self.set_value_into_input_field(self.bulk_add_begin_date_inputbox_locator, begin_date)\n global bulk_add_begin_date\n bulk_add_begin_date = datetime.datetime.strptime(begin_date, \"%m/%d/%Y\")\n return begin_date", "def __init__(self, start: datetime.date) -> None:\n self.start = start\n self.bill = None", "def start(self, start):\n try:\n self._set('start', Timestamp.to_datetime(start))\n except (TypeError, ValueError) as e:\n raise ValidationError(e)", "def set_charge_begin_date(self, charge_begin_date, first_day_of_last_month):\n self.current_charge_begin_date = charge_begin_date\n if first_day_of_last_month is True:\n if charge_begin_date == \"\":\n self.first_day_of_previous_month = self.get_date(first_day_of_last_month=True)\n charge_begin_date = self.first_day_of_previous_month\n else:\n if charge_begin_date == \"\":\n charge_begin_date = self.get_date(current_date=True)\n self.current_charge_begin_date = datetime.datetime.strptime(charge_begin_date, \"%m/%d/%Y\")\n self.set_value_in_grid_column(self.charges_grid_div_id, self.charge_begin_date_column_name, charge_begin_date, True)\n page_header_element = self.wait().until(EC.element_to_be_clickable(self.page_header_locator), 'page header locator not found before specified time out')\n page_header_element.click()", "def start_date(self):\n return self.__start_date" ]
[ "0.86141276", "0.80457026", "0.7868022", "0.7836363", "0.7836363", "0.7836363", "0.7836363", "0.7836363", "0.7836363", "0.7836363", "0.7836363", "0.7836363", "0.7739041", "0.749863", "0.74370885", "0.7398391", "0.729659", "0.7160086", "0.70503116", "0.70105284", "0.69417137", "0.6938817", "0.68402106", "0.6837066", "0.68364817", "0.682576", "0.68250114", "0.6812464", "0.6794182", "0.6727007" ]
0.80691314
1
Implementing set buy and sell management grid settings functionality
def set_buy_and_sell_management_grid_settings(self, grid_settings): self.set_grid_settings(self.buy_and_sell_management_grid_div_id, grid_settings)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_vendor_price_lists_grid_settings(self, grid_settings):\n self.set_grid_settings(self.view_price_list_div_id, grid_settings)", "def set_vendor_price_list_detail_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_rates_grid_div_id, grid_settings)", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def set_values(self):\n super(ResConfigInherit, self).set_values()\n self.env['ir.config_parameter'].sudo().set_param(\n 'sale_stock_restrict.product_restriction', self.product_restriction)\n self.env['ir.config_parameter'].sudo().set_param(\n 'sale_stock_restrict.check_stock', self.check_stock)", "def set_vendor_grid_settings(self, grid_settings):\n self.set_grid_settings(self.vendors_div_id, grid_settings)", "async def economyset_showsettings(self, ctx: commands.Context):\r\n guild = ctx.guild\r\n if await bank.is_global():\r\n conf = self.config\r\n else:\r\n conf = self.config.guild(guild)\r\n await ctx.send(\r\n box(\r\n _(\r\n \"----Economy Settings---\\n\"\r\n \"Minimum slot bid: {slot_min}\\n\"\r\n \"Maximum slot bid: {slot_max}\\n\"\r\n \"Slot cooldown: {slot_time}\\n\"\r\n \"Payday amount: {payday_amount}\\n\"\r\n \"Payday cooldown: {payday_time}\\n\"\r\n \"Amount given at account registration: {register_amount}\\n\"\r\n \"Maximum allowed balance: {maximum_bal}\"\r\n ).format(\r\n slot_min=humanize_number(await conf.SLOT_MIN()),\r\n slot_max=humanize_number(await conf.SLOT_MAX()),\r\n slot_time=humanize_number(await conf.SLOT_TIME()),\r\n payday_time=humanize_number(await conf.PAYDAY_TIME()),\r\n payday_amount=humanize_number(await conf.PAYDAY_CREDITS()),\r\n register_amount=humanize_number(await bank.get_default_balance(guild)),\r\n maximum_bal=humanize_number(await bank.get_max_balance(guild)),\r\n )\r\n )\r\n )", "def _configure(self):\n Values._configure(self)\n self.values = [self.inventory.one, self.inventory.two]\n return", "def set_vendor_price_list_detail_reference_rates_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_reference_rates_grid_div_id, grid_settings)", "def prepare_for_metgrid(self, wps_nml):\n wps_nml['share']['active_grid'] = [True] * len(self.domains)", "def _setup_transactions(\n biz_org,\n biz_admin,\n transaction_currents_amount,\n transaction_price_reported,\n price_actual=None,\n pop_type='rec',\n offer_item_name=\"Test Item\",\n currents_share=40,\n action_type='req'\n):\n offer_item = Item(name=offer_item_name)\n offer_item.save()\n\n offer = Offer(\n org=biz_org,\n item=offer_item,\n currents_share=currents_share\n )\n offer.save()\n\n if price_actual is None:\n price_actual = transaction_price_reported\n\n transaction = Transaction(\n user=biz_admin,\n offer=offer,\n price_reported=transaction_price_reported,\n currents_amount=transaction_currents_amount,\n price_actual=price_actual\n )\n transaction.save()\n\n action = TransactionAction(\n transaction=transaction,\n action_type=action_type\n )\n action.save()", "def initial_values(self):\n\n coin = self.mw.cfg_manager.coin\n pair = self.mw.cfg_manager.pair\n\n self.mw.buy_asset.setText(coin)\n self.mw.sell_asset.setText(coin)\n\n self.mw.limit_total_btc.setText(str(val[\"accHoldings\"][\"BTC\"][\"free\"]) + \" BTC\")\n self.mw.limit_total_coin.setText(str(val[\"accHoldings\"][coin][\"free\"]) + \" \" + coin)\n\n self.mw.limit_buy_label.setText(\"<span style='font-weight: bold; font-size: 12px;'>Buy \" + coin + \"</span>\")\n self.mw.limit_sell_label.setText(\"<span style='font-weight: bold; font-size: 12px;'>Sell \" + coin + \"</span>\")\n\n # self.mw.limit_coin_label_buy.setText(\"<span style='font-weight: bold; color: white;'>\" + coin + \"</span>\")\n # self.mw.limit_coin_label_sell.setText(\"<span style='font-weight: bold; color: white;'>\" + coin + \"</span>\")\n\n # self.mw.limit_buy_input.setText(\"kernoschmaus\")\n self.mw.limit_buy_input.setDecimals(val[\"decimals\"])\n self.mw.limit_buy_input.setSingleStep(float(val[\"coins\"][pair][\"tickSize\"]))\n\n self.mw.limit_sell_input.setDecimals(val[\"decimals\"])\n self.mw.limit_sell_input.setSingleStep(float(val[\"coins\"][pair][\"tickSize\"]))\n\n self.mw.limit_buy_amount.setDecimals(val[\"assetDecimals\"])\n self.mw.limit_buy_amount.setSingleStep(float(val[\"coins\"][pair][\"minTrade\"]))\n\n self.mw.limit_sell_amount.setDecimals(val[\"assetDecimals\"])\n self.mw.limit_sell_amount.setSingleStep(float(val[\"coins\"][pair][\"minTrade\"]))", "def set_buy_sell_deal_account(self, account_list):\n self.multiple_items_selection_from_kendo_dropdown(self.buy_sell_deal_account_dropdown_locator, account_list)\n self.wait_for_ajax_spinner_load()", "def widgetstosettings(self):\n print \"in widgets to settings\"\n self.ABsettings[\"intensity_range\"]=(self.spansliderInt.lowerValue,self.spansliderInt.upperValue)\n self.ABsettings[\"rgb\"]=self.colorBox.getRGB\n self.ABsettings[\"visible\"]=self.abEnabledCB.isChecked()\n self.ABsettings[\"zrange\"]=(self.spansliderZ.lowerValue,self.spansliderZ.upperValue)\n self.ABsettings[\"Antibody\"]=self.ab\n for button in self.radiobuttons:\n if button.isChecked():\n self.ABsettings[\"selected_DAPI_channel\"]=str(button.objectName())[:-3]\n print \"Dapi channel setting is \", self.ABsettings[\"selected_DAPI_channel\"]", "def place_orders(context, data):\r\n log.info(\"*********Monthly flags: %s\" % context.flags)\r\n \r\n context.sell = []\r\n context.buy = []\r\n \r\n # Go through flags to determine buy/sell signals\r\n for asset, flags in context.flags.items():\r\n # If up > down and multiple blue flags, add to buy\r\n if flags['UP'] > flags['DOWN'] and flags['UP'] > 1:\r\n context.buy.append(asset)\r\n \r\n # If down > up and multiple down flags, add to sell\r\n elif flags['DOWN'] > flags['UP'] and flags['DOWN'] > 1:\r\n context.sell.append(asset)\r\n \r\n # If both SPY and QQQ are buys, rebalance weightings and check components\r\n if sid(8554) in context.buy and sid(19920) in context.buy:\r\n rebalance_weightings(context)\r\n \r\n # Reset down sequence\r\n context.first_down_sequence = set()\r\n \r\n # Reset SPY and QQQ to max weightings\r\n context.target_weights[sid(8554)] = context.max_weights[sid(8554)]\r\n context.target_weights[sid(19920)] = context.max_weights[sid(19920)]\r\n \r\n # Convert weights to number of shares \r\n context.target_shares[sid(8554)] = round(context.target_weights[sid(8554)] * context.portfolio.portfolio_value / context.price[sid(8554)])\r\n context.target_shares[sid(19920)] = round(context.target_weights[sid(19920)] * context.portfolio.portfolio_value / context.price[sid(19920)])\r\n \r\n # If not overweighting:\r\n if not context.overweighting:\r\n context.buy.remove(sid(8554))\r\n context.buy.remove(sid(19920))\r\n \r\n # Check components\r\n for asset, ratio in context.up_ratios.items():\r\n # If UP ratio > 1, add to buy\r\n if asset != sid(8554) and asset != sid(19920) and ratio > 1:\r\n context.buy.append(asset)\r\n \r\n # If SPY is a sell, check UP ratios for components\r\n if sid(8554) in context.sell:\r\n for asset, ratio in context.up_ratios.items():\r\n # If UP ratio < 1, add to sell\r\n if asset != sid(8554) and asset != sid(19920) and ratio < 1:\r\n context.sell.append(asset)\r\n \r\n \r\n \r\n # First month at end August 2017: set all other assets to max weighting, except take UP ratio of JKL to be <1 so sell 20% of weighting\r\n if context.first_iteration:\r\n log.info('First iteration')\r\n \r\n # Initialise weightings\r\n rebalance_weightings(context)\r\n context.first_iteration = False\r\n \r\n for asset, weight in context.max_weights.items(): \r\n # JKL\r\n if asset == sid(26451):\r\n context.sell.append(asset)\r\n\r\n context.target_weights[asset] = weight\r\n \r\n # Convert weights to number of shares \r\n context.target_shares[asset] = round(context.target_weights[asset] * context.portfolio.portfolio_value / context.price[asset])\r\n \r\n buy_overweight = []\r\n remaining_cash = context.portfolio.cash\r\n \r\n # Buy components first (before considering overweighting QQQ/SPY)\r\n for asset in sorted(context.buy, reverse=True):\r\n \r\n # This is an up sequence so no subsequent down sequence\r\n if asset in context.first_down_sequence:\r\n context.first_down_sequence.remove(asset) \r\n \r\n # Buy 50% of weighting\r\n log.info('UP flags for %s: Buy 50 percent' % asset)\r\n extra_weight = 0.5 * context.max_weights[asset]\r\n \r\n # Do not exceed max shares by weighting, UNLESS taking from cash from components (overweighting)\r\n if context.target_weights[asset] == context.max_weights[asset] or (context.target_weights[asset] > context.max_weights[asset] and context.overweighting):\r\n buy_overweight.append(asset)\r\n \r\n elif context.target_weights[asset] + extra_weight > context.max_weights[asset]:\r\n context.target_weights[asset] = context.max_weights[asset]\r\n \r\n else:\r\n context.target_weights[asset] += extra_weight\r\n \r\n # Convert weights to number of shares\r\n old_shares = context.target_shares[asset]\r\n context.target_shares[asset] = round(context.target_weights[asset] * context.portfolio.portfolio_value / context.price[asset])\r\n remaining_cash -= (context.target_shares[asset] - old_shares) * context.price[asset]\r\n \r\n for asset in buy_overweight:\r\n if remaining_cash > 0:\r\n # If first overweight or 2 assets to be overweighted, take 50% of available cash\r\n if context.target_weights[asset] > context.max_weights[asset] or len(buy_overweight) > 1:\r\n log.info('Taking half of cash of value: %f' % (remaining_cash * 0.5))\r\n context.target_weights[asset] += 0.5 * remaining_cash / context.portfolio.portfolio_value\r\n \r\n # If second overweight, take all remaining cash\r\n else:\r\n log.info('Taking remaining of cash of value: %f' % (remaining_cash))\r\n context.target_weights[asset] += remaining_cash / context.portfolio.portfolio_value\r\n \r\n else:\r\n # If no cash, ignore\r\n log.info('UP flags for %s: No change' % asset)\r\n continue\r\n \r\n \r\n # For assets in sell list\r\n for asset in context.sell:\r\n \r\n # If asset already has 0 holdings, ignore\r\n if context.target_weights[asset] == 0:\r\n log.info('DOWN flags for %s: No change' % asset)\r\n continue\r\n \r\n # If first multiple down flags, sell 20% of UP weight\r\n elif asset not in context.first_down_sequence:\r\n log.info('First DOWN flags for %s: Sell 20 percent' % asset)\r\n context.target_weights[asset] -= 0.2 * context.max_weights[asset]\r\n context.first_down_sequence.add(asset)\r\n \r\n # If this is a subsequent down flag sequence, sell 40% of UP weight\r\n else:\r\n log.info('DOWN flags for %s: Sell 40 percent' % asset)\r\n context.target_weights[asset] -= 0.4 * context.max_weights[asset]\r\n \r\n # Ensure no short position\r\n if context.target_weights[asset] < 0:\r\n context.target_weights[asset] = 0\r\n \r\n # Convert weights to number of shares \r\n context.target_shares[asset] = round(context.target_weights[asset] * context.portfolio.portfolio_value / context.price[asset])\r\n \r\n print(context.target_weights)", "def change_entries(self, grid_sizer, options, border):\n\n if (self.diag_type is None): return grid_sizer # no data selected, don't change anything\n\n # setup the grid of possible values\n header0 = wx.StaticText(self, -1, \"Add/Remove\")\n header1 = wx.StaticText(self, -1, \"Quantity Code\")\n header2 = wx.StaticText(self, -1, \"Name\")\n header3 = wx.StaticText(self, -1, \"LaTeX Formula\")\n grid_sizer.Add(header0, pos=(0,0), flag=options, border=border)\n grid_sizer.Add(header1, pos=(0,1), flag=options, border=border)\n grid_sizer.Add(header2, pos=(0,2), flag=options, border=border)\n grid_sizer.Add(header3, pos=(0,3), flag=options, border=border)\n grid_sizer.Add(wx.StaticLine(self), pos=(1,0), span=(1,4),\n flag=wx.ALL|wx.EXPAND|wx.GROW, border=border)\n\n self.selected_values = [] # keep track of selected quantities\n\n quantities = self.output_quantities.diagnostic_types[self.diag_type]\n\n # choose a good height/width for formulas\n if (self.diag_type in [\"Linear_Forces\", \"Angular_Momentum\", \"Energy_Flux\",\n \"Induction\", \"Inertial_Forces\", \"Lorentz_Forces\",\n \"Poynting_Flux\", \"TurbKE_Budget\"]):\n width = 100\n elif (self.diag_type == \"Thermal_Equation\"):\n width = 150\n else:\n width = 30\n height = 20\n\n row = 2\n iquant = 0\n for Q in quantities:\n but = wx.ToggleButton(self, Q.code, \"Add\") # build button and place it in second column\n but.Bind(wx.EVT_TOGGLEBUTTON, self.OnToggle)\n grid_sizer.Add(but, pos=(row,0), flag=options, border=border)\n\n q_code = wx.StaticText(self, -1, str(Q.code)) # build other column entries\n q_name = wx.StaticText(self, -1, Q.name) # name\n\n formula = self.RenderTeX(Q, size=(width,height))\n\n # place column entries\n grid_sizer.Add(q_code, pos=(row,1), flag=options, border=border)\n grid_sizer.Add(q_name, pos=(row,2), flag=options, border=border)\n grid_sizer.Add(formula, pos=(row,3), flag=options, border=border)\n\n iquant += 1\n\n # add horizontal line every 5 quantities\n if (iquant % 5 == 0):\n grid_sizer.Add(wx.StaticLine(self), pos=(row+1,0), span=(1,4),\n flag=wx.ALL|wx.EXPAND|wx.GROW, border=border)\n row_inc = 2\n else:\n row_inc = 1\n\n row += row_inc\n\n grid_sizer.AddGrowableCol(2,1) # make the name/formula columns \"1\" growable, i.e., grows as necessary\n grid_sizer.AddGrowableCol(3,1)\n\n return grid_sizer", "def set_values(self):\n super(ResConfigSettings, self).set_values()\n self.env['ir.config_parameter'].sudo().set_param(\n 'pos_all_orders.pos_all_order', self.pos_all_order)\n self.env['ir.config_parameter'].sudo().set_param(\n 'pos_all_orders.n_days',\n self.n_days)", "def settingstowidgets(self):\n\n # disconnect before updating, otherwise\n # the current GUI settings will be reinstated\n # after the first GUI element is updated\n self.disconnect_all_widgets()\n\n self.spansliderInt.setLowerValue(int(self.ABsettings[\"intensity_range\"][0]))\n self.spansliderInt.setUpperValue(int(self.ABsettings[\"intensity_range\"][1]))\n print \"vis setting \",self.ABsettings[\"visible\"]\n if self.ABsettings[\"visible\"]:\n print \"setting \",self.objectName(), \" to visible\"\n self.abEnabledCB.setChecked(True)\n else:\n print \"setting \",self.objectName(), \" to invisible\"\n self.abEnabledCB.setChecked(False)\n self.spansliderZ.setLowerValue(int(self.ABsettings[\"zrange\"][0]))\n self.spansliderZ.setUpperValue(int(self.ABsettings[\"zrange\"][1]))\n #self.ABsettings[\"Antibody\"]=self.ab\n self.colorBox.setRGB(self.ABsettings[\"rgb\"])\n if self.isDAPIPanel:\n for rb in self.radiobuttons:\n print \"radio button \", str(rb.objectName())\n if str(rb.objectName()).split(\"_\")[0]==self.ABsettings[\"selected_DAPI_channel\"]:\n rb.setChecked(True)\n print \"is checked\"\n\n # reconnect everything\n self.connect_all_widgets()\n self.updateSettings()", "def set_vendor_price_list_detail_dial_digits_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_dial_digits_grid_div_id, grid_settings)", "def sell(self):\n self.status = \"sold\"\n return self", "def enableAddOrder(self):\n if self.lstType.selected == -1 or self.lstRes.selected == -1:\n self.btnAddMarketOrder.disable()\n else:\n self.btnAddMarketOrder.enable()\n if self.lstType.getSelectedItem().data == 'sell':\n self.txtMin.readOnly = 0\n else:\n self.txtMax.readOnly = 0", "def set_config(self): # called from button_set object \n self.settings['lights_on'] = self.lights_on.get()\n self.settings['lights_off'] = self.lights_off.get()\n self.settings['ambient_min'] = self.ambient_min.get()\n self.settings['soil_1'] = self.smc1.get()\n self.settings['soil_2'] = self.smc2.get()\n self.settings['soil_3'] = self.smc3.get()\n self.settings['soil_4'] = self.smc4.get()\n self.settings['overhead_level'] = self.overhead_level.get()\n\n # Save settings to config file in case of reboot / power-loss\n print \"UPDATING SETTINGS FILE\"\n with open(self.settings_path, 'w') as jsonfile:\n jsonfile.write(json.dumps(self.settings, indent=4))\n self.active_changes = True # (flag) changes are active!", "def drag_buy_and_sell_management_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.buy_and_sell_management_grid_div_id, source_column, destination_column)", "def create_settings_menu(master: Widget) -> None:\r\n\r\n def create_speed_widgets(master: Widget) -> None:\r\n \"\"\"Create speed widgets with the parent MASTER.\"\"\"\r\n\r\n speed_label = Label(master, text='Speed:', font=self.FONT_NORMAL, bg=self.MAIN_BG)\r\n speed_label.grid(row=0, column=0, sticky=W, padx=self.WIDGET_PAD, \r\n pady=(self.WIDGET_PAD,0))\r\n\r\n self.speed_scale = Scale(\r\n master, from_=0, to=len(self.TIMES_PER_GEN)-1, resolution=1, orient=HORIZONTAL,\r\n bg=self.MAIN_BG, font=self.FONT_SMALL, command=self.on_speed_change)\r\n self.speed_scale.set(self.INITIAL_TIME_PER_GEN)\r\n self.speed_scale.grid(row=0, column=1, sticky=W+E, padx=(0,self.WIDGET_PAD),\r\n pady=(self.WIDGET_PAD,0))\r\n\r\n def create_zoom_widgets(master: Widget) -> None:\r\n \"\"\"Create zoom widgets with the parent MASTER.\"\"\"\r\n\r\n zoom_label = Label(master, text='Zoom:', font=self.FONT_NORMAL, bg=self.MAIN_BG)\r\n zoom_label.grid(row=1, column=0, sticky=W, padx=self.WIDGET_PAD,\r\n pady=(0,self.WIDGET_PAD*2))\r\n\r\n self.zoom_scale = Scale(\r\n master, from_=0, to=len(self.CELL_SIZES)-1, resolution=1, orient=HORIZONTAL,\r\n bg=self.MAIN_BG, font=self.FONT_SMALL, command=self.on_zoom_change)\r\n self.zoom_scale.set(self.INITIAL_ZOOM)\r\n self.zoom_scale.grid(row=1, column=1 ,sticky=W+E, padx=(0,self.WIDGET_PAD),\r\n pady=(0,self.WIDGET_PAD*2))\r\n\r\n def create_rule_widgets(master: Widget) -> None:\r\n \"\"\"Create rule widgets with the parent MASTER.\"\"\"\r\n\r\n rule_label = Label(master, text='Rule:', font=self.FONT_NORMAL, bg=self.MAIN_BG)\r\n rule_label.grid(row=2, column=0, sticky=W, padx=self.WIDGET_PAD,\r\n pady=(0,self.WIDGET_PAD))\r\n\r\n self.rule_entry = Entry(master, font=self.FONT_NORMAL)\r\n self.rule_entry.grid(row=2, column=1, sticky=W+E, padx=(0,self.WIDGET_PAD),\r\n pady=(0,self.WIDGET_PAD))\r\n self.rule_entry.insert(0, self.INITIAL_RULE)\r\n\r\n rule_button = Button(master, text='Set Rule', font=self.FONT_NORMAL, bg=self.MAIN_BG,\r\n command=self.on_set_rule)\r\n rule_button.grid(row=3, column=1, sticky=E, padx=(0,self.WIDGET_PAD),\r\n pady=(0,self.WIDGET_PAD))\r\n\r\n self.settings_menu = LabelFrame(master,text='Settings', font=self.FONT_SMALL,\r\n bg=self.MAIN_BG)\r\n self.settings_menu.pack(side=TOP, pady=self.WIDGET_PAD)\r\n\r\n create_speed_widgets(self.settings_menu)\r\n create_zoom_widgets(self.settings_menu)\r\n create_rule_widgets(self.settings_menu)", "def set_controls(self):\n # Image control\n image = pyxbmct.Image(addonfolder+artsfolder+'/k.png')\n self.placeControl(image, 0, 0, rowspan=8, columnspan=16)\n\n\t\t# KI Plus\n self.k1plus_button = pyxbmct.RadioButton('')\n self.placeControl(self.k1plus_button, 8, 1, rowspan=2, columnspan=4)\n self.connect(self.k1plus_button, self.k1plus_button_update)\n if tools.return_data('TVHWIZARD', 'STRING', 'k1plus', 2) == 1:\n self.k1plus_button.setSelected(True)\n else:\n self.k1plus_button.setSelected(False)\n k1plus = pyxbmct.Image(addonfolder+artsfolder+'/k1plus.png')\n self.placeControl(k1plus, 8, 1, rowspan=2, columnspan=4)\n\n\t\t# KI Pro\n self.k1pro_button = pyxbmct.RadioButton('')\n self.placeControl(self.k1pro_button, 11, 6, rowspan=2, columnspan=4)\n self.connect(self.k1pro_button, self.k1pro_button_update)\n if tools.return_data('TVHWIZARD', 'STRING', 'k1pro', 2) == 1:\n self.k1pro_button.setSelected(True)\n else:\n self.k1pro_button.setSelected(False)\n k1pro = pyxbmct.Image(addonfolder+artsfolder+'/k1pro.png')\n self.placeControl(k1pro, 11, 6, rowspan=2, columnspan=4)\n\n\t\t# KII Pro\n self.k2pro_button = pyxbmct.RadioButton('')\n self.placeControl(self.k2pro_button, 8, 6, rowspan=2, columnspan=4)\n self.connect(self.k2pro_button, self.k2pro_button_update)\n if tools.return_data('TVHWIZARD', 'STRING', 'k2pro', 2) == 1:\n self.k2pro_button.setSelected(True)\n else:\n self.k2pro_button.setSelected(False)\n k2pro = pyxbmct.Image(addonfolder+artsfolder+'/k2pro.png')\n self.placeControl(k2pro, 8, 6, rowspan=2, columnspan=4)\n\n\t\t# KIII Pro\n self.k3pro_button = pyxbmct.RadioButton('')\n self.placeControl(self.k3pro_button, 8, 11, rowspan=2, columnspan=4)\n self.connect(self.k3pro_button, self.k3pro_button_update)\n if tools.return_data('TVHWIZARD', 'STRING', 'k3pro', 2) == 1:\n self.k3pro_button.setSelected(True)\n else:\n self.k3pro_button.setSelected(False)\n k3pro = pyxbmct.Image(addonfolder+artsfolder+'/k3pro.png')\n self.placeControl(k3pro, 8, 11, rowspan=2, columnspan=4)\n\n\t\t# Close button\n self.close_button = pyxbmct.Button('Exit')\n self.placeControl(self.close_button, 13, 15, rowspan=1, columnspan=1)\n self.connect(self.close_button, lambda: self.closepage())", "def use_config_manage_stock(self):\n return self._use_config_manage_stock", "def assignGridInfo(self):\n\n for point in self.gridPoints:\n for house in self.houses:\n if (point.xLocation == house.xLocation and\n point.yLocation == house.yLocation):\n house.gridID = point.ID\n point.cost = [5000, 5000, 5000, 5000, 5000]\n for battery in self.batteries:\n if (point.xLocation == battery.xLocation and\n point.yLocation == battery.yLocation):\n battery.gridID = point.ID\n return True", "def customise_grid(self):\n\n # Agent colours, agent tooltips and grid initialisation\n mapper = linear_cmap(field_name='category',\n palette=['blue', 'red', 'green', 'orange', 'purple'] ,\n low=0,\n high=4)\n TOOLTIPS = [(\"Residential utility\", \"@res_utility\"),\n ('Local composition', '@local_comp'),\n ('Neighbourhood composition', '@n_comp'),\n (\"School utility\", \"@school_utility\"),\n ('Distance', '@dist_school'),\n ('School composition', '@s_comp'),\n ('School composition utility', '@school_comp_utility')]\n hover = HoverTool(names=[\"households\", \"schools\"], tooltips=TOOLTIPS)\n self.grid = figure(x_range=(self.model.grid.x_min - 1,\n self.model.grid.x_max),\n y_range=(self.model.grid.y_min - 1,\n self.model.grid.y_max),\n tools=[hover, 'tap', 'pan',\n WheelZoomTool()],\n tooltips=TOOLTIPS, output_backend=\"webgl\")\n\n # Set WheelZoomTool active by default if not lattice\n if self.params['case'].lower() != 'lattice':\n self.grid.toolbar.active_scroll = self.grid.select_one(\n WheelZoomTool)\n\n # Add a background map using OpenStreetMap (Google Maps is too\n # computationally expensive and cannot zoom properly)\n self.grid.add_tile(get_provider('OSM'))\n\n self.grid.axis.visible = False\n self.grid.grid.visible = False\n # Function to highlight all households that are currently enrolled in\n # the same school.\n self.source.selected.on_change(\"indices\", self.select_households)\n\n # Plot households\n self.grid.circle(x=\"x\",\n y=\"y\",\n size=5,\n view=self.household_view,\n source=self.source,\n fill_color=mapper,\n line_color='black',\n alpha=0.8,\n nonselection_fill_alpha=0.2,\n selection_fill_alpha=1,\n name='households')\n\n # Plot schools\n self.grid.circle(x=\"x\",\n y=\"y\",\n size=7,\n source=self.source,\n view=self.school_view,\n fill_color='yellow',\n line_color='black',\n name='schools')\n\n # Plot neighbourhoods\n self.grid.patches('x',\n 'y',\n source=self.source,\n view=self.neighbourhood_view,\n fill_color=None,\n line_color=\"black\",\n line_width=2,\n hover_alpha=0,\n hover_line_color=None,\n name='neighbourhoods',\n selection_fill_alpha=0)", "def select_buy_and_sell_deal_management_grid_first_row_checkbox(self):\n self.select_grid_row_checkbox(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_checkbox_column_number, True)", "def click_buy_and_sell_deal_management_link(self):\n self.select_static_tab(self.buy_and_sell_deal_management_locator, message=\"buy and sell deal management locator not found before specified time out\")" ]
[ "0.6398558", "0.604387", "0.59920263", "0.5793335", "0.56829715", "0.56757975", "0.5573747", "0.5565601", "0.5519586", "0.54784703", "0.54700017", "0.54370415", "0.5375229", "0.53743625", "0.5347664", "0.53292626", "0.53240794", "0.52728754", "0.52641976", "0.52620107", "0.5251765", "0.5191344", "0.5172993", "0.51595324", "0.5142463", "0.51127005", "0.508476", "0.50781363", "0.50773746", "0.5076457" ]
0.79892987
0
Implementing sort buy and sell management grid column functionality
def sort_buy_and_sell_management_grid_column(self, column_name, descending_order): self.sort_grid_column(self.buy_and_sell_management_grid_div_id, column_name, descending_order)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def OnColumnClick(self, event):\r\n columns = self.data.getParam('columns')\r\n self.SortItems(columns[event.GetColumn()],'INVERT')", "def sort_vendor_price_lists_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.view_price_list_div_id, column_name, descending_order)", "def sort_vendor_price_list_detail_grid_column(self, column_name, descending_order):\n self.wait_for_ajax_spinner_load(300)\n self.sort_grid_column(self.vendor_price_list_detail_rates_grid_div_id, column_name, descending_order)", "def sort_vendor_price_list_detail_reference_rates_grid_column(self, column_name, descending_order):\n self.wait_for_ajax_spinner_load(300)\n self.sort_grid_column(self.vendor_price_list_detail_reference_rates_grid_div_id, column_name, descending_order)", "def get_buy_and_sell_management_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.buy_and_sell_management_grid_div_id)\n return self.column_name_list", "def sort_vendor_price_list_detail_dial_digits_grid_column(self, column_name, descending_order):\n self.wait_for_ajax_spinner_load(300)\n self.sort_grid_column(self.vendor_price_list_detail_dial_digits_grid_div_id, column_name, descending_order)", "def drag_buy_and_sell_management_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.buy_and_sell_management_grid_div_id, source_column, destination_column)", "def sort_rates_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.rates_grid_div_id, column_name, descending_order)", "def sort_vendors_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.vendors_div_id, column_name, descending_order)", "def populate_buttons(self):\n\n # Figure out which index we'll sort on\n if self.sort_group.checkedButton() == self.button_mtime:\n to_sort = self.sort_mtime_idx\n reverse = True\n else:\n to_sort = self.sort_alpha_idx\n reverse = False\n\n # Now add things. This'll automatically shuffle stuff around without\n # us having to worry about removing things first.\n for row, (_, _, button) in enumerate(\n sorted(self.buttons, reverse=reverse, key=lambda i: i[to_sort])\n ):\n self.grid.addWidget(button, row, 0)", "def OnReorder( self, event ):\n column = self.columns[event.GetColumn()]\n if column.sortOn:\n # multiple sorts for the click...\n columns = [ self.columnByAttribute( attr ) for attr in column.sortOn ]\n diff = [ (a,b) for a,b in zip( self.sortOrder, columns ) if b is not a[1]]\n if not diff:\n self.sortOrder[0] = (not self.sortOrder[0][0], column)\n else:\n self.sortOrder = [\n (c.defaultOrder,c) for c in columns \n ] + [ (a,b) for (a,b) in self.sortOrder if b not in columns]\n else:\n if column is self.sortOrder[0][1]:\n # reverse current major order\n self.sortOrder[0] = (not self.sortOrder[0][0], column)\n else:\n self.sortOrder = [(column.defaultOrder,column)] + [\n (a,b) \n for (a,b) in self.sortOrder if b is not column \n ]\n # TODO: store current selection and re-select after sorting...\n self.reorder()\n self.Refresh()", "def drag_vendor_price_lists_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.view_price_list_div_id, source_column, destination_column)", "def sortby(self):\n ...", "def _custom_sorter(self, key1, key2):\n\n col = self._col\n ascending = self._colSortFlag[col]\n real = self.get_real_col(col)\n item1 = self.itemDataMap[key1][real]\n item2 = self.itemDataMap[key2][real]\n\n # Internationalization of string sorting with locale module\n if isinstance(item1, str) and isinstance(item2, str):\n cmpVal = locale.strcoll(item1, item2)\n elif isinstance(item1, bytes) or isinstance(item2, bytes):\n cmpVal = locale.strcoll(str(item1), str(item2))\n else:\n cmpVal = cmp(item1, item2)\n\n # If the items are equal, then pick something else to make the sort value unique\n if cmpVal == 0:\n cmpVal = cmp(*self.GetSecondarySortValues(col, key1, key2))\n\n if ascending:\n return cmpVal\n else:\n return -cmpVal", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def sort_column(self, column):\n if column == 1: # type\n self.sorted_keys = sorted(self.data_dict.keys(),\n key=lambda x: (self.data_dict[x]['type']),\n reverse=self.sorted_type_top)\n # Invert sorting method\n self.sorted_type_top = not self.sorted_type_top\n\n elif column == 2: # Score\n self.sorted_keys = sorted(self.data_dict.keys(),\n key=lambda x: (float(self.data_dict[x]['score'])),\n reverse=self.sorted_score_top)\n # Invert sorting method\n self.sorted_score_top = not self.sorted_score_top\n\n elif column == 4: # Duration\n d = dict()\n for k in self.sorted_keys:\n duration_string = self.data_dict[k]['duration']\n\n # Get amount of episodes\n if 'episode' in duration_string:\n if 'Some' in duration_string:\n episodes = 0\n else:\n episodes = int(duration_string.split(' episodes')[0])\n else:\n episodes = 1\n\n # Get the duration in minutes\n minutes = 0\n if 'min' in duration_string:\n minutes = int(re.findall('([0-9]+)min', duration_string)[0])\n if 'h' in duration_string:\n minutes += int(re.findall('([0-9]+)h', duration_string)[0]) * 60\n\n # Get total duration of the whole show\n minutes *= episodes\n\n # Store it for sorting\n d[k] = minutes\n\n # Sort titles based on duration\n self.sorted_keys = sorted(d.keys(),\n key=lambda x: d[x],\n reverse=self.sorted_duration_top)\n # Invert sorting method\n self.sorted_duration_top = not self.sorted_duration_top\n\n elif column == 5: # release year\n self.sorted_keys = sorted(self.data_dict.keys(),\n key=lambda x: (float(self.data_dict[x]['released'])),\n reverse=self.sorted_year_top)\n # Invert sorting method\n self.sorted_year_top = not self.sorted_year_top\n\n if column != 2:\n # Make sure next time we click to sort by score,\n # the highest score is on top\n self.sorted_score_top = True\n\n # Redraw the table\n self.setup_tableview()", "def get_vendor_price_lists_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.view_price_list_div_id)\n return self.column_name_list", "def drag_vendor_price_list_detail_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_rates_grid_div_id, source_column, destination_column)", "def click_on_sort(self):\r\n self.driver.find_element_by_xpath('//*[@id=\"dataTable\"]/thead/tr[3]/th[4]').click()", "def get_vendor_price_list_detail_grid_column_names_by_order(self):\n self.wait_for_ajax_spinner_load(300)\n self.column_name_list = self.get_grid_column_names_by_order(self.vendor_price_list_detail_rates_grid_div_id)\n return self.column_name_list", "def item_condition(request):\n\n items = Product.objects.all()\n is_used = Product.objects.filter(is_used=True)\n sort = None\n direction = None\n\n if request.GET:\n if 'sort' in request.GET:\n sortkey = request.GET['sort']\n sort = sortkey\n if sortkey == 'name':\n sortkey = 'lower_name'\n items = items.annotate(lower_name=Lower('name'))\n if sortkey == 'category':\n sortkey = 'category__name'\n if 'direction' in request.GET:\n direction = request.GET['direction']\n if direction == 'desc':\n sortkey = f'-{sortkey}'\n items = items.order_by(sortkey)\n\n current_sorting = f'{sort}_{direction}'\n\n context = {\n 'is_used': is_used,\n 'items': items,\n 'current_sorting': current_sorting,\n }\n \n return render(request, 'products/second_hand.html', context)", "def sort(self, col, order):\r\n self.emit(SIGNAL(\"layoutAboutToBeChanged()\"))\r\n self.mylist = sorted(self.mylist,\r\n key=operator.itemgetter(col))\r\n if order == QtCore.Qt.DescendingOrder:\r\n self.mylist.reverse()\r\n self.emit(SIGNAL(\"layoutChanged()\"))", "def sort_fabrics():\n max_shelf_qty = Decimal('240')\n shelves = Shelf.objects.all().order_by('tower', 'name')\n current_shelf_index = 0\n shelf = shelves[current_shelf_index]\n cell_style = \"\"\"\n border-bottom:1px solid #595959;\n border-right:1px solid #595959;\n padding:1em 0.5em;\n text-align:center;\n font-size:1;\n font-family:Tahoma;\n max-height:5em;\n \"\"\"\n header_cell_style = \"\"\"\n border-right:1px solid #595959;\n border-bottom:1px solid #595959;\n border-top:1px solid #595959;\n padding:1em;\n \"\"\"\n item_cell_style = \"\"\"\n padding:0.75em 0.25em;\n \"\"\"\n \n def exceeds_shelf_capacity(shelf, fabric):\n \"\"\"\n Tests whether adding this fabric to shelf will exceed the shelf's maximum \n capacity. Returns a boolean based on the result\n \"\"\"\n shelf_total = Decimal(shelf.fabrics.all().aggregate(Sum('quantity_th'))['quantity_th__sum'] or 0)\n return True if (shelf_total) + fabric.quantity > max_shelf_qty else False\n \n # Reset the shelving arrangements\n Fabric.objects.all().update(shelf=None)\n \n # Loops through the fabrics, organized by patterns so that \n # similar fabrics by patterns are close to each other\n for fabric in Fabric.objects.filter(item__acknowledgement__time_created__gte=date(2014, 1, 1)).distinct().order_by('pattern', 'color'):\n # Only find a shelf if there is fabric to store\n if fabric.quantity > Decimal('0'):\n if not exceeds_shelf_capacity(shelf, fabric):\n fabric.shelf = shelf\n \n else:\n # Loops through all the previous shelves to look for space\n for past_shelf in shelves[0: current_shelf_index]:\n if not exceeds_shelf_capacity(past_shelf, fabric): \n fabric.shelf = past_shelf\n \n try:\n if fabric.shelf is None: \n current_shelf_index += 1\n \n try:\n shelf = shelves[current_shelf_index]\n except (KeyError, IndexError):\n pass#raise ValueError(\"You've run out of space to store fabrics!\")\n \n fabric.shelf = shelf\n \n except Exception:\n current_shelf_index += 1\n \n try:\n shelf = shelves[current_shelf_index]\n except (KeyError, IndexError):\n pass#raise ValueError(\"You've run out of space to store fabrics!\")\n \n fabric.shelf = shelf\n \n fabric.save()\n\n \n \n #return self.message\n return render_to_string('fabric_email.html', {'towers': Tower.objects.all().order_by('id'),\n 'header_style': header_cell_style,\n 'cell_style': cell_style,\n 'item_cell_style': item_cell_style})", "def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)", "def on_sort_column_change(self, model):\n id, order = self.tv.get_model().get_sort_column_id()\n if id is None and order is None:\n return False\n if self.prev_sort_column_id and self.prev_sort_column_id[0][0] == id:\n self.prev_sort_column_id.pop(0)\n self.prev_sort_column_id.insert(0, (id, order))\n if len(self.prev_sort_column_id) > 2:\n self.prev_sort_column_id.pop()", "def sort(self, col, order):\r\n self.emit(SIGNAL(\"layoutAboutToBeChanged()\"))\r\n self.mylist = sorted(self.mylist,\r\n key=operator.itemgetter(col))\r\n if order == Qt.DescendingOrder:\r\n self.mylist.reverse()\r\n self.emit(SIGNAL(\"layoutChanged()\"))", "def products_view(request):\n from_database = Product.objects.all()\n\n sort_type = request.POST.get('sort-selector')\n print(sort_type)\n if sort_type is None:\n from_database = from_database.order_by('title')\n my_context = {\n 'from_database': from_database,\n 'sort_type': sort_type,\n }\n return render(request, \"products.html\", my_context)\n if sort_type == 'name_asc':\n from_database = from_database.order_by('title')\n my_context = {\n 'from_database': from_database,\n 'sort_type': sort_type,\n }\n return render(request, \"products.html\", my_context)\n if sort_type == 'name_desc':\n from_database = from_database.order_by('-title')\n my_context = {\n 'from_database': from_database,\n 'sort_type': sort_type,\n }\n return render(request, \"products.html\", my_context)\n if sort_type == 'price_asc':\n from_database = from_database.order_by('price')\n my_context = {\n 'from_database': from_database,\n 'sort_type': sort_type,\n }\n return render(request, \"products.html\", my_context)\n if sort_type == 'price_desc':\n from_database = from_database.order_by('-price')\n my_context = {\n 'from_database': from_database,\n 'sort_type': sort_type,\n }\n return render(request, \"products.html\", my_context)", "def building_grid(self, instance, value):\n self.books.sort(value)\n # Building a scroll window\n self.recycle.width = Window.width\n self.recycle.height = Window.height - self.headlabel.height - self.warnlabel.height\n self.maingrid.bind(\n minimum_height=self.maingrid.setter('height')\n )\n # Redrawing the list of books\n self.maingrid.clear_widgets()\n for book in self.books:\n self.maingrid.add_widget(\n BookButton(\n book=book,\n top_label=self.headlabel,\n warn_label=self.warnlabel,\n text=str(book),\n size_hint_y=None\n )\n )", "def sort(self, col, order):\n self.layoutAboutToBeChanged.emit()\n self.mylist = sorted(self.mylist,\n key=operator.itemgetter(col))\n if order == Qt.DescendingOrder:\n self.mylist.reverse()\n self.layoutChanged.emit()", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)" ]
[ "0.6722339", "0.65786284", "0.6452577", "0.6041004", "0.6005217", "0.5994471", "0.5969092", "0.5801106", "0.5778006", "0.57584304", "0.57022667", "0.55251795", "0.5473172", "0.5456213", "0.5440098", "0.54038495", "0.53848374", "0.5352643", "0.5340399", "0.5331021", "0.5279471", "0.5275177", "0.526848", "0.5264978", "0.52547777", "0.523983", "0.5232802", "0.520433", "0.518093", "0.516075" ]
0.70848614
0
Implementing get all rows of specific column from buy and sell management grid functionality
def get_all_rows_of_specific_column_from_buy_and_sell_management_grid(self, column_name): self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.buy_and_sell_management_grid_div_id, column_name, self.get_grid_row_count(self.buy_and_sell_management_grid_count_span_locator, 2)) return self.all_row_data
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_all_rows_of_rate_increase_violation_column_from_vendor_price_list_detail_grid(self, column_name):\n self.all_row_data = []\n self.wait_for_ajax_spinner_load()\n row_count = len(self.wait().until(EC.presence_of_all_elements_located(self.vendor_price_list_detail_row_locator), 'vendor price list details row element locator not found before specified time'))\n next_page = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (self.vendor_price_list_detail_rates_grid_div_id, column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n row = 1\n try:\n while row < row_count + 1:\n column_value_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tbody[@role='rowgroup']/tr[%s]/td[%s]\" % (self.vendor_price_list_detail_rates_grid_div_id, row, str(column_index)))\n column_value_element = self.wait(10).until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n column_value = str(column_value_element.get_attribute(\"innerHTML\"))\n if column_value not in self.all_row_data:\n self.all_row_data.append(column_value)\n else:\n row_count += 1\n if row % 50 == 0:\n next_page = True\n next_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-arrow-e']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(next_page_button_locator, True)\n row = 0\n row_count -= 50\n row += 1\n except WebDriverException:\n pass\n if next_page is True:\n first_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-seek-w']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(first_page_button_locator, True)\n return self.all_row_data", "def get_all_rows_of_specific_column_from_vendor_price_lists_grid(self, column_name):\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.view_price_list_div_id, column_name, self.get_grid_row_count(self.price_lists_grid_row_count_span_locator, 7))\n return self.all_row_data", "def get_all_rows_of_specific_column_from_vendor_price_list_detail_grid(self, column_name):\n self.wait_for_ajax_spinner_load(300)\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.vendor_price_list_detail_rates_grid_div_id, column_name)\n return self.all_row_data", "def get_all_rows_of_specific_column_only_from_vendor_price_list_detail_grid(self, div_id, column_name):\n row_count = len(self.wait().until(EC.presence_of_all_elements_located(self.vendor_price_list_detail_row_locator), 'vendor price list details row element locator not found before specified time'))\n all_row_data = []\n next_page = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (div_id, column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n row = 1\n try:\n while row < row_count+1:\n column_value_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tbody[@role='rowgroup']/tr[%s]/td[%s]/descendant::td\" % (div_id, row, str(column_index)))\n column_value_element = self.wait(10).until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n column_value = str(column_value_element.get_attribute(\"innerHTML\"))\n if column_value not in all_row_data:\n all_row_data.append(column_value)\n else:\n row_count += 1\n if row % 50 == 0:\n next_page = True\n next_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-arrow-e']\" % div_id)\n self.click_element(next_page_button_locator, True)\n row = 0\n row_count -= 50\n row += 1\n except WebDriverException:\n pass\n if next_page is True:\n first_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-seek-w']\" % div_id)\n self.click_element(first_page_button_locator, True)\n return all_row_data", "def get_all_rows_of_rate_column_from_vendor_price_list_detail_grid(self, column_name):\n self.wait_for_ajax_spinner_load()\n self.all_row_data = self.get_all_rows_of_specific_column_only_from_vendor_price_list_detail_grid(self.vendor_price_list_detail_rates_grid_div_id, column_name)\n return self.all_row_data", "def get_all_rows_of_specific_column_from_vendor_price_list_detail_reference_rates_grid(self, column_name):\n self.wait_for_ajax_spinner_load(300)\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.vendor_price_list_detail_reference_rates_grid_div_id, column_name)\n return self.all_row_data", "def get_all_rows_of_specific_column_from_vendor_price_list_detail_dial_digits_grid(self, column_name):\n self.wait_for_ajax_spinner_load(300)\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.vendor_price_list_detail_dial_digits_grid_div_id, column_name)\n return self.all_row_data", "def rosterRowData(self):", "def get_buy_and_sell_management_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.buy_and_sell_management_grid_div_id)\n return self.column_name_list", "def get_all_rows_of_specific_column_from_rates_grid(self, column_name):\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.rates_grid_div_id, column_name, self.get_grid_row_count(self.rates_grid_row_count_span_locator, 1))\n return self.all_row_data", "def get_first_row_item_from_price_lists_grid(self, column_name):\n column_locator = (By.XPATH, \"//div[@id='%s']/descendant::th[@data-title='%s']\" % (self.vendor_price_list_grid_div_id, column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n column_value_locator = (By.XPATH, \"//div[@id='%s']/descendant::tbody/descendant::tr[1]/td[%s]/span\" % (self.vendor_price_list_grid_div_id, str(column_index)))\n column_value_element = self.wait().until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n return column_value_element.text", "def get_all_rows_of_specific_column_from_vendors_grid(self, column_name):\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.vendors_div_id, column_name, self.get_grid_row_count(self.vendors_grid_row_count_span_locator, 5))\n return self.all_row_data", "def get_vendor_price_list_detail_grid_column_names_by_order(self):\n self.wait_for_ajax_spinner_load(300)\n self.column_name_list = self.get_grid_column_names_by_order(self.vendor_price_list_detail_rates_grid_div_id)\n return self.column_name_list", "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def get_list_columns(self):\n columns = super(InventoryTransactionAdmin, self).get_list_columns()\n cols = ['total_amount']\n columns = security_util.filter_columns_by_role(\n columns, cols, 'purchase_price_view'\n )\n return columns", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def get_vendor_price_list_detail_reference_rates_grid_column_names_by_order(self):\n self.wait_for_ajax_spinner_load(300)\n self.column_name_list = self.get_grid_column_names_by_order(self.vendor_price_list_detail_reference_rates_grid_div_id)\n return self.column_name_list", "def select_vendor_price_list_grid_row_checkbox(self):\n for row in range(0,2):\n column_locator = (By.XPATH, \"//div[@id='%s']/descendant::th[@data-title='%s']\" % (self.vendor_price_list_grid_div_id, self.price_list_column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n column_value_locator = (By.XPATH, \"//div[@id='%s']/descendant::tbody/descendant::tr[%s]/td[%s]/span\" % (self.vendor_price_list_grid_div_id, (row + 1), str(column_index)))\n column_value_element = self.wait().until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n price_list_item = column_value_element.text\n self.compare_price_lists_list.append(price_list_item)\n self.select_grid_row_checkbox(self.customer_price_list_grid_div_id, price_list_item, self.select_grid_row_checkbox_column_number)", "def get_vendor_price_list_detail_dial_digits_grid_column_names_by_order(self):\n self.wait_for_ajax_spinner_load(300)\n self.column_name_list = self.get_grid_column_names_by_order(self.vendor_price_list_detail_dial_digits_grid_div_id)\n return self.column_name_list", "def select_buy_and_sell_deal_management_grid_first_row_checkbox(self):\n self.select_grid_row_checkbox(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_checkbox_column_number, True)", "def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)", "def get_all_rows_of_specific_column_from_re_analysis_detail_grid(self, column_name):\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.re_analysis_detail_grid_div_id, column_name)\n return self.all_row_data", "def get_vendor_price_lists_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.view_price_list_div_id)\n return self.column_name_list", "def columns(self):\n \n pass", "def get_rows(self):\n raise NotImplementedError('Subclass this to make the rows')", "def get_all_rows_of_specific_column_from_destinations_grid(self, column_name):\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.destinations_grid_div_id, column_name, self.get_grid_row_count(self.destinations_grid_row_count_span_locator, 1))\n return self.all_row_data", "def get_the_sum_of_prices_from_table(table, item_ids):\n\n # your code", "def get_action(self):\n index = []\n action = []\n last_action = 'sell'\n for i, row in self._indicator.iterrows():\n # Buy\n if row[0] < 14.0 and last_action == 'sell':\n index.append(i)\n #action.append(['buy', self._stock.loc[i]['Close']])\n action.append(['buy'] + self._stock.loc[i].values.tolist())\n last_action = 'buy'\n # Sell\n elif row[0] > 17.0 and last_action == 'buy':\n index.append(i)\n #action.append(['sell', self._stock.loc[i]['Close']])\n action.append(['sell'] + self._stock.loc[i].values.tolist())\n last_action = 'sell'\n\n headers = ['Action'] + self._stock.columns.tolist()\n self._action = pd.DataFrame(action, index=index, columns=headers)\n return self._action", "def return_items(self):\n cur = self.cursor\n cur.execute(f\"SELECT * FROM {self.product_name}\")\n products = cur.fetchall()\n return products", "def get_outbound_statement_grid_row_data(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary = {\"Bill Account\": \"\", \"Begin Date\": \"\", \"End Date\": \"\", \"Amount ( EUR )\": \"\"}\n self.outbound_statement_grid_row_data = self.get_grid_row_details(self.outbound_statements_grid_div_id, self.grid_row_details_dictionary)\n return self.outbound_statement_grid_row_data" ]
[ "0.6735556", "0.6720651", "0.67186683", "0.6640583", "0.65773517", "0.6404575", "0.6303615", "0.5945554", "0.5850497", "0.58364195", "0.57948834", "0.57733727", "0.57243025", "0.56459594", "0.56439763", "0.5600407", "0.5567689", "0.547445", "0.5463407", "0.54597545", "0.54290724", "0.53594226", "0.5340059", "0.53175354", "0.52875984", "0.5281515", "0.52622336", "0.5258395", "0.5250748", "0.5214579" ]
0.7284325
0
Implementing drag buy and sell management grid column functionality
def drag_buy_and_sell_management_grid_column(self, source_column, destination_column): self.drag_grid_column(self.buy_and_sell_management_grid_div_id, source_column, destination_column)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def drag_vendor_price_lists_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.view_price_list_div_id, source_column, destination_column)", "def drag_vendor_price_list_detail_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_rates_grid_div_id, source_column, destination_column)", "def drag_rates_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.rates_grid_div_id, source_column, destination_column)", "def _drag_col(self, event):\n x = self._dx + event.x # get dragged column new left x coordinate\n self._visual_drag.place_configure(x=x) # update column preview position\n # if one border of the dragged column is beyon the middle of the\n # neighboring column, swap them\n if (self._dragged_col_neighbor_widths[0] is not None and\n x < self._dragged_col_x - self._dragged_col_neighbor_widths[0] / 2):\n self._swap_columns('left')\n elif (self._dragged_col_neighbor_widths[1] is not None and\n x > self._dragged_col_x + self._dragged_col_neighbor_widths[1] / 2):\n self._swap_columns('right')\n # horizontal scrolling if the cursor reaches the side of the table\n if x < 0 and self.xview()[0] > 0:\n # scroll left and update dragged column x coordinate\n self.xview_scroll(-10, 'units')\n self._dragged_col_x += 10\n elif x + self._dragged_col_width / 2 > self.winfo_width() and self.xview()[1] < 1:\n # scroll right and update dragged column x coordinate\n self.xview_scroll(10, 'units')\n self._dragged_col_x -= 10", "def drag_vendors_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.vendors_div_id, source_column, destination_column)", "def drag_vendor_price_list_detail_reference_rates_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_reference_rates_grid_div_id, source_column, destination_column)", "def _start_drag_col(self, event):\n # identify dragged column\n col = self.identify_column(event.x)\n self._dragged_col = ttk.Treeview.column(self, col, 'id')\n # get column width\n self._dragged_col_width = w = ttk.Treeview.column(self, col, 'width')\n # get x coordinate of the left side of the column\n x = event.x\n while self.identify_region(x, event.y) == 'heading':\n # decrease x until reaching the separator\n x -= 1\n x_sep = x\n w_sep = 0\n # determine separator width\n while self.identify_region(x_sep, event.y) == 'separator':\n w_sep += 1\n x_sep -= 1\n if event.x - x <= self._im_drag.width():\n # start dragging if mouse click was on dragging icon\n x = x - w_sep // 2 - 1\n self._dragged_col_x = x\n # get neighboring column widths\n displayed_cols = self._displayed_cols\n self._dragged_col_index = i1 = displayed_cols.index(self._dragged_col)\n if i1 > 0:\n left = ttk.Treeview.column(self, displayed_cols[i1 - 1], 'width')\n else:\n left = None\n if i1 < len(displayed_cols) - 1:\n right = ttk.Treeview.column(self, displayed_cols[i1 + 1], 'width')\n else:\n right = None\n self._dragged_col_neighbor_widths = (left, right)\n self._dx = x - event.x # distance between cursor and column left border\n # configure dragged column preview\n self._visual_drag.column(self._dragged_col, width=w)\n self._visual_drag.configure(displaycolumns=[self._dragged_col])\n if 'headings' in tuple(str(p) for p in self['show']):\n self._visual_drag.configure(show='headings')\n else:\n self._visual_drag.configure(show='')\n self._visual_drag.place(in_=self, x=x, y=0, anchor='nw',\n width=w + 2, relheight=1)\n self._visual_drag.state(('active',))\n self._visual_drag.update_idletasks()\n self._visual_drag.yview_moveto(self.yview()[0])\n else:\n self._dragged_col = None", "def drag_vendor_price_list_detail_dial_digits_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_dial_digits_grid_div_id, source_column, destination_column)", "def sort_buy_and_sell_management_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.buy_and_sell_management_grid_div_id, column_name, descending_order)", "def OnMouse(self, event):\r\n\r\n # we want to work with logical coords\r\n x, dummy = self._owner.CalcUnscrolledPosition(event.GetX(), 0)\r\n y = event.GetY()\r\n\r\n if event.Moving():\r\n \r\n col = self.XToCol(x)\r\n if col != self._hotTrackCol:\r\n \r\n # Refresh the col header so it will be painted with hot tracking\r\n # (if supported by the native renderer.)\r\n self.RefreshColLabel(col)\r\n\r\n # Also refresh the old hot header\r\n if self._hotTrackCol >= 0:\r\n self.RefreshColLabel(self._hotTrackCol)\r\n\r\n self._hotTrackCol = col\r\n \r\n if event.Leaving() and self._hotTrackCol >= 0:\r\n \r\n # Leaving the window so clear any hot tracking indicator that may be present\r\n self.RefreshColLabel(self._hotTrackCol)\r\n self._hotTrackCol = -1\r\n \r\n if self._isDragging:\r\n\r\n self.SendListEvent(wx.wxEVT_COMMAND_LIST_COL_DRAGGING, event.GetPosition())\r\n\r\n # we don't draw the line beyond our window, but we allow dragging it\r\n # there\r\n w, dummy = self.GetClientSize()\r\n w, dummy = self._owner.CalcUnscrolledPosition(w, 0)\r\n w -= 6\r\n\r\n # erase the line if it was drawn\r\n if self._currentX < w:\r\n self.DrawCurrent()\r\n\r\n if event.ButtonUp():\r\n self._isDragging = False\r\n if self.HasCapture():\r\n self.ReleaseMouse()\r\n self._dirty = True\r\n self.SetColumnWidth(self._column, self._currentX - self._minX)\r\n self.Refresh()\r\n self.SendListEvent(wx.wxEVT_COMMAND_LIST_COL_END_DRAG, event.GetPosition())\r\n else:\r\n self._currentX = max(self._minX + 7, x)\r\n\r\n # draw in the new location\r\n if self._currentX < w:\r\n self.DrawCurrent()\r\n \r\n else: # not dragging\r\n\r\n self._minX = 0\r\n hit_border = False\r\n\r\n # end of the current column\r\n xpos = 0\r\n\r\n # find the column where this event occured\r\n countCol = self.GetColumnCount()\r\n\r\n for column in xrange(countCol):\r\n\r\n if not self.IsColumnShown(column):\r\n continue # do next if not shown\r\n\r\n xpos += self.GetColumnWidth(column)\r\n self._column = column\r\n if abs (x-xpos) < 3 and y < 22:\r\n # near the column border\r\n hit_border = True\r\n break\r\n \r\n if x < xpos:\r\n # inside the column\r\n break\r\n \r\n self._minX = xpos\r\n \r\n if event.LeftDown() or event.RightUp():\r\n if hit_border and event.LeftDown():\r\n self._isDragging = True\r\n self.CaptureMouse()\r\n self._currentX = x\r\n self.DrawCurrent()\r\n self.SendListEvent(wx.wxEVT_COMMAND_LIST_COL_BEGIN_DRAG, event.GetPosition())\r\n else: # click on a column\r\n evt = (event.LeftDown() and [wx.wxEVT_COMMAND_LIST_COL_CLICK] or [wx.wxEVT_COMMAND_LIST_COL_RIGHT_CLICK])[0]\r\n self.SendListEvent(evt, event.GetPosition())\r\n \r\n elif event.LeftDClick() and hit_border:\r\n self.SetColumnWidth(self._column, self._owner.GetBestColumnWidth(self._column))\r\n self.Refresh()\r\n\r\n elif event.Moving():\r\n \r\n if hit_border:\r\n setCursor = self._currentCursor == wx.STANDARD_CURSOR\r\n self._currentCursor = self._resizeCursor\r\n else:\r\n setCursor = self._currentCursor != wx.STANDARD_CURSOR\r\n self._currentCursor = wx.STANDARD_CURSOR\r\n \r\n if setCursor:\r\n self.SetCursor(self._currentCursor)", "def _drag_row(self, event):\n y = self._dy + event.y # get dragged row new upper y coordinate\n self._visual_drag.place_configure(y=y) # update row preview position\n\n if y > self._dragged_row_y:\n # moving downward\n item = self.identify_row(y + self._dragged_row_height)\n if item != '':\n bbox = self.bbox(item)\n if not bbox:\n # the item is not visible so make it visible\n self.see(item)\n self.update_idletasks()\n bbox = self.bbox(item)\n if y > self._dragged_row_y + bbox[3] / 2:\n # the row is beyond half of item, so insert it below\n self._move_dragged_row(item)\n elif item != self.next(self._dragged_row):\n # item is not the lower neighbor of the dragged row so insert the row above\n self._move_dragged_row(self.prev(item))\n elif y < self._dragged_row_y:\n # moving upward\n item = self.identify_row(y)\n if item != '':\n bbox = self.bbox(item)\n if not bbox:\n # the item is not visible so make it visible\n self.see(item)\n self.update_idletasks()\n bbox = self.bbox(item)\n if y < self._dragged_row_y - bbox[3] / 2:\n # the row is beyond half of item, so insert it above\n self._move_dragged_row(item)\n elif item != self.prev(self._dragged_row):\n # item is not the upper neighbor of the dragged row so insert the row below\n self._move_dragged_row(self.next(item))\n self.selection_remove(self._dragged_row)", "def drag_dialed_digits_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.dialed_digits_grid_div_id, source_column, destination_column)", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def drag_destinations_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.destinations_grid_div_id, source_column, destination_column)", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def drag_country_groups_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.country_groups_grid_div_id, source_column, destination_column)", "def OnColumnClick(self, event):\r\n columns = self.data.getParam('columns')\r\n self.SortItems(columns[event.GetColumn()],'INVERT')", "def drag_re_analysis_detail_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.re_analysis_detail_grid_div_id, source_column, destination_column)", "def _on_press(self, event):\n if tk.DISABLED in self.state():\n return\n\n region = self.identify_region(event.x, event.y)\n\n if self._drag_cols and region == 'heading':\n self._start_drag_col(event)\n elif self._drag_rows and region == 'cell':\n self._start_drag_row(event)", "def dragEnterEvent(self, event):", "def drag_re_analysis_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.re_analysis_grid_div_id, source_column, destination_column)", "def OnReorderableGridMotion(self, evt):\n\t\tif not self._potentialRearrange:\n\t\t\tevt.Skip()\n\t\t\treturn\n\n\t\tif not self._enableReorderRows:\n\t\t\tself._potentialRearrange = False\n\t\t\tevt.Skip()\n\t\t\treturn\n\t\t\t\n\t\tx, y = evt.GetX(), evt.GetY()\n\t\trow, col = self.ReorderableGridRowXYToCell(x,y, colheight=0)\n\t\tif row == self._lastRow:\n\t\t\treturn\n\t\t\t\n\t\tself.DeselectRow( self._lastRow )\n\t\t\n\t\tlastRowSave = [self.GetCellValue(self._lastRow, c) for c in range(self.GetNumberCols())]\n\t\tlastRowBackgroundColourSave = [self.GetCellBackgroundColour(self._lastRow, c) for c in range(self.GetNumberCols())]\n\t\tdirection = 1 if row > self._lastRow else -1 if row < self._lastRow else 0\n\t\tfor r in range(self._lastRow, row, direction ):\n\t\t\tself.copyRow( r + direction, r )\n\t\tfor c in range(self.GetNumberCols()):\n\t\t\tself.SetCellValue( row, c, lastRowSave[c] )\n\t\t\tself.SetCellBackgroundColour( row, c, lastRowBackgroundColourSave[c] )\n\t\t\n\t\tself.SelectRow( row, False )\n\t\tself._lastRow = row", "def _place_down_row_( self, a_widget, columnspan, rowspan, sticky = None ):\n #print( \"_place_down_row_ still need to make sticky stick !!\")\n if sticky is None:\n sticky = self.sticky\n #rint( f\"_place_down_row_ row = {self.ix_row} col = {self.ix_col}\" )\n a_widget.grid( row = self.ix_row,\n column = self.ix_col,\n rowspan = rowspan,\n sticky = sticky, )\n\n self.ix_row += rowspan\n if ( self.max > 0 ) and ( self.ix_row >= self.max ):\n print( f\"hit max row {self.max}\" )\n self.ix_col += 1\n self.ix_row = 0", "def _left_button_release_event(self, obj, event):\n #self.OnLeftButtonUp()\n pixel_x, pixel_y = self.parent.vtk_interactor.GetEventPosition()\n #selector = vtk.vtkVisibleCellSelector()\n\n self.picker_points.append((pixel_x, pixel_y))\n\n #print(self.picker_points)\n if len(self.picker_points) == 2:\n p1x, p1y = self.picker_points[0]\n p2x, p2y = self.picker_points[1]\n self.picker_points = []\n xmin = min(p1x, p2x)\n ymin = min(p1y, p2y)\n xmax = max(p1x, p2x)\n ymax = max(p1y, p2y)\n #print(self.picker_points)\n #print('_area_pick_left_button_release', cell_id)\n\n dx = abs(p1x - p2x)\n dy = abs(p1y - p2y)\n self.picker_points = []\n if dx > 0 and dy > 0:\n if self._pick_visible:\n self._pick_visible_ids(xmin, ymin, xmax, ymax)\n else:\n self._pick_depth_ids(xmin, ymin, xmax, ymax)\n self.parent.vtk_interactor.Render()\n self.picker_points = []", "def on_mouse_press(self, x, y, button, modifiers):\n\n # Change the x/y screen coordinates to grid coordinates\n column = int(x // (WIDTH + MARGIN))\n row = int(y // (HEIGHT + MARGIN))\n\n # print(f\"Click coordinates: ({x}, {y}). Grid coordinates: ({row}, {column})\")\n\n # Make sure we are on-grid. It is possible to click in the upper right\n # corner in the margin and go to a grid location that doesn't exist\n # AKA: make sure you are clicking w/in the grid - TH\n if row < ROW_COUNT and column < COLUMN_COUNT:\n # Flip the location between 1 and 0.\n # this will reset value for the recreate grid\n # and change the color - TH\n # if self.grid[row][column] == 0:\n # self.grid[row][column] = self.num_key\n # else:\n # self.grid[row][column] = 0\n self.current_selected = (row, column)\n\n self.recreate_grid()", "def drag_and_drop_by_offset(self, elem, x, y):\n ActionChains(self.driver).drag_and_drop_by_offset(elem, xoffset=x, yoffset=y).perform()", "def slider_dragged(self):\n pass", "def sort_vendor_price_lists_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.view_price_list_div_id, column_name, descending_order)", "def on_tableWidget_Trade_Args_cellClicked(self, row, column):\n # TODO: not implemented yet\n raise NotImplementedError", "def dropEvent(self, QDropEvent):\n srcItems = self.selectedItems()\n dstInd = (self.indexAt(QDropEvent.pos()).row() + 1)\n kbMod = QDropEvent.keyboardModifiers()\n #-- Create New Items --#\n for n, srcItem in enumerate(srcItems):\n itemDict = self.treeParent.getItemDict(srcItem)\n newItem = self.treeParent.on_addVar(index=(dstInd + n))\n self.treeParent.setItem(newItem, **itemDict)\n #-- Remove Items --#\n if not kbMod == QtCore.Qt.ControlModifier:\n for srcItem in srcItems:\n self.takeTopLevelItem(self.indexOfTopLevelItem(srcItem))\n self.treeParent.reindexVar()" ]
[ "0.68649036", "0.6444064", "0.63804555", "0.63166463", "0.62582725", "0.61420965", "0.61303335", "0.59818137", "0.5775787", "0.56232387", "0.5621364", "0.5609542", "0.5536238", "0.55356735", "0.53799516", "0.5366732", "0.53493655", "0.5324431", "0.52932245", "0.525854", "0.5229055", "0.5220589", "0.5187343", "0.5179866", "0.51731324", "0.5171649", "0.51676244", "0.51641", "0.51459086", "0.5106501" ]
0.7623682
0
Implementing get buy and sell management grid column names by order functionality
def get_buy_and_sell_management_grid_column_names_by_order(self): self.column_name_list = self.get_grid_column_names_by_order(self.buy_and_sell_management_grid_div_id) return self.column_name_list
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_vendor_price_lists_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.view_price_list_div_id)\n return self.column_name_list", "def get_vendor_price_list_detail_grid_column_names_by_order(self):\n self.wait_for_ajax_spinner_load(300)\n self.column_name_list = self.get_grid_column_names_by_order(self.vendor_price_list_detail_rates_grid_div_id)\n return self.column_name_list", "def get_vendor_price_list_detail_reference_rates_grid_column_names_by_order(self):\n self.wait_for_ajax_spinner_load(300)\n self.column_name_list = self.get_grid_column_names_by_order(self.vendor_price_list_detail_reference_rates_grid_div_id)\n return self.column_name_list", "def get_vendor_price_list_detail_dial_digits_grid_column_names_by_order(self):\n self.wait_for_ajax_spinner_load(300)\n self.column_name_list = self.get_grid_column_names_by_order(self.vendor_price_list_detail_dial_digits_grid_div_id)\n return self.column_name_list", "def sort_buy_and_sell_management_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.buy_and_sell_management_grid_div_id, column_name, descending_order)", "def get_rates_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.rates_grid_div_id)\n return self.column_name_list", "def sort_vendor_price_lists_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.view_price_list_div_id, column_name, descending_order)", "def sort_vendor_price_list_detail_grid_column(self, column_name, descending_order):\n self.wait_for_ajax_spinner_load(300)\n self.sort_grid_column(self.vendor_price_list_detail_rates_grid_div_id, column_name, descending_order)", "def get_vendors_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.vendors_div_id)\n return self.column_name_list", "def get_re_analysis_detail_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.re_analysis_detail_grid_div_id)\n return self.column_name_list", "def sort_vendor_price_list_detail_reference_rates_grid_column(self, column_name, descending_order):\n self.wait_for_ajax_spinner_load(300)\n self.sort_grid_column(self.vendor_price_list_detail_reference_rates_grid_div_id, column_name, descending_order)", "def OnColumnClick(self, event):\r\n columns = self.data.getParam('columns')\r\n self.SortItems(columns[event.GetColumn()],'INVERT')", "def sort_vendor_price_list_detail_dial_digits_grid_column(self, column_name, descending_order):\n self.wait_for_ajax_spinner_load(300)\n self.sort_grid_column(self.vendor_price_list_detail_dial_digits_grid_div_id, column_name, descending_order)", "def show_orders():\n return 'hehe'", "def get_list_columns(self):\n columns = super(InventoryTransactionAdmin, self).get_list_columns()\n cols = ['total_amount']\n columns = security_util.filter_columns_by_role(\n columns, cols, 'purchase_price_view'\n )\n return columns", "def sort_vendors_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.vendors_div_id, column_name, descending_order)", "def sort_rates_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.rates_grid_div_id, column_name, descending_order)", "def get_re_analysis_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.re_analysis_grid_div_id)\n return self.column_name_list", "def _generate_column_names(self):\n names = []\n # Assuming same number of compounds for every reaction!\n\n names = ['compound_{}'.format(i) for i in range(self.total_compounds)]\n names += ['compound_{}_amount'.format(i)\n for i in range(self.total_compounds)]\n for grid_param in self.params_grid_data.keys():\n names.append(grid_param)\n\n return names", "def columnTitles(self):\n \n pass", "def columnTitles(self):\n \n pass", "def get_dialed_digits_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.dialed_digits_grid_div_id)\n return self.column_name_list", "def order_report():", "def get_destinations_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.destinations_grid_div_id)\n return self.column_name_list", "def init_column_headers(self):\r\n col_headers = {1: \"bid\", 2: \"tricks\", 3: \"score\"}\r\n for player in range(1, 5):\r\n ent_player_name = tk.Entry(self.master, font='courier 10 bold', fg='blue',\r\n borderwidth=2, relief=\"groove\")\r\n ent_player_name.grid(row=0, column=(player - 1) * 3 + 1, columnspan=3,\r\n sticky=tk.W + tk.E, pady=5)\r\n ent_player_name.insert(0, \"Player\" + str(player))\r\n self.ent_player_names.append(ent_player_name)\r\n for key in col_headers:\r\n lbl_column_header = tk.Label(self.master, text=col_headers[key],\r\n font='courier 10 bold', fg='blue',\r\n borderwidth=2, relief=\"groove\")\r\n lbl_column_header.grid(row=1, column=(player - 1) * 3 + key,\r\n sticky=tk.W + tk.E, pady=2)", "def get_sort_columns_raw(self):\n col_sort_orders = self.gridpreference_sort.all().select_related()\n \n return [x.sort_display for x in col_sort_orders]", "def get_country_groups_grid_column_names_by_order(self):\n self.column_name_list = self.get_grid_column_names_by_order(self.country_groups_grid_div_id)\n return self.column_name_list", "def get_order_columns(self):\n return self.order_columns", "def get_all_rows_of_specific_column_from_buy_and_sell_management_grid(self, column_name):\n self.all_row_data = self.get_all_rows_of_specific_column_from_grid(self.buy_and_sell_management_grid_div_id, column_name, self.get_grid_row_count(self.buy_and_sell_management_grid_count_span_locator, 2))\n return self.all_row_data", "def setFieldNames(self, model, lyr): \n #get the fields\n fields = lyr.pendingFields()\n position = 0\n \n #set column names\n for field in fields:\n model.setHorizontalHeaderItem(position, QStandardItem(field.name()))\n position+=1" ]
[ "0.7173671", "0.7164694", "0.6961585", "0.67937434", "0.66460264", "0.64222735", "0.63047063", "0.6221156", "0.61231023", "0.5842353", "0.58139414", "0.57718784", "0.5757523", "0.56256413", "0.5614401", "0.56115836", "0.5588756", "0.55571264", "0.55410564", "0.55111706", "0.55111706", "0.5499306", "0.5453823", "0.5438632", "0.5436244", "0.538761", "0.538091", "0.5367089", "0.53557837", "0.53330827" ]
0.75991935
0
Implementing click buy and sell management grid export to excel button functionality
def click_buy_and_sell_management_grid_export_to_excel_button(self): self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_vendor_price_list_detail_rates_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_rates_grid_div_id)", "def click_vendor_price_list_detail_reference_rates_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_reference_rates_grid_div_id)", "def click_vendor_price_list_detail_dial_digits_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_dial_digits_grid_div_id)", "def click_re_analysis_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.re_analysis_grid_div_id)", "def click_export_to_excel_button(self):\n self.click_element(self.export_to_excel_button_locator, True)", "def click_export_to_excel_button(self):\n self.click_element(self.export_to_excel_button_locator, True)", "def click_country_groups_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.country_groups_grid_div_id)", "def export(self):\r\n self.prices[\"returns\"] = self.returns\r\n self.prices.columns = ['prices', 'returns']\r\n self.prices = self.prices.dropna()\r\n \r\n name = QFileDialog.getSaveFileName(None, 'Save File', filter='*.xlsx')\r\n if(name[0] == ''):\r\n # if name empty\r\n pass\r\n else:\r\n self.prices.to_excel(name[0])", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def export_data(self):\r\n stocks = {}\r\n headings = ['Security', 'Price', 'Change', 'Change %', '52 Week', 'Market Cap']\r\n\r\n for data in range(6):\r\n for items in self.root.main.treeview.get_children():\r\n values = self.root.main.treeview.item(items, 'values')\r\n if headings[data] not in stocks:\r\n stocks[headings[data]] = []\r\n stocks.get(headings[data]).append(values[data])\r\n\r\n df = pd.DataFrame(stocks, columns=headings)\r\n path = tk.filedialog.asksaveasfilename(title='Save File As...',\r\n filetypes=((\"CComma-separated values (.csv)\", \"*.csv\"), (\"Text Document(.txt)\", \"*.txt\")))\r\n\r\n if not path:\r\n return\r\n else:\r\n df.to_excel(path, index=False, header=True)", "def exporter():\n Session = modules.db_connect.connect()\n session = Session()\n report = xlsxwriter.Workbook('perception_report.xlsx')\n top_row_format = report.add_format({'bold': True})\n top_row_format.set_border(style=1)\n top_row_format.set_bg_color('#B8B8B8')\n\n \"\"\"Black row format at the top of each host detailed info\"\"\"\n black_row_format = report.add_format()\n black_row_format.set_border(style=1)\n black_row_format.set_bg_color('#000000')\n\n \"\"\"Detailed host row format\"\"\"\n host_row_format = report.add_format()\n host_row_format.set_border(style=1)\n host_row_format.set_bg_color('#CCCCCC')\n\n \"\"\"Format for text in row with host info\"\"\"\n host_row_wrapped_format = report.add_format()\n host_row_wrapped_format.set_border(style=1)\n host_row_wrapped_format.set_bg_color('#CCCCCC')\n host_row_wrapped_format.set_text_wrap('vjustify')\n\n \"\"\"Format description row in NSE output\"\"\"\n host_nse_output_top_format = report.add_format({'bold': True})\n host_nse_output_top_format.set_border(style=1)\n host_nse_output_top_format.set_bg_color('#B8B8B8')\n\n \"\"\"Format test row in NSE output\"\"\"\n host_nse_output_format = report.add_format()\n host_nse_output_format.set_border(style=1)\n host_nse_output_format.set_bg_color('#CCCCCC')\n\n \"\"\"Build the host_overview_worksheet\"\"\"\n host_overview_worksheet = report.add_worksheet()\n\n \"\"\"Build the host_detail_worksheet\"\"\"\n host_detail_worksheet = report.add_worksheet()\n\n \"\"\"Size up the overview worksheet\"\"\"\n host_overview_worksheet.set_column('B:B', 24)\n host_overview_worksheet.set_column('C:C', 15)\n host_overview_worksheet.set_column('D:D', 15)\n host_overview_worksheet.set_column('E:E', 15)\n host_overview_worksheet.set_column('F:F', 15)\n host_overview_worksheet.set_column('G:G', 20)\n host_overview_worksheet.set_column('H:H', 15)\n\n \"\"\"Size up the detail worksheet\"\"\"\n host_detail_worksheet.set_column('B:B', 38)\n host_detail_worksheet.set_column('C:C', 16)\n host_detail_worksheet.set_column('D:D', 16)\n host_detail_worksheet.set_column('E:E', 28)\n host_detail_worksheet.set_column('F:F', 15)\n host_detail_worksheet.set_column('H:G', 20)\n host_detail_worksheet.set_column('H:H', 25)\n host_detail_worksheet.set_column('I:I', 10)\n\n \"\"\"Description row for host overview\"\"\"\n host_overview_worksheet.write('B2', 'Hostname', top_row_format)\n host_overview_worksheet.write('C2', 'IP v4 Address', top_row_format)\n host_overview_worksheet.write('D2', 'IP v6 Address', top_row_format)\n host_overview_worksheet.write('E2', 'MAC Address', top_row_format)\n host_overview_worksheet.write('F2', 'MAC Vendor', top_row_format)\n host_overview_worksheet.write('G2', 'Operating System', top_row_format)\n host_overview_worksheet.write('H2', 'Host Type', top_row_format)\n\n \"\"\"Query the database for the hosts\"\"\"\n inventory_hosts = session.query(InventoryHost).all()\n\n \"\"\"Build overview worksheet\"\"\"\n overview_row = 2\n overview_col = 1\n for host in inventory_hosts:\n host_overview_worksheet.write(overview_row, overview_col, host.host_name, host_row_format)\n host_overview_worksheet.write(overview_row, overview_col + 1, host.ipv4_addr, host_row_format)\n host_overview_worksheet.write(overview_row, overview_col + 2, host.ipv6_addr, host_row_format)\n host_overview_worksheet.write(overview_row, overview_col + 3, host.macaddr, host_row_format)\n host_overview_worksheet.write(overview_row, overview_col + 4, host.mac_vendor.name, host_row_format)\n host_overview_worksheet.write(overview_row, overview_col + 5, host.product.name, host_row_format)\n host_overview_worksheet.write(overview_row, overview_col + 6, host.host_type, host_row_format)\n overview_row += 1\n\n \"\"\"Build detailed worksheet\"\"\"\n detail_row = 2\n detail_col = 1\n for host in inventory_hosts:\n\n \"\"\"Add the black row to start host detail info\"\"\"\n host_detail_worksheet.set_row(detail_row, 5)\n host_detail_worksheet.write(detail_row, detail_col, '', black_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 1, '', black_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 2, '', black_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 3, '', black_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 4, '', black_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 5, '', black_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 6, '', black_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 7, '', black_row_format)\n detail_row += 1\n\n \"\"\"Add row detail info\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, 'Hostname', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 1, 'IP v4 Address', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 2, 'IP v6 Address', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 3, 'MAC Address', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 4, 'MAC Vendor', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 5, 'Host Type', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 6, 'Operating System', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 7, 'Version', top_row_format)\n detail_row += 1\n\n \"\"\"Add host info\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, host.host_name, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 1, host.ipv4_addr, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 2, host.ipv6_addr, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 3, host.macaddr, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 4, host.mac_vendor.name, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 5, host.host_type, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 6, host.product.name, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 7, host.product.version, host_row_format)\n detail_row += 2\n\n \"\"\"If there is no host nse script, just say so.\"\"\"\n if not host.host_nse_scripts:\n host_detail_worksheet.write(detail_row, detail_col, 'Host NSE Script Name', top_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row, detail_col + 7,\n 'Output', top_row_format)\n detail_row += 1\n host_detail_worksheet.write(detail_row, detail_col, 'No Script Name', host_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row, detail_col + 7,\n 'No Script Output', host_row_wrapped_format)\n detail_row += 2\n else:\n\n \"\"\"Add the row detail\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, 'Host NSE Script Name', top_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row, detail_col + 7,\n 'Output', top_row_format)\n detail_row += 1\n\n \"\"\"Grab all the scripts\"\"\"\n for host_scripts in host.host_nse_scripts:\n\n \"\"\"Count output the lines so we know what to merge\"\"\"\n lines = host_scripts.output.count('\\n')\n\n if lines > 0:\n\n \"\"\"Merge the rows and write the name and output\"\"\"\n host_detail_worksheet.merge_range(detail_row, detail_col, detail_row + lines, detail_col,\n host_scripts.name, host_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row + lines, detail_col + 7,\n host_scripts.output, host_row_wrapped_format)\n detail_row += 1\n else:\n\n \"\"\"Single line output\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, host_scripts.name, host_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row + lines, detail_col + 7,\n host_scripts.output, host_row_wrapped_format)\n detail_row += 1\n\n if not host.inventory_svcs:\n\n \"\"\"If there are no services for this host tell me\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, 'Protocol', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 1, 'Port', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 2, 'Name', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 3, 'Svc Product', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 4, 'Extra Info', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 5, 'Product', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 6, 'Version', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 7, 'Update', top_row_format)\n detail_row += 1\n\n host_detail_worksheet.write(detail_row, detail_col, 'no services', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 1, 'no services', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 2, 'no services', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 3, 'no services', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 4, 'no services', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 5, 'no services', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 6, 'no services', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 7, 'no services', host_row_format)\n detail_row += 1\n\n else:\n for ports in host.inventory_svcs:\n\n \"\"\"Host services row info\"\"\"\n detail_row += 1\n host_detail_worksheet.write(detail_row, detail_col, 'Protocol', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 1, 'Port', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 2, 'Name', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 3, 'Svc Product', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 4, 'Extra Info', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 5, 'Product', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 6, 'Version', top_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 7, 'Update', top_row_format)\n detail_row += 1\n\n \"\"\"Write the service info\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, ports.protocol, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 1, ports.portid, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 2, ports.name, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 3, ports.svc_product, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 4, ports.extra_info, host_row_format)\n try:\n\n \"\"\"There may not be product info, but try.\"\"\"\n host_detail_worksheet.write(detail_row, detail_col + 5, ports.product.name, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 6, ports.product.version, host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 7, ports.product.product_update,\n host_row_format)\n detail_row += 1\n except AttributeError:\n\n \"\"\"Just write unknown if there is no product info\"\"\"\n host_detail_worksheet.write(detail_row, detail_col + 5, 'unknown', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 6, 'unknown', host_row_format)\n host_detail_worksheet.write(detail_row, detail_col + 7, 'unknown', host_row_format)\n detail_row += 1\n\n if not ports.svc_nse_scripts:\n\n \"\"\"If there is no NSE script info just say so.\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, 'Svc NSE Script Name', top_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row, detail_col + 7,\n 'Output', top_row_format)\n detail_row += 1\n host_detail_worksheet.write(detail_row, detail_col, 'No Script Name', host_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row, detail_col + 7,\n 'No Script Output', host_row_wrapped_format)\n detail_row += 2\n\n else:\n\n \"\"\"Service Script row detail\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, 'Svc NSE Script Name', top_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row, detail_col + 7,\n 'Output', top_row_format)\n detail_row += 1\n\n \"\"\"Grab all the scripts\"\"\"\n for nse_scripts in ports.svc_nse_scripts:\n\n \"\"\"Count the lines in the output for merging\"\"\"\n lines = nse_scripts.output.count('\\n')\n\n if lines > 0:\n\n \"\"\"Merge the rows and write the name and output\"\"\"\n host_detail_worksheet.merge_range(detail_row, detail_col, detail_row + lines, detail_col,\n nse_scripts.name, host_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row + lines, detail_col + 7,\n nse_scripts.output, host_row_wrapped_format)\n detail_row += 1\n else:\n\n \"\"\"Single line output\"\"\"\n host_detail_worksheet.write(detail_row, detail_col, nse_scripts.name, host_row_format)\n host_detail_worksheet.merge_range(detail_row, detail_col + 1, detail_row + lines,\n detail_col + 7, nse_scripts.output,\n host_row_wrapped_format)\n detail_row += 1\n\n detail_row += 1\n report.close()\n session.close()", "def onGrid(self, event):\n dlg = wx.FileDialog(self, wildcard=\"*.csv\", style=wx.SAVE)\n if dlg.ShowModal() == wx.ID_OK:\n path = dlg.GetPath()\n self.model.exportToGrid(path)\n dlg.Destroy()", "def __exportMenuTriggered(self, act):\n aw = self.activeWindow()\n if aw:\n exporterFormat = act.data()\n aw.exportFile(exporterFormat)", "def export(self):\n\n rpt_date = datetime.now()\n filename = 'bushfire_indicator_report_{}.xls'.format(rpt_date.strftime('%d%b%Y'))\n response = HttpResponse(content_type='application/vnd.ms-excel')\n response['Content-Disposition'] = 'attachment; filename=' + filename\n\n book = Workbook()\n self.get_excel_sheet(rpt_date, book)\n\n book.add_sheet('Sheet 2')\n book.save(response)\n\n return response", "def export(self):\n\n rpt_date = datetime.now()\n filename = 'bushfire_report_{}.xls'.format(rpt_date.strftime('%d%b%Y'))\n response = HttpResponse(content_type='application/vnd.ms-excel')\n response['Content-Disposition'] = 'attachment; filename=' + filename\n\n book = Workbook()\n self.ministerial.get_excel_sheet(rpt_date, book)\n self.ministerial_auth.get_excel_sheet(rpt_date, book)\n self.ministerial_268.get_excel_sheet(rpt_date, book)\n self.quarterly.get_excel_sheet(rpt_date, book)\n self.by_tenure.get_excel_sheet(rpt_date, book)\n self.by_cause.get_excel_sheet(rpt_date, book)\n self.region_by_tenure.get_excel_sheet(rpt_date, book)\n self.indicator.get_excel_sheet(rpt_date, book)\n self.by_cause_10YrAverage.get_excel_sheet(rpt_date, book)\n\n book.add_sheet('Sheet 1')\n book.save(response)\n\n return response", "def export_data(self):\r\n if len(app.entry6.get()) != 0:\r\n\r\n if app.var.get() == 'xls':\r\n\r\n wb = Workbook()\r\n sheet = wb.add_sheet('Sheet1')\r\n self.columns = ['id', 'Name', 'Section', 'Dept.', 'Gpa', 'MP1', 'MP2', 'MP3', 'MT', 'FINAL']\r\n style = xlwt.easyxf('font: bold 1')\r\n for col in range(10):\r\n sheet.write(0, col, self.columns[col], style)\r\n index=0\r\n for row in range(1,162):\r\n sheet.write(row, 1, open_data.sort_list[index])\r\n index += 1\r\n index1 = -1\r\n for row in range(1,162):\r\n index1 += 1\r\n index2=0\r\n for col in range(10):\r\n if col == 1 or index2 == 1:\r\n index2 += 1\r\n continue\r\n if index2 == 0:\r\n sheet.write(row, col, int(open_data.student[open_data.sort_list[index1]][index2]))\r\n index2 += 1\r\n continue\r\n sheet.write(row, col, open_data.student[open_data.sort_list[index1]][index2])\r\n index2 += 1\r\n file_name=app.entry6.get()\r\n if '.xls' not in file_name:\r\n wb.save(file_name+'.xls')\r\n else:\r\n wb.save(file_name)\r\n\r\n elif app.var.get() == 'txt':\r\n\r\n file_name = app.entry6.get()\r\n if '.txt' not in file_name:\r\n file_name = file_name + '.txt'\r\n file = open(file_name, 'w')\r\n index2 = 0\r\n for key in open_data.student:\r\n for index in range(10):\r\n if index == 0:\r\n file.write(str(int(open_data.student[key][index])))\r\n file.write(', ')\r\n continue\r\n if index == 1:\r\n try:\r\n self.split_names = open_data.sort_list[index2].split(' ')\r\n file.write(self.split_names[0])\r\n file.write(', ')\r\n file.write(self.split_names[1])\r\n file.write(', ')\r\n index2 += 1\r\n except UnicodeEncodeError:\r\n index2 += 1\r\n pass\r\n continue\r\n if index >= 5 and index <= 9:\r\n if open_data.student[key][index] != '':\r\n file.write(str(int(open_data.student[key][index])))\r\n file.write(', ')\r\n else:\r\n file.write('\\n')\r\n break\r\n if index == 9:\r\n file.write('\\n')\r\n continue\r\n try:\r\n file.write(str(open_data.student[key][index]))\r\n file.write(', ')\r\n except UnicodeEncodeError:\r\n pass\r\n file.close()\r\n\r\n\r\n\r\n elif app.var.get() == 'csv':\r\n app.info.configure(text=\"INFO: Type not Supported\")\r\n # The program does not support saving in 'csv' type. If the user selects 'csv' file type, 'Info' Label\r\n # shows the message: 'INFO: Type not Supported'.\r\n\r\n else:\r\n app.info.configure(text='INFO: Type not chosen!')\r\n # Also, If the user presses on 'Export Data' button, with a file name provided, but without choosing a\r\n # file type, 'Info' Label shows the message: 'INFO: Type not chosen'.\r\n\r\n else:\r\n app.info.configure(text=\"INFO: Please provide the name of the file.\")\r\n # Also, if the user presses 'Export Data' button without giving a file name, 'Info' Label shows the message:\r\n # 'INFO: Please provide the name of the file.'\r", "def print_xlsx(self):\n if self.date_from and self.date_to:\n if self.date_from > self.date_to:\n raise ValidationError(\"Date From must be less than Date To\")\n\n # active_record = self._context['id']\n # record = self.env['room.accommodation'].browse(active_record)\n data = {\n 'date_from': self.date_from,\n 'date_to': self.date_to,\n 'guest_id': self.guest_id.id,\n 'model_id': self.id,\n 'check_out': self.check_out,\n 'date_today': fields.Datetime.now()\n }\n\n print(\"XLSX Wizard data : \", data)\n\n return {\n 'type': 'ir.actions.report',\n 'data': {\n 'model': 'accommodation.reporting',\n 'options': json.dumps(data, default=date_utils.json_default),\n 'output_format': 'xlsx',\n 'report_name': 'Accommodation Report'\n },\n 'report_type': 'xlsx'\n }", "def export_order_status_button(cls, store_views):\n pass", "def export(self):\n\n rpt_date = datetime.now()\n filename = 'bushfire_by_cause_report_{}.xls'.format(rpt_date.strftime('%d%b%Y'))\n response = HttpResponse(content_type='application/vnd.ms-excel')\n response['Content-Disposition'] = 'attachment; filename=' + filename\n\n book = Workbook()\n self.get_excel_sheet(rpt_date, book)\n\n book.add_sheet('Sheet 2')\n book.save(response)\n\n return response", "def renderBuySellSignalEMA1226MACD(self, saveFile='', saveOnly=False):\n\n buysignals = ((self.df.ema12gtema26co == True) & (self.df.macdgtsignal == True) & (self.df.obv_pc >= 2)) | ((self.df.ema12gtema26 == True) & (self.df.macdgtsignal == True) & (self.df.obv_pc >= 5)) \n sellsignals = ((self.df.ema12ltema26co == True) & (self.df.macdltsignal == True)) | ((self.df.ema12gtema26 == True) & (self.df.macdltsignal == True) & (self.df.obv_pc < 0))\n df_signals = self.df[(buysignals) | (sellsignals)]\n\n ax1 = plt.subplot(211)\n plt.plot(self.df.close, label=\"price\", color=\"royalblue\")\n plt.plot(self.df.ema12, label=\"ema12\", color=\"orange\")\n plt.plot(self.df.ema26, label=\"ema26\", color=\"purple\")\n plt.ylabel('Price')\n\n action = ''\n last_action = ''\n for idx, row in df_signals.iterrows():\n if row['ema12gtema26co'] == True and row['macdgtsignal'] == True and last_action != 'buy':\n action = 'buy'\n plt.axvline(x=idx, color='green')\n elif row['ema12ltema26'] == True and row['macdltsignal'] == True and action == 'buy':\n action = 'sell'\n plt.axvline(x=idx, color='red')\n\n last_action = action\n\n plt.xticks(rotation=90)\n\n plt.subplot(212, sharex=ax1)\n plt.plot(self.df.macd, label=\"macd\")\n plt.plot(self.df.signal, label=\"signal\")\n plt.legend()\n plt.ylabel('Divergence')\n plt.xticks(rotation=90)\n\n plt.tight_layout()\n plt.legend()\n\n try:\n if saveFile != '':\n plt.savefig(saveFile)\n except OSError:\n raise SystemExit('Unable to save: ', saveFile) \n\n if saveOnly == False:\n plt.show()", "def generate_xlsx_report(self, workbook, data, parts_data):\n worksheet = workbook.add_worksheet(\"daily_parts_issuance_wizard\")\n worksheet.set_column(0, 0, 10)\n worksheet.set_column(1, 1, 15)\n worksheet.set_column(2, 2, 20)\n worksheet.set_column(3, 3, 15)\n worksheet.set_column(4, 4, 10)\n worksheet.set_column(5, 5, 12)\n worksheet.set_column(6, 6, 10)\n worksheet.set_column(7, 7, 10)\n worksheet.set_column(8, 8, 15)\n worksheet.set_column(9, 9, 10)\n worksheet.set_column(10, 10, 15)\n worksheet.set_column(11, 11, 10)\n worksheet.set_column(12, 12, 20)\n worksheet.set_column(13, 13, 5)\n worksheet.set_column(14, 14, 5)\n worksheet.set_column(15, 15, 5)\n\n bold = workbook.add_format(\n {\"bold\": True, \"font_name\": \"Arial\", \"font_size\": \"10\"}\n )\n tot = workbook.add_format(\n {\"border\": 2, \"bold\": True, \"font_name\": \"Arial\", \"font_size\": \"10\"}\n )\n border = workbook.add_format(\n {\"border\": 2, \"font_name\": \"Arial\", \"font_size\": \"10\"}\n )\n merge_format = workbook.add_format({\"border\": 2, \"align\": \"center\"})\n format1 = workbook.add_format(\n {\"border\": 2, \"bold\": True, \"font_name\": \"Arial\", \"font_size\": \"10\"}\n )\n format1.set_bg_color(\"gray\")\n date = workbook.add_format({\"num_format\": \"dd/mm/yy\"})\n\n worksheet.merge_range(\"C3:F3\", \"Merged Cells\", merge_format)\n\n row = 0\n row += 1\n row += 1\n worksheet.write(row, 2, \"DAILY PARTS ISSUANCE\", tot)\n row += 1\n worksheet.write(row, 2, \"Date From:\", tot)\n worksheet.write(row, 3, data[\"form\"][\"date_from\"] or \"\", border)\n worksheet.write(row, 4, \"To:\", tot)\n worksheet.write(row, 5, data[\"form\"][\"date_to\"] or \"\", border)\n row += 2\n worksheet.write(row, 0, \"CMF\", bold)\n row = 3\n\n for objec in self.get_work_order_detail(data[\"form\"]):\n row += 3\n worksheet.write(row, 0, \"DATE ISSUED :\", bold)\n worksheet.write(row, 1, objec.get(\"date\") or \"\", date)\n row += 2\n worksheet.write(row, 0, \"NO.\", format1)\n worksheet.write(row, 1, \"WO NO.\", format1)\n worksheet.write(row, 2, \"VEHICLE ID\", format1)\n worksheet.write(row, 3, \"PART NO.\", format1)\n worksheet.write(row, 4, \"PART NAME\", format1)\n worksheet.write(row, 5, \"VEHICLE MAKE\", format1)\n worksheet.write(row, 6, \"USED\", format1)\n worksheet.write(row, 7, \"UNIT TYPE\", format1)\n worksheet.write(row, 8, \"OLD PART RETURND\", format1)\n worksheet.write(row, 9, \"ISSUED BY\", format1)\n worksheet.write(row, 10, \"REMARKS\", format1)\n line_row = row + 1\n line_col = 0\n counter = 1\n for obj in objec.get(\"value\"):\n worksheet.write(line_row, line_col, counter, border)\n line_col += 1\n worksheet.write(line_row, line_col, obj.get(\"wo_name\") or \"\", border)\n line_col += 1\n worksheet.write(line_row, line_col, obj.get(\"vehicle_id\") or \"\", border)\n line_col += 1\n worksheet.write(line_row, line_col, obj.get(\"part_no\") or \"\", border)\n line_col += 1\n worksheet.write(line_row, line_col, obj.get(\"part_name\") or \"\", border)\n line_col += 1\n worksheet.write(\n line_row, line_col, obj.get(\"vehicle_make\") or \"\", border\n )\n line_col += 1\n worksheet.write(line_row, line_col, obj.get(\"qty\") or \"\", border)\n line_col += 1\n worksheet.write(line_row, line_col, obj.get(\"uom\") or \"\", border)\n line_col += 1\n worksheet.write(\n line_row, line_col, obj.get(\"old_part_return\") or \"\", border\n )\n line_col += 1\n worksheet.write(line_row, line_col, obj.get(\"issued_by\") or \"\", border)\n line_col += 1\n worksheet.write(line_row, line_col, obj.get(\"remarks\") or \"\", border)\n line_col = 0\n line_row += 1\n counter += 1\n worksheet.write(line_row, line_col, \"********\", border)", "def export(self):\n\n rpt_date = datetime.now()\n filename = 'bushfire_by_tenure_report_{}.xls'.format(rpt_date.strftime('%d%b%Y'))\n response = HttpResponse(content_type='application/vnd.ms-excel')\n response['Content-Disposition'] = 'attachment; filename=' + filename\n\n book = Workbook()\n self.get_excel_sheet(rpt_date, book)\n\n book.add_sheet('Sheet 2')\n book.save(response)\n\n return response", "def export(self):\n\n rpt_date = datetime.now()\n filename = 'quarterly_report_{}.xls'.format(rpt_date.strftime('%d%b%Y'))\n response = HttpResponse(content_type='application/vnd.ms-excel')\n response['Content-Disposition'] = 'attachment; filename=' + filename\n\n book = Workbook()\n self.get_excel_sheet(rpt_date, book)\n\n book.add_sheet('Sheet 2')\n book.save(response)\n\n return response", "def on_action_2_triggered(self):\n # TODO: not implemented yet\n model = self.model\n self.doExport(model)", "def click_generate_batch_report_button(self):\n self._basket.click_generate_batch_report_button()", "def export(self):\n rpt_date = datetime.now()\n filename = 'bushfire_regionbytenure_report_{}.xls'.format(rpt_date.strftime('%d%b%Y'))\n response = HttpResponse(content_type='application/vnd.ms-excel')\n response['Content-Disposition'] = 'attachment; filename=' + filename\n\n book = Workbook()\n self.get_excel_sheet(rpt_date, book)\n\n book.add_sheet('Sheet 2')\n book.save(response)\n\n return response", "def export_ho_dan_as_excel_action(fields=None, exclude=None, header=True):\n def export_as_excel(modeladmin, request, queryset):\n opts = modeladmin.model._meta\n field_names = [\"name\", \"status\", \"location\", \"tinh\",\n \"xa\", \"huyen\", \"phone\", \"cuuho\", \"update_time\", \"note\"]\n display_names = [\"Tên hộ dân\", \"Tình trạng\", \"Vị trí\", \"Tỉnh\", \"Xã\",\n \"Huyện\", \"Sdt\", \"hỗ trợ\", \"Thời gian cuối cùng cập nhật\", \"Ghi chú\"]\n file_name = \"Danh_sach_ho_dan\"\n\n output = io.BytesIO()\n\n workbook = xlsxwriter.Workbook(output, {'in_memory': True})\n worksheet = workbook.add_worksheet()\n row = 0\n if header:\n write_a_row(worksheet, row, display_names)\n row += 1\n for obj in queryset:\n arr = []\n for field in field_names:\n if field == \"status\" and obj.status:\n arr.append(obj.status.name)\n elif field == \"update_time\":\n utc_time = getattr(obj, field)\n local_datetime = utc_to_local(utc_time)\n arr.append(local_datetime.strftime(\"%d/%m/%Y %H:%M:%S\"))\n else:\n arr.append(str(getattr(obj, field) or \"\"))\n write_a_row(worksheet, row, arr)\n row += 1\n\n workbook.close()\n\n output.seek(0)\n\n response = HttpResponse(output.read(\n ), content_type=\"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet\")\n response['Content-Disposition'] = f\"attachment; filename={file_name}.xlsx\"\n\n output.close()\n\n return response\n\n export_as_excel.short_description = \"Xuất file excel\"\n return export_as_excel", "def export(self):\n\n rpt_date = datetime.now()\n filename = 'bushfire_by_cause_10yr_average_report_{}.xls'.format(rpt_date.strftime('%d%b%Y'))\n response = HttpResponse(content_type='application/vnd.ms-excel')\n response['Content-Disposition'] = 'attachment; filename=' + filename\n\n book = Workbook()\n self.get_excel_sheet(rpt_date, book)\n\n book.add_sheet('Sheet 2')\n book.save(response)\n\n return response", "def dataframe_to_excel(df, sheet_title, project_constants_lst, \n current_date=str(date.today()), force_flag = False, freeze_column='A'):\n \n project_steps_df, max_title, _, report_requisites_sr, *_ = project_constants_lst\n report_type, export_flag, df_decription = project_steps_df.loc[sheet_title, ['report_type', 'export_to_excel', 'description']].values\n \n # check DataFrame report type to save\n if report_type == 'report':\n report_mark = report_requisites_sr['project_title'] + '_tables'\n else:\n report_mark = report_type\n \n # construct excel filename\n file_name = report_requisites_sr['customer_name'] + '_' + report_mark + '_' + current_date + '.xlsx'\n\n # information string\n info = f'Exporting {sheet_title} table to {report_mark} file'\n print(info, end =\" \")\n file_path = os.path.join(report_requisites_sr['today_report_folder'], file_name)\n \n # save DataFrame to excel file if export_to_excel trigger is ON\n # and DataFrame is not empty\n if (force_flag or export_flag) and not df.empty:\n fsop.create_folder(report_requisites_sr['today_report_folder'], max_title, display_status=False)\n file_mode = 'a' if os.path.isfile(file_path) else 'w'\n df = df.apply(pd.to_numeric, errors='ignore')\n try:\n if_sheet_exists_param = 'replace' if file_mode == 'a' else None\n content_df, item_exist = generate_table_of_contents(file_path, file_mode, sheet_title, df_decription)\n df_flat = drop_multindex(df)\n # write table of contents and data dataframe to the excel file\n with pd.ExcelWriter(file_path, mode=file_mode, if_sheet_exists=if_sheet_exists_param, engine='openpyxl') as writer:\n if file_mode == 'w' or not item_exist:\n content_df.to_excel(writer, sheet_name='Содержание', index=False)\n df_flat.to_excel(writer, sheet_name=sheet_title, startrow=2, index=False)\n # format table of contents and data worksheets\n workbook = openpyxl.load_workbook(file_path)\n format_workbook(workbook, sheet_title, df_decription, freeze_column)\n workbook.save(file_path)\n except PermissionError:\n status_info('fail', max_title, len(info))\n print('\\nPermission denied. Close the file.\\n')\n sys.exit()\n else:\n status_info('ok', max_title, len(info))\n return file_path \n else:\n # if save key is on but DataFrame empty\n if project_steps_df.loc[sheet_title, 'export_to_excel'] and df.empty:\n status_info('no data', max_title, len(info))\n else: \n status_info('skip', max_title, len(info))\n return None", "def export_to_excel(self, worksheet, row_start, col_start, queryset, date_time=timezone.now()):\n\t\tif queryset:\n\t\t\t[row_write, col_write] = self.excel_write_header_and_format(worksheet, row_start, col_start)\n\t\t\tfor q in queryset:\n\t\t\t\t# object_excel_write function---date_time uyiin history objectiig excel -ruu horvuulne\n\t\t\t\t[row_write, col_write] = q.object_excel_write(worksheet, row_write, col_write, date_time=date_time)\n\t\telse:\n\t\t\tworksheet.write_string(row_start, col_start, u'Мэдээлэл байхгүй')" ]
[ "0.750306", "0.74308896", "0.7306206", "0.7060795", "0.68989515", "0.68989515", "0.6491007", "0.6120849", "0.591677", "0.5913262", "0.58993304", "0.56994367", "0.569883", "0.56973416", "0.5684809", "0.56796527", "0.56629676", "0.5620759", "0.5603418", "0.55868685", "0.5576812", "0.55666053", "0.5562649", "0.55548066", "0.5540589", "0.5528343", "0.5526585", "0.5524354", "0.54975826", "0.5487687" ]
0.86254066
0
Implementing select buy and sell deal management grid first row checkbox functionality
def select_buy_and_sell_deal_management_grid_first_row_checkbox(self): self.select_grid_row_checkbox(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_checkbox_column_number, True)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def select_target_buy_list_overrides_grid_row_checkbox(self):\n for i in range(2):\n target_buy_list_overrides_grid_checkbox_locator = (By.XPATH, \"(//div[@data-grid-name='%s']/descendant::input[@type='checkbox'])[%s]\" % (self.target_buy_list_overrides_data_grid_name, str(i + 2)))\n self.click_element(target_buy_list_overrides_grid_checkbox_locator, True)", "def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)", "def GridCheck(Parent,DefaultSelected,Row,Column):\r\n dummyvar = IntVar()\r\n C = Checkbutton(Parent,var=dummyvar)\r\n if DefaultSelected == 1:\r\n C.select()\r\n C.grid(row=Row,column=Column)\r\n C.isChecked = dummyvar\r\n return C", "def select_vendor_price_list_grid_row_checkbox(self):\n for row in range(0,2):\n column_locator = (By.XPATH, \"//div[@id='%s']/descendant::th[@data-title='%s']\" % (self.vendor_price_list_grid_div_id, self.price_list_column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n column_value_locator = (By.XPATH, \"//div[@id='%s']/descendant::tbody/descendant::tr[%s]/td[%s]/span\" % (self.vendor_price_list_grid_div_id, (row + 1), str(column_index)))\n column_value_element = self.wait().until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n price_list_item = column_value_element.text\n self.compare_price_lists_list.append(price_list_item)\n self.select_grid_row_checkbox(self.customer_price_list_grid_div_id, price_list_item, self.select_grid_row_checkbox_column_number)", "def select_adjustment_charge_grid_first_row_checkbox(self):\n self.click_element(self.adjustment_charges_grid_first_row_checkbox_locator, True, True)", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def DoCheck(self,event):\r\n index = event.GetSelection()\r\n item = self.items[index]\r\n if self.list.IsChecked(index):\r\n self.data.check(item)\r\n else:\r\n self.data.uncheck(item)\r\n #self.list.SetSelection(index)\r", "def select_add_disputes_grid_row_checkbox(self, list_of_audit_account):\n self.select_grid_row_checkbox(self.add_disputes_grid_div_id, list_of_audit_account, self.add_disputes_grid_checkbox_column_number)\n self.wait_for_ajax_spinner_load()", "def enableAddOrder(self):\n if self.lstType.selected == -1 or self.lstRes.selected == -1:\n self.btnAddMarketOrder.disable()\n else:\n self.btnAddMarketOrder.enable()\n if self.lstType.getSelectedItem().data == 'sell':\n self.txtMin.readOnly = 0\n else:\n self.txtMax.readOnly = 0", "def select_charges_grid_row_checkbox(self, charge_name):\n self.select_grid_row_checkbox(self.charges_grid_div_id, charge_name, self.charges_grid_checkbox_column_number)", "def select_statement_details_disputes_grid_first_row_checkbox(self):\n self.click_element(self.statement_details_dispute_grid_first_row_checkbox_locator, True, True)", "def populate_filter(self, src, col, tbl):\r\n sub_cats = set([c for l in kit.SQL_pull(col, tbl) for c in l])\r\n select = {g: tk.BooleanVar() for g in sub_cats}\r\n for key in select:\r\n select[key].set(True)\r\n \r\n line = 1\r\n for g in sub_cats:\r\n line += 1\r\n tk.Checkbutton(src,\r\n text = g,\r\n variable = select[g],\r\n command = lambda x = tbl,\r\n y = False,\r\n z = select: self.checked(x, y, z),\r\n **jt.filter_style).grid(row = line, \r\n sticky = tk.W)\r\n \r\n return select", "def select_transactions_grid_row_checkbox(self, row_identifier_text):\n self.select_grid_row_checkbox(self.transactions_grid_div_id, row_identifier_text, self.transactions_grid_checkbox_column_number)\n self.click_element(self.page_header_locator)", "def change_entries(self, grid_sizer, options, border):\n\n if (self.diag_type is None): return grid_sizer # no data selected, don't change anything\n\n # setup the grid of possible values\n header0 = wx.StaticText(self, -1, \"Add/Remove\")\n header1 = wx.StaticText(self, -1, \"Quantity Code\")\n header2 = wx.StaticText(self, -1, \"Name\")\n header3 = wx.StaticText(self, -1, \"LaTeX Formula\")\n grid_sizer.Add(header0, pos=(0,0), flag=options, border=border)\n grid_sizer.Add(header1, pos=(0,1), flag=options, border=border)\n grid_sizer.Add(header2, pos=(0,2), flag=options, border=border)\n grid_sizer.Add(header3, pos=(0,3), flag=options, border=border)\n grid_sizer.Add(wx.StaticLine(self), pos=(1,0), span=(1,4),\n flag=wx.ALL|wx.EXPAND|wx.GROW, border=border)\n\n self.selected_values = [] # keep track of selected quantities\n\n quantities = self.output_quantities.diagnostic_types[self.diag_type]\n\n # choose a good height/width for formulas\n if (self.diag_type in [\"Linear_Forces\", \"Angular_Momentum\", \"Energy_Flux\",\n \"Induction\", \"Inertial_Forces\", \"Lorentz_Forces\",\n \"Poynting_Flux\", \"TurbKE_Budget\"]):\n width = 100\n elif (self.diag_type == \"Thermal_Equation\"):\n width = 150\n else:\n width = 30\n height = 20\n\n row = 2\n iquant = 0\n for Q in quantities:\n but = wx.ToggleButton(self, Q.code, \"Add\") # build button and place it in second column\n but.Bind(wx.EVT_TOGGLEBUTTON, self.OnToggle)\n grid_sizer.Add(but, pos=(row,0), flag=options, border=border)\n\n q_code = wx.StaticText(self, -1, str(Q.code)) # build other column entries\n q_name = wx.StaticText(self, -1, Q.name) # name\n\n formula = self.RenderTeX(Q, size=(width,height))\n\n # place column entries\n grid_sizer.Add(q_code, pos=(row,1), flag=options, border=border)\n grid_sizer.Add(q_name, pos=(row,2), flag=options, border=border)\n grid_sizer.Add(formula, pos=(row,3), flag=options, border=border)\n\n iquant += 1\n\n # add horizontal line every 5 quantities\n if (iquant % 5 == 0):\n grid_sizer.Add(wx.StaticLine(self), pos=(row+1,0), span=(1,4),\n flag=wx.ALL|wx.EXPAND|wx.GROW, border=border)\n row_inc = 2\n else:\n row_inc = 1\n\n row += row_inc\n\n grid_sizer.AddGrowableCol(2,1) # make the name/formula columns \"1\" growable, i.e., grows as necessary\n grid_sizer.AddGrowableCol(3,1)\n\n return grid_sizer", "def select_non_traffic_charges_grid_row_checkbox(self, row_identifier_text):\n self.select_grid_row_checkbox(self.non_traffic_charges_grid_div_id, row_identifier_text, self.non_traffic_charges_grid_checkbox_column_number, first_row=True)", "def onTradeSelected(self, item):\n if self.lstTrade.getMultiSelectedItems() == []:\n self.btnCancelTrade.disable()\n else:\n self.btnCancelTrade.enable()", "def click_statement_fields_grid_first_row_checkbox(self):\n self.click_element(self.statement_fields_grid_first_row_checkbox_locator, True)", "def select_received_charges_grid_row_checkbox(self, row_identifier_text):\n self.select_grid_row_checkbox(self.received_charges_grid_div_id, row_identifier_text, self.received_charges_grid_checkbox_column_number)", "def on_row_selected(self, obj_name):\n if obj_name == 'none':\n for act in self.app.ui.menuobjects.actions():\n act.setChecked(False)\n return\n\n # get the name of the selected objects and add them to a list\n name_list = []\n for obj in self.get_selected():\n name_list.append(obj.options['name'])\n\n # set all actions as unchecked but the ones selected make them checked\n for act in self.app.ui.menuobjects.actions():\n act.setChecked(False)\n if act.text() in name_list:\n act.setChecked(True)", "def onCheckbuttonChecked(self):\n if self.CheckButton.isChecked() == True:\n if not self.LLN_Node:\n self.Stats.scalarSelector.setCurrentNode(None)\n self.NativeT1_Selector.setCurrentNode(None) \n self.Stats.scalarSelector.enabled = False\n self.NativeT1_Selector.enabled = False\n else:\n self.Stats.scalarSelector.setCurrentNode(self.T1_LLN_Node)\n self.NativeT1_Selector.setCurrentNode(self.T1_LLN_Node)\n self.Stats.scalarSelector.enabled = False\n self.NativeT1_Selector.enabled = False\n if self.LLE_Node:\n self.Stats.scalarSelector2.setCurrentNode(self.T1_LLE_Node)\n self.EnhancedT1_Selector.setCurrentNode(self.T1_LLE_Node)\n self.Stats.scalarSelector2.enabled = False\n self.EnhancedT1_Selector.enabled = False \n else:\n self.Stats.scalarSelector2.setCurrentNode(None)\n self.EnhancedT1_Selector.setCurrentNode(None) \n self.Stats.scalarSelector2.enabled = False\n self.EnhancedT1_Selector.enabled = False \n else:\n self.Stats.scalarSelector.setCurrentNode(None)\n self.Stats.scalarSelector2.setCurrentNode(None)\n self.NativeT1_Selector.setCurrentNode(None)\n self.EnhancedT1_Selector.setCurrentNode(None)\n self.Stats.scalarSelector.enabled = True\n self.NativeT1_Selector.enabled = True\n self.Stats.scalarSelector2.enabled = True\n self.EnhancedT1_Selector.enabled = True", "def click_vendor_price_list_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.vendor_price_list_grid_div_id, None, self.view_price_list_column_number, True)", "def _generateCellCheckedState(self, obj, **args):\n result = []\n if self._script.utilities.hasMeaningfulToggleAction(obj):\n oldRole = self._overrideRole(pyatspi.ROLE_CHECK_BOX, args)\n result.extend(self.generate(obj, **args))\n self._restoreRole(oldRole, args)\n\n return result", "def click_target_buy_policy_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, column_number=2, first_row=True)\n self.wait_for_ajax_spinner_load()", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def click_target_buy_policies_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, self.target_buy_policies_grid_inline_action_column_number, first_row=True)", "def is_reference_rates_grid_accept_column_checkbox(self):\n is_checkbox = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (self.vendor_price_list_detail_reference_rates_grid_div_id, self.accept_column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n column_type_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tr/td[%s]/input\" % (self.vendor_price_list_detail_reference_rates_grid_div_id, str(column_index)))\n column_type_element = self.wait().until(EC.presence_of_element_located(column_type_locator), 'column type locator not found before specified time out')\n column_type = column_type_element.get_attribute('type')\n if column_type == \"checkbox\":\n is_checkbox = True\n return is_checkbox", "def onIndustrySelected(self, item):\n if self.lstIndustry.getMultiSelectedItems() == []:\n self.btnRemoveIndustry.disable()\n self.btnUpgradeIndustry.disable()\n else:\n self.enableButtons(self.lstIndustry, [self.btnRemoveIndustry, self.btnUpgradeIndustry])", "def on_source_checked(self, source_item):\n # don't bother for non-checked events\n if not source_item.has_checked_changed():\n return\n # leave checked for as long as search is running\n if self.is_searching():\n source_item.setCheckState(Qt.Checked)\n return\n # keep to the same if exporting\n if self.is_exporting():\n if source_item.is_checked:\n source_item.setCheckState(Qt.Checked)\n else:\n source_item.setCheckState(Qt.Unchecked)\n return\n is_checked = source_item.current_state()\n for key, geometry in source_item.geometries.iteritems():\n if is_checked:\n geometry.enable_source(source_item.title)\n else:\n geometry.disable_source(source_item.title)\n for key, type_entry in source_item.type_entries.iteritems():\n if is_checked:\n type_entry.enable_source(source_item.title, self.geometries)\n else:\n type_entry.disable_source(source_item.title, self.geometries)\n source_item.update_checked()", "def onToSystemSelected(self, item):\n if not item:\n self.btnAddTradeRoute.disable()\n self.btnTradeAllRoute.disable()\n self.btnAddOneTimeTrade.disable()\n else:\n self.btnAddTradeRoute.enable()\n self.btnTradeAllRoute.enable()\n self.btnAddOneTimeTrade.enable()", "def onSelected(self, item):\n if not item:\n self.clearBidData()\n else:\n self.enableAddOrder()" ]
[ "0.63728535", "0.6276276", "0.62166536", "0.61831915", "0.6053796", "0.59584445", "0.5842777", "0.5797149", "0.57607424", "0.5707139", "0.5612971", "0.5531448", "0.5506551", "0.5497594", "0.5481814", "0.54728425", "0.54640925", "0.5431332", "0.538145", "0.5380802", "0.53794587", "0.53445077", "0.5343553", "0.53397954", "0.531875", "0.53083813", "0.5303139", "0.5292009", "0.52885085", "0.52568805" ]
0.80763716
0
Implementing click buy and sell deal bulk edit button functionality
def click_buy_and_sell_deal_bulk_edit_button(self): self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def edit_tools(self, e):\n #GETTING SELECTION\n\n self.selected_item = self.user_inventory.selection()\n self.select_name = self.user_inventory.item([i for i in self.selected_item], \"values\")[0]\n self.select_entdate = self.user_inventory.item([i for i in self.selected_item], \"values\")[3]\n\n self.df_same_name = self.df_user.query(\"title == @self.select_name\")\n #this is the selected one for sure\n self.df_the_selected_item = self.df_same_name.loc[self.df_same_name[\"entry date\"] == self.select_entdate]\n\n #GETTING THE INDEX NUMBER OF THE SELECTION IN .CSV FILE\n self.index_select = self.df_the_selected_item.index\n self.index_select_number = self.index_select.tolist()\n\n #bottom buttons appear:\n self.changing_item_label.config(text=\"Now editing \"+self.select_name+\" that added on \"+self.select_entdate+\":\")\n\n self.delete_but = Button (self.bottom_frame, text=\"DELETE\", command=self.delete_button)\n self.delete_but.place(relx=0.1, rely=0.7, relwidth=0.28, anchor=\"w\")\n\n self.servings_list = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]\n self.serv_drop = Combobox(self.bottom_frame, value=self.servings_list, state=\"readonly\")\n self.serv_drop.place(relx=0.5, rely=0.7, relwidth=0.2, anchor=CENTER)\n\n \n self.serv_but = Button(self.bottom_frame, text=\"CHANGE AMOUNT\", command=self.change_amount_button, state=\"disabled\")\n self.serv_but.place(relx=0.9, rely=0.7, relwidth=0.28, anchor=\"e\")\n\n self.serv_drop.bind(\"<<ComboboxSelected>>\", self.activate_button)", "def click_bulk_edit_button(self):\n self.click_element(self.bulk_edit_button_locator, True)", "def click_statement_fields_bulk_edit_button(self):\n self.click_element(self.statement_fields_bulk_edit_button_locator, True)", "def click_bulk_edit_submit_button(self):\n self.click_element(self.bulk_edit_submit_button_locator, True)", "def click_bulk_edit_submit_button(self):\n self.click_element(self.bulk_edit_submit_button_locator)", "def edit(self, **kwargs):\n ...", "def edit(self):\n\n pass", "def editChange(self,editBtn):\n if self.edit ==True:\n self.updateDetails(\"normal\") #update details column\n self.edit = False #switch boolean\n self.editBtn.setText(\"Edit\") #update button text\n else:\n self.updateDetails(\"edit\") #update details column\n self.edit= True #switch boolean\n self.editBtn.setText(\"Stop Editing\") #update button text", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)", "def __actions__(self, obj):\n bool_ultimo = obj.bool_ultimo \n primary_fields = self.__provider__.get_primary_fields(self.__entity__)\n pklist = '/'.join(map(lambda x: str(getattr(obj, x)), primary_fields))\n\n if bool_ultimo == 1:\n cod_item = obj.cod_item\n value = '<div>'\n if has_permission('editar_item'):\n value = value + '<div><a class=\"edit_link\" href=\"'+pklist+'/edit\" style=\"text-decoration:none\">edit</a></div>'\n if has_permission('eliminar_relacion'):\n value = value + '<div><form method=\"POST\" action=\"'+pklist+'\" class=\"button-to\"><input type=\"hidden\" name=\"_method\" value=\"DELETE\" /><input class=\"delete-button\" onclick=\"return confirm(\\'Are you sure?\\');\" value=\"delete\" type=\"submit\" style=\"background-color: transparent; float:left; border:0; color: #286571; display: inline; margin: 0; padding: 0;\"/></form></div>'\n value = value + '<div><a class=\"relacion_link\" href=\"../relacions/?iid='+pklist+'\">Relaciones </a><br/><a class=\"versiones_link\" href=\"./?codi='+cod_item+'\">Revertir</a></div></div>'\n \n else:\n id_item_rev = DBSession.query(Item).filter_by(cod_item = obj.cod_item, bool_ultimo = 1).one().id_item\n ids = str(pklist) + \"-\" + str(id_item_rev)\n href = \"./revertir/?ids=\" + ids\n value = '<div><div><a class=\"edit_link\" href=\"'+pklist+'/edit\" style=\"text-decoration:none\">edit</a>'\\\n '</div><div>'\\\n '<form method=\"POST\" action=\"'+pklist+'\" class=\"button-to\">'\\\n '<input type=\"hidden\" name=\"_method\" value=\"DELETE\" />'\\\n '<input class=\"delete-button\" onclick=\"return confirm(\\'Are you sure?\\');\" value=\"delete\" type=\"submit\" '\\\n 'style=\"background-color: transparent; float:left; border:0; color: #286571; display: inline; margin: 0; padding: 0;\"/>'\\\n '</form>'\\\n '<a class=\"relacion_link\" href=\"../relacions/?iid='+pklist+'\">Relaciones </a>'\\\n '<a class=\"volver_link\" href=\"'+href+'\">Volver a</a>'\\\n '</div></div>'\n\n return value", "def on_edit_clicked(self,button):\n\t\tself.list_partitions.edit_partition()", "def click_target_buy_list_overrides_bulk_edit_button(self):\n self.click_element(self.target_buy_list_overrides_bulk_edit_button_locator, error_message='target buy list overrides bulk edit button locator not found before specified time out')", "def button_confirm(self):\n for record in self:\n res = super(PurchaseOrder, self).button_confirm()\n if not record.sale_order_id and record.customer_id:\n sale_order_line_obj = self.env['sale.order.line']\n attachment_ids = []\n sale_order_obj = self.env['sale.order']\n for attchment in record.attachment_ids:\n attachment_ids.append((0, 0, {\n 'name': attchment.name,\n 'datas': attchment.datas,\n \"description\": attchment.description,\n \"mimetype\": attchment.mimetype,\n 'index_content': attchment.index_content,\n \"create_uid\": attchment.create_uid.id,\n }))\n\n vals = {\n \"partner_id\": record.customer_id.id,\n \"vendor_id\": record.partner_id.id,\n \"purchase_order_id\": record.id,\n \"attachment_ids\": attachment_ids,\n \"colour_instructions\": record.colour_instructions,\n \"packing\": record.packing,\n \"name\": record.name,\n \"face_stamp\": record.face_stamp,\n \"selvedge\": record.selvedge,\n \"shipping_mark\": record.shipping_mark,\n \"shipping_sample_book\": record.shipping_sample_book,\n \"notes\": record.notes,\n \"shipment_date\": record.shipment_date,\n \"destination_id\": record.destination_id.id,\n \"currency_id\": record.currency_id.id,\n }\n sale_order = sale_order_obj.create(vals)\n record.sale_order_id = sale_order.id\n for line in record.order_line:\n taxes = line.product_id.taxes_id\n fpos = record.fiscal_position_id\n taxes_id = fpos.map_tax(taxes, line.product_id, record.partner_id) if fpos else taxes\n if taxes_id:\n taxes_id = taxes_id.filtered(lambda x: x.company_id.id == record.company_id.id)\n sale_order_line = sale_order_line_obj.create({'product_id': line.product_id.id,\n 'name': line.name,\n 'tax_id': [(6, 0, taxes_id.ids)],\n 'product_uom_qty': line.product_qty,\n \"product_uom\": line.product_uom.id,\n 'price_unit': line.price_unit,\n \"order_id\": sale_order.id,\n # \"discount\": line.discount,\n \"purchase_order_line_id\": line.id,\n \"actual_qty\": line.actual_qty\n })\n line.sale_order_line_id = sale_order_line.id\n\n return res", "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def edit():", "def item_update(request):\n if request.method == 'POST':\n item_to_update = get_object_or_404(StockItem, pk=request.POST['id'])\n item_to_update.name = request.POST['name']\n item_to_update.count = int(request.POST['count'])\n item_to_update.date_of_expiration = request.POST['exp']\n item_to_update.fk_category = Category.objects.get(name=request.POST['cat'])\n item_to_update.fk_subcategory = SubCategory.objects.get(name=request.POST['subcat'])\n item_to_update.notes = request.POST['notes']\n item_to_update.save()\n return HttpResponse(status=200)", "def edit_item(self):\n for item in self.selection():\n origin_name = self.set(item)[\"1\"]\n origin_url = self.set(item)[\"2\"]\n popup = GetItemURLDialogue(self, \"Edit Item\", origin_name, origin_url)\n\n self.item(item, values=(popup.name, popup.url, self.set(item)[\"3\"]))\n self.set(item)[\"2\"] = popup.url\n\n # Edit the item - backend\n for row in s.item:\n if row[\"item\"] == origin_name and row[\"url\"] == origin_url:\n s.item.remove(row)\n s.updateItem({\"item\": popup.name, \"url\": popup.url})", "def edit_from_list(id=None,item_id=None):\n setExits()\n #import pdb;pdb.set_trace()\n \n item_id=cleanRecordID(item_id)\n item_rec = None\n rec = None\n warehouses = Warehouse(g.db).select()\n trx_types = get_site_config().get('trx_types',['Add','Remove',])\n transaction = Transaction(g.db)\n trx_id = cleanRecordID(id)\n if trx_id > 0:\n rec = transaction.get(trx_id)\n \n if rec:\n item_id = rec.item_id\n else:\n rec = transaction.new()\n rec.created = local_datetime_now()\n if 'last_trx' in session:\n transaction.update(rec,session['last_trx'])\n \n # Handle Response?\n if request.form:\n #import pdb;pdb.set_trace()\n error_list=[]\n transaction.update(rec,request.form)\n if save_record(rec,error_list):\n return \"success\" # the success function looks for this...\n else:\n pass\n \n \n if item_id > 0:\n item_rec = Item(g.db).get(item_id)\n \n if not item_rec:\n flash(\"This is not a valid item id\")\n return \"failure: This is not a valid item id.\"\n else:\n rec.item_id=item_id\n \n \n return render_template('trx_edit_from_list.html',rec=rec,current_item=item_rec,warehouses=warehouses,trx_types=trx_types)", "def trobz_button_upgrade(self):\n _logger.info(\"Trobz_button_upgrade is processing.........\")\n upgrade_ids = self.ids\n # check whether \"trobz_base\" is in the list\n check_trobz_base = self.search([('name', '=', 'trobz_base'),\n ('id', 'in', upgrade_ids)])\n if check_trobz_base:\n # get all installed module with author \"Trobz\"\n installed_trobz_modules = self.search([('state', '=', 'installed'),\n ('author', 'ilike', 'Trobz')])\n upgrade_ids.extend(installed_trobz_modules.ids)\n \"\"\"\n uniquifying the ids to avoid:\n Error: \"One of the records you are trying to modify has\n already been deleted (Document type: %s)\"\n if exist an duplicate id in ids\n \"\"\"\n upgrade_ids = list(set(upgrade_ids))\n _logger.info(\"Trobz_button_upgrade ids of modules \"\n \"that need to upgrade: %s\" % upgrade_ids)\n _logger.info(\"Trobz_button_upgrade call native \"\n \"native_button_upgrade...\")\n # call super\n upgrade_modules = self.browse(upgrade_ids)\n native_button_upgrade(upgrade_modules)", "def click_bulk_edit_cancel_button(self):\n self.click_element(self.bulk_edit_cancel_button_locator, True)", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def on_edit(self, dataobj):", "def save_change(self):\n try:\n self.customer.budget.income = abs(float(self.incomeItem.text()))\n self.incomeItem.setText(f\"{self.customer.budget.income}\")\n fixed_expenses = {\n \"rent\": abs(float(self.listOfExpensesSEK.item(4).text())),\n \"subscription\": abs(float(self.listOfExpensesSEK.item(2).text())),\n \"insurance\": abs(float(self.listOfExpensesSEK.item(3).text())),\n \"others\": abs(float(self.listOfExpensesSEK.item(5).text()))\n }\n variable_expenses = {\n \"food\": abs(float(self.listOfExpensesSEK.item(11).text())),\n \"bills\": abs(float(self.listOfExpensesSEK.item(12).text())),\n \"transportation\": abs(float(self.listOfExpensesSEK.item(13).text())),\n \"hygien\": abs(float(self.listOfExpensesSEK.item(14).text())),\n \"clothes\": abs(float(self.listOfExpensesSEK.item(15).text())),\n \"entertainment\": abs(float(self.listOfExpensesSEK.item(16).text())),\n \"others\": abs(float(self.listOfExpensesSEK.item(17).text()))\n }\n self.customer.budget.set_budget(self.customer.budget.income,\n variable_expenses, fixed_expenses)\n # update instead of set\n DB.update_variable_expenses(self.customer.email, variable_expenses)\n DB.update_fixed_expenses(self.customer.email, fixed_expenses)\n DB.update_income(self.customer.budget.income, self.customer.email)\n total_fix, total_var = self.customer.budget.get_expenses()\n self.listOfExpensesSEK.item(1).setText(total_fix)\n self.listOfExpensesSEK.item(10).setText(total_var)\n self.customer.budget.set_buffert(abs(float(\n self.listOfExpensesSEK.item(20).text()\n )))\n DB.update_buffert(self.customer.email, abs(float(self.listOfExpensesSEK.item(20).text())))\n self.label_3.setText(str(self.customer.budget.income -\n self.customer.budget.get_total_expenses())\n )\n except Exception:\n self.popUp.exec_()", "def update_item(self, table, item):", "def sell():\n if request.method == \"GET\":\n portf = db.execute(\"SELECT * FROM portfolio WHERE id=:id\", id = session[\"user_id\"])\n return render_template(\"sell.html\",portfolio = portf)\n else:\n\n quote = lookup(request.form.get('stocklist'))\n print(str(quote))\n # Remove the stock frm user's portfolio\n # taking no of shares provided by user in form\n shares = int(request.form.get(\"no_of_shares\"))\n\n # Taking the price of that share\n\n price = db.execute(\"SELECT price FROM portfolio WHERE symbol=:symbol AND id=:id\", symbol = quote[\"symbol\"], id = session[\"user_id\"])\n\n # totla_price\n total_remove_price = shares * quote[\"price\"]\n # Now updating\n print(total_remove_price)\n # Taking total no of shares from portfolio\n share = db.execute(\"SELECT shares FROM portfolio WHERE id=:id AND symbol=:symbol\",symbol = quote[\"symbol\"],\n id = session[\"user_id\"])\n total = db.execute(\"SELECT total FROM portfolio WHERE id=:id AND symbol=:symbol\",symbol = quote[\"symbol\"],\n id = session[\"user_id\"])\n\n # if share provided by user in form is less than or equal to total shares owned then only transaction will processed\n print(share[0][\"shares\"])\n print(shares)\n if (shares < share[0][\"shares\"]):\n # Remove stock and price and no of stocks stocks = stocks - n\n real_total = total[0][\"total\"].split(\"$\")\n\n new_total1 = real_total[1][2:]\n new_total2 = real_total[1][:1]\n yup_final = new_total1 + new_total2\n print(yup_final)\n db.execute(\"UPDATE portfolio set total=:total, shares=:shares WHERE id=:id\", total = float(yup_final) - total_remove_price\n , shares = int(share[0][\"shares\"]) - shares , id=session[\"user_id\"])\n # current selling price = price * stocks and add this to user's cash\n elif (shares == share[0][\"shares\"]):\n db.execute(\"DELETE FROM portfolio WHERE id=:id AND symbol=:symbol\", id = session[\"user_id\"], symbol = quote['symbol'])\n else:\n return apology(\"Unable to process request\", 404)\n return redirect(\"/\")", "def sell():\n\n if request.method == \"POST\":\n symbol = request.form.get(\"symbol\")\n if not symbol:\n return apology(\"Must enter a symbol\", 400)\n num_shares = request.form.get(\"shares\")\n if not num_shares:\n return apology(\"Must enter some number of shares to sell\", 400)\n company_quote = lookup(symbol)\n if company_quote == None:\n return apology(\"Invalid Symbol\", 400)\n num_shares = int(num_shares)\n if num_shares <= 0:\n return apology(\"Must enter a positve number of shares to sell\", 400)\n\n rows = db.execute(\"SELECT id, symbol, numshares FROM totalshares WHERE id = :id AND symbol = :symbol\",\n id=session[\"user_id\"], symbol=company_quote[\"symbol\"])\n if len(rows) != 1:\n return apology(\"You do not have shares of \" + symbol, 400)\n if num_shares > rows[0][\"numshares\"]:\n return apology(\"You cannot sell more shares than you have\", 400)\n\n sale_value = num_shares * company_quote[\"price\"]\n\n balance = db.execute(\"SELECT cash FROM users WHERE id = :id\",\n id=session['user_id'])\n balance = balance[0][\"cash\"]\n new_balance = balance + sale_value\n date_time = datetime.now().strftime(\"%d/%m/%Y %H:%M:%S\")\n # Update history in history table\n return_val = db.execute(\"INSERT INTO 'history' (id, symbol, shares, price, transacted) VALUES (:id, :symbol, :shares, :price, :transacted)\",\n id=session[\"user_id\"], symbol=company_quote[\"symbol\"], shares=num_shares*-1, price=company_quote[\"price\"], transacted = date_time)\n if return_val == None:\n return apology(\"something went wrong\", 403)\n\n #Update total number and value of each shares (symbol) held in totalshares table\n rows = db.execute(\"SELECT id, symbol, numshares, totalvalue FROM totalshares WHERE id = :id AND symbol = :symbol\",\n id=session[\"user_id\"], symbol=company_quote[\"symbol\"])\n new_numshares = rows[0][\"numshares\"] - num_shares\n new_totalvalue = rows[0][\"totalvalue\"] - sale_value\n return_val = db.execute(\"UPDATE totalshares SET numshares = :new_numshares, totalvalue = :new_totalvalue WHERE id = :id AND symbol = :symbol\",\n new_numshares=new_numshares, new_totalvalue=new_totalvalue, id=session[\"user_id\"], symbol=company_quote[\"symbol\"])\n if return_val == None:\n return apology(\"something went wrong\", 403)\n\n #Update balance in users table\n return_val = db.execute(\"UPDATE users SET cash = :cash WHERE id = :id\", cash=new_balance, id=session[\"user_id\"])\n if return_val != None:\n return redirect(\"/\")\n else:\n return apology(\"something went wrong\", 403)\n else:\n rows = db.execute(\"SELECT symbol, numshares FROM totalshares WHERE id = :id\", id=session[\"user_id\"])\n symbol_options = []\n if rows != None and len(rows) > 0:\n for row in rows:\n if row[\"numshares\"] > 0:\n symbol_options.append(row[\"symbol\"])\n return render_template(\"sell.html\", symbol_options=symbol_options)", "def DoAction(self,event):\r\n selections = self.list.GetSelections()\r\n if not selections: return bell()\r\n itemDex = selections[0]\r\n item = self.items[itemDex]\r\n self.data.action(item)", "def done(self, cr, uid, ids, context=None):\n \n voucher_obj = self.pool.get('account.voucher')\n voucher_line_obj = self.pool.get('account.voucher.line')\n admin_affairs_model_obj = self.pool.get('admin.affairs.model')\n affairs_account_obj = self.pool.get('admin_affairs.account') \n model_id = admin_affairs_model_obj.search(cr, uid, [('model','=','environment.and.safety')], context=context)[0] \n affairs_account = affairs_account_obj.search(cr, uid, [('model_id','=',model_id)], context=context)\n if not affairs_account:\n raise osv.except_osv(_('Warning !'), _('Please insert account configuration For Environment and safety'))\n affairs_account_id = affairs_account[0]\n \n affairs_account_record = affairs_account_obj.browse(cr, uid, affairs_account_id,context=context) \n for record in self.browse(cr, uid, ids, context=context):\n if not record.allowances_lines_after :\n raise osv.except_osv(_('Partner Amount !'), _('Sorry no partner Amount After Rate To Transfer!'))\n notes = _(\"Enviroment and Safety allowances Contract: %s\")%(record.name)\n \n journal_id = affairs_account_record.journal_id\n analytic_id = affairs_account_record.analytic_id\n account_id = affairs_account_record.account_id\n\n\t\t# Creating Voucher / Ratitication\n voucher_id = voucher_obj.create(cr, uid, {\n 'amount': record.amount_total,\n 'type': 'ratification',\n 'date': time.strftime('%Y-%m-%d'),\n 'partner_id': record.partner_id.id,\n 'journal_id': journal_id and journal_id.id , \n 'state': 'draft',\n\t\t\t\t\t 'notes':record.notes,\n\t\t\t\t\t 'narration':notes ,\n \t 'company_id':record.company_id.id,\n })\n \t# Creating Voucher / Ratitication Lines\n for line in record.allowances_lines_after:\n '''account_id =line.category_id.account_id\n if not account_id:\n account_id = line.category_id.parent_id.account_id\n \n if not account_id:\n account_id = affairs_account_record.account_id \n\n if not account_id:\n raise osv.except_osv(_('Invalid action !'), _('Please insert Account configuration For Environment and safety Service')) ''' \n \n account_analytic_id =line.category_id.analytic_id\n if not account_analytic_id:\n account_analytic_id = line.category_id.parent_id.analytic_id \n \n if not account_analytic_id:\n account_analytic_id = affairs_account_record.analytic_id\n \n vocher_line_id = voucher_line_obj.create(cr, uid, {\n 'amount': record.amount_total,\n 'voucher_id': voucher_id,\n\t\t\t\t\t 'account_id':account_id and account_id.id,\n\t\t\t\t\t 'account_analytic_id':account_analytic_id and account_analytic_id.id ,\n 'type': 'dr',\n 'name':'environment and Safety allowances :' + record.name,\n })\n\t\t\n\t\t# Selecting Voucher Number / Refernece \n\n voucher_number = self.pool.get('account.voucher').browse(cr,uid,voucher_id)\n\n copy_attachments(self,cr,uid,[record.id],'services.contracts.archive',voucher_id,'account.voucher', context)\n self.write(cr, uid, ids, {'state':'done','transfer':True,'voucher_no':voucher_number.number}) \n return True", "def edit_button_clicked(self, obj):\n handle = self.get_selected()\n if handle:\n note = self.dbstate.db.get_note_from_handle(handle)\n try:\n from .. import EditNote\n EditNote(self.dbstate, self.uistate, self.track, note,\n callertitle = self.callertitle,\n extratype = [self.notetype] )\n except WindowActiveError:\n pass" ]
[ "0.66338634", "0.66057515", "0.6131386", "0.60525894", "0.59958154", "0.58884054", "0.58270955", "0.5821739", "0.57709527", "0.5718494", "0.57100034", "0.5659034", "0.56481487", "0.56460947", "0.5622122", "0.55970424", "0.55949754", "0.5559805", "0.5537368", "0.5531731", "0.5488313", "0.54579365", "0.5449568", "0.5434454", "0.5429909", "0.5417879", "0.53878057", "0.53674006", "0.53337747", "0.5317123" ]
0.8322212
0
Implementing set bulk edit end date functionality
def set_bulk_edit_end_date(self, end_date, future_date): if future_date is True: end_date = self.get_date(future_date=True, number_of_days_to_add=30) self.set_value_into_input_field(self.bulk_edit_end_date_inputbox_locator, end_date)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_bulk_add_end_date(self, end_date):\n global bulk_add_begin_date\n if end_date != \"\" and datetime.datetime.strptime(end_date, \"%m/%d/%Y\") > bulk_add_begin_date:\n self.set_value_into_input_field(self.bulk_add_end_date_inputbox_locator, str(end_date))\n elif end_date == \"\":\n end_date = self.get_date(last_day_of_current_month=True)\n self.set_value_into_input_field(self.bulk_add_end_date_inputbox_locator, str(end_date))\n else:\n end_date = bulk_add_begin_date + datetime.timedelta(days=7)\n end_date = \"%d/%d/%d\" % (end_date.month, end_date.day, end_date.year)\n self.set_value_into_input_field(self.bulk_add_end_date_inputbox_locator, str(end_date))\n return str(end_date)", "def set_end_date(self, date):\n pass", "def change_end_date(self, new_end_date):\n self.end_date = new_end_date", "def end_date(self, end_date):\n self._end_date = end_date", "def model_end_date(self, model_end_date):\n\n self._model_end_date = model_end_date", "def end_date(self, end_date):\n\n self._end_date = end_date", "def end_date(self, end_date):\n\n self._end_date = end_date", "def end_date(self, end_date):\n\n self._end_date = end_date", "def end_date(self, end_date):\n\n self._end_date = end_date", "def end_date(self, end_date):\n\n self._end_date = end_date", "def end_date(self, end_date):\n\n self._end_date = end_date", "def end_date(self, end_date):\n\n self._end_date = end_date", "def end_date(self, end_date):\n\n self._end_date = end_date", "def correct_list_to_end_date(self):\n if len(self.change_events):\n event_index = len(self.change_events) - 1\n while ((event_index > -1) and (self.change_events[event_index].date_ordinal >\n self.end_date_ordinal)):\n event = self.change_events[event_index]\n student_key = event.student_key\n if event.add_to_section():\n # add is in forward direction - for reverse we\n # remove it\n if self.all_students_dict.has_key(student_key):\n del self.all_students_dict[student_key] \n else:\n self.all_students_dict[student_key] = \\\n db.get(student_key)\n event_index -= 1", "def set_adjustment_charge_end_date(self, end_date):\n self.set_value_into_input_field(self.end_date_locator, end_date)", "def set_statement_end_date(self, end_date):\n end_date_to_set = None\n if end_date != \"\":\n end_date_to_set = end_date\n else:\n end_date_to_set = self.get_date(last_day_of_last_month=True)\n self.set_value_into_input_field(self.statement_end_date_locator, end_date_to_set)\n return end_date_to_set", "def set_end_date_for_search(self, end_date, allow_past_date):\n if allow_past_date is True and end_date != \"\":\n end_date_to_set = end_date\n else:\n if (end_date == \"\" or time.strptime(end_date, \"%m/%d/%Y\") < time.strptime(self.get_current_date(), \"%m/%d/%Y\")):\n end_date_to_set = self.get_date(current_date=True)\n else:\n end_date_to_set = end_date\n self.set_value_into_input_field(self.end_date_locator, end_date_to_set)", "def end_date(self, end_date):\n if end_date is None:\n end_date = datetime.utcnow()\n\n self._end_date = dt_utils.parse_date(end_date)", "def SubstitudedEndDate(obj):\n if not (obj.end_date) and obj.replaced_id:\n raise interface.Invalid(\n _(\"If a person is substituted End Date must be set\"), \n \"replaced_id\", \n \"end_date\")", "def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n set_fields_to_required(self, ['end_date'])", "def end_date(self, end_date):\n if end_date is None:\n raise ValueError(\"Invalid value for `end_date`, must not be `None`\")\n\n self._end_date = end_date", "def test_set_control_end_date(self):\n control = factories.ControlFactory()\n\n self.api.put(control, control.id, {\"end_date\": \"2015-10-10\"})\n\n control = db.session.query(all_models.Control).get(control.id)\n self.assertIsNone(control.end_date)", "def end(self, end):\n if end is None:\n self._set('end', end)\n else:\n try:\n self._set('end', Timestamp.to_datetime(end))\n except (TypeError, ValueError) as e:\n raise ValidationError(e)", "def end_date(self, value):\n\n if not isinstance(value, datetime):\n raise TypeError(_pretty_message(\n '''\n end_date must be an instance of datetime.datetime, not %s\n ''',\n _type_name(value)\n ))\n\n self._end_date = value", "def set_bulk_add_begin_date(self, begin_date):\n if begin_date == \"\":\n begin_date = self.get_date(current_date=True)\n self.set_value_into_input_field(self.bulk_add_begin_date_inputbox_locator, begin_date)\n global bulk_add_begin_date\n bulk_add_begin_date = datetime.datetime.strptime(begin_date, \"%m/%d/%Y\")\n return begin_date", "def adjust_end(self):\n # The current last end dates\n current_dict = {\n name: self._ensure_date(tracker.last_end_date())\n for (name, tracker) in self._tracker_dict.items()}\n # Adjusted end date\n adjusted_str = max(current_dict.values()).strftime(self.DATE_FORMAT)\n for (name, _) in self._tracker_dict.items():\n try:\n self.add(end_date=adjusted_str, name=name)\n except ValueError:\n pass\n return self", "def end_date_time(self, end_date_time):\n\n self._end_date_time = end_date_time", "def upgrade():\n\n op.execute(\"\"\"\n UPDATE task_group_tasks\n SET start_date = CURDATE(), end_date=CURDATE()\n WHERE (start_date IS NOT NULL AND start_date < \"1900-01-01\") OR\n (end_date IS NOT NULL AND end_date < \"1900-01-01\")\n \"\"\")", "def edit_date(entry):\n entry.date = get_date()\n entry.save()\n input(\"Edit successful. \")\n return entry", "def end_date(self):\n return self.__end_date" ]
[ "0.7402478", "0.7031963", "0.67778796", "0.6563116", "0.64243233", "0.63864547", "0.63864547", "0.63864547", "0.63864547", "0.63864547", "0.63864547", "0.63864547", "0.63864547", "0.63433135", "0.629675", "0.6284176", "0.61299545", "0.604149", "0.6033477", "0.59961516", "0.59430873", "0.59345084", "0.59264565", "0.5909958", "0.58784324", "0.58698535", "0.5828401", "0.56969357", "0.56830335", "0.5657813" ]
0.74479747
0
Implementing click bulk edit submit button functionality
def click_bulk_edit_submit_button(self): self.click_element(self.bulk_edit_submit_button_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_bulk_edit_submit_button(self):\n self.click_element(self.bulk_edit_submit_button_locator, True)", "def click_bulk_edit_button(self):\n self.click_element(self.bulk_edit_button_locator, True)", "def click_statement_fields_bulk_edit_button(self):\n self.click_element(self.statement_fields_bulk_edit_button_locator, True)", "def edit(self):\n\n pass", "def edit(self, **kwargs):\n ...", "def click_bulk_edit_cancel_button(self):\n self.click_element(self.bulk_edit_cancel_button_locator, True)", "def submit_row(context):\r\n ctx = original_submit_row(context)\r\n\r\n if context.get('readonly', False):\r\n ctx.update({\r\n 'show_delete_link': False,\r\n 'show_save_as_new': False,\r\n 'show_save_and_add_another': False,\r\n 'show_save_and_continue': False,\r\n 'show_save': False,\r\n })\r\n else:\r\n return ctx", "def edit():", "def custom_actions(self, form_entry, request=None):", "def submit_row_frontendadmin(context):\r\n opts = context['opts']\r\n change = context['change']\r\n is_popup = context['is_popup']\r\n save_as = context['save_as']\r\n return {\r\n #'show_delete_link': (not is_popup and context['has_delete_permission']\r\n # and (change or context['show_delete'])),\r\n 'show_save_as_new': not is_popup and change and save_as,\r\n 'show_save_and_add_another': context['has_add_permission'] and\r\n not is_popup and (not save_as or context['add']),\r\n 'show_save_and_continue': not is_popup and context['has_change_permission'],\r\n 'is_popup': is_popup,\r\n 'show_save': True\r\n }", "def after_successful_edit(self):\n pass", "def edit(self, *args, **kw):\n tmpl_context.widget = self.edit_form\n #pks = self.provider.get_primary_fields(self.model)\n \n log.debug(\"soyRomperLB= %s\" %kw)\n\n ###########################################\n pks = self.provider.get_primary_fields(self.model)\n \n ###########################################\n kw = {}\n for i, pk in enumerate(pks):\n kw[pk] = args[i]\n value = self.edit_filler.get_value(kw)\n value['_method'] = 'PUT'\n return dict(value=value, model=self.model.__name__, pk_count=len(pks))", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def on_click(self) -> None:\n self.editing = True", "def __actions__(self, obj):\n bool_ultimo = obj.bool_ultimo \n primary_fields = self.__provider__.get_primary_fields(self.__entity__)\n pklist = '/'.join(map(lambda x: str(getattr(obj, x)), primary_fields))\n\n if bool_ultimo == 1:\n cod_item = obj.cod_item\n value = '<div>'\n if has_permission('editar_item'):\n value = value + '<div><a class=\"edit_link\" href=\"'+pklist+'/edit\" style=\"text-decoration:none\">edit</a></div>'\n if has_permission('eliminar_relacion'):\n value = value + '<div><form method=\"POST\" action=\"'+pklist+'\" class=\"button-to\"><input type=\"hidden\" name=\"_method\" value=\"DELETE\" /><input class=\"delete-button\" onclick=\"return confirm(\\'Are you sure?\\');\" value=\"delete\" type=\"submit\" style=\"background-color: transparent; float:left; border:0; color: #286571; display: inline; margin: 0; padding: 0;\"/></form></div>'\n value = value + '<div><a class=\"relacion_link\" href=\"../relacions/?iid='+pklist+'\">Relaciones </a><br/><a class=\"versiones_link\" href=\"./?codi='+cod_item+'\">Revertir</a></div></div>'\n \n else:\n id_item_rev = DBSession.query(Item).filter_by(cod_item = obj.cod_item, bool_ultimo = 1).one().id_item\n ids = str(pklist) + \"-\" + str(id_item_rev)\n href = \"./revertir/?ids=\" + ids\n value = '<div><div><a class=\"edit_link\" href=\"'+pklist+'/edit\" style=\"text-decoration:none\">edit</a>'\\\n '</div><div>'\\\n '<form method=\"POST\" action=\"'+pklist+'\" class=\"button-to\">'\\\n '<input type=\"hidden\" name=\"_method\" value=\"DELETE\" />'\\\n '<input class=\"delete-button\" onclick=\"return confirm(\\'Are you sure?\\');\" value=\"delete\" type=\"submit\" '\\\n 'style=\"background-color: transparent; float:left; border:0; color: #286571; display: inline; margin: 0; padding: 0;\"/>'\\\n '</form>'\\\n '<a class=\"relacion_link\" href=\"../relacions/?iid='+pklist+'\">Relaciones </a>'\\\n '<a class=\"volver_link\" href=\"'+href+'\">Volver a</a>'\\\n '</div></div>'\n\n return value", "def edit(self, *args, **kw):\n\t\t\ttmpl_context.widget = self.edit_form\n\t\t\tpks \t\t= self.provider.get_primary_fields(self.model)\n\t\t\tkw \t\t\t= {}\n\n\t\t\tfor i, pk in enumerate(pks):\n\t\t\t\tkw[pk] \t\t= args[i]\n\n\t\t\tvalue \t\t= self.edit_filler.get_value(kw)\n\t\t\tvalue['_method'] \t= 'PUT'\n\n\t\t\treturn dict(value = value, model = self.model.__name__, pk_count = len(pks))", "def admin():\n return render_template('bulkform.html')", "def on_edit(self, dataobj):", "def click_bulk_add_button(self):\n self.click_element(self.bulk_add_button_locator, True)", "def __actions__(self, obj):\n\t\t\tprimary_fields \t= self.__provider__.get_primary_fields(self.__entity__)\n\t\t\tpklist \t\t= '/'.join(map(lambda x: str(getattr(obj, x)), primary_fields))\n\n\t\t\tvalue \t\t= '<div>'\n\t\t\tif has_permission('editar_LB'):\n\t\t\t\tvalue = value + '<div><a class=\"edit_link\" href=\"'+pklist+'/edit\" style=\"text-decoration:none\">edit</a></div>'\n\t\t\tif has_permission('eliminar_LB'):\n\t\t\t\tvalue = value + '<div><form method=\"POST\" action=\"'+pklist+'\" class=\"button-to\"><input type=\"hidden\" name=\"_method\" value=\"DELETE\" /><input class=\"delete-button\" onclick=\"return confirm(\\'Est&aacute; seguro que desea eliminar?\\');\" value=\"delete\" type=\"submit\" style=\"background-color: transparent; float:left; border:0; color: #286571; display: inline; margin: 0; padding: 0;\"/></form></div>'\n\t\t\tvalue = value + '</div>'\n\t\t\treturn value", "def editChange(self,editBtn):\n if self.edit ==True:\n self.updateDetails(\"normal\") #update details column\n self.edit = False #switch boolean\n self.editBtn.setText(\"Edit\") #update button text\n else:\n self.updateDetails(\"edit\") #update details column\n self.edit= True #switch boolean\n self.editBtn.setText(\"Stop Editing\") #update button text", "def on_edit_clicked(self,button):\n\t\tself.list_partitions.edit_partition()", "def edit(self, *args, **kw):\n tmpl_context.widget = self.edit_form\n pks = self.provider.get_primary_fields(self.model)\n kw = {}\n for i, pk in enumerate(pks):\n kw[pk] = args[i]\n\n value = self.edit_filler.get_value(kw)\n value['_method'] = 'PUT'\n \n return dict(value=value, model=self.model.__name__, pk_count=len(pks))", "def DoEdit(self,event):\r\n raise UncodedError", "def commit(self,form_list):\n raise NotImplementedError", "def _do_request_update_all(self, __button):\n return self._dtc_data_controller.request_do_update_all()", "def edit_button_clicked(self, obj):\n handle = self.get_selected()\n if handle:\n note = self.dbstate.db.get_note_from_handle(handle)\n try:\n from .. import EditNote\n EditNote(self.dbstate, self.uistate, self.track, note,\n callertitle = self.callertitle,\n extratype = [self.notetype] )\n except WindowActiveError:\n pass", "def click_the_edit_button_that_appears(driver):\n driver.find_element_by_xpath(xpaths.users.eric_Edit_Button).click()", "def test_otoroshi_controllers_adminapi_tcp_service_api_controller_bulk_update_action(self):\n pass", "def edit_tools(self, e):\n #GETTING SELECTION\n\n self.selected_item = self.user_inventory.selection()\n self.select_name = self.user_inventory.item([i for i in self.selected_item], \"values\")[0]\n self.select_entdate = self.user_inventory.item([i for i in self.selected_item], \"values\")[3]\n\n self.df_same_name = self.df_user.query(\"title == @self.select_name\")\n #this is the selected one for sure\n self.df_the_selected_item = self.df_same_name.loc[self.df_same_name[\"entry date\"] == self.select_entdate]\n\n #GETTING THE INDEX NUMBER OF THE SELECTION IN .CSV FILE\n self.index_select = self.df_the_selected_item.index\n self.index_select_number = self.index_select.tolist()\n\n #bottom buttons appear:\n self.changing_item_label.config(text=\"Now editing \"+self.select_name+\" that added on \"+self.select_entdate+\":\")\n\n self.delete_but = Button (self.bottom_frame, text=\"DELETE\", command=self.delete_button)\n self.delete_but.place(relx=0.1, rely=0.7, relwidth=0.28, anchor=\"w\")\n\n self.servings_list = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]\n self.serv_drop = Combobox(self.bottom_frame, value=self.servings_list, state=\"readonly\")\n self.serv_drop.place(relx=0.5, rely=0.7, relwidth=0.2, anchor=CENTER)\n\n \n self.serv_but = Button(self.bottom_frame, text=\"CHANGE AMOUNT\", command=self.change_amount_button, state=\"disabled\")\n self.serv_but.place(relx=0.9, rely=0.7, relwidth=0.28, anchor=\"e\")\n\n self.serv_drop.bind(\"<<ComboboxSelected>>\", self.activate_button)" ]
[ "0.80722886", "0.7676083", "0.72128063", "0.6668404", "0.6572236", "0.65535694", "0.63675964", "0.63469726", "0.63170797", "0.62994254", "0.6182403", "0.61548156", "0.61191946", "0.6086712", "0.60528755", "0.60109603", "0.59184855", "0.5897436", "0.58945984", "0.58505696", "0.5842662", "0.58424354", "0.5778383", "0.5760711", "0.57459337", "0.57073754", "0.5682217", "0.5675151", "0.5647281", "0.5622662" ]
0.8029734
1
Implementing click buy and sell deal management grid first row inline action button functionality
def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self): self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def click_target_buy_policy_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, column_number=2, first_row=True)\n self.wait_for_ajax_spinner_load()", "def click_vendor_price_list_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.vendor_price_list_grid_div_id, None, self.view_price_list_column_number, True)", "def click_target_buy_policies_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, self.target_buy_policies_grid_inline_action_column_number, first_row=True)", "def click_adjustment_charges_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.adjustment_charges_grid_div_id, None, self.adjustment_charges_grid_inline_action_column_number, True)", "def click_vendors_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.vendors_div_id, None, self.grid_column_number, True)", "def click_buy_page_inline_action_button(self, vendor):\n self.click_inline_action_button(self.vendors_div_id, vendor, self.grid_column_number)", "def click_view_price_list_detail_first_row_inline_action_button(self):\n self.click_inline_action_button(self.view_price_list_div_id, None, self.view_price_list_column_number, True)\n self.wait_for_ajax_spinner_load()", "def click_adjustments_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.adjustments_grid_div_id, None, self.adjustments_grid_inline_action_column_number, True)", "def click_disputes_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.statement_details_dispute_grid_div_id, None, self.disputes_grid_inline_action_column_number, True)", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def select_buy_and_sell_deal_management_grid_first_row_checkbox(self):\n self.select_grid_row_checkbox(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_checkbox_column_number, True)", "def click_line_item_disputes_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.line_item_disputes_grid_div_id, None, self.line_item_disputes_grid_inline_action_column_number, True)", "def click_outbound_statement_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, None, self.outbound_statements_grid_inline_action_column_number, True)", "def click_outbound_statements_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, None, self.outbound_statements_grid_inline_action_column_number, True)", "def click_inbound_statement_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.inbound_statement_grid_div_id, None, self.inbound_statement_grid_inline_action_column_number, True)", "def click_re_analysis_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.re_analysis_grid_div_id, None, self.re_analysis_grid_inline_action_column_number, True)", "def click_inbound_statement_grid_inline_action_button(self, row_identifier_text):\n self.click_inline_action_button(self.inbound_statement_grid_div_id, row_identifier_text, self.inbound_statement_grid_inline_action_column_number)", "def click_statement_validation_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.statement_validation_grid_div_id, None, self.statement_validation_grid_inline_action_column_number, True)", "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def click_resolution_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.resolution_grid_div_id, None, self.resolution_grid_inline_column_number, True)", "def click_outbound_statements_grid_inline_action_button(self, row_identifier_text):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, row_identifier_text, self.outbound_statements_grid_inline_action_column_number)", "def trade_action(self, BUY_QTY):\n BUY_QTY = 4500\n self.trade(BUY_QTY)\n #self.show()", "def click_view_price_list_detail_page_inline_action_button(self, price_list_item):\n self.click_inline_action_button(self.view_price_list_div_id, price_list_item, self.view_price_list_column_number)\n self.wait_for_ajax_spinner_load()", "def select_bill_and_audit_module_grid_inline_item(self, item_name):\n self.select_inline_action_item(item_name, module_name = self.bill_audit_module_name)\n self.wait_for_ajax_spinner_load()", "def click_buy_and_sell_deal_management_link(self):\n self.select_static_tab(self.buy_and_sell_deal_management_locator, message=\"buy and sell deal management locator not found before specified time out\")", "def button_confirm(self):\n for record in self:\n res = super(PurchaseOrder, self).button_confirm()\n if not record.sale_order_id and record.customer_id:\n sale_order_line_obj = self.env['sale.order.line']\n attachment_ids = []\n sale_order_obj = self.env['sale.order']\n for attchment in record.attachment_ids:\n attachment_ids.append((0, 0, {\n 'name': attchment.name,\n 'datas': attchment.datas,\n \"description\": attchment.description,\n \"mimetype\": attchment.mimetype,\n 'index_content': attchment.index_content,\n \"create_uid\": attchment.create_uid.id,\n }))\n\n vals = {\n \"partner_id\": record.customer_id.id,\n \"vendor_id\": record.partner_id.id,\n \"purchase_order_id\": record.id,\n \"attachment_ids\": attachment_ids,\n \"colour_instructions\": record.colour_instructions,\n \"packing\": record.packing,\n \"name\": record.name,\n \"face_stamp\": record.face_stamp,\n \"selvedge\": record.selvedge,\n \"shipping_mark\": record.shipping_mark,\n \"shipping_sample_book\": record.shipping_sample_book,\n \"notes\": record.notes,\n \"shipment_date\": record.shipment_date,\n \"destination_id\": record.destination_id.id,\n \"currency_id\": record.currency_id.id,\n }\n sale_order = sale_order_obj.create(vals)\n record.sale_order_id = sale_order.id\n for line in record.order_line:\n taxes = line.product_id.taxes_id\n fpos = record.fiscal_position_id\n taxes_id = fpos.map_tax(taxes, line.product_id, record.partner_id) if fpos else taxes\n if taxes_id:\n taxes_id = taxes_id.filtered(lambda x: x.company_id.id == record.company_id.id)\n sale_order_line = sale_order_line_obj.create({'product_id': line.product_id.id,\n 'name': line.name,\n 'tax_id': [(6, 0, taxes_id.ids)],\n 'product_uom_qty': line.product_qty,\n \"product_uom\": line.product_uom.id,\n 'price_unit': line.price_unit,\n \"order_id\": sale_order.id,\n # \"discount\": line.discount,\n \"purchase_order_line_id\": line.id,\n \"actual_qty\": line.actual_qty\n })\n line.sale_order_line_id = sale_order_line.id\n\n return res", "def set_shortfall_grid_actions(self, action_item):\n self.single_selection_from_kendo_in_grid(self.shortfall_grid_div_id, self.actions_column_name, action_item)", "def OnLeftUp_ClickButton(self, event):\r\n \r\n self._hover_button = None\r\n\r\n if self._action_part:\r\n self.RefreshButton(self._action_part)\r\n\r\n # make sure we're still over the item that was originally clicked\r\n if self._action_part == self.HitTest(*event.GetPosition()):\r\n \r\n # fire button-click event\r\n e = AuiManagerEvent(wxEVT_AUI_PANE_BUTTON)\r\n e.SetManager(self)\r\n e.SetPane(self._action_part.pane)\r\n e.SetButton(self._action_part.button.button_id)\r\n self.ProcessMgrEvent(e)", "def click_add_financial_charges_button(self):\n self.number_of_financial_charges_before_create = self.get_number_of_financial_charges_in_the_grid()\n self.click_element(self.add_financial_charges_button_locator, hover=True)" ]
[ "0.7165491", "0.69654083", "0.69372815", "0.69077474", "0.66696733", "0.6508445", "0.6368367", "0.6347989", "0.63158387", "0.627581", "0.6269957", "0.61755913", "0.61733496", "0.6149201", "0.6137883", "0.6119374", "0.58740306", "0.5799829", "0.57629937", "0.57190955", "0.566324", "0.566295", "0.55592984", "0.55453634", "0.5487429", "0.5470098", "0.5457542", "0.5430609", "0.5386233", "0.5291522" ]
0.8185726
0
Implementing is account label displayed correctly functionality
def is_account_label_displayed_correctly(self): is_displayed_correctly = False if self.get_text_from_element(self.account_label_locator) == self.account_label: is_displayed_correctly = True return is_displayed_correctly
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def is_account_manager_label_displayed_correctly(self):\n is_displayed_correctly = False\n if self.get_text_from_element(self.account_manager_label_locator) == self.account_manager_label:\n is_displayed_correctly = True\n return is_displayed_correctly", "def Label(self) -> str:", "def _add_user_label(self):\n user_label = tk.Label(parent, text='Пользователь: ' +\n self.user_info.ShortUserName + ' Версия ' + __version__,\n font=('Arial', 8))\n user_label.pack(side=tk.RIGHT, anchor=tk.NE)", "def label_from_instance(obj):\n if len(obj.first_name) > 0 and len(obj.last_name) > 0:\n return \"{} {}\".format(obj.first_name, obj.last_name)\n else:\n return \"<{}>\".format(obj.username)", "def test_useraccount_display(self):\n self.assertEqual(\"Partagez, échangez\", self.driver.title)", "def account_summary(self):\n pass", "def get_label(settings):", "def _generateLabel(self, obj, **args):\n result = []\n label = self._script.utilities.displayedLabel(obj)\n if label:\n result.append(label)\n return result", "def display_account(account):\n if 'accountName' not in account and 'emailAddress' not in account:\n account_template = '{accountId}'\n elif 'emailAddress' not in account:\n account_template = '{accountName} ({accountId})'\n elif 'accountName' not in account:\n account_template = '{emailAddress} ({accountId})'\n else:\n account_template = '{accountName}, {emailAddress} ({accountId})'\n return account_template.format(**account)", "def show_account(self, obj): # pylint: disable=no-self-use\n return '%s\\\\%s' % (obj.domain, obj.username)", "def display_name(self):", "def get_true_label(self):\n return gettext_lazy('Yes')", "def acutalizeRolelabel (self, data):\n if data.has_key(StatusMsg.label_role):\n self.rolelabel.setText(self.roleDecoder.get(data.get(StatusMsg.label_role)))", "def _get_label ( self ):\n if self._label is not None:\n return self._label\n return user_name_for( self.name )", "def __init__(self, label, account):\n self.label = label\n self.account = account", "def item_label(self):\r\n label_to_print = (\"\\nYou have selected \" + self.name_i\r\n + \" in tier \" + str(self.tier) + \".\")\r\n return label_to_print", "def get_label(self):\n\n auth = self.authorizations[0]\n return auth.label", "def label(self):\n return ''", "def test_render_label(self):\n label = self.block.meta.label\n self.assertEqual(label, 'Google Calendar', 'The labels are not the same')", "def label(self, margin):\n if self.alphaL == None or self.alphaR == None:\n self.label = \"N\"\n elif abs(self.alphaL - self.alphaR) <= margin:\n self.label = \"S\"\n elif (self.alphaL - self.alphaR) > margin:\n self.label = \"L\"\n elif -(self.alphaL - self.alphaR) > margin:\n self.label = \"R\"\n else:\n self.label = \"N\"", "def account():\n return render_template('user/account.html')", "def has_label(self):\n return self.label is not None", "def test_get_form_label_email(self):\n self.assertEqual(\n self.user.get_form_label(email=True),\n 'testuser <[email protected]>',\n )", "def __str__(self):\n\n return \"\\nBasic Account\\nAccount Name: {self.name}\\nAccount Number: {self.acNum:03d}\\nAvailable Balance: £{self.balance:.2f}\".format(self=self)", "def labels_validation(ele,actultext):\r\n msg, status = \"\", True\r\n\r\n try:\r\n 'Here we will pass objectidentifier for required element'\r\n text_heading = ui_controls.text_view(get_obj_identifier(ele))\r\n print \"Current label returned by UI is==>\"+str(text_heading)\r\n\r\n 'To check if it is returning empty value here'\r\n if not text_heading:\r\n print str(actultext)+\" label does not exist and it is returning empty value.\"\r\n return False, msg\r\n\r\n 'Comparing text retrieved from UI with validation text'\r\n if actultext.strip() == text_heading.strip():\r\n print (str(actultext)+\" label has been found!!!\")\r\n else:\r\n print(\"Sorry!!!lable has been mismatched,it should be \"+str(actultext))\r\n print (\"label shown at UI is==>\"+str(text_heading))\r\n return False, msg \r\n\r\n except Exception as excp:\r\n traceback.print_exc()\r\n msg += str(excp)\r\n status = False\r\n return status, msg", "def account():\n\n return render_template('account_page.html', title='Account')", "def display_accounts_details():\n return Credentials.display_credentials()", "def format_account(account):\n return(f'{account[\"name\"]} from {account[\"country\"]} is a {account[\"description\"]}')", "def update_label(self):\n text = _lbl_text_only_active if self.only_show_active else _lbl_text_allmods\n\n self._label.setText(\n text.format(\n shown=self._filter.rowCount(),\n total=self._srcmodel.rowCount()))", "def final_level_text_is_displayed(self):\n final_level_text = self.driver.find_element_by_name(self.FINAL_LEVEL_TEXT_NAME)\n return final_level_text.is_displayed()" ]
[ "0.6368619", "0.6344278", "0.6144431", "0.61396897", "0.6069333", "0.5957378", "0.591541", "0.5847986", "0.5835125", "0.5799273", "0.5767016", "0.57317346", "0.5720798", "0.5717911", "0.5690789", "0.5643161", "0.5638978", "0.5619242", "0.5607045", "0.5593206", "0.55726165", "0.5523999", "0.551239", "0.5500891", "0.548203", "0.54618955", "0.5460251", "0.5456734", "0.5431958", "0.54282874" ]
0.74997246
0
Implementing is account manager label displayed correctly functionality
def is_account_manager_label_displayed_correctly(self): is_displayed_correctly = False if self.get_text_from_element(self.account_manager_label_locator) == self.account_manager_label: is_displayed_correctly = True return is_displayed_correctly
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def is_account_label_displayed_correctly(self):\n is_displayed_correctly = False\n if self.get_text_from_element(self.account_label_locator) == self.account_label:\n is_displayed_correctly = True\n return is_displayed_correctly", "def _add_user_label(self):\n user_label = tk.Label(parent, text='Пользователь: ' +\n self.user_info.ShortUserName + ' Версия ' + __version__,\n font=('Arial', 8))\n user_label.pack(side=tk.RIGHT, anchor=tk.NE)", "def Label(self) -> str:", "def acutalizeRolelabel (self, data):\n if data.has_key(StatusMsg.label_role):\n self.rolelabel.setText(self.roleDecoder.get(data.get(StatusMsg.label_role)))", "def displayMl(self, *args):\n\t\tself.callMethod(('ManialinkManager', 'displayManialinkToLogin'), *args)", "def test_useraccount_display(self):\n self.assertEqual(\"Partagez, échangez\", self.driver.title)", "def show_account(self, obj): # pylint: disable=no-self-use\n return '%s\\\\%s' % (obj.domain, obj.username)", "def display_name(self):", "def get_label(settings):", "def get_label(self):\n\n auth = self.authorizations[0]\n return auth.label", "def account_summary(self):\n pass", "def get_true_label(self):\n return gettext_lazy('Yes')", "def store_name_print(self):\r\n store_label = (\"You are in \" + self.name_s\r\n + \". These are the items in the store: \")\r\n return store_label", "def display_accounts_details():\n return Credentials.display_credentials()", "def _get_label ( self ):\n if self._label is not None:\n return self._label\n return user_name_for( self.name )", "def label_from_instance(obj):\n if len(obj.first_name) > 0 and len(obj.last_name) > 0:\n return \"{} {}\".format(obj.first_name, obj.last_name)\n else:\n return \"<{}>\".format(obj.username)", "def displayName(self):\r\n return self.tr('Lockdown Liveability Tool')", "def account():\n return render_template('user/account.html')", "def accounts():", "def label(self):\n return ''", "def _generateLabel(self, obj, **args):\n result = []\n label = self._script.utilities.displayedLabel(obj)\n if label:\n result.append(label)\n return result", "def account():\n\n return render_template('account_page.html', title='Account')", "def manage_info():", "def update_label(self):\n text = _lbl_text_only_active if self.only_show_active else _lbl_text_allmods\n\n self._label.setText(\n text.format(\n shown=self._filter.rowCount(),\n total=self._srcmodel.rowCount()))", "def on_UseDefaultA1Account_clicked(self):\n # TODO: not implemented yet\n raise NotImplementedError", "def showHint():\n\t#First get the master pod\n\tcurrentMasterPod=masterPod.currentMasterPod\n\t#Get the hint\n\thint=currentMasterPod.hint\n\t#Check if the pod is disabled\n\tif hasattr(currentMasterPod,\"locked\"):\n\t\tcurrentTime = getCurrentTime()\n\t\tif currentMasterPod.locked < currentTime:\n\t\t\tloginAttemptVar.set(hint)\n\n\telse:\n\t\t#Update the control variable\n\t\tloginAttemptVar.set(hint)", "def __init__(self, label, account):\n self.label = label\n self.account = account", "def manager_info(self, manager_cn):\n manager_login = manager_cn[4:manager_cn.find(\",\")]\n manager_info = self.locate_user(manager_login)\n try:\n return manager_info[0][1]['cn'][0] + \\\n \" (\" + manager_info[0][1]['uid'][0] + \")\"\n except IndexError:\n return \"(LDAP record removed) %s\" % manager_cn", "def __display_login_info(self):\n print(f'\\nYour card has been created\\n'\n f'Your card number:\\n'\n # f'{self.__card_display()}\\n' # uncomment this line and comment out line below for pretty display\n f'{self.card_number}\\n'\n f'Your card PIN:\\n'\n f'{self.__account_pin}\\n', )", "def label(self, margin):\n if self.alphaL == None or self.alphaR == None:\n self.label = \"N\"\n elif abs(self.alphaL - self.alphaR) <= margin:\n self.label = \"S\"\n elif (self.alphaL - self.alphaR) > margin:\n self.label = \"L\"\n elif -(self.alphaL - self.alphaR) > margin:\n self.label = \"R\"\n else:\n self.label = \"N\"" ]
[ "0.660459", "0.6117637", "0.5810706", "0.579244", "0.5719326", "0.5605203", "0.55454856", "0.55295485", "0.5480238", "0.5474781", "0.5458673", "0.5429711", "0.541652", "0.54080826", "0.53880817", "0.5370094", "0.53372633", "0.533329", "0.53238416", "0.5259748", "0.5249081", "0.5244948", "0.52415377", "0.5239572", "0.5230557", "0.5211616", "0.5211341", "0.5204124", "0.51853186", "0.5176338" ]
0.74036354
0
Implementing click vendor grid add destination button functionality
def click_vendor_grid_add_destination_button(self): self.click_element(self.vendor_grid_add_destination_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_customer_grid_add_destination_button(self):\n self.click_element(self.customer_grid_add_destination_locator)", "def set_destination_to_add_destination(self, destination):\n self.multiple_items_selection_from_kendo_dropdown(self.destination_multiple_kendo_dropdown_locator, destination)\n self.click_element(self.new_destination_header_locator)", "def add_destination(self):\n pass", "def extend_button_event(self, srcevent, push):\n srcevent.targets_sch = self.find_at(srcevent.x, srcevent.y)\n srcevent.push = push\n x, y = self.schcoords(srcevent.x, srcevent.y)\n srcevent.x_sch = x\n srcevent.y_sch = y\n return srcevent", "def click_bulk_add_button(self):\n self.click_element(self.bulk_add_button_locator, True)", "def drag_buy_and_sell_management_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.buy_and_sell_management_grid_div_id, source_column, destination_column)", "def LayersButton(self, event):\n pass", "def click_add_destination_save_button(self):\n self.click_element(self.add_destination_save_button_locator, True)\n try:\n self.wait().until(EC.visibility_of_element_located(self.confirmation_popup_locator), 'confirmation popup locator not found before specified time out')\n self.click_element(self.ok_button_locator, True)\n except:\n raise", "def set_source_to_add_destination(self, source_name):\n self.single_selection_from_static_kendo_dropdown(self.source_kendo_dropdown_arrow_locator, source_name)", "def set_route_class_to_add_destination(self, route_class):\n self.multiple_items_selection_from_kendo_dropdown(self.route_class_multiple_kendo_dropdown_locator, route_class)\n self.click_element(self.new_destination_header_locator)", "def drag_vendor_price_lists_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.view_price_list_div_id, source_column, destination_column)", "def drag_vendor_price_list_detail_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_rates_grid_div_id, source_column, destination_column)", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def drag_vendors_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.vendors_div_id, source_column, destination_column)", "def click_add_resolution_button(self):\n self.click_element(self.add_resolution_button_locator)", "def addToolBarButtons(self):", "def drag_destinations_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.destinations_grid_div_id, source_column, destination_column)", "def on_pushButton_11_clicked(self):\n # TODO: not implemented yet\n raise NotImplementedError", "def drag_vendor_price_list_detail_reference_rates_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_reference_rates_grid_div_id, source_column, destination_column)", "def add_grid_button(self, w):\n self.event_emitters[(w.__class__.__name__, w.type)].append(w)\n self.grid[(w.x, w.y)] = w\n self.note_buttons[w.note] = w", "def on_mouse_click(self, event: Event):\r\n self.control.add_gem(self.row, self.column)", "def _on_click_browse_to_pt_project(self):\n pass", "def on_pushButton_12_clicked(self):\n # TODO: not implemented yet\n raise NotImplementedError", "def click_vendor_price_list_detail_dial_digits_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_dial_digits_grid_div_id)", "def createGlobalMoverButton(self, name, parent, uiInstance):\n\n part = name.partition(self.name)[2]\n\n # create the icon\n pixmap = QtGui.QPixmap(20, 15)\n pixmap.fill(QtGui.QColor(\"yellow\"))\n icon = QtGui.QIcon(pixmap)\n\n # create the button\n self.outlinerWidgets[name + \"_globalMoverBtn\"] = QtWidgets.QPushButton(icon, \"\")\n self.outlinerWidgets[name + \"_globalMoverBtn\"].setMinimumSize(QtCore.QSize(20, 15))\n self.outlinerWidgets[name + \"_globalMoverBtn\"].setMaximumSize(QtCore.QSize(20, 15))\n uiInstance.treeWidget.setItemWidget(parent, 1, self.outlinerWidgets[name + \"_globalMoverBtn\"])\n\n # connect and add to list\n self.outlinerWidgets[name + \"_globalMoverBtn\"].clicked.connect(\n partial(self.selectMover, part, True, False, False, self.outlinerWidgets[name + \"_globalMoverBtn\"]))\n self.outlinerControls.append([name + \"_globalMoverBtn\", name + \"_mover\", icon])", "def onAddButtonPress(self, button):\n\t\twp_x = float(self.traj_to_x_entry.get_text())\n\t\twp_y = float(self.traj_to_y_entry.get_text())\n\t\twp_z = float(self.traj_to_z_entry.get_text())\n\t\twp_yaw = float(self.traj_to_yaw_entry.get_text())\n\n\t\t# add waypoint to list\n\t\twaypoints_gui.append([wp_x, wp_y, wp_z, wp_yaw])\n\n\t\t# reset entry fields\n\t\tself.traj_to_x_entry.set_text('')\n\t\tself.traj_to_y_entry.set_text('')\n\t\tself.traj_to_z_entry.set_text('')\n\t\tself.traj_to_yaw_entry.set_text('')", "def click(self):\r\n pass", "def placeCalendarButton(data,row,target,path,alts,**kwargs):\n# printPretty(\"args: %s %s %s %s\" % (data,row,target,path))\n datebut = gtk.Button()\n datebut.show()\n image = gtk.Image()\n image.set_from_file(\"img/date.png\")\n datebut.set_image(image)\n datebut.unset_flags(gtk.CAN_FOCUS)\n datebut.connect(\"clicked\",dateChoose,target,data,path,alts,kwargs)\n datebut.set_tooltip_text(\"Click to choose date from calendar\")\n row.pack_start(datebut,0,0,2)", "def on_toolButton_clicked(self):\n # TODO: not implemented yet\n raise NotImplementedError", "def on_pushButton_6_clicked(self):\n # TODO: not implemented yet\n raise NotImplementedError" ]
[ "0.8016771", "0.646783", "0.62058467", "0.59663063", "0.5802453", "0.5654677", "0.558464", "0.55153704", "0.5512954", "0.54436165", "0.5412148", "0.5398465", "0.5385228", "0.5373527", "0.5345784", "0.53445095", "0.53421956", "0.53135604", "0.5309377", "0.5302662", "0.52934057", "0.52860004", "0.5279597", "0.52751553", "0.5271698", "0.52713376", "0.5269765", "0.52665323", "0.52602655", "0.5244735" ]
0.8421994
0
Implementing set source to add destination functionality
def set_source_to_add_destination(self, source_name): self.single_selection_from_static_kendo_dropdown(self.source_kendo_dropdown_arrow_locator, source_name)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def add_destination(self):\n pass", "def set_source(self, source):\n self.data['source'] = source", "def attach(self, destination): \r\n self.destination= destination", "def source(self, other):\n raise NotImplementedError", "def set_destination_to_add_destination(self, destination):\n self.multiple_items_selection_from_kendo_dropdown(self.destination_multiple_kendo_dropdown_locator, destination)\n self.click_element(self.new_destination_header_locator)", "def add(self, dest, source=None):\n raise NotImplementedYet()", "def attach(self, destination): \r\n self.destination=destination", "def __init__(self, src, dest):\n self.src = src\n self.dest = dest", "def setDestination(self, node):\n self.dest_node = node", "def __init__( self, source, label, destination ):\n self.source = source;\n self.label = label;\n self.destination = destination;", "def transfer(self,source,target):\n self.__transfer_dict[tuple([source,target])]()", "def add_link (self, src, dst):\n raise NotImplementedError", "def AddSource (self, name, source, filename):\n self.containments [name] = source, filename, False", "def add_source_achors():\n pass", "def add_source_address(self, srcAddr):\n self.source.address = srcAddr", "def add_source_address(self, srcAddr):\n self.source.address = srcAddr", "def source(self, source):\n\n self._source = source", "def source(self, source):\n\n self._source = source", "def source(self, source):\n\n self._source = source", "def source(self, source):\n\n self._source = source", "def source(self, source):\n\n self._source = source", "def source(self, source):\n\n self._source = source", "def source(self, source):\n\n self._source = source", "def addedge(self, source, target, **attrs):\n\n raise NotImplementedError", "def add_new_source_sink(self):\n source = self.source()\n sink = self.sink()\n for arc in self.out_arcs_lists[source]:\n self.arc_info[arc][\"lower_bound\"] = 0\n self.arc_info[arc][\"upper_bound\"] = float('inf')\n for arc in self.in_arcs_lists[sink]:\n self.arc_info[arc][\"lower_bound\"] = 0\n self.arc_info[arc][\"upper_bound\"] = float('inf')\n for vert in self.vertices:\n if vert != source and vert != sink:\n if self.get_arc(source, vert) is None:\n self.add_inexact_edge(source, vert, 0, float('inf'))\n if self.get_arc(vert, sink) is None:\n self.add_inexact_edge(vert, sink, 0, float('inf'))", "def add_source(self, name, position):#)*args, **kwargs):\n return self._add_object(name, Source, position)#*args, **kwargs)", "def add_sglink (self, src, dst):\n raise NotImplementedError", "def __init__(self, source=None, dest=None):\n if source == None:\n self._source_list = []\n else:\n self._source_list = source\n if dest == None:\n self._dest_list = []\n else:\n self._dest_list = dest", "def destination(self, destination):\n\n self._destination = destination", "def __init__(self, origin, destination):\n self.origin = origin\n self.destination = destination" ]
[ "0.7595484", "0.68353415", "0.6798214", "0.67796534", "0.67778414", "0.67471397", "0.67426056", "0.6698542", "0.64227223", "0.64126635", "0.63714236", "0.6344426", "0.6306878", "0.628486", "0.6280195", "0.6280195", "0.62796885", "0.62796885", "0.62796885", "0.62796885", "0.62796885", "0.62796885", "0.62796885", "0.622716", "0.6223678", "0.6203589", "0.62027794", "0.61968184", "0.6186988", "0.6185319" ]
0.71872395
1
Implementing set destination to add destination functionality
def set_destination_to_add_destination(self, destination): self.multiple_items_selection_from_kendo_dropdown(self.destination_multiple_kendo_dropdown_locator, destination) self.click_element(self.new_destination_header_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def add_destination(self):\n pass", "def attach(self, destination): \r\n self.destination= destination", "def attach(self, destination): \r\n self.destination=destination", "def destination(self, destination):\n\n self._destination = destination", "def do_destination(self, args):\n self.destination = int(args)", "def setDestination(self, node):\n self.dest_node = node", "def handle_set_destination(self, data):\n #If the origin_id is 0, it has not been specified and we must find\n #the closest node to where we are now\n self.dest_node = data.dest_id\n if data.origin_id == 0:\n #Will set self.current_node\n self.get_nearest_node(data.dest_id)\n else:\n self.current_node = data.origin_id\n if self.current_node == data.dest_id:\n self.at_dest = True\n msg = (\"We're already there!\")\n return srvs.SetDestinationResponse(True, msg)\n rospy.wait_for_service('/get_trajectory')\n get_traj = rospy.ServiceProxy('/get_trajectory', srvs.GetTrajectory)\n trajectory = get_traj(False, self.current_node, data.dest_id).trajectory\n self.np_trajectory = to_numpy_trajectory(trajectory)\n self.loop = False\n self.at_dest = False\n msg = (\"Trajectory to destination of vehicle #%i \" % self.vehicle_id +\n \"successfully set.\")\n return srvs.SetDestinationResponse(True, msg)", "def set_destination(self, value):\n try:\n self._destination = value\n self._repository = hg.repository(ui.ui(), self._destination)\n except:\n self._repository = None", "def SetDestination(self, point):\n\t\tif type(point) != Point:\n\t\t\traise TypeError(\"Incorrect Type\")\n\n\t\tself.destination = point", "def destination_config(self, destination_config):\n self._destination_config = destination_config", "def __init__(self, origin, destination):\n self.origin = origin\n self.destination = destination", "def __init__(self, destination: Type[Model], **kwargs):\n self.destination = destination", "def set_data_dest(self, destination_id):\n self.data_dest = destination_id", "def Destination(self) -> _n_0_t_1:", "def destination_id(self, destination_id):\n\n self._destination_id = destination_id", "def setup_signal_destination(cls, dest):\n cls.signal_destination = dest", "def _route_to_dest(self):\n # Ask the network\n self.route = self.network.determine_route(self.start, self.dest)\n # Set the index to where we are now\n self.route_index = 0", "def destinations(self, destinations):\n\n self._destinations = destinations", "def destination(self, value: Union[str, Station]):\n if type(value) is str:\n # Passed a str\n if len(value) <= 9:\n self._destination = value\n else:\n raise ValueError(\"Destination must be a maximum of 9 character\")\n elif type(value) is Station:\n # Passed a Station\n self._destination = value\n elif value is None:\n # Passed None\n self._destination = None\n else:\n raise TypeError(\"Destination must be of type 'str' and a maximum of 9 characters\")", "def visit_move(self, move):\n dest_id = self.event_json['destination']['id']\n destination = self.world.entities[dest_id]\n move.destination = destination", "def post(self, destination_id, add_to_beginning=\"False\",clear_other_waypoints=\"False\",datasource=\"tranquility\",**kwargs):\n kwargs_dict ={\n\"destination_id\" : destination_id, \"add_to_beginning\" : add_to_beginning, \"clear_other_waypoints\" : clear_other_waypoints, \"datasource\" : datasource, \n }\n kwargs_dict.update(kwargs)\n return EsiRequestObject(self.base_url, self.post_responses) \\\n .post(**kwargs_dict)", "def set_dests(self, increment=1000000):\n modified = 0\n pb = Progress(len(self.graph.routers), 'Setting destinations', increment=increment, callback=lambda: 'Modified {:,d}'.format(modified))\n for router in pb.iterator(self.graph.routers.values()):\n for interface in router.interfaces:\n # Copy destination ASes to avoid messing up original\n idests: Set[int] = set(interface.dests)\n # If last hop, interface has non-IXP AS mapping, and interface has destination ASes\n if not router.succ and idests and interface.asn > 0:\n origin = interface.asn\n # Interface must have exactly 2 destination ASes and one must be its origin AS\n if len(idests) == 2 and origin in idests:\n other_asn = peek(idests - {origin}) # other AS\n # If other AS is likely customer of interface origin AS, and it's a small AS\n if self.bgp.conesize[origin] > self.bgp.conesize[other_asn] and self.bgp.conesize[other_asn] < 5:\n idests.discard(origin)\n modified += 1\n # Add all remaining destination ASes to the router destination AS set\n router.dests.update(idests)", "def set_source_to_add_destination(self, source_name):\n self.single_selection_from_static_kendo_dropdown(self.source_kendo_dropdown_arrow_locator, source_name)", "def destination(self) -> pulumi.Input['DestinationArgs']:\n return pulumi.get(self, \"destination\")", "def _adjust_destination(self, destProps):\n super(PHEME_http_receiverTransferAgent, self).\\\n _adjust_destination(destProps)\n for prop in destProps.iter(tag='property'):\n if prop.attrib['name'] == 'host':\n prop.text = self.options.input_dir", "def destination(self):\n return self._destination", "def destination_code(self, destination_code):\n if destination_code is None:\n raise ValueError(\"Invalid value for `destination_code`, must not be `None`\")\n\n self._destination_code = destination_code", "def __init__(self, src, dest):\n self.src = src\n self.dest = dest", "def add_link (self, src, dst):\n raise NotImplementedError", "def add(self, destination: n, weight: w):\n self.connections[destination] = weight" ]
[ "0.8411674", "0.7463169", "0.7400204", "0.7207934", "0.70057255", "0.6809444", "0.6675045", "0.66749346", "0.66615766", "0.6569928", "0.65571934", "0.6486881", "0.64785564", "0.64413476", "0.6326941", "0.631957", "0.63072693", "0.630284", "0.61267376", "0.6122438", "0.61160386", "0.6082256", "0.605486", "0.6028217", "0.6021821", "0.60098314", "0.5995255", "0.5944578", "0.5926774", "0.59223235" ]
0.7659271
1
Implementing set deal rate to add destination functionality
def set_deal_rate_to_add_destination(self, deal_rate): self.set_value_into_input_field(self.deal_rate_textbox_locator, deal_rate, use_win32com=True) self.click_element(self.new_destination_header_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_rates_destination(self):\n self.set_value_into_input_field(self.rates_tab_destination_textbox_locator, self.buy_page_excel_data_dictionary[\"Destination\"])", "def __call__(self, rate:'kW'):\n self.rate = rate\n self.cost = self.price * rate", "def _do_set_rate(self, rate):\n self.set_remote_status(1)\n if rate == 0:\n self.set_to_slow()\n elif rate == 1:\n self.set_to_fast()\n self.set_remote_status(3)\n print(self._do_get_rate())", "def set_current_rate(self, rate_to_set):\n pass", "def test_add_taxation_strategy_to_rate_plan(self):\n pass", "def test_create_rate_plan(self):\n pass", "def test_update_rate_plan(self):\n pass", "def add_shipping_rate(self, country, shipping_company_id, description, base_rate, addon_rate, product_ids):\n self.login()\n self.form_request('shipping/create/', data={\n 'country': country,\n 'shipping_company': shipping_company_id,\n 'description': description,\n 'base_rate': \"%.2f\" % base_rate,\n 'addon_rate': \"%.2f\" % addon_rate,\n 'products': product_ids,\n })", "def test_retire_rate_plan(self):\n pass", "def rate(self, rate):\n\n self._rate = rate", "def rate(self, rate):\n\n self._rate = rate", "def set_costs(self) -> None:\n self[\"glider cost\"] = (\n self[\"glider base mass\"] * self[\"glider cost slope\"]\n + self[\"glider cost intercept\"]\n )\n self[\"lightweighting cost\"] = (\n self[\"glider base mass\"]\n * self[\"lightweighting\"]\n * self[\"glider lightweighting cost per kg\"]\n )\n self[\"electric powertrain cost\"] = (\n self[\"electric powertrain cost per kW\"] * self[\"electric power\"]\n )\n self[\"combustion powertrain cost\"] = (\n self[\"combustion power\"] * self[\"combustion powertrain cost per kW\"]\n )\n self[\"fuel cell cost\"] = self[\"fuel cell power\"] * self[\"fuel cell cost per kW\"]\n self[\"power battery cost\"] = (\n self[\"battery power\"] * self[\"power battery cost per kW\"]\n )\n self[\"energy battery cost\"] = (\n self[\"energy battery cost per kWh\"] * self[\"electric energy stored\"]\n )\n self[\"fuel tank cost\"] = self[\"fuel tank cost per kg\"] * self[\"fuel mass\"]\n # Per km\n self[\"energy cost\"] = self[\"energy cost per kWh\"] * self[\"TtW energy\"] / 3600\n\n # For battery, need to divide cost of electricity\n # at battery by efficiency of charging\n # to get costs at the \"wall socket\".\n\n _ = lambda x: np.where(x == 0, 1, x)\n self[\"energy cost\"] /= _(self[\"battery charge efficiency\"])\n\n self[\"component replacement cost\"] = (\n self[\"energy battery cost\"] * self[\"battery lifetime replacements\"]\n + self[\"fuel cell cost\"] * self[\"fuel cell lifetime replacements\"]\n )\n\n with open(DATA_DIR / \"purchase_cost_params.yaml\", \"r\") as stream:\n to_markup = yaml.safe_load(stream)[\"markup\"]\n\n self[to_markup] *= self[\"markup factor\"]\n\n # calculate costs per km:\n self[\"lifetime\"] = self[\"lifetime kilometers\"] / self[\"kilometers per year\"]\n\n with open(DATA_DIR / \"purchase_cost_params.yaml\", \"r\") as stream:\n purchase_cost_params = yaml.safe_load(stream)[\"purchase\"]\n\n self[\"purchase cost\"] = self[purchase_cost_params].sum(axis=2)\n # per km\n amortisation_factor = self[\"interest rate\"] + (\n self[\"interest rate\"]\n / (\n (np.array(1) + self[\"interest rate\"]) ** self[\"lifetime kilometers\"]\n - np.array(1)\n )\n )\n self[\"amortised purchase cost\"] = (\n self[\"purchase cost\"] * amortisation_factor / self[\"kilometers per year\"]\n )\n\n # per km\n self[\"maintenance cost\"] = (\n self[\"maintenance cost per glider cost\"]\n * self[\"glider cost\"]\n / self[\"kilometers per year\"]\n )\n\n # simple assumption that component replacement\n # occurs at half of life.\n self[\"amortised component replacement cost\"] = (\n (\n self[\"component replacement cost\"]\n * (\n (np.array(1) - self[\"interest rate\"]) ** self[\"lifetime kilometers\"]\n / 2\n )\n )\n * amortisation_factor\n / self[\"kilometers per year\"]\n )\n\n self[\"total cost per km\"] = (\n self[\"energy cost\"]\n + self[\"amortised purchase cost\"]\n + self[\"maintenance cost\"]\n + self[\"amortised component replacement cost\"]\n )", "def save(self, *args, **kwargs):\n self.item.rates_total += 1\n self.item.average_rate += (self.item.average_rate + self.rate) / self.item.rates_total\n self.item.save()\n super(Rate, self).save(*args, **kwargs)", "def __init__(self, rate, **kwargs):\n super(Dropout, self).__init__(**kwargs)\n self.rate = rate\n self.inplace = kwargs.get('inplace', False)", "def withAdaptRate(self, value):\n self._refinement.setAdaptRate(value)\n return self", "def _set_costs(self):\n plant_size_kw = (self.sam_sys_inputs[\"resource_potential\"]\n / self._RESOURCE_POTENTIAL_MULT) * 1000\n\n cc_per_kw = self.sam_sys_inputs.pop(\"capital_cost_per_kw\", None)\n if cc_per_kw is not None:\n capital_cost = cc_per_kw * plant_size_kw\n logger.debug(\"Setting the capital_cost to ${:,.2f}\"\n .format(capital_cost))\n self.sam_sys_inputs[\"capital_cost\"] = capital_cost\n\n dc_per_well = self.sam_sys_inputs.pop(\"drill_cost_per_well\", None)\n num_wells = self.sam_sys_inputs.pop(\"prod_and_inj_wells_to_drill\",\n None)\n if dc_per_well is not None:\n if num_wells is None:\n msg = ('Could not determine number of wells to be drilled. '\n 'No drilling costs added!')\n logger.warning(msg)\n warn(msg)\n else:\n capital_cost = self.sam_sys_inputs[\"capital_cost\"]\n drill_cost = dc_per_well * num_wells\n logger.debug(\"Setting the drilling cost to ${:,.2f} \"\n \"({:.2f} wells at ${:,.2f} per well)\"\n .format(drill_cost, num_wells, dc_per_well))\n self.sam_sys_inputs[\"capital_cost\"] = capital_cost + drill_cost\n\n foc_per_kw = self.sam_sys_inputs.pop(\"fixed_operating_cost_per_kw\",\n None)\n if foc_per_kw is not None:\n fixed_operating_cost = foc_per_kw * plant_size_kw\n logger.debug(\"Setting the fixed_operating_cost to ${:,.2f}\"\n .format(capital_cost))\n self.sam_sys_inputs[\"fixed_operating_cost\"] = fixed_operating_cost", "def _cost(self, action):\n raise NotImplementedError", "def add_reservation(self, src, dst,duration, bandwidth):\n \n # locks the self.current_reservations data structure. This is done\n # because there is a thread that could access it concurrently.\n with self.update_lock:\n\n # PART 1, TASK 3.4 check if there is an existing reservation for (src,dst). \n # you can use the self.current_reservations dictionary to check it.\n # If the reservation exists get the path and bw and update the links capacity \n # data structure using `self.add_link_capacity(path, bw)`\n \n # PART 1, TASK 3.1. Once get_available_path is implemented call it to get a path.\n path = self.get_available_path(src, dst, bandwidth)\n\n # PART 1, TASK 3.2 If there is an available path \n if path: \n pass\n # PART 1, TASK 3.2 Get mpls stack of labels\n\n # PART 1, TASK 3.3 get:\n # 1) ingress switch name\n # 2) action name using `mpls_ingress_x_hop` set x as number of labels\n # 3) src and dst ips (your match)\n # 4) make sure all your labels are strings and use them as action parameters\n\n # PART 1, TASK 3.4\n\n # check if its a new or an existing reservation (to update)\n\n # add entry or modify\n # PART 2 TASK 1.4 Configure the associated meter properly.\n\n # update controllers data structures: self.current_reservation & self.links_capacity\n \n\n # PART 1, TASK 3.2 otherwise we print no path available\n else:\n # PART 1, task 4.3 if we dont find a path but the reservation existed\n # you have to erase it while making sure you update links_capacity accordingly \n print(\"\\033[91mRESERVATION FAILURE: no bandwidth available!\\033[0m\")", "def add_destination(self):\n pass", "def setDataRate(self, DataRate):\n \n self.DataRate = DataRate", "def __init__(self, origin, destination):\n self.origin = origin\n self.destination = destination", "def _set_rate(self):\r\n interval = self.data.iloc[2, 0] - self.data.iloc[1, 0]\r\n self.rate = int(1 / interval)", "def rate(self, rate):\n # Get the sign of the rates before calculating\n x_sign = copysign(1, self.x_rate)\n y_sign = copysign(1, self.y_rate)\n self._rate = rate\n # Multiply by the original sign to retain direction\n self.x_rate = x_sign * fabs(rate * cos(self._angle))\n self.y_rate = y_sign * fabs(rate * sin(self._angle))", "def rates(self):\n raise NotImplementedError(\"Must be implemented by subclass.\")", "def exchange_rate(self, exchange_rate):\n\n self._exchange_rate = exchange_rate", "def rates(self, rates):\n\n self._rates = rates", "def rates(self, rates):\n\n self._rates = rates", "def flowrate(self, values):\n self._qao = float(values.get('ao', self._qao))\n self._qmv = float(values.get('mv', self._qmv))\n self._qper = float(values.get('per', self._qper))\n if self._lvad is not None:\n q_lvad = values.get('lvad', self._lvad.flowrate['lvad'])\n self._lvad.flowrate = {'lvad': q_lvad}", "def flowrate(self, values):\n self._qao = float(values.get('ao', self._qao))\n self._qmv = float(values.get('mv', self._qmv))\n self._qper = float(values.get('per', self._qper))\n if self._lvad is not None:\n q_lvad = values.get('lvad', self._lvad.flowrate['lvad'])\n self._lvad.flowrate = {'lvad': q_lvad}", "def set_costs_table(self) -> None:\n self.costs[\"B\"] = 2\n self.costs[\"A\"] = 6\n self.costs[\"fin\"] = float(\"inf\")" ]
[ "0.66945755", "0.6256623", "0.58632624", "0.56280166", "0.5582999", "0.5560925", "0.5540805", "0.5522839", "0.5397975", "0.5368135", "0.5368135", "0.53364104", "0.5259004", "0.5257495", "0.52336127", "0.52244794", "0.5215375", "0.5210986", "0.5177353", "0.51745504", "0.51631546", "0.51562047", "0.51294494", "0.5120258", "0.51017624", "0.5090379", "0.5090379", "0.508901", "0.508901", "0.5073798" ]
0.74631596
0
Implementing click customer grid add destination button functionality
def click_customer_grid_add_destination_button(self): self.click_element(self.customer_grid_add_destination_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_vendor_grid_add_destination_button(self):\n self.click_element(self.vendor_grid_add_destination_locator)", "def set_destination_to_add_destination(self, destination):\n self.multiple_items_selection_from_kendo_dropdown(self.destination_multiple_kendo_dropdown_locator, destination)\n self.click_element(self.new_destination_header_locator)", "def add_destination(self):\n pass", "def click_bulk_add_button(self):\n self.click_element(self.bulk_add_button_locator, True)", "def click_add_destination_save_button(self):\n self.click_element(self.add_destination_save_button_locator, True)\n try:\n self.wait().until(EC.visibility_of_element_located(self.confirmation_popup_locator), 'confirmation popup locator not found before specified time out')\n self.click_element(self.ok_button_locator, True)\n except:\n raise", "def handle_add(self, controller):\n \n controller.customer.CreatePizza()\n controller.show_frame(PageOne)", "def choose_new_address_delivery(self) -> \"CheckoutPage\":\n self.accordion = DeliveryDetailsUser(self.driver)\n self.accordion.btn_new_address_delivery.click()\n return self", "def post(self, destination_id, add_to_beginning=\"False\",clear_other_waypoints=\"False\",datasource=\"tranquility\",**kwargs):\n kwargs_dict ={\n\"destination_id\" : destination_id, \"add_to_beginning\" : add_to_beginning, \"clear_other_waypoints\" : clear_other_waypoints, \"datasource\" : datasource, \n }\n kwargs_dict.update(kwargs)\n return EsiRequestObject(self.base_url, self.post_responses) \\\n .post(**kwargs_dict)", "def handle_add(self, controller):\n \n try:\n pizza = controller.customer.pizza ## get a reference to pizza object of the customer\n \n except Exception:\n showinfo(title='Pop-up', message=\"No Pizza Created Yet.\")\n return\n \n else:\n # create an order if not exist, and add pizza to order\n c = controller.customer\n self.onPress(c) ## update requested data\n if not c.my_order:\n c.my_order = Order(c.name, c.address, c.id)\n \n c.AddToOrder()\n controller.show_frame(PageTwo) ## go to my order page", "def set_route_class_to_add_destination(self, route_class):\n self.multiple_items_selection_from_kendo_dropdown(self.route_class_multiple_kendo_dropdown_locator, route_class)\n self.click_element(self.new_destination_header_locator)", "def LayersButton(self, event):\n pass", "def customer_add_view_new(h):\n global html\n html = h\n common_elements = customer_common_elements()\n \n css_list = common_elements[\"css_list\"]\n\n javascript_list = [\"js/commonFunctions.js\"]\n javascript_list += common_elements[\"javascript_list\"]\n\n all_btn = common_elements[\"all_btn\"]\n\n html.new_header(\"Add Customers\", \"customer_management.py\", all_btn, css_list, javascript_list)\n\n customer_add_html = \"\"\"\n <table width=\"100%\" class=\"content_tbl\" border=\"0\" cellspacing=\"0\" cellpadding=\"0\">\n <tbody>\n <tr>\n <th>Customer Details</th>\n </tr>\n <tr>\n <td>\n <table width=\"100%\" class=\"deatil_tbl\" border=\"0\" cellspacing=\"0\" cellpadding=\"0\">\n <tbody>\n <tr>\n <td class=\"label\">Organization Name</td>\n <td colspan=\"3\"><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr>\n <td>Telephone Number</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td class=\"label\">Fax</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr>\n <td>Website</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td>Email</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr>\n <td>Business Type</td>\n <td><select class=\"txt_bx\"><option> Auction </option></select></td>\n <td>Submission</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr>\n <td>Address</td>\n <td colspan=\"3\"><textarea rows=\"4\" cols=\"27\"></textarea></td>\n </tr>\n </tbody>\n </table>\n </td>\n </tr>\n <tr>\n <td>\n <table width=\"100%\" id=\"iDetails_table\" class=\"individual_tbl\" border=\"0\" cellspacing=\"0\" cellpadding=\"0\">\n <tbody>\n <tr>\n <th colspan=\"4\">Individual Details <a href=\"\" class=\"addChildTr\"><img class=\"add\" src=\"images/add-gray.png\" align=\"right\" alt=\"\" border=\"0\"></a></th>\n </tr>\n <tr class=\"findInnerHTML\"><td colspan=\"4\"></td></tr>\n <tr class=\"findInnerHTML\">\n <td class=\"label\">Contact Person Name</td>\n <td colspan=\"3\"><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr class=\"findInnerHTML\">\n <td>Email</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td class=\"label\">Contact Number</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr class=\"findInnerHTML\">\n <td>Department</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td>Designation</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr><td colspan=\"4\"></td></tr>\n </tbody>\n </table>\n </td>\n </tr>\n <tr>\n <td>\n <table width=\"100%\" class=\"individual_tbl\" border=\"0\" cellspacing=\"0\" cellpadding=\"0\">\n <tbody>\n <tr>\n <th colspan=\"4\">Customer Portal Configration</th>\n </tr>\n <tr><td colspan=\"4\"></td></tr>\n <tr>\n <td class=\"label\">Portal ID</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td class=\"label\">Portal Link</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr>\n <td>Set Theme</td>\n <td><select class=\"txt_bx\"><option> Theme </option></select></td>\n <td>Port Number</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr><td colspan=\"4\"></td></tr>\n <tr>\n <td colspan=\"4\">\n <table border=\"0\" class=\"add_user_tbl\" width=\"90%\" cellpadding=\"0\" cellspacing=\"0\">\n <tbody>\n <tr>\n <th colspan=\"6\">Add Customer Users <a href=\"#\" class=\"addChildTr\"><img class=\"add\" src=\"images/add-gray.png\" align=\"right\" alt=\"\" border=\"0\"></a></th>\n </tr>\n <tr><td colspan=\"6\"></td></tr>\n <tr>\n <td>User Name</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td>Password</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td>Email</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr class=\"findInnerHTML\">\n <td>User Name</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td>Password</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td>Email</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr><td colspan=\"6\"></td></tr>\n </tbody>\n </table>\n </td>\n </tr>\n <tr><td>&nbsp;</td></tr>\n </tbody></table>\n </td>\n </tr>\n <tr>\n <td>\n <table width=\"100%\" id=\"billing_dept_table\" class=\"individual_tbl endof_tbl\" border=\"0\" cellspacing=\"0\" cellpadding=\"0\">\n <tbody><tr>\n <th colspan=\"4\">Billing Departments</th>\n </tr>\n <tr><td colspan=\"4\"></td></tr>\n <tr>\n <td class=\"width_150\">\n Select Billing Department \n <br> \n <em class=\"fs10\">(If TPS selected)</em>\n </td>\n <td colspan=\"3\">\n <select id=\"billing_determinant_select_box\" class=\"txt_bx\">\n <option>TPS</option>\n <option>Bandwidth</option>\n <option>Sim Subscription</option>\n </select>\n </td>\n </tr>\n <tr>\n <td class=\"width_150\">Max TPS</td>\n <td colspan=\"3\"><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr>\n <td class=\"width_150\">Billing Cycle</td>\n <td colspan=\"3\"><select class=\"txt_bx\"><option> Monthly </option></select></td>\n </tr>\n <tr>\n <td class=\"width_150\">Billing Start Date</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n <td class=\"width_150\">Billing Amount</td>\n <td><input type=\"text\" value=\"\" class=\"txt_bx\"></td>\n </tr>\n <tr><td colspan=\"4\"></td></tr>\n </tbody></table>\n </td>\n </tr>\n </tbody></table>\n \n \"\"\"\n html.write(customer_add_html)\n html.write(\"\"\"\n <script>\n customer_add_view_new();\n </script>\n \"\"\")\n html.new_footer()", "def OnAdd(self, controller):\n pass", "def customer_add_view(h):\n global html\n html = h\n \n common_elements = customer_common_elements()\n \n css_list = common_elements[\"css_list\"]\n\n javascript_list = common_elements[\"javascript_list\"]\n\n all_btn = common_elements[\"all_btn\"]\n\n html.new_header(\"Add Customers\", \"customer_management.py\", all_btn, css_list, javascript_list)\n customer_string = \"\"\"\n <div id=\"grid_view_div\">\n <div class=\"yo-tabs\">\n <ul>\n <li>\n <a class=\"active\" href=\"#content_1\" id=\"active_host_tab\">Customer Application Form</a>\n </li>\n <li>\n <a href=\"#content_2\" id=\"disable_host_tab\">Bulk Upload</a>\n </li>\n <li>\n <a href=\"#content_3\" id=\"discovered_host_tab\">CRM Import</a>\n </li>\n </ul>\n <div id=\"content_1\" class=\"tab-content\" style=\"display:block;height:100%;\">\n <form action=\"customer_post_ajax.py\" method=\"get\" id=\"add_customer_form\" name=\"add_customer_form\" autocomplete=\"on\" >\n <div class=\"form-div\" style=\"top:30px;\">\n <div class=\"form-body\">\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"username\">Customer User Name</label>\n <input type=\"text\" id=\"user_name\" name=\"user_name\" \n title=\"Choose Unique User Name. <br/>Must be at least 5 characters.\" />\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"password\">Password</label>\n <input type=\"password\" id=\"password\" name=\"password\" title=\"Must be at least 8 characters. \"/>\n </div>\n \"\"\"\n customer_string += \"\"\"\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"groups\">Customer Organisation(Role)</label>\n \"\"\"\n customer_string += (customer_group_customer_widget())\n customer_string += \"\"\"\n </div>\n \"\"\"\n customer_string += \"\"\"\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"first_name\">First Name</label>\n <input type=\"text\" id=\"first_name\" name=\"first_name\" title=\"Please Enter First name.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"last_name\">Last Name</label>\n <input type=\"text\" id=\"last_name\" name=\"last_name\" title=\"Please Enter Last name.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"company_name\">Company</label>\n <input type=\"text\" id=\"company_name\" name=\"company_name\" title=\"Please Enter Company Name.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"designation\">Designation</label>\n <input type=\"text\" id=\"designation\" name=\"designation\" title=\"Please Enter Designation.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"mobile_no\">Mobile Number</label>\n <input type=\"text\" id=\"mobile_no\" name=\"mobile_no\" \n title=\"Please Enter Mobile Number<br/> Don't include +91 or 0.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"telephone_no\">Telephone Number</label>\n <input type=\"text\" id=\"telephone_no\" name=\"telephone_no\" \n title=\"Please Enter Mobile Number<br/> Don't include +91 or 0.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"fax\">Fax</label>\n <input type=\"text\" id=\"fax\" name=\"fax\" \n title=\"Please Enter Mobile Number<br/> Don't include +91 or 0.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"email_id\">E-Mail ID</label>\n <input type=\"text\" id=\"email_id\" name=\"email_id\" title=\"Please Enter E-Mail ID.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"city_id\">City</label>\n <input type=\"text\" id=\"city_id\" name=\"city_id\" title=\"Please Enter City Name.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"state_id\">State</label>\n <input type=\"text\" id=\"state_id\" name=\"state_id\" title=\"Please Enter State.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"country_id\">Country</label>\n <input type=\"text\" id=\"country_id\" name=\"country_id\" title=\"Please Enter Country.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"usage\">Usage</label>\n <select id='usage' name='usage'>\n <option value=0>Personal</option>\n <option value=1>Commercial</option>\n </select>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"purpose\">Purpose</label>\n <input type=\"text\" id=\"purpose\" name=\"purpose\" title=\"Please Enter Purpose.\"/>\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"address\">Address</label>\n <textarea id=\"address\" name=\"address\" title=\"Please Enter own Address.\"></textarea>\n </div>\n </div>\n </div>\n <div class=\"form-div-footer\">\n <button type=\"submit\" class=\"yo-small yo-button\"><span class=\"add\">Save</span></button>\n <button type=\"reset\" class=\"yo-small yo-button\" id=\"close_add_user\"><span class=\"cancel\">Cancel</span></button>\n </div>\n </form>\n </div>\n <!-- container tab 2 -->\n <div id=\"content_2\" class=\"tab-content\" style=\"display:block;height:100%;\">\n <form action=\"#\" method=\"post\">\n <label for=\"bulk\">Upload Customers</label>\n <input name=\"bulk\" id=\"bulk\" type=\"file\" />\n </form>\n </div>\n <!-- container tab 3 -->\n <div id=\"content_3\" class=\"tab-content\" style=\"display:block;height:100%;\">\n <form class=\"form-body\" id=\"crm_conn\" action=\"#\" method=\"get\">\n <div class=\"form-div\" style=\"top:30px;\">\n <div class=\"form-body\">\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"crm_link\">CRM Link Location</label>\n <input type=\"text\" name=\"crm_link\" id=\"crm_link\" />\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"crm_user\">CRM Username</label>\n <input type=\"text\" name=\"crm_user\" id=\"crm_user\" />\n </div>\n <div class=\"row-elem\">\n <label class=\"lbl lbl-big\" for=\"crm_pass\">CRM Password</label>\n <input type=\"password\" name=\"crm_pass\" id=\"crm_pass\" />\n </div>\n </div>\n </div>\n <div class=\"form-div-footer\">\n <button type=\"submit\" class=\"yo-small yo-button\"><span class=\"add\">Test Connection</span></button>\n <button type=\"reset\" class=\"yo-small yo-button\" id=\"close_add_user\"><span class=\"add\">Start Import</span></button>\n </div>\n </form>\n </div>\n </div>\n </div>\n \"\"\" \n customer_string += \"\"\"\n <script>\n post_customers();\n </script>\n \"\"\"\n html.write(customer_string)\n html.new_footer()", "def drag_buy_and_sell_management_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.buy_and_sell_management_grid_div_id, source_column, destination_column)", "def choose_new_address(self) -> \"CheckoutPage\":\n self.accordion = BillingDetailsUser(self.driver)\n self.accordion.btn_new_address.click()\n return self", "def set_source_to_add_destination(self, source_name):\n self.single_selection_from_static_kendo_dropdown(self.source_kendo_dropdown_arrow_locator, source_name)", "def drag_vendor_price_list_detail_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_rates_grid_div_id, source_column, destination_column)", "def click_add_new_record_button(self):\n self.click_element(self.add_new_record_button_locator, True)", "def drag_vendor_price_list_detail_reference_rates_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_reference_rates_grid_div_id, source_column, destination_column)", "def onAddButtonPress(self, button):\n\t\twp_x = float(self.traj_to_x_entry.get_text())\n\t\twp_y = float(self.traj_to_y_entry.get_text())\n\t\twp_z = float(self.traj_to_z_entry.get_text())\n\t\twp_yaw = float(self.traj_to_yaw_entry.get_text())\n\n\t\t# add waypoint to list\n\t\twaypoints_gui.append([wp_x, wp_y, wp_z, wp_yaw])\n\n\t\t# reset entry fields\n\t\tself.traj_to_x_entry.set_text('')\n\t\tself.traj_to_y_entry.set_text('')\n\t\tself.traj_to_z_entry.set_text('')\n\t\tself.traj_to_yaw_entry.set_text('')", "def click_add_financial_charges_button(self):\n self.number_of_financial_charges_before_create = self.get_number_of_financial_charges_in_the_grid()\n self.click_element(self.add_financial_charges_button_locator, hover=True)", "def drag_vendor_price_list_detail_dial_digits_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_dial_digits_grid_div_id, source_column, destination_column)", "def drag_destinations_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.destinations_grid_div_id, source_column, destination_column)", "def drag_vendor_price_lists_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.view_price_list_div_id, source_column, destination_column)", "def custom_actions(self, form_wizard_entry, request=None):", "def customer_created_handler(event):\n obj = event.obj\n\n # submit customer after creation\n obj.workflow.submit()", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def add_view(self, *args, **kwargs):\r\n if not kwargs.get(\"extra_context\"):\r\n kwargs[\"extra_context\"] = {}\r\n kwargs[\"extra_context\"].update({\r\n \"insert_classes\": self.admin_site.insert_classes,\r\n \"form_url\": \"herp\"\r\n })\r\n return super(ServeeModelAdmin, self).add_view(*args, **kwargs)", "def import_orders_button(cls, store_views):\n pass" ]
[ "0.80682164", "0.6506563", "0.611456", "0.56259584", "0.5613796", "0.56112826", "0.55190563", "0.55092216", "0.5477043", "0.54032", "0.5402372", "0.53741306", "0.53735375", "0.5371906", "0.5339226", "0.52918637", "0.5284191", "0.5281223", "0.52697504", "0.52324903", "0.5203534", "0.5191478", "0.51899236", "0.5159754", "0.5152351", "0.51389784", "0.5132878", "0.512404", "0.5116337", "0.51065403" ]
0.8646041
0
Implementing click add destination save button functionality
def click_add_destination_save_button(self): self.click_element(self.add_destination_save_button_locator, True) try: self.wait().until(EC.visibility_of_element_located(self.confirmation_popup_locator), 'confirmation popup locator not found before specified time out') self.click_element(self.ok_button_locator, True) except: raise
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_vendor_grid_add_destination_button(self):\n self.click_element(self.vendor_grid_add_destination_locator)", "def add_destination(self):\n pass", "def click_customer_grid_add_destination_button(self):\n self.click_element(self.customer_grid_add_destination_locator)", "def set_destination_to_add_destination(self, destination):\n self.multiple_items_selection_from_kendo_dropdown(self.destination_multiple_kendo_dropdown_locator, destination)\n self.click_element(self.new_destination_header_locator)", "def _create_actions(self):\n self.save_button.clicked.connect(self._save)", "def save(self, *args):\n # need to do!!\n pass", "def save(self, *args, **kwargs):\n return", "def save(self, *args, **kwargs):\n pass", "def save(self):\n # TODO (Pierre): code", "def save (self):\n pass", "def save():", "def saveAsButtonMethod(self):\n return AbstractAction(\"Save As\")", "def save(self):\n pass", "def save(self):\n pass", "def save(self):\n pass", "def save(self):\n pass", "def save(self):\n pass", "def add_option_save(self):\n logger.debug(\"Adding save option\")\n btnsave = ttk.Button(self.optsframe,\n image=get_images().icons[\"save\"],\n command=self.save_items)\n btnsave.pack(padx=2, side=tk.RIGHT)\n Tooltip(btnsave,\n text=\"Save {}(s) to file\".format(self.tabname),\n wraplength=200)", "def onSaveNotesButton(self, button):\n pass", "def save():\n pass", "def saveButtonMethod(self):\n return AbstractAction(\"Save\")", "def on_save_button_clicked(self, widget):\n active_page = self.get_active_tab()\n active_page.save()", "def onSaveMarkersButton(self):\n fileName = self.seedsPath + self.fileNameSeedsLineEdit.text\n markupsNode = slicer.util.getNode( slicer.modules.markups.logic().GetActiveListID() )\n \n if self.fileNameSeedsComboBox.findText(self.fileNameSeedsLineEdit.text) == -1:\n self.fileNameSeedsComboBox.addItem(self.fileNameSeedsLineEdit.text)\n \n with open(fileName, \"w+\") as fp:\n fp.write(\"# columns = name, x, y, z, label\\n\")\n for i in range(markupsNode.GetNumberOfFiducials()): \n point_ras = [0, 0, 0]\n markupsNode.GetNthFiducialPosition(i, point_ras)\n name = markupsNode.GetNthFiducialLabel(i)\n label = int(markupsNode.GetNthControlPointDescription(i))\n fp.write(\"{};{};{};{};{}\\n\".format(name, point_ras[0], point_ras[1], point_ras[2], label))\n \n logging.info('Markers saved in ' + fileName)", "def on_press_save(self):\n\n if self.dbChk.isChecked():\n self.processed_fields['db'] = self.dbPath.text()\n self.dbForm = DBFormWindow(self.processed_fields, self)\n self.dbForm.show()\n\n if self.savePathChk.isChecked():\n if self.savePath.text():\n shutil.copy(self.tempFile[1], os.path.join(self.savePath.text(), self.saveName.text()+'.wav'))\n os.close(self.tempFile[0])\n os.remove(self.tempFile[1])\n\n QMessageBox.information(self, 'Saved', f'Saved to: {os.path.join(self.savePath.text(), self.saveName.text()+\".wav\")}')\n\n self.saveBtn.setEnabled(False)\n self.deleteBtn.setEnabled(False)\n self.recordBtn.setEnabled(True)\n self.recordBtn.setIcon(QIcon(r'.\\assets\\record.png'))\n self.recordShortCut.setEnabled(True)\n self.inputDropDown.setEnabled(True)\n self.canvasStack.setCurrentWidget(self.microphoneCanvas)\n self.state = 'Waiting'", "def save(self):\n\n pass", "def save_file(self):\n if self.select_path.text() != \"\":\n filepath = self.select_path.text()\n road_network = self.map_selection.currentText()\n if self.map_selection.currentText() == \"User Defined\":\n road_network = self.map_selection_user_defined.text()\n gen_xml = GenerateXML(filepath, road_network)\n gen_xml.main()\n # remember Road Network for future\n set_metadata(road_network_filepath=road_network)\n else:\n message = \"No export path was selected\"\n iface.messageBar().pushMessage(\"Warning\", message, level=Qgis.Warning)\n QgsMessageLog.logMessage(message, level=Qgis.Warning)", "def tool_save_clicked(self, widget, data=None):\n self.communicator.save_project()", "def save_journey():\n destination = request.form.get('destination_id', ''), request.form.get('destination_name', '')\n origin = request.form.get('origin_id', ''), request.form.get('origin_name', '')\n if '' not in destination or '' not in origin:\n trip_db: Cache = g.trip_db\n trip_db.read_db()\n trip_db.write_db((origin, destination))\n print(trip_db.data)\n return redirect('/')", "def save(self, *args, **kwargs) -> None:\n pass", "def save(self, *args, **kwargs) -> None:\n pass" ]
[ "0.6717724", "0.67038506", "0.6494957", "0.64152914", "0.6392956", "0.62395084", "0.6176817", "0.6160083", "0.6086105", "0.6084194", "0.60714966", "0.6068572", "0.60467327", "0.60467327", "0.60467327", "0.60467327", "0.60467327", "0.60350305", "0.599849", "0.5995645", "0.597675", "0.5969995", "0.5949585", "0.5948958", "0.5942456", "0.59373987", "0.59267485", "0.5925426", "0.5916673", "0.5916673" ]
0.712687
0
Implementing compare vendor price list detail rates grid destination functionality
def compare_vendor_price_list_detail_rates_grid_destination(self): self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary() is_compared = False rates_grid_destination = self.get_specific_column_value_from_grid(self.vendor_price_list_detail_rates_grid_div_id, 1, self.destination_column_name) if self.buy_page_excel_data_dictionary["Destination"] == rates_grid_destination: is_compared = True return is_compared
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def compare_vendor_price_list_detail_rates_grid_rate(self):\n self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary()\n is_compared = False\n rates_grid_rate_column_element = self.wait().until(EC.presence_of_element_located(self.rates_grid_rate_column_locator))\n if self.buy_page_excel_data_dictionary[\"Rate1\"] == (rates_grid_rate_column_element.text).strip():\n is_compared = True\n return is_compared", "def compare_vendor_price_list_detail_dial_digits_grid_destination(self):\n self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary()\n is_compared = False\n dial_digits_grid_destination = self.get_specific_column_value_from_grid(self.vendor_price_list_detail_dial_digits_grid_div_id, 1, self.destination_column_name)\n if self.buy_page_excel_data_dictionary[\"Destination\"] == dial_digits_grid_destination:\n is_compared = True\n return is_compared", "def click_vendor_price_list_detail_reference_rates_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_reference_rates_grid_div_id)", "def click_vendor_price_list_detail_rates_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_rates_grid_div_id)", "def compare_rates_with_excel_data(self):\n is_compared = True\n self.grid_row_data.clear()\n self.grid_row_data.update({\"Rate1\": \"\", \"Rate2\": \"\", \"Rate3\": \"\", \"Begin Date\": \"\", \"Rating Method\": \"\"})\n self.buy_page_excel_data_dictionary[\"Rating Method\"] = self.buy_page_excel_data_dictionary[\"RatingMethod\"]\n self.buy_page_excel_data_dictionary[\"Begin Date\"] = self.buy_page_excel_data_dictionary[\"EffectiveDate\"]\n rates_grid_row_data = self.get_vendor_profile_page_grid_row_details(self.rates_grid_div_id, self.grid_row_data)\n for key in rates_grid_row_data:\n if rates_grid_row_data[key] != self.buy_page_excel_data_dictionary[key]:\n is_compared = False\n break\n return is_compared", "def select_vendor_price_list_detail_reference_rates_tab(self):\n self.click_element(self.vendor_price_list_details_reference_rates_tab_locator, True)", "def set_vendor_price_list_detail_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_rates_grid_div_id, grid_settings)", "def set_vendor_price_list_detail_reference_rates_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_reference_rates_grid_div_id, grid_settings)", "def test_get_rate_plan_by_product_and_rate_plan(self):\n pass", "def get_vendor_price_lists_details(self):\n try:\n self.vendor_price_lists_dict = self.get_grid_row_details(self.customer_price_list_grid_div_id, self.vendor_price_lists_dict)\n return True\n except:\n return False", "def test_get_rate_plan_by_product(self):\n pass", "def drag_vendor_price_list_detail_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_rates_grid_div_id, source_column, destination_column)", "def drag_vendor_price_list_detail_reference_rates_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_reference_rates_grid_div_id, source_column, destination_column)", "def compare_destination_with_excel_data(self):\n is_compared = False\n self.grid_row_data.clear()\n self.grid_row_data.update({\"Destination\": \"\", \"Begin Date\": \"\"})\n destination_grid_row_data = self.get_vendor_profile_page_grid_row_details(self.destinations_grid_div_id, self.grid_row_data)\n if self.buy_page_excel_data_dictionary[\"Destination\"] == destination_grid_row_data[\"Destination\"] and self.buy_page_excel_data_dictionary[\"EffectiveDate\"] == destination_grid_row_data[\"Begin Date\"]:\n is_compared = True\n return is_compared", "def click_vendor_price_list_detail_dial_digits_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_dial_digits_grid_div_id)", "def get_prices(uuid, card_format, price_source, price_list, card_type_order, price_data_json):\n if price_source not in price_data_json[uuid][card_format]:\n pass\n #print(f'Price source value of {price_source} is not available for {card_format} and {uuid}')\n else:\n source = price_data_json[uuid][card_format][price_source]\n if price_list not in source:\n pass\n #print(f'Price list value of {price_list} is not available for {price_source} and {uuid}')\n else:\n retail = source[price_list]\n for type in card_type_order:\n if type in retail:\n return retail[type]", "def sub_tax_cost_transfer_pricing_eur(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel\"])\n # df_circuit = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel_circuit\"])\n # df_other = manager.get_dataframe(tables[\"dwc_bok_t_canco_other\"])\n # df_transfer = manager.get_dataframe(tables[\"dwc_bok_t_canco_transfer\"])\n # df_endow = manager.get_dataframe(tables[\"dwc_bok_t_canco_endowments\"])\n # df_extra = manager.get_dataframe(tables[\"dwc_bok_t_canco_extra\"])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\", \"invoicing_company\", \"creation_date\",\n \"booking_currency\")\n\n df_hotel = sub_tax_transfer_pricing_eur_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_tax_transfer_pricing_eur_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_tax_transfer_pricing_eur_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_tax_transfer_pricing_eur_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_tax_transfer_pricing_eur_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_tax_transfer_pricing_eur_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canco = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canco = df_impuesto_canco.groupBy(\"operative_incoming\", \"booking_id\") \\\n .agg({'impuesto_canco': 'sum'}).withColumnRenamed(\"SUM(impuesto_canco)\", \"impuesto_canco\")\n\n df_impuesto_canco = df_impuesto_canco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n df_fields = df_fields.join(df_impuesto_canco, [df_fields.operative_incoming == df_impuesto_canco.seq_rec,\n df_fields.booking_id == df_impuesto_canco.seq_res],\n 'left_outer').drop(\"seq_rec\", \"seq_res\")\n\n df_addcanco = sub_transfer_pricing_aux_add_canco(manager, df_fields, seq_recs, seq_reservas, EUR)\n\n df_addcanco = df_addcanco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add add_impuesto_canco\n df_fields = df_fields.join(df_addcanco, [df_fields.operative_incoming == df_addcanco.seq_rec,\n df_fields.booking_id == df_addcanco.seq_res],\n \"left_outer\").drop(df_addcanco.seq_rec).drop(df_addcanco.seq_res)\n\n df_fields = df_fields.na.fill({'impuesto_canco': 0, 'add_impuesto_canco': 0})\n\n df_fields = df_fields.withColumn(\"Tax_Cost_Transfer_pricing_EUR\",\n df_fields.impuesto_canco + df_fields.add_impuesto_canco) \\\n .drop(\"impuesto_canco\", \"add_impuesto_canco\")\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canco, df_addcanco\n\n return df_fields", "def get_all_rows_of_rate_increase_violation_column_from_vendor_price_list_detail_grid(self, column_name):\n self.all_row_data = []\n self.wait_for_ajax_spinner_load()\n row_count = len(self.wait().until(EC.presence_of_all_elements_located(self.vendor_price_list_detail_row_locator), 'vendor price list details row element locator not found before specified time'))\n next_page = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (self.vendor_price_list_detail_rates_grid_div_id, column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n row = 1\n try:\n while row < row_count + 1:\n column_value_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tbody[@role='rowgroup']/tr[%s]/td[%s]\" % (self.vendor_price_list_detail_rates_grid_div_id, row, str(column_index)))\n column_value_element = self.wait(10).until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n column_value = str(column_value_element.get_attribute(\"innerHTML\"))\n if column_value not in self.all_row_data:\n self.all_row_data.append(column_value)\n else:\n row_count += 1\n if row % 50 == 0:\n next_page = True\n next_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-arrow-e']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(next_page_button_locator, True)\n row = 0\n row_count -= 50\n row += 1\n except WebDriverException:\n pass\n if next_page is True:\n first_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-seek-w']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(first_page_button_locator, True)\n return self.all_row_data", "def verify_rates_grid_columns(self):\n is_verified = True\n if(self.is_element_present(self.rates_grid_destination_column_locator) == False):\n is_verified = False\n if (is_verified != False):\n for column_name in self.compare_price_lists_list:\n rates_grid_new_column_locator = (By.XPATH, \"//div[@id='divComaprePriceListRate']/descendant::span[contains(text(),'%s')]\" % column_name)\n if (self.is_element_present(rates_grid_new_column_locator) == False):\n is_verified = False\n break\n return is_verified", "def compare_qos_with_excel_data(self):\n self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary()\n is_compared = True\n self.grid_row_data.clear()\n self.grid_row_data.update({\"MNP\": \"\", \"Comments\": \"\"})\n qos_grid_row_data = self.get_vendor_profile_page_grid_row_details(self.qos_grid_div_id, self.grid_row_data)\n for key in qos_grid_row_data:\n if qos_grid_row_data[key] != self.buy_page_excel_data_dictionary[key]:\n is_compared = False\n break\n return is_compared", "def sub_tax_cost_transfer_pricing(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel\"])\n # df_circuit = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel_circuit\"])\n # df_other = manager.get_dataframe(tables[\"dwc_bok_t_canco_other\"])\n # df_transfer = manager.get_dataframe(tables[\"dwc_bok_t_canco_transfer\"])\n # df_endow = manager.get_dataframe(tables[\"dwc_bok_t_canco_endowments\"])\n # df_extra = manager.get_dataframe(tables[\"dwc_bok_t_canco_extra\"])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\", \"invoicing_company\", \"creation_date\",\n \"booking_currency\")\n\n df_hotel = sub_tax_cost_transfer_pricing_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_tax_cost_transfer_pricing_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_tax_cost_transfer_pricing_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_tax_cost_transfer_pricing_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_tax_cost_transfer_pricing_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_tax_cost_transfer_pricing_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canco = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canco = df_impuesto_canco.groupBy(\"operative_incoming\", \"booking_id\") \\\n .agg({'impuesto_canco': 'sum'}).withColumnRenamed(\"SUM(impuesto_canco)\", \"impuesto_canco\")\n\n df_impuesto_canco = df_impuesto_canco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add impuesto_canco\n df_fields = df_fields.join(df_impuesto_canco, [df_fields.operative_incoming == df_impuesto_canco.seq_rec,\n df_fields.booking_id == df_impuesto_canco.seq_res],\n 'left_outer').drop(\"seq_rec\", \"seq_res\")\n\n df_addcanco = sub_transfer_pricing_aux_add_canco(manager, df_fields, seq_recs, seq_reservas)\n\n df_addcanco = df_addcanco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add add_impuesto_canco\n df_fields = df_fields.join(df_addcanco, [df_fields.operative_incoming == df_addcanco.seq_rec,\n df_fields.booking_id == df_addcanco.seq_res],\n \"left_outer\").drop(df_addcanco.seq_rec).drop(df_addcanco.seq_res)\n\n df_fields = df_fields.na.fill({'impuesto_canco': 0, 'add_impuesto_canco': 0})\n\n df_fields = df_fields.withColumn(\"Tax_Cost_Transfer_pricing\",\n df_fields.impuesto_canco + df_fields.add_impuesto_canco) \\\n .drop(\"impuesto_canco\", \"add_impuesto_canco\")\n\n df_fields = df_fields.withColumn(\"Tax_Cost_Transfer_pricing\", udf_round_ccy(df_fields.Tax_Cost_Transfer_pricing,\n df_fields.booking_currency))\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canco, df_addcanco, df_aux\n\n return df_fields", "def sub_tax_sales_transfer_pricing(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables['dwc_bok_t_canco_hotel'])\n # df_circuit = manager.get_dataframe(tables['dwc_bok_t_canco_hotel_circuit'])\n # df_other = manager.get_dataframe(tables['dwc_bok_t_canco_other'])\n # df_transfer = manager.get_dataframe(tables['dwc_bok_t_canco_transfer'])\n # df_endow = manager.get_dataframe(tables['dwc_bok_t_canco_endowments'])\n # df_extra = manager.get_dataframe(tables['dwc_bok_t_canco_extra'])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\")\n\n df_hotel = sub_tax_sales_transfer_pricing_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_tax_sales_transfer_pricing_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_tax_sales_transfer_pricing_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_tax_sales_transfer_pricing_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_tax_sales_transfer_pricing_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_tax_sales_transfer_pricing_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canal = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canal = df_impuesto_canal.groupBy(\"seq_rec\", \"seq_reserva\") \\\n .agg({'impuesto_canal': 'sum'}).withColumnRenamed(\"SUM(impuesto_canal)\", \"Tax_Sales_Transfer_pricing\")\n\n df_fields = df_fields.join(df_impuesto_canal, [df_fields.operative_incoming == df_impuesto_canal.seq_rec,\n df_fields.booking_id == df_impuesto_canal.seq_reserva],\n 'left_outer').drop(df_impuesto_canal.seq_rec).drop(df_impuesto_canal.seq_reserva)\n\n df_fields = df_fields.na.fill({\"Tax_Sales_Transfer_pricing\": 0})\n\n df_fields = df_fields.withColumn(\"Tax_Sales_Transfer_pricing\",\n udf_round_ccy(df_fields.Tax_Sales_Transfer_pricing,\n df_fields.booking_currency))\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canal\n\n return df_fields", "def test_visualize_price_breakdown(self):\n pass", "def get_queryset(self): # NOQA\n rates = Rate.objects.filter(\n inventory__date__gte=self.kwargs.get('checkin'),\n inventory__date__lte=self.kwargs.get('checkout'),\n room__hotel__public_id=self.kwargs.get('hotel_id')\n ).values('room__public_id', 'price', \"inventory__date\")\n return rates", "def sub_tax_transfer_pricing_eur(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel\"])\n # df_circuit = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel_circuit\"])\n # df_other = manager.get_dataframe(tables[\"dwc_bok_t_canco_other\"])\n # df_transfer = manager.get_dataframe(tables[\"dwc_bok_t_canco_transfer\"])\n # df_endow = manager.get_dataframe(tables[\"dwc_bok_t_canco_endowments\"])\n # df_extra = manager.get_dataframe(tables[\"dwc_bok_t_canco_extra\"])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\", \"invoicing_company\", \"creation_date\",\n \"booking_currency\")\n\n df_hotel = sub_tax_transfer_pricing_eur_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_tax_transfer_pricing_eur_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_tax_transfer_pricing_eur_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_tax_transfer_pricing_eur_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_tax_transfer_pricing_eur_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_tax_transfer_pricing_eur_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canco = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canco = df_impuesto_canco.groupBy(\"operative_incoming\", \"booking_id\") \\\n .agg({'impuesto_canco': 'sum'}).withColumnRenamed(\"SUM(impuesto_canco)\", \"impuesto_canco\")\n\n df_impuesto_canco = df_impuesto_canco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n df_fields = df_fields.join(df_impuesto_canco, [df_fields.operative_incoming == df_impuesto_canco.seq_rec,\n df_fields.booking_id == df_impuesto_canco.seq_res],\n 'left_outer').drop(\"seq_rec\", \"seq_res\")\n\n df_addcanco = sub_transfer_pricing_aux_add_canco(manager, df_fields, seq_recs, seq_reservas, EUR)\n\n df_addcanco = df_addcanco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add add_impuesto_canco\n df_fields = df_fields.join(df_addcanco, [df_fields.operative_incoming == df_addcanco.seq_rec,\n df_fields.booking_id == df_addcanco.seq_res],\n \"left_outer\").drop(df_addcanco.seq_rec).drop(df_addcanco.seq_res)\n\n df_fields = df_fields.na.fill({'impuesto_canco': 0, 'add_impuesto_canco': 0})\n\n df_fields = df_fields.withColumn(\"Tax_Transfer_pricing_EUR\",\n df_fields.impuesto_canco + df_fields.add_impuesto_canco) \\\n .drop(\"impuesto_canco\", \"add_impuesto_canco\")\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canco, df_addcanco\n\n return df_fields", "def verify_vendor_price_lists_details(self, row_data):\n return self.verify_grid_row_details(self.customer_price_list_grid_div_id, row_data)", "def calc_Cinv_boiler(Q_design_W, technology_type, boiler_cost_data):\n Capex_a_Boiler_USD = 0.0\n Opex_a_fix_Boiler_USD = 0.0\n Capex_Boiler_USD = 0.0\n\n if Q_design_W > 0.0:\n boiler_cost_data = boiler_cost_data[boiler_cost_data['code'] == technology_type]\n # if the Q_design is below the lowest capacity available for the technology, then it is replaced by the least\n # capacity for the corresponding technology from the database\n if Q_design_W < boiler_cost_data.iloc[0]['cap_min']:\n Q_design_W = boiler_cost_data.iloc[0]['cap_min']\n max_boiler_size = boiler_cost_data.iloc[0]['cap_max']\n\n if Q_design_W <= max_boiler_size:\n\n boiler_cost_data = boiler_cost_data[\n (boiler_cost_data['cap_min'] <= Q_design_W) & (boiler_cost_data['cap_max'] > Q_design_W)]\n\n Inv_a = boiler_cost_data.iloc[0]['a']\n Inv_b = boiler_cost_data.iloc[0]['b']\n Inv_c = boiler_cost_data.iloc[0]['c']\n Inv_d = boiler_cost_data.iloc[0]['d']\n Inv_e = boiler_cost_data.iloc[0]['e']\n Inv_IR = boiler_cost_data.iloc[0]['IR_%']\n Inv_LT = boiler_cost_data.iloc[0]['LT_yr']\n Inv_OM = boiler_cost_data.iloc[0]['O&M_%'] / 100.0\n\n InvC = Inv_a + Inv_b * (Q_design_W) ** Inv_c + (Inv_d + Inv_e * Q_design_W) * log(Q_design_W)\n\n Capex_a_Boiler_USD = calc_capex_annualized(InvC, Inv_IR, Inv_LT)\n Opex_a_fix_Boiler_USD = InvC * Inv_OM\n Capex_Boiler_USD = InvC\n\n else:\n number_of_boilers = int(ceil(Q_design_W / max_boiler_size))\n Q_nom_W = Q_design_W / number_of_boilers\n\n boiler_cost_data = boiler_cost_data[\n (boiler_cost_data['cap_min'] <= Q_nom_W) & (boiler_cost_data['cap_max'] > Q_nom_W)]\n\n Inv_a = boiler_cost_data.iloc[0]['a']\n Inv_b = boiler_cost_data.iloc[0]['b']\n Inv_c = boiler_cost_data.iloc[0]['c']\n Inv_d = boiler_cost_data.iloc[0]['d']\n Inv_e = boiler_cost_data.iloc[0]['e']\n Inv_IR = boiler_cost_data.iloc[0]['IR_%']\n Inv_LT = boiler_cost_data.iloc[0]['LT_yr']\n Inv_OM = boiler_cost_data.iloc[0]['O&M_%'] / 100.0\n\n InvC = (Inv_a + Inv_b * (Q_nom_W) ** Inv_c + (Inv_d + Inv_e * Q_nom_W) * log(Q_nom_W)) * number_of_boilers\n\n Capex_a_Boiler_USD = calc_capex_annualized(InvC, Inv_IR, Inv_LT)\n Opex_a_fix_Boiler_USD = InvC * Inv_OM\n Capex_Boiler_USD = InvC\n\n return Capex_a_Boiler_USD, Opex_a_fix_Boiler_USD, Capex_Boiler_USD", "def compare_eia_heat_rates_to_ampl_projs(year):\n\n db_gen_projects = pull_generation_projects_data(gen_scenario_id=1).rename(\n columns={'name':'Plant Name', 'gen_tech':'Prime Mover'})\n db_gen_projects.loc[:,'Prime Mover'].replace(\n {\n 'Coal_Steam_Turbine':'ST',\n 'Gas_Steam_Turbine':'ST',\n 'Gas_Combustion_Turbine':'GT',\n 'Gas_Combustion_Turbine_Cogen':'GT',\n 'CCGT':'CC',\n 'DistillateFuelOil_Combustion_Turbine':'GT',\n 'DistillateFuelOil_Internal_Combustion_Engine':'IC',\n 'Geothermal':'ST',\n 'Gas_Internal_Combustion_Engine':'IC',\n 'Bio_Gas_Internal_Combustion_Engine':'IC',\n 'Bio_Gas_Steam_Turbine':'ST'\n },\n inplace=True)\n eia_gen_projects = filter_plants_by_region_id(13, year)\n\n df = pd.merge(db_gen_projects, eia_gen_projects,\n on=['Plant Name','Prime Mover'], how='left').loc[:,[\n 'Plant Name','gen_tech','energy_source','full_load_heat_rate',\n 'Best Heat Rate','Prime Mover','Energy Source','Energy Source 2','Operating Year']]\n df = df[df['full_load_heat_rate']>0]\n\n print \"\\nPrinting intersection of DB and EIA generation projects that have a specified heat rate to heat_rate_comparison.tab\"\n \n fpath = os.path.join('processed_data','heat_rate_comparison.tab')\n with open(fpath, 'w') as outfile:\n df.to_csv(outfile, sep='\\t', header=True, index=False)\n\n return df", "def sub_transfer_pricing(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel\"])\n # df_circuit = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel_circuit\"])\n # df_other = manager.get_dataframe(tables[\"dwc_bok_t_canco_other\"])\n # df_transfer = manager.get_dataframe(tables[\"dwc_bok_t_canco_transfer\"])\n # df_endow = manager.get_dataframe(tables[\"dwc_bok_t_canco_endowments\"])\n # df_extra = manager.get_dataframe(tables[\"dwc_bok_t_canco_extra\"])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\", \"invoicing_company\",\n \"creation_date\", \"booking_currency\")\n\n df_hotel = sub_transfer_pricing_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_transfer_pricing_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_transfer_pricing_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_transfer_pricing_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_transfer_pricing_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_transfer_pricing_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canco = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canco = df_impuesto_canco.groupBy(\"operative_incoming\", \"booking_id\") \\\n .agg({'impuesto_canco': 'sum'}).withColumnRenamed(\"SUM(impuesto_canco)\", \"impuesto_canco\")\n\n df_impuesto_canco = df_impuesto_canco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add impuesto_canco\n df_fields = df_fields.join(df_impuesto_canco, [df_fields.operative_incoming == df_impuesto_canco.seq_rec,\n df_fields.booking_id == df_impuesto_canco.seq_res],\n 'left_outer').drop(\"seq_rec\", \"seq_res\")\n\n df_addcanco = sub_transfer_pricing_aux_add_canco(manager, df_fields, seq_recs, seq_reservas)\n\n df_addcanco = df_addcanco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add add_impuesto_canco\n df_fields = df_fields.join(df_addcanco, [df_fields.operative_incoming == df_addcanco.seq_rec,\n df_fields.booking_id == df_addcanco.seq_res],\n \"left_outer\").drop(df_addcanco.seq_rec).drop(df_addcanco.seq_res)\n\n df_fields = df_fields.na.fill({'impuesto_canco': 0, 'add_impuesto_canco': 0})\n\n df_fields = df_fields.withColumn(\"Transfer_pricing\", df_fields.impuesto_canco + df_fields.add_impuesto_canco) \\\n .drop(\"impuesto_canco\", \"add_impuesto_canco\")\n\n df_fields = df_fields.withColumn(\"Transfer_pricing\", udf_round_ccy(df_fields.Transfer_pricing,\n df_fields.booking_currency))\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canco, df_addcanco, df_aux\n\n return df_fields", "def compareData(self) : \n logger.info('Comparing data ...')\n showAll = self.ui.showAll_checkBox.isChecked()\n selData = self.getSelectedRowData()\n\n if selData : \n assetName = selData[self.setCols.index('Asset Name')]\n rootLoc, rootExists = self.getRootLoc(assetName)\n \n if rootExists : \n abcAssetHero = str(self.ui.asset_lineEdit.text())\n abcShotHero = str(self.ui.shot_lineEdit.text())\n add = None \n remove = None\n\n if self.ui.compareCurrent_checkBox.isChecked() : \n if abcAssetHero : \n if self.ui.asset_radioButton.isChecked() : \n add, remove = sd_utils.compareLoc(rootLoc, abcAssetHero)\n\n if abcShotHero : \n if self.ui.shot_radioButton.isChecked() : \n add, remove = sd_utils.compareLoc(rootLoc, abcShotHero)\n\n else : \n add, remove = sd_utils.compareAbc(abcShotHero, abcAssetHero)\n\n self.ui.compare_listWidget.clear()\n \n if not showAll : \n if add : \n print 'add', add\n for item in add : \n self.addListWidgetItem(item, color=self.green)\n\n if remove : \n print 'remove', remove\n for item in remove : \n self.addListWidgetItem(item, color=self.red)\n\n if showAll : \n rootLocs = sd_utils.getSceneLocator(rootLoc)\n\n for item in rootLocs : \n color = [0, 0, 0]\n\n if item in remove : \n color = self.red \n\n self.addListWidgetItem(item, color=color)\n\n if add : \n for item in add : \n self.addListWidgetItem(item, color=self.green)\n\n else : \n logger.info('No set found')" ]
[ "0.7282243", "0.66426474", "0.63583803", "0.6355037", "0.5928445", "0.58797383", "0.57747257", "0.57692534", "0.5715802", "0.567752", "0.5522854", "0.5438436", "0.54336697", "0.54138416", "0.5389457", "0.5359198", "0.5304087", "0.52684253", "0.5221472", "0.5190703", "0.5189784", "0.51789516", "0.51517564", "0.51222533", "0.510655", "0.51061505", "0.5101934", "0.50556505", "0.50440097", "0.49854994" ]
0.7443731
0
Implementing compare vendor price list detail dial digits grid destination functionality
def compare_vendor_price_list_detail_dial_digits_grid_destination(self): self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary() is_compared = False dial_digits_grid_destination = self.get_specific_column_value_from_grid(self.vendor_price_list_detail_dial_digits_grid_div_id, 1, self.destination_column_name) if self.buy_page_excel_data_dictionary["Destination"] == dial_digits_grid_destination: is_compared = True return is_compared
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def compare_vendor_price_list_detail_rates_grid_destination(self):\n self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary()\n is_compared = False\n rates_grid_destination = self.get_specific_column_value_from_grid(self.vendor_price_list_detail_rates_grid_div_id, 1, self.destination_column_name)\n if self.buy_page_excel_data_dictionary[\"Destination\"] == rates_grid_destination:\n is_compared = True\n return is_compared", "def compare_vendor_price_list_detail_rates_grid_rate(self):\n self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary()\n is_compared = False\n rates_grid_rate_column_element = self.wait().until(EC.presence_of_element_located(self.rates_grid_rate_column_locator))\n if self.buy_page_excel_data_dictionary[\"Rate1\"] == (rates_grid_rate_column_element.text).strip():\n is_compared = True\n return is_compared", "def click_vendor_price_list_detail_dial_digits_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_dial_digits_grid_div_id)", "def get_vendor_price_lists_details(self):\n try:\n self.vendor_price_lists_dict = self.get_grid_row_details(self.customer_price_list_grid_div_id, self.vendor_price_lists_dict)\n return True\n except:\n return False", "def select_vendor_price_list_detail_dial_digits_tab(self):\n self.click_element(self.vendor_price_list_details_dial_digits_tab_locator, True)", "def click_vendor_price_list_detail_reference_rates_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_reference_rates_grid_div_id)", "def verify_vendor_price_lists_details(self, row_data):\n return self.verify_grid_row_details(self.customer_price_list_grid_div_id, row_data)", "def click_vendor_price_list_detail_rates_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_rates_grid_div_id)", "def select_vendor_price_list_detail_reference_rates_tab(self):\n self.click_element(self.vendor_price_list_details_reference_rates_tab_locator, True)", "def test_visualize_price_breakdown(self):\n pass", "def test_get_rate_plan_by_product_and_rate_plan(self):\n pass", "def compare_price(self):\n if self.__product_price < self.__alert_price:\n #print(\"price drop...\")\n self.__alert_client = True\n self.__price_difference = self.__product_price - self.__alert_price\n else:\n #print(\"Price not reduced...\")\n self.__alert_client = False\n self.__price_difference = self.__product_price - self.__alert_price", "def compare_qos_with_excel_data(self):\n self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary()\n is_compared = True\n self.grid_row_data.clear()\n self.grid_row_data.update({\"MNP\": \"\", \"Comments\": \"\"})\n qos_grid_row_data = self.get_vendor_profile_page_grid_row_details(self.qos_grid_div_id, self.grid_row_data)\n for key in qos_grid_row_data:\n if qos_grid_row_data[key] != self.buy_page_excel_data_dictionary[key]:\n is_compared = False\n break\n return is_compared", "def test_callPrice(self):\n call_price1 = calculator.BlackScholes.call_price(**self.params_1)\n call_price2 = calculator.BlackScholes.call_price(**self.params_2)\n self.assertAlmostEqual(call_price1,10.45,delta=0.01)\n self.assertAlmostEqual(call_price2,7.965,delta=0.01)", "def get_prices(uuid, card_format, price_source, price_list, card_type_order, price_data_json):\n if price_source not in price_data_json[uuid][card_format]:\n pass\n #print(f'Price source value of {price_source} is not available for {card_format} and {uuid}')\n else:\n source = price_data_json[uuid][card_format][price_source]\n if price_list not in source:\n pass\n #print(f'Price list value of {price_list} is not available for {price_source} and {uuid}')\n else:\n retail = source[price_list]\n for type in card_type_order:\n if type in retail:\n return retail[type]", "def compare_destination_with_excel_data(self):\n is_compared = False\n self.grid_row_data.clear()\n self.grid_row_data.update({\"Destination\": \"\", \"Begin Date\": \"\"})\n destination_grid_row_data = self.get_vendor_profile_page_grid_row_details(self.destinations_grid_div_id, self.grid_row_data)\n if self.buy_page_excel_data_dictionary[\"Destination\"] == destination_grid_row_data[\"Destination\"] and self.buy_page_excel_data_dictionary[\"EffectiveDate\"] == destination_grid_row_data[\"Begin Date\"]:\n is_compared = True\n return is_compared", "def drag_vendor_price_list_detail_dial_digits_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_dial_digits_grid_div_id, source_column, destination_column)", "def drag_vendor_price_list_detail_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_rates_grid_div_id, source_column, destination_column)", "def query_part_info(parts, distributors, currency):\n solved = set()\n # Loop through the parts looking for those sourced by local distributors\n # that won't be found online. Place any user-added info for these parts\n # (such as pricing) into the part dictionary.\n for p in parts:\n # Find the manufacturer's part number if it exists.\n pn = p.fields.get('manf#') # Returns None if no manf# field.\n\n # Now look for catalog number, price list and webpage link for this part.\n for dist in distributors:\n cat_num = p.fields.get(dist + ':cat#')\n pricing = p.fields.get(dist + ':pricing')\n link = p.fields.get(dist + ':link')\n avail = p.fields.get(dist + ':avail')\n if cat_num is None and pricing is None and link is None:\n continue\n\n cat_num = cat_num or pn or make_unique_catalog_number(p, dist)\n p.fields[dist + ':cat#'] = cat_num # Store generated cat#.\n # Get the DistData for this distributor\n dd = p.dd.get(dist, DistData())\n dd.part_num = cat_num\n\n if link:\n url_parts = list(urlsplit(link))\n if url_parts[0] == '':\n url_parts[0] = u'http'\n link = urlunsplit(url_parts)\n else:\n # This happens when no part URL is found.\n debug_obsessive('No part URL found for local \\'{}\\' distributor!'.format(dist))\n dd.url = link\n\n price_tiers = {}\n try:\n local_currency = re.findall('[a-zA-Z]{3}', pricing)[0].upper()\n except Exception:\n local_currency = currency\n old_pricing = pricing\n pricing = re.sub('[^0-9.;:]', '', pricing) # Keep only digits, decimals, delimiters.\n for qty_price in pricing.split(';'):\n splitted = qty_price.split(SEPRTR)\n if len(splitted) == 2:\n qty, price = splitted\n if local_currency:\n dd.currency = local_currency\n try:\n price_tiers[int(qty)] = float(price)\n except ValueError:\n warning(W_BADPRICE, 'Malformed pricing number: `{}` at {}'.format(old_pricing, p.refs))\n else:\n warning(W_BADPRICE, 'Malformed pricing entry: `{}` at {}'.format(qty_price, p.refs))\n # dd.moq = min(price_tiers.keys())\n if not price_tiers:\n # This happens when no pricing info is found.\n debug_obsessive('No pricing information found for local \\'{}\\' distributor!'.format(dist))\n dd.price_tiers = price_tiers\n\n # Availability\n if avail is not None:\n dd.qty_avail = avail\n\n # Update the DistData for this distributor\n p.dd[dist] = dd\n # We have data for this distributor. Avoid marking normal distributors.\n if dist in dist_local_template.api_distributors:\n solved.add(dist)\n return solved", "def test_get_rate_plan_by_product(self):\n pass", "def action_number(self, *args):\n # obj_ret = self.browse()\n if self.type == 'out_invoice':\n self._cr.execute(\n 'SELECT id, number '\n 'FROM account_wh_src '\n 'WHERE id IN (' + ','.join([str(item) for item in self.ids]) + ')')\n\n for (aws_id, number) in self._cr.fetchall():\n if not number:\n number = self.env['ir.sequence'].get(\n 'account.wh.src.%s' % self.type)\n self._cr.execute('UPDATE account_wh_src SET number=%s '\n 'WHERE id=%s', (number, aws_id))\n\n return True", "def dr14comp(a,b,av,bv):\n load=apload.ApLoad(apred='r11')\n dr14=apload.ApLoad(dr='dr14')\n\n i1,i2=match.match(a['APOGEE_ID'],b['APOGEE_ID'])\n gd = np.where((a['NVISITS'][i1] == b['NVISITS'][i2]) & (a['SNR'][i1]>75) )[0]\n a=a[i1[gd]]\n b=b[i2[gd]]\n \n j=np.argsort(a['VHELIO_AVG']-b['VHELIO_AVG'])\n \n fig,ax=plots.multi(1,3,hspace=0.3) \n pfig,pax=plots.multi(1,3,hspace=0.3) \n wfig,wax=plots.multi(1,3,hspace=0.3) \n chips=['a','b','c']\n for jj in j :\n j1=np.where(av['APOGEE_ID'] == a['APOGEE_ID'][jj])[0]\n j2=np.where(bv['APOGEE_ID'] == a['APOGEE_ID'][jj])[0]\n print(a['APOGEE_ID'][jj],a['RV_TEFF'][jj],b['RV_TEFF'][jj],a['SNR'][jj],b['SNR'][jj])\n for jjj,kkk in zip(j1,j2) : \n print(av['MJD'][jjj],av['PLATE'][jjj],av['FIELD'][jjj],av['SNR'][jjj],av['FIBERID'][jjj],av['VHELIO'][jjj],av['ESTVHELIO'][jjj])\n print(bv['MJD'][kkk],bv['PLATE'][kkk],bv['FIELD'][kkk],bv['SNR'][kkk],bv['FIBERID'][kkk],bv['VHELIO'][kkk],bv['ESTVHELIO'][kkk])\n va=load.apPlate(int(av['PLATE'][jjj]),av['MJD'][jjj])\n vsum=load.apVisitSum(int(av['PLATE'][jjj]),av['MJD'][jjj])[1].data\n f=np.where(vsum['FIBERID'] == av['FIBERID'][jjj])[0]\n print(vsum['RV_TEFF'][f])\n applot.chip(va,ax=ax,row=300-av['FIBERID'][jjj],color='r')\n applot.chip(va,ax=pax,row=300-av['FIBERID'][jjj],color='r',pixel=True)\n vb={}\n for chip in chips :\n tmp=fits.open(os.environ['APOGEE_REDUX']+'/r8/apo25m/{:04d}/{:05d}/apPlate-{:s}-{:04d}-{:05d}.fits'.format(\n int(bv['PLATE'][kkk]),bv['MJD'][kkk],chip,int(bv['PLATE'][kkk]),bv['MJD'][kkk]))\n vb[chip] = tmp\n vsum=fits.open(os.environ['APOGEE_REDUX']+'/r8/fields/apo25m/{:04d}/apVisitSum-{:04d}-{:05d}.fits'.format(\n int(bv['LOCATION_ID'][kkk]),int(bv['PLATE'][kkk]),bv['MJD'][kkk]))[1].data\n f=np.where(vsum['FIBERID'] == bv['FIBERID'][kkk])[0]\n print(vsum['RV_TEFF'][f])\n applot.chip(vb,ax=ax,row=300-bv['FIBERID'][kkk],color='b')\n applot.chip(vb,ax=pax,row=300-bv['FIBERID'][kkk],color='b',pixel=True)\n for ichip,chip in enumerate(chips) :\n wax[ichip].plot(va[chip][4].data[300-av['FIBERID'][jjj],:]-vb[chip][4].data[300-bv['FIBERID'][kkk],:])\n plt.show()\n pdb.set_trace()\n\n for ichip in range(3) :\n ax[ichip].cla()\n pax[ichip].cla()\n wax[ichip].cla()", "def set_vendor_price_list_detail_dial_digits_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_dial_digits_grid_div_id, grid_settings)", "def validate(self, entity, value, tracker, dispatcher, events_custom = None):\n global pincodes_available\n\n df = pd.read_excel('SampleModelSerialGEA.xlsx')\n print(entity, value)\n\n if entity == 'appliance': ###IMP :: can reduce appliance value to one allowed here\n print('Checking appliance')\n if value not in appliances:\n print('error')\n value = None\n\n else:\n idx_lis = df.index[df.loc[:, 'Product Line'].map(lambda x : x.lower()) == value.lower()].tolist()\n\n if tracker.get_slot('serialnumber') and not tracker.get_slot('serialnumber').upper() in [df.loc[i, 'Serial Number'] for i in idx_lis]:\n dispatcher.utter_message(\"For appliance %s serial number %s does not match !!!\" % (value, tracker.get_slot('serialnumber')))\n value = None\n\n elif tracker.get_slot('modelnumber') and not tracker.get_slot('modelnumber').upper() in [df.loc[i, 'Model Number'] for i in idx_lis]:\n dispatcher.utter_message(\"For appliance %s model number %s does not match !!!\" % (value, tracker.get_slot('modelnumber')))\n value = None\n\n\n if entity == 'modelnumber':\n print('Checking modelnumber')\n if not any(df.loc[:, 'Model Number'] == value.upper()):\n print('error')\n dispatcher.utter_message(\"Model no %s not in database !!!\" % tracker.get_slot('modelnumber'))\n value = None\n\n else:\n idx_lis = df.index[df.loc[:, 'Model Number'] == value.upper()].tolist()\n idx = idx_lis[0]\n\n if tracker.get_slot('appliance') and not df.loc[idx, 'Product Line'].lower() == tracker.get_slot('appliance'):\n dispatcher.utter_message(\"For model number %s appliance %s does not match !!!\" % (value, tracker.get_slot('appliance')))\n value = None\n\n elif not tracker.get_slot('appliance'):\n events_custom.extend([SlotSet(\"appliance\", df.loc[idx, 'Product Line'].lower())]) \n dispatcher.utter_message(\"For model number %s appliance %s added from database !!!\" % (value, df.loc[idx, 'Product Line'].lower()))\n\n if tracker.get_slot('serialnumber') and not tracker.get_slot('serialnumber').upper() in [df.loc[i, 'Serial Number'] for i in idx_lis]:\n dispatcher.utter_message(\"For model number %s serial number %s does not match !!!\" % (value, tracker.get_slot('serialnumber')))\n value = None\n\n\n if entity == 'serialnumber':\n print('Checking serialnumber')\n \n # if df[idx]['Serial Number'] == value.upper():\n # print('error')\n # value = None\n if not any(df.loc[:, 'Serial Number'] == value.upper()):\n print('error')\n dispatcher.utter_message(\"Serial no %s not in database !!!\" % tracker.get_slot('serialnumber'))\n value = None\n else:\n idx = df.index[df.loc[:, 'Serial Number'] == value.upper()].tolist()[0]\n \n if tracker.get_slot('modelnumber') and not df.loc[idx, 'Model Number'] == tracker.get_slot('modelnumber').upper():\n dispatcher.utter_message(\"For serial number %s model number %s does not match !!!\" % (value, tracker.get_slot('modelnumber')))\n value = None\n \n elif not tracker.get_slot('modelnumber'):\n print('Adding modelnumber given serialnumber')\n events_custom.extend([SlotSet('modelnumber', df.loc[idx, 'Model Number'].lower())]) \n dispatcher.utter_message(\"For serial number %s model number %s added from database !!!\" % (value, df.loc[idx, 'Model Number'].lower()))\n\n if tracker.get_slot('appliance') and not df.loc[idx, 'Product Line'].lower() == tracker.get_slot('appliance'):\n dispatcher.utter_message(\"For serial number %s appliance %s does not match !!!\" % (value, tracker.get_slot('appliance')))\n value = None\n\n elif not tracker.get_slot('appliance'):\n print('Adding appliance given serialnumber')\n events_custom.extend([SlotSet('appliance', df.loc[idx, 'Product Line'].lower())]) \n dispatcher.utter_message(\"For serial number %s appliance %s added from database !!!\" % (value, df.loc[idx, 'Product Line'].lower()))\n \n\n if entity == 'pincode':\n print(value)\n #print(value.type)\n if value not in pincodes_available:\n print('error')\n dispatcher.utter_message(\"The provided pincode %s is unavailable for service. We are sorry for the inconvinence.\" % value)\n value = None\n\n if entity == 'phonenumber':\n try:\n print('Checking phone number')\n r = requests.get('http://apilayer.net/api/validate?access_key=%s&number=%s&country_code=IN&format=1'%(PHONE_NO_API_KEY, value)).json()\n print(r)\n if not r['valid']:\n value = None\n except Exception as e:\n print(str(e))\n dispatcher.utter_message(\"Phone No is unable to be confirmed !!!! Believing entered info is correct\")\n\n return value", "def calc_Cinv_boiler(Q_design_W, technology_type, boiler_cost_data):\n Capex_a_Boiler_USD = 0.0\n Opex_a_fix_Boiler_USD = 0.0\n Capex_Boiler_USD = 0.0\n\n if Q_design_W > 0.0:\n boiler_cost_data = boiler_cost_data[boiler_cost_data['code'] == technology_type]\n # if the Q_design is below the lowest capacity available for the technology, then it is replaced by the least\n # capacity for the corresponding technology from the database\n if Q_design_W < boiler_cost_data.iloc[0]['cap_min']:\n Q_design_W = boiler_cost_data.iloc[0]['cap_min']\n max_boiler_size = boiler_cost_data.iloc[0]['cap_max']\n\n if Q_design_W <= max_boiler_size:\n\n boiler_cost_data = boiler_cost_data[\n (boiler_cost_data['cap_min'] <= Q_design_W) & (boiler_cost_data['cap_max'] > Q_design_W)]\n\n Inv_a = boiler_cost_data.iloc[0]['a']\n Inv_b = boiler_cost_data.iloc[0]['b']\n Inv_c = boiler_cost_data.iloc[0]['c']\n Inv_d = boiler_cost_data.iloc[0]['d']\n Inv_e = boiler_cost_data.iloc[0]['e']\n Inv_IR = boiler_cost_data.iloc[0]['IR_%']\n Inv_LT = boiler_cost_data.iloc[0]['LT_yr']\n Inv_OM = boiler_cost_data.iloc[0]['O&M_%'] / 100.0\n\n InvC = Inv_a + Inv_b * (Q_design_W) ** Inv_c + (Inv_d + Inv_e * Q_design_W) * log(Q_design_W)\n\n Capex_a_Boiler_USD = calc_capex_annualized(InvC, Inv_IR, Inv_LT)\n Opex_a_fix_Boiler_USD = InvC * Inv_OM\n Capex_Boiler_USD = InvC\n\n else:\n number_of_boilers = int(ceil(Q_design_W / max_boiler_size))\n Q_nom_W = Q_design_W / number_of_boilers\n\n boiler_cost_data = boiler_cost_data[\n (boiler_cost_data['cap_min'] <= Q_nom_W) & (boiler_cost_data['cap_max'] > Q_nom_W)]\n\n Inv_a = boiler_cost_data.iloc[0]['a']\n Inv_b = boiler_cost_data.iloc[0]['b']\n Inv_c = boiler_cost_data.iloc[0]['c']\n Inv_d = boiler_cost_data.iloc[0]['d']\n Inv_e = boiler_cost_data.iloc[0]['e']\n Inv_IR = boiler_cost_data.iloc[0]['IR_%']\n Inv_LT = boiler_cost_data.iloc[0]['LT_yr']\n Inv_OM = boiler_cost_data.iloc[0]['O&M_%'] / 100.0\n\n InvC = (Inv_a + Inv_b * (Q_nom_W) ** Inv_c + (Inv_d + Inv_e * Q_nom_W) * log(Q_nom_W)) * number_of_boilers\n\n Capex_a_Boiler_USD = calc_capex_annualized(InvC, Inv_IR, Inv_LT)\n Opex_a_fix_Boiler_USD = InvC * Inv_OM\n Capex_Boiler_USD = InvC\n\n return Capex_a_Boiler_USD, Opex_a_fix_Boiler_USD, Capex_Boiler_USD", "def _get_prix_tarif(self,cout,pricelist):\n cr = self._cr\n product=cout.name\n prix_tarif=0\n date=time.strftime('%Y-%m-%d') # Date du jour\n if pricelist:\n #Convertion du lot_mini de US vers UA\n min_quantity = self.env['product.uom']._compute_qty(cout.name.uom_id.id, cout.name.lot_mini, cout.name.uom_po_id.id)\n #TODO : Pour contourner un bug d'arrondi (le 31/01/2017)\n min_quantity=min_quantity+0.00000000001\n #TODO en utilisant la fonction repr à la place de str, cela ne tronque pas les décimales\n SQL=\"\"\"\n select ppi.price_surcharge\n from product_pricelist_version ppv inner join product_pricelist_item ppi on ppv.id=ppi.price_version_id\n where ppv.pricelist_id=\"\"\"+str(pricelist.id)+ \"\"\" \n and min_quantity<=\"\"\"+repr(min_quantity)+\"\"\"\n and (ppv.date_start <= '\"\"\"+date+\"\"\"' or ppv.date_start is null)\n and (ppv.date_end >= '\"\"\"+date+\"\"\"' or ppv.date_end is null)\n\n and ppi.product_id=\"\"\"+str(product.id)+ \"\"\" \n and (ppi.date_start <= '\"\"\"+date+\"\"\"' or ppi.date_start is null)\n and (ppi.date_end >= '\"\"\"+date+\"\"\"' or ppi.date_end is null)\n order by ppi.sequence\n limit 1\n \"\"\"\n cr.execute(SQL)\n result = cr.fetchall()\n for row in result:\n coef=1\n if min_quantity:\n coef=cout.name.lot_mini/min_quantity\n prix_tarif=row[0]/coef\n\n\n\n return prix_tarif", "def compareData(self) : \n logger.info('Comparing data ...')\n showAll = self.ui.showAll_checkBox.isChecked()\n selData = self.getSelectedRowData()\n\n if selData : \n assetName = selData[self.setCols.index('Asset Name')]\n rootLoc, rootExists = self.getRootLoc(assetName)\n \n if rootExists : \n abcAssetHero = str(self.ui.asset_lineEdit.text())\n abcShotHero = str(self.ui.shot_lineEdit.text())\n add = None \n remove = None\n\n if self.ui.compareCurrent_checkBox.isChecked() : \n if abcAssetHero : \n if self.ui.asset_radioButton.isChecked() : \n add, remove = sd_utils.compareLoc(rootLoc, abcAssetHero)\n\n if abcShotHero : \n if self.ui.shot_radioButton.isChecked() : \n add, remove = sd_utils.compareLoc(rootLoc, abcShotHero)\n\n else : \n add, remove = sd_utils.compareAbc(abcShotHero, abcAssetHero)\n\n self.ui.compare_listWidget.clear()\n \n if not showAll : \n if add : \n print 'add', add\n for item in add : \n self.addListWidgetItem(item, color=self.green)\n\n if remove : \n print 'remove', remove\n for item in remove : \n self.addListWidgetItem(item, color=self.red)\n\n if showAll : \n rootLocs = sd_utils.getSceneLocator(rootLoc)\n\n for item in rootLocs : \n color = [0, 0, 0]\n\n if item in remove : \n color = self.red \n\n self.addListWidgetItem(item, color=color)\n\n if add : \n for item in add : \n self.addListWidgetItem(item, color=self.green)\n\n else : \n logger.info('No set found')", "def click_compare_price_lists_button(self):\n self.click_element(self.compare_price_lists_button_locator, True)", "def drag_vendor_price_list_detail_reference_rates_grid_column(self, source_column, destination_column):\n self.wait_for_ajax_spinner_load(300)\n self.drag_grid_column(self.vendor_price_list_detail_reference_rates_grid_div_id, source_column, destination_column)", "def set_vendor_price_list_detail_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_rates_grid_div_id, grid_settings)" ]
[ "0.69088066", "0.66299367", "0.6117149", "0.6081349", "0.60306233", "0.5784268", "0.5770761", "0.5732137", "0.5709735", "0.54425746", "0.53622025", "0.53092676", "0.5295223", "0.5292526", "0.5235184", "0.5194165", "0.51922876", "0.51855004", "0.5147623", "0.51423776", "0.5084434", "0.50533056", "0.5026011", "0.5015003", "0.5000142", "0.499607", "0.49760154", "0.4968857", "0.49687827", "0.49442196" ]
0.7398431
0
Implementing compare vendor price list detail rates grid rate functionality
def compare_vendor_price_list_detail_rates_grid_rate(self): self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary() is_compared = False rates_grid_rate_column_element = self.wait().until(EC.presence_of_element_located(self.rates_grid_rate_column_locator)) if self.buy_page_excel_data_dictionary["Rate1"] == (rates_grid_rate_column_element.text).strip(): is_compared = True return is_compared
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def compare_rates_with_excel_data(self):\n is_compared = True\n self.grid_row_data.clear()\n self.grid_row_data.update({\"Rate1\": \"\", \"Rate2\": \"\", \"Rate3\": \"\", \"Begin Date\": \"\", \"Rating Method\": \"\"})\n self.buy_page_excel_data_dictionary[\"Rating Method\"] = self.buy_page_excel_data_dictionary[\"RatingMethod\"]\n self.buy_page_excel_data_dictionary[\"Begin Date\"] = self.buy_page_excel_data_dictionary[\"EffectiveDate\"]\n rates_grid_row_data = self.get_vendor_profile_page_grid_row_details(self.rates_grid_div_id, self.grid_row_data)\n for key in rates_grid_row_data:\n if rates_grid_row_data[key] != self.buy_page_excel_data_dictionary[key]:\n is_compared = False\n break\n return is_compared", "def test_get_rate_plan_by_product_and_rate_plan(self):\n pass", "def compare_vendor_price_list_detail_rates_grid_destination(self):\n self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary()\n is_compared = False\n rates_grid_destination = self.get_specific_column_value_from_grid(self.vendor_price_list_detail_rates_grid_div_id, 1, self.destination_column_name)\n if self.buy_page_excel_data_dictionary[\"Destination\"] == rates_grid_destination:\n is_compared = True\n return is_compared", "def test_get_rate_plan_by_product(self):\n pass", "def select_vendor_price_list_detail_reference_rates_tab(self):\n self.click_element(self.vendor_price_list_details_reference_rates_tab_locator, True)", "def click_vendor_price_list_detail_rates_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_rates_grid_div_id)", "def set_vendor_price_list_detail_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_rates_grid_div_id, grid_settings)", "def rates(self):\n raise NotImplementedError(\"Must be implemented by subclass.\")", "def set_vendor_price_list_detail_reference_rates_grid_settings(self, grid_settings):\n self.wait_for_ajax_spinner_load(300)\n self.set_grid_settings(self.vendor_price_list_detail_reference_rates_grid_div_id, grid_settings)", "def test_visualize_price_breakdown(self):\n pass", "def click_vendor_price_list_detail_reference_rates_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.vendor_price_list_detail_reference_rates_grid_div_id)", "def get_vendor_price_lists_details(self):\n try:\n self.vendor_price_lists_dict = self.get_grid_row_details(self.customer_price_list_grid_div_id, self.vendor_price_lists_dict)\n return True\n except:\n return False", "def calc_Cinv_boiler(Q_design_W, technology_type, boiler_cost_data):\n Capex_a_Boiler_USD = 0.0\n Opex_a_fix_Boiler_USD = 0.0\n Capex_Boiler_USD = 0.0\n\n if Q_design_W > 0.0:\n boiler_cost_data = boiler_cost_data[boiler_cost_data['code'] == technology_type]\n # if the Q_design is below the lowest capacity available for the technology, then it is replaced by the least\n # capacity for the corresponding technology from the database\n if Q_design_W < boiler_cost_data.iloc[0]['cap_min']:\n Q_design_W = boiler_cost_data.iloc[0]['cap_min']\n max_boiler_size = boiler_cost_data.iloc[0]['cap_max']\n\n if Q_design_W <= max_boiler_size:\n\n boiler_cost_data = boiler_cost_data[\n (boiler_cost_data['cap_min'] <= Q_design_W) & (boiler_cost_data['cap_max'] > Q_design_W)]\n\n Inv_a = boiler_cost_data.iloc[0]['a']\n Inv_b = boiler_cost_data.iloc[0]['b']\n Inv_c = boiler_cost_data.iloc[0]['c']\n Inv_d = boiler_cost_data.iloc[0]['d']\n Inv_e = boiler_cost_data.iloc[0]['e']\n Inv_IR = boiler_cost_data.iloc[0]['IR_%']\n Inv_LT = boiler_cost_data.iloc[0]['LT_yr']\n Inv_OM = boiler_cost_data.iloc[0]['O&M_%'] / 100.0\n\n InvC = Inv_a + Inv_b * (Q_design_W) ** Inv_c + (Inv_d + Inv_e * Q_design_W) * log(Q_design_W)\n\n Capex_a_Boiler_USD = calc_capex_annualized(InvC, Inv_IR, Inv_LT)\n Opex_a_fix_Boiler_USD = InvC * Inv_OM\n Capex_Boiler_USD = InvC\n\n else:\n number_of_boilers = int(ceil(Q_design_W / max_boiler_size))\n Q_nom_W = Q_design_W / number_of_boilers\n\n boiler_cost_data = boiler_cost_data[\n (boiler_cost_data['cap_min'] <= Q_nom_W) & (boiler_cost_data['cap_max'] > Q_nom_W)]\n\n Inv_a = boiler_cost_data.iloc[0]['a']\n Inv_b = boiler_cost_data.iloc[0]['b']\n Inv_c = boiler_cost_data.iloc[0]['c']\n Inv_d = boiler_cost_data.iloc[0]['d']\n Inv_e = boiler_cost_data.iloc[0]['e']\n Inv_IR = boiler_cost_data.iloc[0]['IR_%']\n Inv_LT = boiler_cost_data.iloc[0]['LT_yr']\n Inv_OM = boiler_cost_data.iloc[0]['O&M_%'] / 100.0\n\n InvC = (Inv_a + Inv_b * (Q_nom_W) ** Inv_c + (Inv_d + Inv_e * Q_nom_W) * log(Q_nom_W)) * number_of_boilers\n\n Capex_a_Boiler_USD = calc_capex_annualized(InvC, Inv_IR, Inv_LT)\n Opex_a_fix_Boiler_USD = InvC * Inv_OM\n Capex_Boiler_USD = InvC\n\n return Capex_a_Boiler_USD, Opex_a_fix_Boiler_USD, Capex_Boiler_USD", "def _generate_currency_rates(self, parsed_data):\n\n for line in self:\n rate_info = parsed_data.get(line.move_id.currency_id.name, None)\n\n if not rate_info:\n raise UserError(_(\"Your main currency (%s) is not supported by this exchange rate provider. Please choose another one.\", company.currency_id.name))\n\n base_currency_rate = rate_info[0]\n\n for currency, (rate, date_rate) in parsed_data.items():\n rate_value = rate/base_currency_rate\n if currency == line.user_currency_id.name:\n line.exchange_rate = rate_value\n line.price_unit = rate_value * line.user_amount", "def __call__(self, rate:'kW'):\n self.rate = rate\n self.cost = self.price * rate", "def test_company_EU_GR_vies_tax(self):\n self.assertEqual(self.policy.get_tax_rate(\"123456\", \"GR\"), (24, False))", "def getActiveCurrencies():", "def test_currency_rate(self):\n currency_name = ['USD'] * 4\n rates = [3.67, 4.07, 3.04, 3.89]\n helper.currency_loop_helper(get_historical_currency_rate, TestHistoricalRates.dates_rate,\n rates, currency_name)", "def indexa_generate_currency_rates(self, parsed_data):\n Currency = self.env['res.currency']\n CurrencyRate = self.env['res.currency.rate']\n\n today = fields.Date.today()\n for company in self:\n for currency, (rate, date_rate) in parsed_data.items():\n rate_value = 1/rate\n\n currency_object = Currency.search([('name','=',currency)])\n already_existing_rate = CurrencyRate.search([\n ('currency_id', '=', currency_object.id),\n ('name', '=', date_rate),\n ('company_id', '=', company.id)\n ])\n if already_existing_rate:\n already_existing_rate.rate = rate_value\n else:\n CurrencyRate.create({'currency_id': currency_object.id,\n 'rate': rate_value,\n 'name': date_rate,\n 'company_id': company.id})", "def get_prices(self):\n pass", "def sub_tax_cost_transfer_pricing_eur(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel\"])\n # df_circuit = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel_circuit\"])\n # df_other = manager.get_dataframe(tables[\"dwc_bok_t_canco_other\"])\n # df_transfer = manager.get_dataframe(tables[\"dwc_bok_t_canco_transfer\"])\n # df_endow = manager.get_dataframe(tables[\"dwc_bok_t_canco_endowments\"])\n # df_extra = manager.get_dataframe(tables[\"dwc_bok_t_canco_extra\"])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\", \"invoicing_company\", \"creation_date\",\n \"booking_currency\")\n\n df_hotel = sub_tax_transfer_pricing_eur_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_tax_transfer_pricing_eur_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_tax_transfer_pricing_eur_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_tax_transfer_pricing_eur_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_tax_transfer_pricing_eur_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_tax_transfer_pricing_eur_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canco = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canco = df_impuesto_canco.groupBy(\"operative_incoming\", \"booking_id\") \\\n .agg({'impuesto_canco': 'sum'}).withColumnRenamed(\"SUM(impuesto_canco)\", \"impuesto_canco\")\n\n df_impuesto_canco = df_impuesto_canco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n df_fields = df_fields.join(df_impuesto_canco, [df_fields.operative_incoming == df_impuesto_canco.seq_rec,\n df_fields.booking_id == df_impuesto_canco.seq_res],\n 'left_outer').drop(\"seq_rec\", \"seq_res\")\n\n df_addcanco = sub_transfer_pricing_aux_add_canco(manager, df_fields, seq_recs, seq_reservas, EUR)\n\n df_addcanco = df_addcanco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add add_impuesto_canco\n df_fields = df_fields.join(df_addcanco, [df_fields.operative_incoming == df_addcanco.seq_rec,\n df_fields.booking_id == df_addcanco.seq_res],\n \"left_outer\").drop(df_addcanco.seq_rec).drop(df_addcanco.seq_res)\n\n df_fields = df_fields.na.fill({'impuesto_canco': 0, 'add_impuesto_canco': 0})\n\n df_fields = df_fields.withColumn(\"Tax_Cost_Transfer_pricing_EUR\",\n df_fields.impuesto_canco + df_fields.add_impuesto_canco) \\\n .drop(\"impuesto_canco\", \"add_impuesto_canco\")\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canco, df_addcanco\n\n return df_fields", "def get_all_rows_of_rate_increase_violation_column_from_vendor_price_list_detail_grid(self, column_name):\n self.all_row_data = []\n self.wait_for_ajax_spinner_load()\n row_count = len(self.wait().until(EC.presence_of_all_elements_located(self.vendor_price_list_detail_row_locator), 'vendor price list details row element locator not found before specified time'))\n next_page = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (self.vendor_price_list_detail_rates_grid_div_id, column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n row = 1\n try:\n while row < row_count + 1:\n column_value_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tbody[@role='rowgroup']/tr[%s]/td[%s]\" % (self.vendor_price_list_detail_rates_grid_div_id, row, str(column_index)))\n column_value_element = self.wait(10).until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n column_value = str(column_value_element.get_attribute(\"innerHTML\"))\n if column_value not in self.all_row_data:\n self.all_row_data.append(column_value)\n else:\n row_count += 1\n if row % 50 == 0:\n next_page = True\n next_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-arrow-e']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(next_page_button_locator, True)\n row = 0\n row_count -= 50\n row += 1\n except WebDriverException:\n pass\n if next_page is True:\n first_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-seek-w']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(first_page_button_locator, True)\n return self.all_row_data", "def test_companies_company_id_data_tax_rates_get(self):\n pass", "def sub_tax_sales_transfer_pricing(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables['dwc_bok_t_canco_hotel'])\n # df_circuit = manager.get_dataframe(tables['dwc_bok_t_canco_hotel_circuit'])\n # df_other = manager.get_dataframe(tables['dwc_bok_t_canco_other'])\n # df_transfer = manager.get_dataframe(tables['dwc_bok_t_canco_transfer'])\n # df_endow = manager.get_dataframe(tables['dwc_bok_t_canco_endowments'])\n # df_extra = manager.get_dataframe(tables['dwc_bok_t_canco_extra'])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\")\n\n df_hotel = sub_tax_sales_transfer_pricing_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_tax_sales_transfer_pricing_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_tax_sales_transfer_pricing_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_tax_sales_transfer_pricing_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_tax_sales_transfer_pricing_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_tax_sales_transfer_pricing_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canal = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canal = df_impuesto_canal.groupBy(\"seq_rec\", \"seq_reserva\") \\\n .agg({'impuesto_canal': 'sum'}).withColumnRenamed(\"SUM(impuesto_canal)\", \"Tax_Sales_Transfer_pricing\")\n\n df_fields = df_fields.join(df_impuesto_canal, [df_fields.operative_incoming == df_impuesto_canal.seq_rec,\n df_fields.booking_id == df_impuesto_canal.seq_reserva],\n 'left_outer').drop(df_impuesto_canal.seq_rec).drop(df_impuesto_canal.seq_reserva)\n\n df_fields = df_fields.na.fill({\"Tax_Sales_Transfer_pricing\": 0})\n\n df_fields = df_fields.withColumn(\"Tax_Sales_Transfer_pricing\",\n udf_round_ccy(df_fields.Tax_Sales_Transfer_pricing,\n df_fields.booking_currency))\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canal\n\n return df_fields", "def getActiveCurrency():", "def evaluate_price(params, predictor=False):\n # def evaluate_price(alpha,theta,mu,sigma,lambda_,b,m,a,A0,predictor=False):\n alpha = params[0]\n\n theta = params[1]\n\n mu = params[2]\n\n sigma = params[3]\n\n lambda_ = params[4]\n\n b = params[5]\n m = params[6]\n a = params[7]\n A0 = params[8]\n\n call_data = data[data['is_call'] == 1]\n r = 0\n target_list = (call_data['Ask_price'].values +\n call_data['Bid_price'].values) / 2\n\n def eval_row(row): return call_price(\n row['S0'], row['K'], r, row['T'] / 360, alpha, theta, mu, sigma, lambda_, b, m, a, row['T'] / 360, A0=A0)\n pred_list = call_data.apply(eval_row, axis=1).values\n\n rmse = np.sqrt(np.mean((target_list - pred_list)**2))\n\n print(rmse)\n # plt.plot(target_list)\n # plt.plot(pred_list)\n plt.scatter(call_data['K'].values, 0.5 * call_data['Ask_price'] +\n 0.5 * call_data['Bid_price'], label='Real Data')\n plt.scatter(call_data['K'].values, pred_list,\n label='SV Time change', marker='+')\n plt.xlabel('K')\n plt.ylabel('Option Price')\n plt.legend()\n plt.show()\n if predictor:\n return rmse, pred_list\n return rmse", "def execQ4():\n frame = pan.DataFrame(data, columns=['Product', 'Price', 'Period'])\n tuna = frame[dFrame.Series_title_1 == \"Tuna - canned (supermarket only), 185g\"]\n expensive = tuna.sort_values(by=\"Price\", ascending=False).head(1)\n return expensive", "def sub_tax_transfer_pricing_eur(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel\"])\n # df_circuit = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel_circuit\"])\n # df_other = manager.get_dataframe(tables[\"dwc_bok_t_canco_other\"])\n # df_transfer = manager.get_dataframe(tables[\"dwc_bok_t_canco_transfer\"])\n # df_endow = manager.get_dataframe(tables[\"dwc_bok_t_canco_endowments\"])\n # df_extra = manager.get_dataframe(tables[\"dwc_bok_t_canco_extra\"])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\", \"invoicing_company\", \"creation_date\",\n \"booking_currency\")\n\n df_hotel = sub_tax_transfer_pricing_eur_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_tax_transfer_pricing_eur_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_tax_transfer_pricing_eur_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_tax_transfer_pricing_eur_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_tax_transfer_pricing_eur_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_tax_transfer_pricing_eur_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canco = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canco = df_impuesto_canco.groupBy(\"operative_incoming\", \"booking_id\") \\\n .agg({'impuesto_canco': 'sum'}).withColumnRenamed(\"SUM(impuesto_canco)\", \"impuesto_canco\")\n\n df_impuesto_canco = df_impuesto_canco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n df_fields = df_fields.join(df_impuesto_canco, [df_fields.operative_incoming == df_impuesto_canco.seq_rec,\n df_fields.booking_id == df_impuesto_canco.seq_res],\n 'left_outer').drop(\"seq_rec\", \"seq_res\")\n\n df_addcanco = sub_transfer_pricing_aux_add_canco(manager, df_fields, seq_recs, seq_reservas, EUR)\n\n df_addcanco = df_addcanco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add add_impuesto_canco\n df_fields = df_fields.join(df_addcanco, [df_fields.operative_incoming == df_addcanco.seq_rec,\n df_fields.booking_id == df_addcanco.seq_res],\n \"left_outer\").drop(df_addcanco.seq_rec).drop(df_addcanco.seq_res)\n\n df_fields = df_fields.na.fill({'impuesto_canco': 0, 'add_impuesto_canco': 0})\n\n df_fields = df_fields.withColumn(\"Tax_Transfer_pricing_EUR\",\n df_fields.impuesto_canco + df_fields.add_impuesto_canco) \\\n .drop(\"impuesto_canco\", \"add_impuesto_canco\")\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canco, df_addcanco\n\n return df_fields", "def sub_tax_cost_transfer_pricing(manager, df_fields, seq_recs, seq_reservas):\n # df_hotel = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel\"])\n # df_circuit = manager.get_dataframe(tables[\"dwc_bok_t_canco_hotel_circuit\"])\n # df_other = manager.get_dataframe(tables[\"dwc_bok_t_canco_other\"])\n # df_transfer = manager.get_dataframe(tables[\"dwc_bok_t_canco_transfer\"])\n # df_endow = manager.get_dataframe(tables[\"dwc_bok_t_canco_endowments\"])\n # df_extra = manager.get_dataframe(tables[\"dwc_bok_t_canco_extra\"])\n\n df_aux = df_fields.select(\"operative_incoming\", \"booking_id\", \"invoicing_company\", \"creation_date\",\n \"booking_currency\")\n\n df_hotel = sub_tax_cost_transfer_pricing_aux(manager, df_hotelt, seq_recs, seq_reservas, df_aux)\n df_circuit = sub_tax_cost_transfer_pricing_aux(manager, df_circuitt, seq_recs, seq_reservas, df_aux)\n df_other = sub_tax_cost_transfer_pricing_aux(manager, df_othert, seq_recs, seq_reservas, df_aux)\n df_transfer = sub_tax_cost_transfer_pricing_aux(manager, df_transfert, seq_recs, seq_reservas, df_aux)\n df_endow = sub_tax_cost_transfer_pricing_aux(manager, df_endowt, seq_recs, seq_reservas, df_aux)\n df_extra = sub_tax_cost_transfer_pricing_aux_extra(manager, df_extrat, seq_recs, seq_reservas, df_aux)\n\n df_impuesto_canco = df_hotel.union(df_circuit).union(df_other).union(df_transfer).union(df_endow).union(\n df_extra)\n\n df_impuesto_canco = df_impuesto_canco.groupBy(\"operative_incoming\", \"booking_id\") \\\n .agg({'impuesto_canco': 'sum'}).withColumnRenamed(\"SUM(impuesto_canco)\", \"impuesto_canco\")\n\n df_impuesto_canco = df_impuesto_canco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add impuesto_canco\n df_fields = df_fields.join(df_impuesto_canco, [df_fields.operative_incoming == df_impuesto_canco.seq_rec,\n df_fields.booking_id == df_impuesto_canco.seq_res],\n 'left_outer').drop(\"seq_rec\", \"seq_res\")\n\n df_addcanco = sub_transfer_pricing_aux_add_canco(manager, df_fields, seq_recs, seq_reservas)\n\n df_addcanco = df_addcanco.withColumnRenamed(\"operative_incoming\", \"seq_rec\") \\\n .withColumnRenamed(\"booking_id\", \"seq_res\")\n\n # add add_impuesto_canco\n df_fields = df_fields.join(df_addcanco, [df_fields.operative_incoming == df_addcanco.seq_rec,\n df_fields.booking_id == df_addcanco.seq_res],\n \"left_outer\").drop(df_addcanco.seq_rec).drop(df_addcanco.seq_res)\n\n df_fields = df_fields.na.fill({'impuesto_canco': 0, 'add_impuesto_canco': 0})\n\n df_fields = df_fields.withColumn(\"Tax_Cost_Transfer_pricing\",\n df_fields.impuesto_canco + df_fields.add_impuesto_canco) \\\n .drop(\"impuesto_canco\", \"add_impuesto_canco\")\n\n df_fields = df_fields.withColumn(\"Tax_Cost_Transfer_pricing\", udf_round_ccy(df_fields.Tax_Cost_Transfer_pricing,\n df_fields.booking_currency))\n\n del df_hotel, df_circuit, df_other, df_transfer, df_endow, df_extra, df_impuesto_canco, df_addcanco, df_aux\n\n return df_fields", "def verify_vendor_price_lists_details(self, row_data):\n return self.verify_grid_row_details(self.customer_price_list_grid_div_id, row_data)" ]
[ "0.63370734", "0.60845774", "0.60595584", "0.602486", "0.58634794", "0.5762692", "0.5760641", "0.5752293", "0.57230544", "0.56713235", "0.56500137", "0.5597601", "0.5579603", "0.55513275", "0.5518962", "0.551297", "0.5499407", "0.5466408", "0.546308", "0.54610395", "0.5433152", "0.5384661", "0.5376861", "0.53698915", "0.53301775", "0.5326263", "0.5315522", "0.530767", "0.53066987", "0.53045225" ]
0.7548665
0
Implementing get reference rates tab analyzed rate functionality
def get_reference_rates_tab_analyzed_rate(self): return self.get_text_from_element(self.reference_rates_tab_analyzed_tab_column_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def rates(self):\n raise NotImplementedError(\"Must be implemented by subclass.\")", "def get_current_rate(self):\n pass", "def select_reference_rates_tab(self):\n self.select_static_tab(self.reference_rates_tab_locator, True)", "def calculateDataRate(self):\n pass", "def rate(self):\n return self.brate / FAC", "def rates(self):\n return self._rates", "def data_rate(self):\n return self._data_rate", "def get_rates(table_id):\n fields = [\"0\",\"0\",\"0\",\"0\",\"0\",\"0\"]\n for pos, name in enumerate(rates_key_list):\n full_table_id = RATES_TABLE_PREFIX + table_id\n counter_data = self.db.get(self.db.COUNTERS_DB, full_table_id, name)\n if counter_data is None:\n fields[pos] = STATUS_NA\n elif fields[pos] != STATUS_NA:\n fields[pos] = float(counter_data)\n cntr = RateStats._make(fields)\n return cntr", "def rate(self):\n return self.__rate", "def rate(self):\n return self._rate", "def get_rate(parent=None):\n dialog = RateDialog(parent)\n dialog.exec_()\n rate = dialog.rate\n return rate", "def exchange_rate(self):\n res = r.get(self.url + self.current_rate)\n return self.execute(res)", "def get_rate(currency, date):\n status = 400\n while status != 200:\n url = (\"http://api.nbp.pl/api/exchangerates/rates/A/%s/%d-%02d-%02d?format=json\" %\n (currency, date.year, date.month, date.day))\n\n response = requests.get(url)\n status = response.status_code\n if status != 200:\n date = date - datetime.timedelta(1)\n\n tree = json.loads(response.content)\n assert len(tree['rates']) == 1\n print_rate_info(tree['rates'])\n return (tree['rates'][0]['mid'], date)", "def _do_get_rate(self):\n rate = {\n 1: \"1 : Helium Probe in FAST rate\",\n 0: \"0 : Helium Probe in SLOW rate\"\n }\n result = self._execute('X')\n return rate.get(int(format(int(result[5:7]), '08b')[6]), \"Unknown\")", "def get_scan_rate(self):\n raise NotImplementedError", "def getActiveCurrencies():", "def select_vendor_price_list_detail_reference_rates_tab(self):\n self.click_element(self.vendor_price_list_details_reference_rates_tab_locator, True)", "def test_retire_rate_plan(self):\n pass", "def getRate(self) -> int:\n if (self._total_stake.get() + self._daily_reward.get()) == 0:\n rate = DENOMINATOR\n else:\n rate = (self._total_stake.get() + self._daily_reward.get()) * DENOMINATOR // self.sICX_score.totalSupply()\n return rate", "def api_call(cls, currency):\n headers = {\"x-accept-version\": \"2.0.0\", \"Accept\": \"application/json\"}\n r = requests.get(cls.API_URL + currency, headers=headers)\n r.raise_for_status()\n return r.json()[\"data\"][\"rate\"]", "def getMostRecentRelevantRate(self, currency_rates, reference_date=QDate.currentDate().toPyDate()):\n\n try:\n for date in currency_rates:\n if QDate.fromString(date, \"yyyy-MM-dd\").toPyDate() <= reference_date and currency_rates[date] != 'N/A':\n return currency_rates[date]\n except Exception as e:\n print('Could not retrieve any relevant rate')\n print(e)", "def test_companies_company_id_data_tax_rates_get(self):\n pass", "def getActiveCurrency():", "def get_rate(self, t):\n return self.rates[bisect.bisect(self.change_times, t) - 1]", "def get_rates(src, dst):\n if not dst:\n dst = ''\n\n request = '{}{}'.format(api_url, rates_query.replace(':?', '\"{}{}\"'.format(src, dst)))\n\n response = web.get(urllib.quote(request, ':/?&=*'))\n\n response.raise_for_status()\n\n rates = response.json()\n\n rate_resp = rates['query']['results']['rate']\n\n if rate_resp['Rate'] == 'N/A':\n return -1\n\n return Decimal(rate_resp['Rate'])", "def test_get_all_rate_plans(self):\n pass", "def GetOpsRates():\n return GetDataFromCsvFile('ops_rates.csv')", "def _set_rate(self):\r\n interval = self.data.iloc[2, 0] - self.data.iloc[1, 0]\r\n self.rate = int(1 / interval)", "def select_rates_tab(self):\n self.select_static_tab(self.rates_tab_locator, True)", "def GetCountRates (self, arguments=None) :\n\t\treturn \tself.lib.PH_GetCountRate (self.DEV_INDX, ctypes.c_int(0)),\n\t\t\t\tself.lib.PH_GetCountRate (self.DEV_INDX, ctypes.c_int(1))" ]
[ "0.6990552", "0.66156983", "0.6201779", "0.61440897", "0.6085955", "0.59836334", "0.598278", "0.58565754", "0.57975405", "0.5742415", "0.5740475", "0.56852716", "0.56798387", "0.56295514", "0.5606371", "0.56015044", "0.5600266", "0.5587584", "0.5586825", "0.55694264", "0.5565084", "0.55453175", "0.55059975", "0.5481953", "0.54731077", "0.54479", "0.54418874", "0.54168284", "0.54146886", "0.54144406" ]
0.7216807
0
Implementing get reference rates tab effective date functionality
def get_reference_rates_tab_effective_date(self): return self.get_specific_column_value_from_grid(self.vendor_price_list_detail_reference_rates_grid_div_id, 1, column_name=self.effective_date_column_name)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _rate_dates(self, common_object):\n if common_object.IsKindOf(acm.FCashFlow):\n start_date = common_object.StartDate()\n elif common_object.IsKindOf(acm.FReset):\n start_date = common_object.Day()\n else:\n message = \"Rate dates for {0} object are not defined\".format(\n type(common_object))\n raise ProvisionHandlerError(message)\n\n end_date = acm.Time().DateAddDelta(start_date, 0, 3, 0)\n end_date = self._adjust_to_banking_day(end_date)\n\n return (start_date, end_date)", "def get_reference_rates_tab_analyzed_rate(self):\n return self.get_text_from_element(self.reference_rates_tab_analyzed_tab_column_locator)", "def getMostRecentRelevantRate(self, currency_rates, reference_date=QDate.currentDate().toPyDate()):\n\n try:\n for date in currency_rates:\n if QDate.fromString(date, \"yyyy-MM-dd\").toPyDate() <= reference_date and currency_rates[date] != 'N/A':\n return currency_rates[date]\n except Exception as e:\n print('Could not retrieve any relevant rate')\n print(e)", "def _get_prorata_interval_rate(self, cr, uid, change_date, context=None):\n month_days = calendar.monthrange(change_date.year,\n change_date.month)[1]\n start_date = add_months(change_date, 1)\n end_date = start_date.replace(day=month_days)\n used_days = month_days - change_date.day\n ptx = self._prorata_rate(used_days, month_days)\n\n return start_date, end_date, ptx", "def select_reference_rates_tab(self):\n self.select_static_tab(self.reference_rates_tab_locator, True)", "def valuation(self, from_date=None):\n import pandas_datareader.data as pdr\n import datetime\n to_date = datetime.date.today()\n if not from_date: from_date = to_date - datetime.timedelta(days=1)\n px = pdr.DataReader(self.ticker, 'yahoo', from_date, to_date)\n\n f = self.Fundamentals\n\n print(\"OF COURSE \", 7, f, px)\n # for i in set(f.perod_end_date):", "def _get_sum_detail(self, cr, uid, ids=None, number_month=None,\n context=None):\n fy_obj = self.pool.get('account.fiscalyear')\n period_obj = self.pool.get('account.period')\n context = context and dict(context) or {}\n cx = context.copy()\n res = 0.0\n\n if not cx.get('fiscalyear'):\n cx['fiscalyear'] = fy_obj.find(cr, uid)\n\n fy_id = cx['fiscalyear']\n\n brw = self.browse(cr, uid, ids)\n\n if brw.acc_val == 'init':\n if cx.get('whole_fy', False):\n cx['periods'] = period_obj.search(cr, uid, [\n ('fiscalyear_id', '=', fy_id), ('special', '=', True)])\n else:\n period_from = period_obj.search(cr, uid, [\n ('fiscalyear_id', '=', fy_id), ('special', '=', True)])\n # Case when the period_from is the first non-special period\n # of the fiscalyear\n if period_obj.browse(cr, uid, cx['period_from']).date_start ==\\\n fy_obj.browse(cr, uid, fy_id).date_start:\n cx['period_to'] = period_from[0]\n else:\n cx['period_to'] = period_obj.previous(\n cr, uid, cx['period_from'])\n cx['period_from'] = period_from[0]\n elif brw.acc_val == 'var':\n # it is going to be the one sent by the previous cx\n if cx.get('whole_fy', False):\n cx['periods'] = period_obj.search(cr, uid, [\n ('fiscalyear_id', '=', fy_id), ('special', '=', False)])\n else:\n # it is going to be from the fiscalyear's beginning\n if cx.get('whole_fy', False):\n cx['periods'] = period_obj.search(cr, uid, [\n ('fiscalyear_id', '=', fy_id)])\n else:\n period_from = period_obj.search(cr, uid, [\n ('fiscalyear_id', '=', fy_id), ('special', '=', True)])\n cx['period_from'] = period_from[0]\n cx['periods'] = \\\n period_obj.build_ctx_periods(cr, uid, cx['period_from'],\n cx['period_to'])\n\n if brw.type == 'detail':\n # Si es de tipo detail\n # If we have to only take into account a set of Journals\n cx['journal_ids'] = [aj_brw.id for aj_brw in brw.journal_ids]\n cx['analytic'] = [an.id for an in brw.analytic_ids]\n cx['ifrs_tax'] = [tx.id for tx in brw.tax_code_ids]\n cx['ifrs_partner'] = [p_brw.id for p_brw in brw.partner_ids]\n cx['ifrs_query'] = self._get_ifrs_query(cr, uid, brw, context)\n\n # NOTE: This feature is not yet been implemented\n # cx['partner_detail'] = cx.get('partner_detail')\n\n # Refreshing record with new context\n brw = self.browse(cr, uid, ids, context=cx)\n\n for aa in brw.cons_ids:\n # Se hace la sumatoria de la columna balance, credito o debito.\n # Dependiendo de lo que se escoja en el wizard\n if brw.value == 'debit':\n res += aa.debit\n elif brw.value == 'credit':\n res += aa.credit\n else:\n res += aa.balance\n return res", "def current_effective_deadline(cls) -> float:", "def get_rate(currency, date):\n status = 400\n while status != 200:\n url = (\"http://api.nbp.pl/api/exchangerates/rates/A/%s/%d-%02d-%02d?format=json\" %\n (currency, date.year, date.month, date.day))\n\n response = requests.get(url)\n status = response.status_code\n if status != 200:\n date = date - datetime.timedelta(1)\n\n tree = json.loads(response.content)\n assert len(tree['rates']) == 1\n print_rate_info(tree['rates'])\n return (tree['rates'][0]['mid'], date)", "def get_price_history_dates(access_token,ticker,start_date,end_date,frequencyType,frequency):\r\n \r\n price_url = 'https://api.tdameritrade.com/v1/marketdata/{}/pricehistory'.format(ticker)\r\n\r\n #The header for getting a quote needs to define the input type (json)\r\n headers = {'Authorization':'Bearer {}'.format(access_token),\r\n 'Content-Type':'application/json'}\r\n\r\n #Parameters for period of time and frequency of data to get\r\n params = {'startDate':start_date,\r\n 'endDate': end_date,\r\n 'frequencyType': frequencyType,\r\n 'frequency': frequency}\r\n \r\n #Make the get request to TD Ameritrade\r\n price_history_json = requests.get(url=price_url,headers=headers,params=params)\r\n return price_history_json.json()", "def rate_between(self, from_date, to_date):\n print(\"override the above\")", "def get_current_rate(self):\n pass", "def date(self):", "def get_exchange_rate_data(self, source_currency, exchanged_currency, valuation_date):\n raise NotImplementedError", "def expense_history(request):\n qs: QuerySet = Expense.objects.by_user(request.user.id)\n file_title: str = \"Latest_150_Expenses\"\n form = ExpenseHistory(request.GET)\n if form.is_valid():\n cd: dict = form.cleaned_data\n target: str = cd[\"target\"]\n user_id = request.user.id\n if target == \"date\":\n qs = Expense.objects.filter(date=cd[\"date1\"], user_id=user_id)\n file_title = f'For_{cd[\"date1\"]}'\n elif target == \"each_month\":\n qs = Expense.objects.filter(date__month=cd[\"month\"], user_id=user_id)\n file_title = f\"Every_{calendar.month_name[cd['month']]}_Month\"\n elif target == \"months\":\n qs = Expense.objects.last_n_months_expense(cd[\"p_months\"], user_id)\n file_title = f\"Last_{cd['p_months']}_months\"\n elif target == \"month\":\n qs = Expense.objects.month_expense(cd[\"month\"], cd[\"year\"], user_id)\n file_title = f'For_{calendar.month_name[cd[\"month\"]]}-{cd[\"year\"]}'\n elif target == \"year\":\n qs = Expense.objects.year_expense(cd[\"year\"], user_id)\n file_title = f\"{cd['year']}\"\n elif target == \"between\":\n qs = Expense.objects.filter(date__gte=cd[\"date1\"], date__lte=cd[\"date2\"],\n user__id=user_id)\n file_title = f'Between_{cd[\"date1\"]}_{cd[\"date2\"]}'\n qs = qs.order_by(\"-date\", \"-id\").values_list(\n \"date\", \"description\", \"category__name\", \"method\", \"app\", \"amount\",\n )\n if not form.is_valid():\n qs = qs[:150]\n qs_list = []\n if qs:\n for q in qs:\n qs_list.append([\n q[0], q[1], q[2], METHOD_DICT[q[3]], APP_DICT.get(q[4], \"Other\"), q[5]\n ])\n file_title = f\"{date.today()}_\" + file_title\n return render(request, \"tracker/history.html\",\n {\"qs\": qs_list, \"file_title\": file_title, \"form\": form})", "def rate(self):\n return self.brate / FAC", "def rates(self):\n raise NotImplementedError(\"Must be implemented by subclass.\")", "def get_price_history_lookback(access_token,ticker,periodType,period,frequencyType,frequency):\r\n \r\n price_url = 'https://api.tdameritrade.com/v1/marketdata/{}/pricehistory'.format(ticker)\r\n\r\n #The header for getting a quote needs to define the input type (json)\r\n headers = {'Authorization':'Bearer {}'.format(access_token),\r\n 'Content-Type':'application/json'}\r\n\r\n #Parameters for period of time and frequency of data to get\r\n params = {'periodType':periodType,\r\n 'period': period,\r\n 'frequencyType': frequencyType,\r\n 'frequency': frequency}\r\n \r\n #Make the get request to TD Ameritrade\r\n price_history_json = requests.get(url=price_url,headers=headers,params=params)\r\n return price_history_json.json()", "def tbl_restrns_date(self,*expect_restrns):\n\n for count,restrn in enumerate(self.final_dataframe.keys()):\n \n if 'No Key Found' not in expect_restrns:\n \n if count in (2,3,4,5) and restrn in expect_restrns:\n \n \"\"\"\n 2 == FTNT, 3 == FARERULE, 4 == ALTRULE, 5 == GENRULE\n \"\"\"\n \n if len(self.final_dataframe[restrn]) != 0 and (self.final_dataframe[restrn].NOAPPL.isnull().any()):\n \"\"\"\n if table restrictions are not empty and NOAPLL is null for e.g. FR has records with NOAPP is null,\n then call to restrn_date function to caputre restriction dates\n \"\"\"\n \n self.restrn_date(restrn)\n \n elif count == 6:\n pass\n \n elif 'No Key Found' in expect_restrns:\n \n if count in (2,3,4,5) and restrn in expect_restrns and len(self.final_dataframe[restrn]) != 0 and (self.final_dataframe[restrn].NOAPPL.notnull().any()):\n \n self.restrn_date(restrn)\n \n elif count == 6:\n pass", "def get_furniture():", "def getActiveCurrency():", "def new_get_historical_price(base, target, date):\n return {base: {target: 10}}", "def select_vendor_price_list_detail_reference_rates_tab(self):\n self.click_element(self.vendor_price_list_details_reference_rates_tab_locator, True)", "def onchange_date(self, cr, user, ids, date, context=None):\n res = {}\n if context is None:\n context = {}\n period_pool = self.pool.get('account.period')\n ctx = dict(context, account_period_prefer_normal=True)\n pids = period_pool.find(cr, user, date, context=ctx)\n if pids:\n res.update({\n 'period_id':pids[0]\n })\n context.update({\n 'period_id':pids[0]\n })\n return {\n 'value':res,\n 'context':context,\n }", "def get_data(end_date, n, local, foreign):\n URL = \"https://api.exchangeratesapi.io/history\"\n PARAMS = {'start_at': str(get_weekday_n_days_ago(end_date, n)),\n 'end_at': str(end_date),\n 'symbols': foreign,\n 'base': local}\n r = requests.get(url=URL, params=PARAMS)\n data = r.json()\n input_data = []\n for day in data['rates']:\n input_data.append([datetime.strptime(day, '%Y-%m-%d').date(),\n float(\"{:.8f}\".format(data['rates'][day][foreign]))])\n input_data.sort(key=lambda x: x[0])\n return input_data[-n:]", "def getData(self):\n\n url = 'https://www.ecb.europa.eu/stats/eurofxref/eurofxref-hist.zip'\n try:\n file, _ = urlretrieve(url)\n zip_file_object = zipfile.ZipFile(file, 'r')\n first_file = zip_file_object.namelist()[0]\n file = zip_file_object.open(first_file)\n\n file_handler = []\n for row in file:\n file_handler.append(row.decode())\n\n # getting the currency headers into header_list\n header_list = []\n notFound = True\n x = 0\n while notFound:\n if file_handler[x].startswith('Date'):\n header = file_handler[x].split(',')\n for col in header:\n header_list.append(col.strip())\n notFound = False\n x += 1\n self.currencies = list(filter(None, header_list))\n self.currencies.append('EUR')\n self.currencies = self.currencies[1:] # Removing the \"Date\" entry\n\n data = []\n for row in file_handler[x:]:\n if row.startswith('`\\n'):\n break\n else:\n data.append(list(filter(None, [x.replace('\\n', '') for x in row.split(',')]))) # Removing any empty extra columns at the end of each rows\n\n # filling my self.rates with the currency in the format {CURR: {date: rate, ...}, ...}\n for row in data:\n for i in range(len(self.currencies)):\n try:\n if self.currencies[i] not in self.rates:\n self.rates[self.currencies[i]] = {row[0]: row[i + 1]}\n else:\n self.rates[self.currencies[i]].update({row[0]: row[i + 1]})\n except IndexError:\n # We reached the EUR section\n if self.currencies[i] not in self.rates:\n self.rates[self.currencies[i]] = {row[0]: '1.0000'}\n else:\n self.rates[self.currencies[i]].update({row[0]: '1.0000'})\n\n self.currencies.sort()\n\n except Exception as e:\n print('Failed to process the data')\n print(e)\n finally:\n file.close()", "def historical(self, date, base='USD'):\n try:\n resp = self.client.get(self.ENDPOINT_HISTORICAL %\n date.strftime(\"%Y-%m-%d\"),\n params={'base': base})\n resp.raise_for_status()\n except requests.exceptions.RequestException as e:\n raise OpenExchangeRatesClientException(e)\n return resp.json(parse_int=decimal.Decimal,\n parse_float=decimal.Decimal)", "def test_get_historical_gold_rate(self):\n rates = [153.50, 162.49, 123.86, 155.10]\n helper.gold_loop_helper(get_historical_gold_rate, TestHistoricalRates.dates_rate, rates)", "def calculate_demo(self, request, parent_lookup_client, pk, format=None):\n\n retirement_plan = self.get_object()\n tickers = Ticker.objects.filter(~Q(state=Ticker.State.CLOSED.value))\n portfolio = []\n projection = []\n for idx, ticker in enumerate(tickers[:10]):\n percent = 0\n if idx <= 9:\n # 10% each for first 10 tickers\n percent = 10\n portfolio.append([ticker.id, percent])\n # grab 50 evenly spaced time points between dob and current time\n today = timezone.now().date()\n last_day = retirement_plan.client.date_of_birth + relativedelta(years=retirement_plan.selected_life_expectancy)\n day_interval = (last_day - today) / 49\n income_start = 20000\n assets_start = 100000\n for i in range(50):\n income = income_start + (i * 50)\n assets = assets_start + (i * 1000)\n dt = today + i * day_interval\n projection.append([d2ed(dt), assets, income])\n return Response({'portfolio': portfolio, 'projection': projection})", "def compute_costs(timesheet, biller, date1=None, date2=None): \n # Slice\n f = slice_by_dates(timesheet, date1, date2)\n\n # Resample and add start/end dates\n if biller.freq is not None:\n freq = biller.freq\n f = timesheet.set_index('date')[['duration']].resample(freq).sum()\n f = f.reset_index()\n f['period'] = f['date'].map(lambda x: pd.Period(x, freq))\n f['start_date'] = f['period'].map(lambda x: x.start_time)\n f['end_date'] = f['period'].map(lambda x: x.end_time)\n else:\n start_date, end_date = f['date'].min(), f['date'].max()\n f['start_date'] = start_date\n f['end_date'] = end_date\n\n # Get bins for aggregating\n if biller.base_fee:\n bins = [0] + biller.bins\n else:\n bins = biller.bins\n\n def my_agg(group):\n d = OrderedDict()\n d['start_date'] = group['start_date'].iat[0]\n d['end_date'] = group['end_date'].iat[0]\n t = group['duration'].iat[0]\n d['duration'] = pd.Series(decompose(t, bins))\n c1 = d['duration'].cumsum().map(biller)\n c2 = c1.shift(1).fillna(0)\n cost = c1 - c2\n d['rate'] = cost/d['duration']\n d['cost'] = cost\n return pd.DataFrame(d)\n \n f = f.groupby('date').apply(my_agg\n ).reset_index().drop(['level_1', 'date'], axis=1)\n\n # Drop NaN rate items\n f = f.dropna(subset=['rate'])\n\n return f" ]
[ "0.61450714", "0.6118674", "0.59151864", "0.570882", "0.54793197", "0.5440474", "0.5394293", "0.53898215", "0.53862345", "0.5362355", "0.53242594", "0.5290705", "0.5275283", "0.5234759", "0.5224288", "0.52232665", "0.5206688", "0.52063805", "0.51902455", "0.5173077", "0.51618296", "0.5157474", "0.51360965", "0.5134611", "0.5131753", "0.5129398", "0.51117337", "0.50873107", "0.5086376", "0.50781316" ]
0.8118622
0
Implementing is target buy policies tab displayed functionality
def is_target_buy_policies_tab_displayed(self): return self.is_element_visible(self.target_buy_policies_tab_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def click_buy_and_sell_deal_management_link(self):\n self.select_static_tab(self.buy_and_sell_deal_management_locator, message=\"buy and sell deal management locator not found before specified time out\")", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def select_buy_dashboard_tab(self):\n self.select_static_tab(self.buy_dashboard_tab_locator, True)", "def is_buy_dashboard_tab_present(self):\n return self.is_element_present(self.buy_dashboard_tab_locator)", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def select_qos_tab(self):\n self.click_element(self.qos_tab_locator, True)", "def tabSelected(self):", "def tabSelected(self):", "def is_target_buy_details_section_present_in_vendor_profile_page(self):\n return self.is_element_present(self.target_buy_details_section_locator)", "def tabSelected(self):\r\n self.transactionMenuWidget.tabSelected()", "def is_target_buy_list_overrides_screen_loaded(self):\n return self.is_element_visible(self.target_buy_list_overrides_page_header_locator)", "def select_adjustments_tab(self):\n self.click_element(self.adjustments_tab_locator)", "def rf_libTab(self):\n if self.rbScript.isChecked():\n self.flScript.setVisible(True)\n self.flSaveSpacer.setVisible(False)\n else:\n self.flScript.setVisible(False)\n self.flSaveSpacer.setVisible(True)", "def _test_display_up_button(self):\n return (self.product_displays.top_index > 0)", "def click_save_target_buy_policy_button(self):\n self.click_element(self.save_target_buy_policy_button_locator, error_message='save target buy policy button locator not found before specified time out')\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def buy_or_sell(self, debug=True):\n if self.will_plot:\n self.plot()\n self.predict()\n self.check_sma()\n self.check_bollinger()\n self.volume()\n self.rsi()\n self.sharpe()\n self.extrema()\n self.net_gains()\n self.beta_and_alpha()\n self.decision()\n if debug: print(self.debug)\n if self.will_plot:\n plt.show()", "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def select_vendor_price_list_detail_reference_rates_tab(self):\n self.click_element(self.vendor_price_list_details_reference_rates_tab_locator, True)", "def draw_menu(self, context):\n if context.engine == 'RPR':\n layout = self.layout\n layout.popover('RPR_VIEW3D_PT_panel')", "def show(self):\n #print(\" ===== I am in show function ----\")\n if self.brightness > 0.99:\n global myItemTab\n for i in range(len(cfg.myItemTabHandler)):\n pen = QPen(QColor(self.stripTab[i]))\n brush = QBrush(pen.color())\n #brush = QBrush(pen.color().darker(100))\n cfg.myItemTabHandler[i].setPen(pen)\n cfg.myItemTabHandler[i].setBrush(brush)\n\n\n else:\n pass", "def policy(agent):", "def select_tab_of_view_price_list_detail_page(self, tab_name):\n self.wait_for_ajax_spinner_load(300)\n view_price_list_detail_page_tab_locator = (By.XPATH, self.vendor_profile_page_tab_locator_string + \"[text()='%s']\" % tab_name)\n self.select_static_tab(view_price_list_detail_page_tab_locator, 'tab locator not found')", "def CashMode(self):\n self.cred_left = 0\n self.is_member = False\n self.cred_id = ''\n self.cred_card = ''\n self.builder.get_object('GuiMode').set_label(\"Payment in Cash\")", "def switch_to_buy(self):\n self.switch_to_window()\n self.accept_ssl_certificate()", "def trade_action(self, BUY_QTY):\n BUY_QTY = 4500\n self.trade(BUY_QTY)\n #self.show()", "def plans():", "def productactivate():\n pass" ]
[ "0.70824105", "0.6161778", "0.6106716", "0.5919944", "0.57483226", "0.57384634", "0.5682125", "0.5653956", "0.5572523", "0.55241346", "0.55241346", "0.5510905", "0.5445419", "0.54413915", "0.5437206", "0.5395717", "0.5387464", "0.53720695", "0.53677285", "0.5356148", "0.531586", "0.53106815", "0.5196825", "0.51738054", "0.51650816", "0.5135971", "0.51007175", "0.5095864", "0.5091751", "0.5086096" ]
0.750019
0
Implementing select target buy policies tab functionality
def select_target_buy_policies_tab(self): self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def select_buy_dashboard_tab(self):\n self.select_static_tab(self.buy_dashboard_tab_locator, True)", "def select_adjustments_tab(self):\n self.click_element(self.adjustments_tab_locator)", "def select_qos_tab(self):\n self.click_element(self.qos_tab_locator, True)", "def click_buy_and_sell_deal_management_link(self):\n self.select_static_tab(self.buy_and_sell_deal_management_locator, message=\"buy and sell deal management locator not found before specified time out\")", "async def on_select_trade(self, payload):\n\n self.source._to_trade = self.current_page\n await self.show_page(self.current_page)", "def select_transactions_tab(self):\n self.click_element(self.transactions_tab_locator)", "def is_target_buy_policies_tab_displayed(self):\n return self.is_element_visible(self.target_buy_policies_tab_locator)", "def click_save_target_buy_policy_button(self):\n self.click_element(self.save_target_buy_policy_button_locator, error_message='save target buy policy button locator not found before specified time out')\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def tabSelected(self):\r\n self.transactionMenuWidget.tabSelected()", "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def tabSelected(self):", "def tabSelected(self):", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def select_vendor_price_list_detail_reference_rates_tab(self):\n self.click_element(self.vendor_price_list_details_reference_rates_tab_locator, True)", "def click_cancel_edited_target_buy_policy_button(self):\n self.click_element(self.cancel_edited_target_buy_policy_button_locator)", "def click_create_new_target_buy_policy_button(self):\n self.click_element(self.create_new_target_buy_policy_button_locator, script_executor=True, error_message='create new target buy policy button locator not found before specified time out')", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def click_save_edited_target_buy_policy_button(self):\n self.click_element(self.save_edited_target_buy_policy_button_locator)\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def select_disputes_tab(self):\n self.click_element(self.disputes_tab_locator)", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def set_buy_sell_deal_account(self, account_list):\n self.multiple_items_selection_from_kendo_dropdown(self.buy_sell_deal_account_dropdown_locator, account_list)\n self.wait_for_ajax_spinner_load()", "def select_tab_of_view_price_list_detail_page(self, tab_name):\n self.wait_for_ajax_spinner_load(300)\n view_price_list_detail_page_tab_locator = (By.XPATH, self.vendor_profile_page_tab_locator_string + \"[text()='%s']\" % tab_name)\n self.select_static_tab(view_price_list_detail_page_tab_locator, 'tab locator not found')", "def select_action(self):\n pass", "def choose_action(self):\r\n pass", "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def state_choose_do(cfg, app, win, events):", "def select_rates_tab(self):\n self.select_static_tab(self.rates_tab_locator, True)", "def select_vendors_tab(self):\n self.select_static_tab(self.vendors_tab_locator, 'vendors tab not found before specified time')" ]
[ "0.6704803", "0.6659049", "0.62752223", "0.612902", "0.6105099", "0.60682184", "0.5983011", "0.59394455", "0.59090525", "0.58154", "0.579811", "0.5768202", "0.57265127", "0.57265127", "0.5684236", "0.56841564", "0.55653906", "0.55247366", "0.55149025", "0.5494688", "0.54541177", "0.54045665", "0.53248566", "0.53083706", "0.5299271", "0.5295442", "0.5288878", "0.52853066", "0.52830845", "0.5263585" ]
0.76746756
0
Implementing is target buy policies grid column present functionality
def is_target_buy_policies_grid_column_present(self, column_name): column_locator = (By.XPATH, "//div[contains(@id, '%s')]/descendant::th[@data-title='%s']" % (self.target_buy_policies_grid_div_id, column_name)) return self.is_element_present(column_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def is_target_buy_policies_tab_displayed(self):\n return self.is_element_visible(self.target_buy_policies_tab_locator)", "def populate_buy_trend(self, dataframe: DataFrame, metadata: dict) -> DataFrame:\n\n dataframe.loc[\n (\n (qtpylib.crossed_above(dataframe['ema'],dataframe['ema2']))\n ),'buy'] = 1\n\n return dataframe", "def front_column_model_p_gain():", "def get_display_price_data(self, source, commitment):\n overage, included = self.get_price_data(source, commitment)\n if self.name == settings.BILLING_DEFAULT_PLAN_NAME:\n included = OFFICIAL_BUILDER_LIMITS[source]\n return overage, included", "def populate_buy_trend(dataframe: DataFrame, metadata: dict) -> DataFrame:\n conditions = []\n\n conditions.append(\n ((dataframe['bull'] > 0) & qtpylib.crossed_below(dataframe['rsi'], params['bull-buy-rsi-value'])) |\n (~(dataframe['bull'] > 0) & qtpylib.crossed_below(dataframe['rsi'], params['bear-buy-rsi-value']))\n )\n\n conditions.append(dataframe['volume'] > 0)\n\n dataframe.loc[\n reduce(lambda x, y: x & y, conditions),\n 'buy'] = 1\n\n return dataframe", "def click_buy_and_sell_management_grid_export_to_excel_button(self):\n self.click_grid_export_to_excel_button(self.buy_and_sell_management_grid_div_id)", "def verify_rates_grid_columns(self):\n is_verified = True\n if(self.is_element_present(self.rates_grid_destination_column_locator) == False):\n is_verified = False\n if (is_verified != False):\n for column_name in self.compare_price_lists_list:\n rates_grid_new_column_locator = (By.XPATH, \"//div[@id='divComaprePriceListRate']/descendant::span[contains(text(),'%s')]\" % column_name)\n if (self.is_element_present(rates_grid_new_column_locator) == False):\n is_verified = False\n break\n return is_verified", "def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)", "def drag_buy_and_sell_management_grid_column(self, source_column, destination_column):\n self.drag_grid_column(self.buy_and_sell_management_grid_div_id, source_column, destination_column)", "def select_buy_and_sell_deal_management_grid_first_row_checkbox(self):\n self.select_grid_row_checkbox(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_checkbox_column_number, True)", "def click_target_buy_policies_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, self.target_buy_policies_grid_inline_action_column_number, first_row=True)", "def gridDisplay(self):\n\n if self.griddButton.isCheckable():\n self.photo_grid.setVisible(False)\n self.griddButton.setCheckable(False)\n self.griddButton.setDown(False)\n self.statustext.setText(\"Hide Grid\")\n else:\n self.griddButton.setCheckable(True)\n self.photo_grid.setVisible(True)\n self.griddButton.setDown(True)\n self.statustext.setText(\"Display Grid - Rule of thirds\")", "def change_entries(self, grid_sizer, options, border):\n\n if (self.diag_type is None): return grid_sizer # no data selected, don't change anything\n\n # setup the grid of possible values\n header0 = wx.StaticText(self, -1, \"Add/Remove\")\n header1 = wx.StaticText(self, -1, \"Quantity Code\")\n header2 = wx.StaticText(self, -1, \"Name\")\n header3 = wx.StaticText(self, -1, \"LaTeX Formula\")\n grid_sizer.Add(header0, pos=(0,0), flag=options, border=border)\n grid_sizer.Add(header1, pos=(0,1), flag=options, border=border)\n grid_sizer.Add(header2, pos=(0,2), flag=options, border=border)\n grid_sizer.Add(header3, pos=(0,3), flag=options, border=border)\n grid_sizer.Add(wx.StaticLine(self), pos=(1,0), span=(1,4),\n flag=wx.ALL|wx.EXPAND|wx.GROW, border=border)\n\n self.selected_values = [] # keep track of selected quantities\n\n quantities = self.output_quantities.diagnostic_types[self.diag_type]\n\n # choose a good height/width for formulas\n if (self.diag_type in [\"Linear_Forces\", \"Angular_Momentum\", \"Energy_Flux\",\n \"Induction\", \"Inertial_Forces\", \"Lorentz_Forces\",\n \"Poynting_Flux\", \"TurbKE_Budget\"]):\n width = 100\n elif (self.diag_type == \"Thermal_Equation\"):\n width = 150\n else:\n width = 30\n height = 20\n\n row = 2\n iquant = 0\n for Q in quantities:\n but = wx.ToggleButton(self, Q.code, \"Add\") # build button and place it in second column\n but.Bind(wx.EVT_TOGGLEBUTTON, self.OnToggle)\n grid_sizer.Add(but, pos=(row,0), flag=options, border=border)\n\n q_code = wx.StaticText(self, -1, str(Q.code)) # build other column entries\n q_name = wx.StaticText(self, -1, Q.name) # name\n\n formula = self.RenderTeX(Q, size=(width,height))\n\n # place column entries\n grid_sizer.Add(q_code, pos=(row,1), flag=options, border=border)\n grid_sizer.Add(q_name, pos=(row,2), flag=options, border=border)\n grid_sizer.Add(formula, pos=(row,3), flag=options, border=border)\n\n iquant += 1\n\n # add horizontal line every 5 quantities\n if (iquant % 5 == 0):\n grid_sizer.Add(wx.StaticLine(self), pos=(row+1,0), span=(1,4),\n flag=wx.ALL|wx.EXPAND|wx.GROW, border=border)\n row_inc = 2\n else:\n row_inc = 1\n\n row += row_inc\n\n grid_sizer.AddGrowableCol(2,1) # make the name/formula columns \"1\" growable, i.e., grows as necessary\n grid_sizer.AddGrowableCol(3,1)\n\n return grid_sizer", "def get_list_columns(self):\n columns = super(InventoryTransactionAdmin, self).get_list_columns()\n cols = ['total_amount']\n columns = security_util.filter_columns_by_role(\n columns, cols, 'purchase_price_view'\n )\n return columns", "def compare_vendor_price_list_detail_rates_grid_rate(self):\n self.buy_page_excel_data_dictionary = self.get_excel_data_dictionary()\n is_compared = False\n rates_grid_rate_column_element = self.wait().until(EC.presence_of_element_located(self.rates_grid_rate_column_locator))\n if self.buy_page_excel_data_dictionary[\"Rate1\"] == (rates_grid_rate_column_element.text).strip():\n is_compared = True\n return is_compared", "def click_target_buy_policy_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, column_number=2, first_row=True)\n self.wait_for_ajax_spinner_load()", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def is_reference_rates_grid_accept_column_checkbox(self):\n is_checkbox = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (self.vendor_price_list_detail_reference_rates_grid_div_id, self.accept_column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n column_type_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tr/td[%s]/input\" % (self.vendor_price_list_detail_reference_rates_grid_div_id, str(column_index)))\n column_type_element = self.wait().until(EC.presence_of_element_located(column_type_locator), 'column type locator not found before specified time out')\n column_type = column_type_element.get_attribute('type')\n if column_type == \"checkbox\":\n is_checkbox = True\n return is_checkbox", "def set_buy_and_sell_management_grid_settings(self, grid_settings):\n self.set_grid_settings(self.buy_and_sell_management_grid_div_id, grid_settings)", "def testPolicy( \t\t \t \t\t \t \t\t\t \t\t \t\t\t \t \t \t\t \t\t \t\n self, \t\t \t \t\t \t \t\t\t \t\t \t\t\t \t \t \t\t \t\t \t\n symbol=\"jpm\", \t\t \t \t\t \t \t\t\t \t\t \t\t\t \t \t \t\t \t\t \t\n sd=dt.datetime(2008, 1, 1), \t\t \t \t\t \t \t\t\t \t\t \t\t\t \t \t \t\t \t\t \t\n ed=dt.datetime(2009, 12, 31), \t\t \t \t\t \t \t\t\t \t\t \t\t\t \t \t \t\t \t\t \t\n sv=10000, \t\t \t \t\t \t \t\t\t \t\t \t\t\t \t \t \t\t \t\t \t\n ): \t\t \t \t\t \t \t\t\t \t\t \t\t\t \t \t \t\t \t\t \t\n \t\t \t \t\t \t \t\t\t \t\t \t\t\t \t \t \t\t \t\t \t\n dates = pd.date_range(sd,ed)\n df_prices = ind.get_price(symbol, dates)\n\n daily_rets = (df_prices / df_prices.shift(1)) - 1\n daily_rets = daily_rets[1:]\n\n\n sd_older = sd - dt.timedelta(days=365)\n dates_older = pd.date_range(sd_older,ed)\n df_prices_older = ind.get_price(symbol, dates_older)\n sd_key = df_prices.index[0]\n sd_index = df_prices_older.index.get_loc(sd_key)\n\n\n df_holdings = df_prices.copy()\n df_holdings['Holdings'] = np.nan\n del df_holdings[symbol]\n # print(df_holdings)\n\n cum_ret_prev = 0\n iters = 0\n\n\n num_bins = len(self.bins)\n\n _,_,ind1 = ind.get_BB(df_prices_older, self.lookback)\n ind2 = ind.get_CCI(df_prices_older, self.lookback)\n _,_,ind3 = ind.get_SMA_Cross(self.lookback, 100, df_prices_older)\n ind4 = ind.get_momentum(df_prices_older, self.lookback)\n _,_,ind5 = ind.get_MACD(df_prices_older)\n BB = ind1.iloc[sd_index:].values\n CCI = ind2.iloc[sd_index:].values\n SMA_Cross = ind3.iloc[sd_index:].values\n Momentum = ind4.iloc[sd_index:].values\n MACD = ind5.iloc[sd_index:].values\n _,self.x0bins = pd.qcut(BB[:,0], num_bins,labels=False,retbins=True)\n _,self.x1bins = pd.qcut(CCI[:,0],num_bins,labels=False,retbins=True)\n _,self.x2bins = pd.qcut(SMA_Cross[:,0],num_bins,labels=False,retbins=True)\n _,self.x3bins = pd.qcut(Momentum[:,0],num_bins,labels=False,retbins=True)\n _,self.x4bins = pd.qcut(MACD[:,0],num_bins,labels=False,retbins=True)\n x_0 = np.digitize(BB[:,0], self.x0bins[1:-1])\n x_1 = np.digitize(CCI[:,0], self.x1bins[1:-1])\n x_2 = np.digitize(SMA_Cross[:,0], self.x2bins[1:-1])\n x_3 = np.digitize(Momentum[:,0], self.x3bins[1:-1])\n x_4 = np.digitize(MACD[:,0], self.x4bins[1:-1])\n state = x_0 + x_3*10 + x_4*100\n\n\n\n\n\n self.learner.rar = 0\n\n action = self.learner.querysetstate(state[0])\n\n daily_return = daily_rets.iloc[0][symbol]\n df_holdings.iloc[0]['Holdings'] = 0\n\n\n for day_idx in range(1,daily_rets.shape[0]):\n\n # implement action\n cur_price = df_prices.iloc[day_idx-1][symbol]\n next_price = df_prices.iloc[day_idx][symbol]\n action = self.learner.querysetstate(state[day_idx])\n df_holdings.iloc[day_idx]['Holdings'],_ = self.take_action(df_holdings.iloc[day_idx-1]['Holdings'], action, cur_price, next_price)\n\n\n df_holdings.iloc[-1]['Holdings'] = 0\n df_trades = df_holdings.diff()\n df_trades['Trades'] = df_trades['Holdings']\n del df_trades['Holdings']\n df_trades.iloc[0]['Trades'] = 0\n return df_trades", "def set_order_conditions(self, df_lt_spm: pd.DataFrame, df_lt_repair: pd.DataFrame, procurement_mode: int = 1):\n\n # Set procurement types based on any match found\n self.is_spm = any([material.is_spm for material in self.materials])\n self.is_repairable = any([material.is_repairable for material in self.materials])\n self.is_buyable = any([material.is_buyable for material in self.materials])\n self.is_dismountable = any([material.is_dismountable for material in self.materials])\n\n # If no procurement type set as buyable\n self.has_procurement_type = self.is_dismountable or self.is_buyable or self.is_spm or self.is_repairable\n if not self.has_procurement_type:\n self.is_buyable = True\n\n # --------------\n\n # Set unique values (and override if needed)\n\n # If CORE VOZ, set dismountable instead of repairable\n if procurement_mode == 0:\n self.procurement_type = 'Buyable'\n self.leadtime = max([material.leadtime for material in self.materials if material.is_buyable])\n self.leadtime_sd = max([material.leadtime_sd for material in self.materials if material.is_buyable])\n return\n elif procurement_mode == 1:\n if self.domain == 'CORE VOZ' and not self.is_spm and not self.is_buyable\\\n and self.is_dismountable and self.is_repairable:\n self.procurement_type = 'Dismountable'\n self.leadtime = 90\n self.leadtime_sd = 0\n return\n\n if self.is_spm:\n self.procurement_type = 'SPM'\n\n # Override\n if (self.domain, self.brand) in df_lt_spm.index:\n try:\n new_leadtime = df_lt_spm.loc[(self.domain, self.brand)]['leadtime_spm']\n new_leadtime = float(new_leadtime)\n\n self.leadtime_override = True\n self.leadtime = new_leadtime\n self.leadtime_sd = 0\n return\n except:\n self.leadtime = 2\n self.leadtime_sd = 0\n return\n else:\n self.leadtime = 2\n self.leadtime_sd = 0\n return\n\n if self.is_repairable:\n self.procurement_type = 'Repairable'\n\n # Override\n if self.domain in df_lt_repair.index:\n try:\n new_leadtime = df_lt_repair.loc[self.domain]['leadtime_reparable']\n new_leadtime = float(new_leadtime)\n\n self.leadtime_override = True\n self.leadtime = new_leadtime\n self.leadtime_sd = 0\n return\n except:\n self.leadtime = 30\n self.leadtime_sd = 0\n return\n else:\n self.leadtime = 30\n self.leadtime_sd = 0\n return\n\n if self.is_buyable:\n self.procurement_type = 'Buyable'\n try:\n self.leadtime = max([material.leadtime for material in self.materials if material.is_buyable])\n self.leadtime_sd = max([material.leadtime_sd for material in self.materials if material.is_buyable])\n return\n except:\n self.leadtime = 90\n self.leadtime_sd = 0\n\n # Else, return Dismountable\n self.procurement_type = 'Dismountable'\n self.leadtime = 90\n self.leadtime_sd = 0\n return\n else:\n raise Exception(\"procurement_mode not valid.\")", "def select_target_buy_list_overrides_grid_row_checkbox(self):\n for i in range(2):\n target_buy_list_overrides_grid_checkbox_locator = (By.XPATH, \"(//div[@data-grid-name='%s']/descendant::input[@type='checkbox'])[%s]\" % (self.target_buy_list_overrides_data_grid_name, str(i + 2)))\n self.click_element(target_buy_list_overrides_grid_checkbox_locator, True)", "def buy_and_hold(dataframe: pd.DataFrame) -> pd.DataFrame:\n dataframe[PandasEnum.ALLOCATION.value] = 1.0\n return dataframe", "def get_active_features(summary_df, slots_offered): # prev -> getActiveFeatures\n disc_cols = [col+'_Discount' for col in slots_offered]\n eco_cols = [col+'_Eco' for col in slots_offered]\n gr_cols = [col+'_Eco' for col in slots_offered]\n features = summary_df.loc[:, disc_cols+eco_cols+gr_cols]\n features = features.loc[:, features.sum(axis=0) > 0]\n for i in reversed(['NO_PURCHASE']+slots_offered):\n features.insert(0, i+'_Asc', value=1)\n return features, disc_cols, eco_cols, gr_cols", "def __addColumnConstraints(self):\n for x in range(self.width):\n plusTarget = self.columnPlusCounts[x]\n minusTarget = self.columnMinusCounts[x]\n plusTotal = 0\n minusTotal = 0\n for y in range(self.height):\n g = self.grid[(x, y)]\n plusTotal = plusTotal + If(g == Magnets.PLUS, 1, 0)\n minusTotal = minusTotal + If(g == Magnets.MINUS, 1, 0)\n if plusTarget != None:\n self.solver.add(plusTotal == plusTarget)\n if minusTarget != None:\n self.solver.add(minusTotal == minusTarget)", "def prepare_for_metgrid(self, wps_nml):\n wps_nml['share']['active_grid'] = [True] * len(self.domains)", "def _test_display_up_button(self):\n return (self.product_displays.top_index > 0)", "def sort_buy_and_sell_management_grid_column(self, column_name, descending_order):\n self.sort_grid_column(self.buy_and_sell_management_grid_div_id, column_name, descending_order)", "def get_all_rows_of_rate_increase_violation_column_from_vendor_price_list_detail_grid(self, column_name):\n self.all_row_data = []\n self.wait_for_ajax_spinner_load()\n row_count = len(self.wait().until(EC.presence_of_all_elements_located(self.vendor_price_list_detail_row_locator), 'vendor price list details row element locator not found before specified time'))\n next_page = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (self.vendor_price_list_detail_rates_grid_div_id, column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n row = 1\n try:\n while row < row_count + 1:\n column_value_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tbody[@role='rowgroup']/tr[%s]/td[%s]\" % (self.vendor_price_list_detail_rates_grid_div_id, row, str(column_index)))\n column_value_element = self.wait(10).until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n column_value = str(column_value_element.get_attribute(\"innerHTML\"))\n if column_value not in self.all_row_data:\n self.all_row_data.append(column_value)\n else:\n row_count += 1\n if row % 50 == 0:\n next_page = True\n next_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-arrow-e']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(next_page_button_locator, True)\n row = 0\n row_count -= 50\n row += 1\n except WebDriverException:\n pass\n if next_page is True:\n first_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-seek-w']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(first_page_button_locator, True)\n return self.all_row_data" ]
[ "0.65470636", "0.5555432", "0.5375789", "0.53040683", "0.5265732", "0.5251825", "0.52483046", "0.5247218", "0.52333355", "0.5223402", "0.5211076", "0.51632774", "0.5090065", "0.50898117", "0.5086829", "0.5062219", "0.5055388", "0.50354195", "0.50350845", "0.50128216", "0.5012641", "0.50038457", "0.49882057", "0.49787086", "0.49786305", "0.49580047", "0.49406242", "0.4925437", "0.49201298", "0.49055472" ]
0.6384165
1
Implementing click target buy policies grid first row inline action button functionality
def click_target_buy_policies_grid_first_row_inline_action_button(self): self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, self.target_buy_policies_grid_inline_action_column_number, first_row=True)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_target_buy_policy_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, column_number=2, first_row=True)\n self.wait_for_ajax_spinner_load()", "def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)", "def click_vendor_price_list_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.vendor_price_list_grid_div_id, None, self.view_price_list_column_number, True)", "def click_adjustment_charges_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.adjustment_charges_grid_div_id, None, self.adjustment_charges_grid_inline_action_column_number, True)", "def click_vendors_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.vendors_div_id, None, self.grid_column_number, True)", "def click_outbound_statement_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, None, self.outbound_statements_grid_inline_action_column_number, True)", "def click_adjustments_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.adjustments_grid_div_id, None, self.adjustments_grid_inline_action_column_number, True)", "def click_outbound_statements_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, None, self.outbound_statements_grid_inline_action_column_number, True)", "def click_inbound_statement_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.inbound_statement_grid_div_id, None, self.inbound_statement_grid_inline_action_column_number, True)", "def click_disputes_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.statement_details_dispute_grid_div_id, None, self.disputes_grid_inline_action_column_number, True)", "def click_line_item_disputes_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.line_item_disputes_grid_div_id, None, self.line_item_disputes_grid_inline_action_column_number, True)", "def click_re_analysis_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.re_analysis_grid_div_id, None, self.re_analysis_grid_inline_action_column_number, True)", "def click_view_price_list_detail_first_row_inline_action_button(self):\n self.click_inline_action_button(self.view_price_list_div_id, None, self.view_price_list_column_number, True)\n self.wait_for_ajax_spinner_load()", "def click_statement_validation_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.statement_validation_grid_div_id, None, self.statement_validation_grid_inline_action_column_number, True)", "def click_resolution_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.resolution_grid_div_id, None, self.resolution_grid_inline_column_number, True)", "def click_buy_page_inline_action_button(self, vendor):\n self.click_inline_action_button(self.vendors_div_id, vendor, self.grid_column_number)", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def click_inbound_statement_grid_inline_action_button(self, row_identifier_text):\n self.click_inline_action_button(self.inbound_statement_grid_div_id, row_identifier_text, self.inbound_statement_grid_inline_action_column_number)", "def click_outbound_statements_grid_inline_action_button(self, row_identifier_text):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, row_identifier_text, self.outbound_statements_grid_inline_action_column_number)", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def select_buy_and_sell_deal_management_grid_first_row_checkbox(self):\n self.select_grid_row_checkbox(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_checkbox_column_number, True)", "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def click(self):\r\n pass", "def click_view_price_list_detail_page_inline_action_button(self, price_list_item):\n self.click_inline_action_button(self.view_price_list_div_id, price_list_item, self.view_price_list_column_number)\n self.wait_for_ajax_spinner_load()", "def set_shortfall_grid_actions(self, action_item):\n self.single_selection_from_kendo_in_grid(self.shortfall_grid_div_id, self.actions_column_name, action_item)", "def _ClickPrimaryActionButton(self):\n self._ExecuteOobeApi('Oobe.clickGaiaPrimaryButtonForTesting')", "def select_adjustment_charge_grid_first_row_checkbox(self):\n self.click_element(self.adjustment_charges_grid_first_row_checkbox_locator, True, True)", "def OnLeftUp_ClickButton(self, event):\r\n \r\n self._hover_button = None\r\n\r\n if self._action_part:\r\n self.RefreshButton(self._action_part)\r\n\r\n # make sure we're still over the item that was originally clicked\r\n if self._action_part == self.HitTest(*event.GetPosition()):\r\n \r\n # fire button-click event\r\n e = AuiManagerEvent(wxEVT_AUI_PANE_BUTTON)\r\n e.SetManager(self)\r\n e.SetPane(self._action_part.pane)\r\n e.SetButton(self._action_part.button.button_id)\r\n self.ProcessMgrEvent(e)", "def click_statement_fields_grid_first_row_checkbox(self):\n self.click_element(self.statement_fields_grid_first_row_checkbox_locator, True)", "def select_bill_and_audit_module_grid_inline_item(self, item_name):\n self.select_inline_action_item(item_name, module_name = self.bill_audit_module_name)\n self.wait_for_ajax_spinner_load()" ]
[ "0.8105151", "0.794805", "0.73137873", "0.7264641", "0.7102072", "0.7087844", "0.70674783", "0.7059369", "0.7035884", "0.70264035", "0.68555933", "0.6739181", "0.67389554", "0.6670417", "0.65871423", "0.6563748", "0.6447696", "0.64369726", "0.6267657", "0.61293465", "0.59965545", "0.5843443", "0.56700474", "0.5652095", "0.56238323", "0.5586356", "0.55849177", "0.55566823", "0.55511916", "0.5548367" ]
0.81812817
0
Implementing is inline action item present functionality
def is_inline_action_item_present(self, item_name): inline_action_item_locator = (By.XPATH, "//ul[@id='Actions_listbox']/li[text()='%s']" % item_name) return self.is_element_present(inline_action_item_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def action(self,item):\r\n pass", "def _action(self):\n pass", "def action(self):\n pass", "def action(self):\n pass", "def onShowed(self):\n self.parent.actionTagTwo=\"\"\n pass", "def __actions__(self, obj):\n primary_fields = self.__provider__.get_primary_fields(self.__entity__)\n pklist = '/'.join(map(lambda x: str(getattr(obj, x)), primary_fields))\n #if has_permission('manage'):############\n \n historial = DBSession.query(Item.nrohistorial).filter_by(id=pklist).first()\n idlineabase = DBSession.query(Item.idLineaBase).filter_by(nrohistorial=historial, ultimaversion=1).first()\n lineabase = DBSession.query(LineaBase).filter_by(id=idlineabase).first()\n \n value = '<div></div>'\n \n if lineabase != None:\n if str(lineabase.estado).__eq__('abierta'):\n value = '<div><a class=\"loginlogout\" href=\"'+pklist+'/edit\" style=\"text-decoration:none\">Revertir</a></div>'\n else:\n value = '<div><a class=\"loginlogout\" href=\"'+pklist+'/edit\" style=\"text-decoration:none\">Revertir</a></div>'\n \n return value", "def custom_actions(self, form_entry, request=None):", "def manage_actions(self):\n # selelect item actions\n self.action_item_read.setDisabled(not self.selected_item)\n self.action_item_shared.setDisabled(not self.selected_item)\n self.action_item_starred.setDisabled(not self.selected_item)\n if self.selected_item:\n self.action_item_read.setChecked(not self.selected_item.unread)\n self.action_item_read.setDisabled(not self.selected_item.can_unread)\n self.action_item_shared.setChecked(self.selected_item.shared)\n self.action_item_starred.setChecked(self.selected_item.starred)\n \n # current feed actions\n self.action_mark_all_read.setDisabled(not (self.current_feed and self.current_feed.unread and not self.current_feed.is_loading))\n self.action_fetch_more.setDisabled(not (self.current_feed and self.can_fetch_more and not self.current_feed.is_loading))\n self.action_refresh.setDisabled(not (self.current_feed and not self.current_feed.is_loading))\n \n # display show mode\n if self.show_unread_only:\n self.action_show_unread_only.setChecked(True)\n else:\n self.action_show_all.setChecked(True)", "def display_item_process(self):\n raise NotImplementedError()", "def is_action(self) -> bool:\n return self.is_action_str(self.content)", "def InlineEditor(item: Item, view, pos: tuple[int, int] | None = None) -> bool:\n return False", "def __actions__(self, obj):\n value = '<div>'\n clase = 'actions'\n id = str(obj.id_atributos_por_tipo_item)\n id_tipo = UrlParser.parse_id(request.url, \"tipositems\")\n \n \n if PoseePermiso('redefinir tipo item',\n id_tipo_item=obj.id_tipo_item).is_met(request.environ):\n value += '<div>' + \\\n '<a href=\"./' + id + '/edit\" ' + \\\n 'class=\"' + clase + '\">Modificar</a>' + \\\n '</div><br />'\n\n if obj.puede_eliminarse():\n if PoseePermiso('redefinir tipo item',\n id_tipo_item=obj.id_tipo_item).is_met(request.environ):\n value += '<div><form method=\"POST\" action=\"' + id + '\" class=\"button-to\">'+\\\n '<input type=\"hidden\" name=\"_method\" value=\"DELETE\" />' +\\\n '<input onclick=\"return confirm(\\'Está seguro?\\');\" value=\"Eliminar\" type=\"submit\" '+\\\n 'style=\"background-color: transparent; float:left; border:0; color: #286571;'+\\\n 'display: inline; margin: 0; padding: 0; margin-left:-3px;\" class=\"' + clase + '\"/>'+\\\n '</form></div><br />'\n value += '</div>'\n return value", "def action_type(self):", "def is_inline_action_popup_loaded_properly(self):\n return self.is_element_present(self.vendor_profile_inline_item_locator)", "def _do_action(self):\n pass", "def _do_action(self):\n pass", "def DoAction(self,event):\r\n selections = self.list.GetSelections()\r\n if not selections: return bell()\r\n itemDex = selections[0]\r\n item = self.items[itemDex]\r\n self.data.action(item)", "def named_item_inline_editor(item, view, pos=None) -> bool:\n\n @transactional\n def update_text(text):\n item.subject.name = text\n return True\n\n subject = item.subject\n if not subject:\n return False\n\n if isinstance(item, LinePresentation):\n box = item.middle_shape_size\n i2v = view.get_matrix_i2v(item)\n x, y = i2v.transform_point(box.x, box.y)\n w, h = i2v.transform_distance(box.width, box.height)\n box = Rectangle(x, y, w, h)\n else:\n box = view.get_item_bounding_box(view.selection.hovered_item)\n name = subject.name or \"\"\n entry = popup_entry(name, update_text)\n\n def escape():\n subject.name = name\n\n show_popover(entry, view, box, escape)\n\n return True", "def actions(self):\n raise NotImplementedError", "def act(self):\n pass", "def choose_action(self):\r\n pass", "def is_inline(self):\n return True", "def get_action(self, context):\n pass", "def select_action(self):\n pass", "def act(self, infoset):\n assert self.action in infoset.legal_actions\n return self.action", "def demonstrate(self,**kwargs):\n\n members = self.bl.getAllSavedActions()\n entries={}\n\n for param in members:\n entries[str(param)] = self.executeAction # save param names in entries\n\n# entries['search colour for position'] = self.search_menu\n entries['move block to position'] = self.move_block_menu\n entries['move arm to position'] = self.move_menu\n self.mm.addGenericMenu(\"actionMenu\",self.mm.cur_page,\"Select the action to demonstrate\", entries)\n self.mm.loadMenu(\"actionMenu\")", "def select_inline_action_item_with_module_name(self, inline_item):\n self.select_inline_action_item(inline_item, self.buy_module_name)\n self.wait_for_ajax_spinner_load()", "def act(self):\n raise NotImplementedError", "def can_add_to_toolbar ( self, action ):\r\n return True", "def on_action_clicked(self, content):\n checked = content['checked']\n self.set_guarded(checked=checked)\n self.clicked(checked)" ]
[ "0.7179313", "0.6199474", "0.6192554", "0.6192554", "0.6180556", "0.61662686", "0.6146612", "0.6127133", "0.604029", "0.6010713", "0.6008397", "0.6008368", "0.58624053", "0.5829709", "0.5829346", "0.5829346", "0.5772425", "0.57424104", "0.5731849", "0.56981766", "0.56816936", "0.5678207", "0.5657569", "0.56371707", "0.5634718", "0.56224155", "0.56151223", "0.5614266", "0.558944", "0.5586091" ]
0.70441365
1
Implementing click create new target buy policy button functionality
def click_create_new_target_buy_policy_button(self): self.click_element(self.create_new_target_buy_policy_button_locator, script_executor=True, error_message='create new target buy policy button locator not found before specified time out')
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def click_save_target_buy_policy_button(self):\n self.click_element(self.save_target_buy_policy_button_locator, error_message='save target buy policy button locator not found before specified time out')\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def click_save_edited_target_buy_policy_button(self):\n self.click_element(self.save_edited_target_buy_policy_button_locator)\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def click_request_new_deal_button(self):\n self.click_element(self.request_new_deal_button_locator)", "def click_cancel_edited_target_buy_policy_button(self):\n self.click_element(self.cancel_edited_target_buy_policy_button_locator)", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def goto_create(self):\n\n self.create.click()", "def click_create_new_statement_button(self):\n self.click_element(self.create_new_statement_button_locator)", "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def btn_create_order_pro(self):\n\t\tprint()\n\t\tprint('treatment - btn_create_order_pro')\n\n\t\t# Search Partner\n\t\tpartner = tre_funcs.get_partner(self, self.patient.name)\n\n\t\t# Search pricelist\n\t\tpricelist = tre_funcs.get_pricelist(self)\n\n\t\t# Search product\n\t\t# Create Product tuple\n\t\tproduct_tup = []\n\t\t#for service in self.service_all_ids:\n\t\tfor service in self.service_ids:\n\t\t\t#print()\n\t\t\t#print('* Create Product tuple')\n\t\t\t#print(service)\n\t\t\t#print(service.service)\n\t\t\t#print(service.service.name)\n\t\t\t#print(service.qty)\n\t\t\t#print(service.service.list_price)\n\t\t\t\n\t\t\t# Init\n\t\t\tproduct_template = service.service\n\t\t\tname = service.service.name\n\t\t\tqty = service.qty\n\t\t\tprice = service.service.list_price\n\t\t\t\n\t\t\t# Check Exceptions\n\t\t\ttry:\n\t\t\t\tprice_list = '2019'\n\t\t\t\tproduct = tre_funcs.get_product_product(self, name, price_list)\n\t\t\t\tproduct_tup.append((product, qty, price))\n\n\t\t\texcept Exception:\n\t\t\t\tprint('ERROR - Treatment - Product not in 2019 price_list !')\n\t\t\t\tprint('Search in other price_lists')\n\n\t\t\t\ttry:\n\t\t\t\t\tprice_list = False\n\t\t\t\t\tproduct = tre_funcs.get_product(self, name, price_list)\n\t\t\t\t\tprint(product)\n\t\t\t\t\tproduct_tup.append((product, qty, price))\n\n\t\t\t\texcept Exception:\n\t\t\t\t\tprint('ERROR - Treatment - Product Not Available at all !!!!!')\n\n\t\t\t#else:\n\t\t\t#\tprint('jx - Else !')\n\t\t\t\t#pass\n\n\n\t\t\t# Check \n\t\t\ttre_funcs.check_product(self, '2019', product, product_template)\n\t\t\n\t\t# Create order \n\t\torder = pl_creates.create_order(self, partner.id, pricelist.id, product_tup)\n\t\tprint(order)\n\n\t\t# Open Order\n\t\treturn action_funcs.open_order(order)", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def click_buy_and_sell_deal_save_button(self):\n self.click_element(self.save_vendor_profile_locator, True)\n try:\n self.wait().until(EC.visibility_of_element_located(self.confirmation_popup_locator), 'confirmation popup locator not found before specified time out')\n self.click_element(self.ok_button_locator, True)\n except:\n raise", "def test_create_warranty(self):\n pass", "def click_request_new_buy_and_sell_deal(self):\n self.click_element(self.request_new_buy_and_sell_deal_locator, error_message=\"request new buy and sell deal locator not found before specified time out\")", "def click_buy_and_sell_deal_management_link(self):\n self.select_static_tab(self.buy_and_sell_deal_management_locator, message=\"buy and sell deal management locator not found before specified time out\")", "def awaiting_payment(self):", "def set_target_buy_distribution_link_code(self):\n self.set_value_into_input_field(self.target_buy_distribution_link_code_textbox_locator, self.random_string_generator(6))\n self.wait_for_ajax_spinner_load()", "def create_order_con(self, target, price_list):\n\tprint()\n\tprint('OH - pl_create_order_con')\n\tprint(self)\n\tprint(target)\n\tprint(price_list)\n\n\t# Search Partner\n\tprint()\n\tprint('Search partner')\n\tpartner = self.env['res.partner'].search([\n\t\t\t\t\t\t\t\t\t\t\t\t('name', '=', self.patient.name),\n\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\tlimit=1,)\n\tprint(partner)\n\n\t# Search\n\tprint()\n\tprint('Search pricelist')\n\tpricelist = self.env['product.pricelist'].search([\n\t\t\t\t\t\t\t\t\t\t\t#('active', 'in', [True]),\n\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\t#order='x_serial_nr asc',\n\t\t\t\t\t\t\t\t\t\t\tlimit=1,\n\t\t\t\t\t\t\t\t\t\t)\n\tprint(pricelist)\n\n\n\t# Create Order\n\tprint()\n\tprint('Create order')\n\tprint(partner.id)\n\tprint(self.patient.id)\n\tprint(self.patient.x_id_doc)\n\tprint(self.patient.x_id_doc_type)\n\tprint(self.physician.id)\n\torder = self.env['sale.order'].create({\n\t\t\t\t\t\t\t\t\t\t\t'patient': self.patient.id,\n\t\t\t\t\t\t\t\t\t\t\t'x_id_doc': self.patient.x_id_doc,\n\t\t\t\t\t\t\t\t\t\t\t'x_id_doc_type': self.patient.x_id_doc_type,\n\t\t\t\t\t\t\t\t\t\t\t'x_doctor': self.physician.id,\n\t\t\t\t\t\t\t\t\t\t\t'state':'draft',\n\t\t\t\t\t\t\t\t\t\t\t'partner_id': partner.id,\n\t\t\t\t\t\t\t\t\t\t\t'x_family': 'consultation',\n\t\t\t\t\t\t\t\t\t\t\t'treatment': self.id,\n\n\t\t\t\t\t\t\t\t\t\t\t'pricelist_id': pricelist.id,\n\t\t\t\t\t\t\t\t\t\t})\n\tprint(order)\n\n\t# Init\n\t_dic_con = {\n\t\t\t\t\t'medical':\t\t'CONSULTA MEDICA',\n\t\t\t\t\t'gynecology':\t'CONSULTA GINECOLOGICA',\n\t\t\t\t\t'premium':\t\t'CONSULTA MEDICA DR. CHAVARRI',\n\t}\n\tname = _dic_con[target]\n\n\t# Search\n\tprint()\n\tprint('Search product')\n\tproduct = self.env['product.product'].search([\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t('name', 'in', [name]),\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t('pl_price_list', 'in', [price_list]),\n\t\t\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t#order='date_begin asc',\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t#limit=1,\n\t\t\t\t\t\t\t\t\t\t\t\t)\n\tprint(product)\n\tprint(product.name)\n\n\t# Create Order Line\n\tprint()\n\tprint('Create order line')\n\tol = order.order_line.create({\n\t\t\t\t\t\t\t\t\t'name': \t\t\tproduct.name,\n\t\t\t\t\t\t\t\t\t'product_id': \t\tproduct.id,\n\t\t\t\t\t\t\t\t\t'order_id': \t\torder.id,\n\t\t\t\t\t\t\t\t})\n\treturn order", "def test_create_custom_button(self):\n pass", "def create_deb_for_target(ctx, target=target):\n pass", "def click_add_financial_charges_button(self):\n self.number_of_financial_charges_before_create = self.get_number_of_financial_charges_in_the_grid()\n self.click_element(self.add_financial_charges_button_locator, hover=True)", "def click_create_adjustment_charge_save_button(self):\n self.click_element(self.create_adjustment_charge_save_button_locator)\n try:\n self.wait().until(EC.visibility_of_element_located(self.success_message_locator), 'success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, True)\n except:\n raise\n self.wait_for_ajax_spinner_load()", "def purchase(self, item_type):", "def click_create_vendor_button(self):\n create_vendor_element = self.wait().until(EC.element_to_be_clickable(self.create_vendor_locator), \"create vendor locator not found before specified time out\")\n create_vendor_element.click()\n self.wait_for_ajax_spinner_load()", "def test_buy_now(self):\n catalog_page = CatalogPage(self.driver)\n product_page = ProductPage(self.driver)\n payment_page = PaymentPage(self.driver)\n payment_review_page = PaymentReviewPage(self.driver)\n payment_info_page = PaymentInfoPage(self.driver)\n success_page = SuccessPage(self.driver)\n # buy the new product\n navigate_to(self.driver, ProductPage.URL(self.new_product['product']['title']))\n product_page.add_to_cart.click()\n # by an old product\n catalog_page.catalog.click()\n # Sort products to move the newly created to last page\n catalog_page.sorting_order.select_by_visible_text(\"Date, old to new\")\n catalog_page.image.random_click()\n product = product_page.product.get_text()\n product_page.add_to_cart.click()\n catalog_page.catalog.click()\n catalog_page.cart.click()\n payment_dic = {\n 'address' : f'{randint(1, 99999)} {random_name(5, 8)}',\n 'city' : \"San Francisco\",\n 'email_or_mobile_phone_number_input' : random_name(8) + \"@gmail.com\",\n 'last_name' : random_name(3, 12),\n 'zip_code' : '94107',\n }\n if randint(0, 1):\n payment_dic['first_name'] = random_name(4, 16)\n if randint(0, 1):\n payment_dic['address2'] = random_name(5)\n for _ in payment_dic:\n exec(f\"payment_page.{_}.enter(payment_dic['{_}'])\")\n payment_page.continue_to_shipping.click()\n payment_review_page.continue_to_payment.click()\n payment_info_page.full_address.get_text()\n # validate address\n for _ in ['address', 'city', 'zip_code']:\n assert_and_log(payment_dic[_] in payment_info_page.full_address.get_text(),\n f\"{_} in full address\")\n payment_info_page.enter_bogus_payment(1)\n assert_and_log(success_page.thank_you.find_visible_element(),\n \"'Thank you' appeared as a sign of successful transaction\",\n continue_on_error=False)\n validate(success_page.basic_validation_list)", "def click_target_buy_overrides_button(self):\n self.click_element(self.target_buy_overrides_button_locator, script_executor=True, error_message='target buy overrides button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def set_route_class_to_create_new_target_buy_policy(self, route_class):\n self.single_selection_from_static_kendo_dropdown(self.route_class_single_kendo_dropdown_arrow_locator, route_class)" ]
[ "0.7408507", "0.7157741", "0.6822164", "0.6584997", "0.65837705", "0.627723", "0.6117332", "0.6084861", "0.60220706", "0.59835017", "0.5927379", "0.5761761", "0.57569414", "0.57379687", "0.5736414", "0.5709918", "0.5691938", "0.56675476", "0.5583201", "0.5517948", "0.5517535", "0.5505093", "0.5491316", "0.54294497", "0.542886", "0.53930825", "0.5390926", "0.538549", "0.538316", "0.53798056" ]
0.7672479
0
Implementing set target buy policy name functionality
def set_target_buy_policy_name(self, prefix): global target_buy_policy target_buy_policy = (prefix + self.random_string_generator(size=4)) self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_target_buy_policy_abbreviation(self, prefix):\n self.set_value_into_input_field(self.target_buy_policy_abbreviation_input_field_locator, value=(prefix + self.random_string_generator(size=4)))", "def use(target, name):\n return \"You use the \" + name + \" to pay \" + target.name", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def set_policyname(self, policyname):\n self.options[\"policyname\"] = policyname", "def set_policyname(self, policyname):\n self.options['policyname'] = policyname", "def policy_name(self, policy_name):\n\n self._policy_name = policy_name", "def policy_name(self, policy_name):\n\n self._policy_name = policy_name", "def set_buy_and_sell_deal_name(self, buy_and_sell_deal_prefix):\n global buy_and_sell_deal\n buy_and_sell_deal = buy_and_sell_deal_prefix + self.random_string_generator(size=4)\n self.set_value_into_input_field(self.buy_and_sell_deal_textbox_locator, buy_and_sell_deal)", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def set_target_policy(self, policy):\n self.target_policy = policy", "def name(self, name: str):\n self.inst['targetname'] = name", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def target_name(self, target_name):\n\n self._target_name = target_name", "def target_name(self, target_name):\n\n self._target_name = target_name", "def set_target_buy_distribution_receiving_member(self, member_prefix):\n member_name = member_prefix + self.random_string_generator(size=4)\n self.set_value_into_input_field(self.target_buy_distribution_receiving_member_textbox_locator, member_name)\n self.wait_for_ajax_spinner_load()", "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def policy_alias(self):", "def policy(agent):", "def set_policy(self, name, policy):\n client = self.connect(VAULT_TOKEN)\n client.set_policy(name, policy)", "def setName(self, name: str, /) -> Any:\n ...", "def use(target, name):\n return \"You insert the \" + name + \" into \" + target.name", "def __init__(__self__, *,\n name: pulumi.Input['SkuName']):\n pulumi.set(__self__, \"name\", name)", "def use(target, name):\n return \"The \" + name + \" takes effect on \" + target.name", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def set_target_buy_distribution_link_code(self):\n self.set_value_into_input_field(self.target_buy_distribution_link_code_textbox_locator, self.random_string_generator(6))\n self.wait_for_ajax_spinner_load()", "def policy_target_version(self, policy_target_version):\n\n self._policy_target_version = policy_target_version", "def use(target, name):\n out = target.damage() + \"\\n\"\n return out + \"You swing the \" + name + \" at \" + target.name", "def test_basic_change_policy_name(self):\n args = self.get_args()\n config = self.create_config_file()\n mac = '00:11:22:33:33:33'\n ip = '3.4.3.4'\n self.write_config_file(config, args)\n collector = execute_tool(args, test_mode=True)\n time.sleep(4)\n self.assertTrue(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertTrue(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n\n config = self.create_diff_epg_config_file()\n self.write_config_file(config, args)\n collector.reload_config()\n\n time.sleep(4)\n\n self.assertFalse(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertFalse(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertTrue(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg2'))\n self.assertTrue(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg2'))", "def use(target, name):", "def update_policy(self, *args, **kwargs):\r\n pass" ]
[ "0.70902216", "0.65581495", "0.6447595", "0.63000184", "0.6270888", "0.6120589", "0.6120589", "0.6109877", "0.604", "0.5914904", "0.5862381", "0.58405066", "0.5839304", "0.5839304", "0.56366056", "0.55916286", "0.5582298", "0.556557", "0.55630815", "0.5550732", "0.55316234", "0.5488869", "0.5473841", "0.5392743", "0.53885436", "0.5384199", "0.5380518", "0.537215", "0.535304", "0.52839744" ]
0.79451764
0
Implementing get target buy policy functionality
def get_target_buy_policy(self): global target_buy_policy return target_buy_policy
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def policy(agent):", "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def getAction(self, state):\n # Pick Action\n legalActions = self.getLegalActions(state)\n action = None\n\n \"\"\"Description:\n Use util.flipCoin, if return true then randomly choice from legalAction\n if flase, then sue getPolicy to get best policy action\n \"\"\"\n \"\"\" YOUR CODE HERE \"\"\"\n if len(legalActions) == 0:\n return action # None\n \n if util.flipCoin(self.epsilon):\n ''' exploration function (not work well)''' \n# posPol = util.Counter()\n# for a in legalActions:\n# if self.getQValue(state,a) >= 0:\n# posPol[a] = -1*self.getQValue(state, a) + (1000/(self.vitCount[(state,a)]+0.0001))\n# #print \"posPol[\", a, \"]= \",posPol[a]\n# #posPol[a] = (self.getQValue(state, a) * self.epsilon** self.vitCount[(state,a)]) + ( self.epsilon/(self.vitCount[(state,a)]+0.1) )\n# if len(posPol) == 0:\n# action = random.choice(legalActions)\n# else:\n# action = posPol.argMax() # random.choice(posPol.keys())\n ''' Random exploration '''\n action = random.choice(legalActions)\n else:\n action = self.getPolicy(state)\n \n \"\"\" END CODE \"\"\"\n\n return action", "def getPolicy(self, state):\n \"\"\"Description:\n Find all of q-values of current state, and choose the action \n with the hight q-value as optimal policy\n \"\"\"\n \"\"\" YOUR CODE HERE \"\"\"\n legalActions = self.getLegalActions(state)\n action = None\n policy = util.Counter() # use counter to store action and its q-value\n \n if len(legalActions) == 0:\n return action\n \n for a in legalActions:\n policy[a] = self.getQValue(state, a)\n action = policy.argMax()\n return action\n\n \"\"\" END CODE \"\"\"", "def get_policy(self, *args, **kwargs):\r\n pass", "def target_buy_price(self):\n if self.period_tick == 0:\n return random.randint(1, 10)\n elif self.period_tick % self.perseverance == 0:\n # Player runs out of patience and decides to change target price.\n (avg_price,\n max_price,\n min_price) = self.market.get_stock_price_last_period()\n\n power = self.period_tick / self.perseverance\n target_price = min(min_price + power, self.money_balance * 0.5)\n return target_price\n else:\n return None", "def get_payoff(model, attack_policy, defense_policy):\n ave_discount_reward = get_payoff_mixed(model, [attack_policy], [defense_policy], [1.0], [1.0])\t\n return ave_discount_reward", "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def getPolicy(self, state):\n \"*** YOUR CODE HERE ***\"\n # OUR CODE HERE\n possibleActions = self.mdp.getPossibleActions(state)\n #checking for terminal state (no possible actions)\n if len(possibleActions) is 0: \n return None\n \n #attempt at using the Counter\n eValsActions = util.Counter()\n for action in possibleActions:\n for transitionState, probability in self.mdp.getTransitionStatesAndProbs(state, action):\n eValsActions[action] += probability * (self.mdp.getReward( state, action, transitionState) + self.discount * self.values[transitionState])\n \n return eValsActions.argMax()\n \n #fail attempt using lists :(\n \"\"\"\n #list to hold the expected value of the actions\n eValsActions = []\n #iterate through all actions and their transtion states\n for action in possibleActions:\n for transitionState, probability in self.mdp.getTransitionStatesAndProbs(state, action):\n #expected value of reward with discount * the value of the transitions\n eValsActions[action] += probability * (self.mdp.getReward( state, action, transitionState) + self.discount * self.values[transitionState])\n \n #now iterate through and find the action with the best value\n #(that will be the best action)\n maxVal = -float(\"inf\")\n bestAction = None\n for action in possibleActions:\n if eValsActions[action] > maxVal:\n maxVal = eValsAction[action]\n bestAction = action\n \"\"\"\n return action\n # END OUR CODE", "def target_sell_price(self):\n return super(Player, self).target_sell_price", "def get_target_market():\n return choice(GraphParameters.TARGET_MARKETS)", "def calculate_appropriate_target(self):\n pass", "def calculate_appropriate_target(self):\n pass", "def buy_and_pay(self):\n return self.price", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def get_cost(org, target, amount):\n rep, _ = target.Dominion.reputations.get_or_create(organization=org)\n base = 200\n if amount > 0:\n base -= rep.respect + rep.affection\n else:\n base += rep.respect + rep.affection\n if base < 0:\n base = 0\n return base * abs(amount)", "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def get_target_policy(self, state):\n values = self.get_all_state_action_values(state)\n return self.target_policy(values)", "def buy_one_get_one(products):\n if 'p1' in products and products['p1'] >= 2:\n return -20\n else:\n return 0", "def execute_policy_and_get_cost(curr_node, reward_machines, policy_bank, tester, new_task_rm, new_task_u1,\n bound=np.inf):\n game = copy.deepcopy(curr_node.parent_state)\n num_features = len(game.get_features())\n s1, s1_features = game.get_state_and_features()\n curr_policy = curr_node.policy\n curr_policy_rm = reward_machines[curr_policy[0]]\n\n bonus = []\n for t in range(tester.testing_params.num_steps):\n a = policy_bank.get_best_action(curr_policy[0], curr_policy[1],\n s1_features.reshape((1, num_features)), add_noise=False)\n game.execute_action(a)\n # game.render()\n s2, s2_features = game.get_state_and_features()\n curr_policy_u2 = curr_policy_rm.get_next_state(curr_policy[1], game.get_true_propositions())\n new_task_u2 = new_task_rm.get_next_state(new_task_u1, game.get_true_propositions())\n\n desired_next_state = curr_policy_rm.get_next_state(curr_policy[1], curr_policy[2])\n\n r = new_task_rm.get_reward(new_task_u1, new_task_u2, s1, a, s2)\n if curr_policy_u2 == desired_next_state:\n logger.info(\"EXECUTED ACTION {}, CAN GO TO NEXT LEVEL\".format(curr_policy[2]))\n return t + 1, game, new_task_u2, r, bonus\n elif curr_policy_u2 == curr_policy[1]:\n logger.info(\"STILL FOLLOWING CURRENT POLICY {}, CONTINUE\".format(curr_policy[2]))\n if new_task_u2 != new_task_u1:\n logger.info(\n \"ENCOUNTERED EVENT {} WHILE FOLLOWING {}\".format(game.get_true_propositions(), curr_policy[2]))\n bonus.append(game.get_true_propositions())\n # else:\n # curr_policy_u2 = curr_policy[1]\n # print(game.get_true_propositions())\n # print(\"OOPS, WRONG WAY\")\n # return np.inf, game, new_task_u1, r, bonus\n\n if game.is_env_game_over() or t + 1 >= bound:\n return np.inf, game, new_task_u2, r, bonus\n\n s1, s1_features = s2, s2_features\n new_task_u1 = new_task_u2\n\n return np.inf, game, new_task_u1, 0, bonus", "def policy(self):\r\n actions = self.domain.get_actions()\r\n # Random policy (please uncomment the following line to test it, and comment the \"always accelerate\" policy)\r\n #index = random.randint(0, 1)\r\n # \"Always accelerate\" policy\r\n index = 1\r\n return actions[index]", "def getPolicy(self, state):\n \"*** YOUR CODE HERE ***\"\n possibleActions = self.mdp.getPossibleActions(state)\n if len(possibleActions) == 0: return None\n results = []\n for action in possibleActions:\n total = 0\n for (nextState, prob) in self.mdp.getTransitionStatesAndProbs(state,action):\n total += (prob * self.values[nextState])\n results.append(total)\n maxIndex = max(enumerate(results), key=lambda x: x[1])[0]\n #print(\"here\")\n return possibleActions[maxIndex]", "def prepare_trade(self, input_, prediction):\n if prediction > input_:\n # Price will go up, so we should buy\n # amount = self.amount\n amount = self.amount\n allowed, amount_ret = self.check_risk('buy', amount)\n assert amount == amount_ret or amount == 'max', \"Mistake in check_risk function\"\n if allowed:\n return 'buy', amount_ret\n else:\n return False, amount_ret\n elif prediction < input_:\n # Sell, short or hold?\n amount = -1 * self.amount\n allowed, amount_ret = self.check_risk('buy', amount)\n assert amount == amount_ret, \"Mistake in check_risk function\"\n if allowed:\n return 'sell', amount_ret\n else:\n return False, amount_ret", "def policy(self, s):\r\n if s.dealer_sum >= 16:\r\n return Action.STICK\r\n else:\r\n return Action.HIT", "def test_get_available_taxation_strategies_for_rate_plan(self):\n pass", "def awaiting_payment(self):", "def purchase(self, item_type):", "def best_buy(self):\n return Library.functions.best_buy(self._book)", "def get_policy(usage_id):\r\n return policy.get(policy_key(usage_id), {})", "def test_get_hyperflex_auto_support_policy_by_moid(self):\n pass" ]
[ "0.62555337", "0.6185256", "0.61499", "0.61307913", "0.60904336", "0.6055906", "0.595635", "0.5928054", "0.5914833", "0.5885227", "0.58382344", "0.5782354", "0.5782354", "0.57409716", "0.5726041", "0.56711364", "0.5637255", "0.563183", "0.56234705", "0.5622425", "0.5594815", "0.5589983", "0.558783", "0.5585256", "0.5551527", "0.5538434", "0.5509124", "0.54890233", "0.5461867", "0.54330665" ]
0.7827906
0
Implementing set target buy policy abbreviation functionality
def set_target_buy_policy_abbreviation(self, prefix): self.set_value_into_input_field(self.target_buy_policy_abbreviation_input_field_locator, value=(prefix + self.random_string_generator(size=4)))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def use(target, name):\n return \"You use the \" + name + \" to pay \" + target.name", "def test_add_taxation_strategy_to_rate_plan(self):\n pass", "def _get_set(cost):\n if any(i in cost for i in [\"_cap\", \"depreciation_rate\", \"purchase\", \"area\"]):\n return \"loc_techs_investment_cost\"\n elif any(i in cost for i in [\"om_\", \"export\"]):\n return \"loc_techs_om_cost\"\n else:\n return \"loc_techs\"", "def set_buy_and_sell_deal_name(self, buy_and_sell_deal_prefix):\n global buy_and_sell_deal\n buy_and_sell_deal = buy_and_sell_deal_prefix + self.random_string_generator(size=4)\n self.set_value_into_input_field(self.buy_and_sell_deal_textbox_locator, buy_and_sell_deal)", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def set_target_buy_distribution_link_code(self):\n self.set_value_into_input_field(self.target_buy_distribution_link_code_textbox_locator, self.random_string_generator(6))\n self.wait_for_ajax_spinner_load()", "def test_get_available_taxation_strategies_for_rate_plan(self):\n pass", "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def policy_alias(self):", "def policy(agent):", "def test_remove_taxation_strategy_from_rate_plan(self):\n pass", "def getPolicy(self, state):\n \"\"\"Description:\n Find all of q-values of current state, and choose the action \n with the hight q-value as optimal policy\n \"\"\"\n \"\"\" YOUR CODE HERE \"\"\"\n legalActions = self.getLegalActions(state)\n action = None\n policy = util.Counter() # use counter to store action and its q-value\n \n if len(legalActions) == 0:\n return action\n \n for a in legalActions:\n policy[a] = self.getQValue(state, a)\n action = policy.argMax()\n return action\n\n \"\"\" END CODE \"\"\"", "def abbreviation(self, abbreviation):\n self._abbreviation = abbreviation", "def left_refactor(self,target_rule_name,stop_at_set):\n name_suffix = \".post.{}\".format(target_rule_name)\n\n # Map a rule name X to a set of rules Y where X appears\n # as a first nonterminal in one of Y's options.\n appears_first_in = defaultdict(set)\n for name, rule in self.rules.items():\n for option in rule.as_container():\n first = option.as_container()[0]\n if first.is_symbol_name():\n appears_first_in[first.content].add(name)\n #print(\"appears first dict\\n{}\\n\\n\".format(appears_first_in))\n\n po = PrintOption()\n po.is_canonical = self.is_canonical\n po.inline_synthetic = False\n candidates = set(self.rules.keys())\n while len(candidates) > 0:\n for A in list(candidates):\n candidates.remove(A)\n if A in stop_at_set:\n continue\n rule = self.rules[A]\n (starts,others,terms,empties) = rule.partition(target_rule_name)\n if len(starts) > 0 and (len(others)+len(terms)+len(empties) == 0):\n #print(\"processing {}\".format(A))\n # Create the new rule.\n new_rule_name = \"{}{}\".format(A,name_suffix)\n # Form alpha1 ... alphaN\n new_options = []\n for option in rule:\n if len(option.as_container()) == 1:\n new_options.append(self.MakeEmpty())\n else:\n assert option.is_container() and (len(option)>1)\n new_options.append(self.MakeSeq(option[1:]))\n self.rules[new_rule_name] = self.MakeChoice(new_options)\n\n # Rewrite A itself.\n self_parts = [self.MakeSymbolName(x) for x in [target_rule_name,new_rule_name]]\n self.rules[A] = self.MakeChoice([self.MakeSeq(self_parts)])\n\n # Update bookkeeping for appears_first_in\n for option in new_options:\n first = option.as_container()[0]\n if first.is_symbol_name():\n appears_first_in[first.content].add(new_rule_name)\n\n # Replace the old rule everywhere it appears in the first\n # position\n for parent_name in list(appears_first_in[A]):\n if parent_name == A:\n # Already processed above\n continue\n parent = self.rules[parent_name]\n (starts,others,terms,empties) = parent.partition(A)\n new_options = []\n for option in starts:\n parts = []\n parts.append(self.MakeSymbolName(target_rule_name))\n parts.append(self.MakeSymbolName(new_rule_name))\n parts.extend(option.as_container()[1:])\n new_options.append(self.MakeSeq(parts))\n new_options.extend(others+terms+empties)\n self.rules[parent_name] = self.MakeChoice(new_options)\n appears_first_in[A].remove(parent_name)\n appears_first_in[target_rule_name].add(parent_name)\n # Set up transitive closure.\n candidates.add(parent_name)\n\n #print()\n #print()\n #print()\n\n #self.absorb_post(target_rule_name)\n self.remove_unused_rules()", "def get_target_market():\n return choice(GraphParameters.TARGET_MARKETS)", "def use(target, name):\n return \"You insert the \" + name + \" into \" + target.name", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def target_option(s):\n return s", "def calculate_appropriate_target(self):\n pass", "def calculate_appropriate_target(self):\n pass", "def __get_settlement_type():\n return 'Beneficial Ownership'", "def set_abbreviation(self, abbreviation_name_prefix):\n abbreviation_name = abbreviation_name_prefix + self.vendor_unique_id\n self.set_value_into_input_field(self.abbreviation_textbox_locator, abbreviation_name)\n self.wait_for_ajax_spinner_load()\n return abbreviation_name", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def test_reasonable_auction(self):\n # I have no preferences\n bids = [Cost(ITEM1, ACTOR1, 1000),\n Cost(ITEM2, ACTOR1, 1000),\n Cost(ITEM3, ACTOR1, 1000),\n Cost(ITEM4, ACTOR1, 1000),\n Cost(ITEM5, ACTOR1, 1000),\n\n # I have linear preferences\n Cost(ITEM1, ACTOR2, 700),\n Cost(ITEM2, ACTOR2, 800),\n Cost(ITEM3, ACTOR2, 1000),\n Cost(ITEM4, ACTOR2, 1200),\n Cost(ITEM5, ACTOR2, 1300),\n\n # I have non-linear preferences\n Cost(ITEM1, ACTOR3, 400),\n Cost(ITEM2, ACTOR3, 800),\n Cost(ITEM3, ACTOR3, 1000),\n Cost(ITEM4, ACTOR3, 1200),\n Cost(ITEM5, ACTOR3, 1600),\n\n # I have arbitrary preference\n Cost(ITEM1, ACTOR4, 2435),\n Cost(ITEM2, ACTOR4, 305),\n Cost(ITEM3, ACTOR4, 310),\n Cost(ITEM4, ACTOR4, 1725),\n Cost(ITEM5, ACTOR4, 225),\n\n # I have strong preferences\n Cost(ITEM1, ACTOR5, 0),\n Cost(ITEM2, ACTOR5, 0),\n Cost(ITEM3, ACTOR5, 0),\n Cost(ITEM4, ACTOR5, 0),\n Cost(ITEM5, ACTOR5, 5000)]\n result = self.splitter.split(ITEMS[:5], ACTORS[:5], bids)\n expected = [(ITEM1, ACTOR4, None),\n (ITEM2, ACTOR1, None),\n (ITEM3, ACTOR3, None),\n (ITEM4, ACTOR2, None),\n (ITEM5, ACTOR5, None)]\n item_assignments_present(self, result, expected)", "async def _edittarget_heist(self, ctx, *, target: str):\r\n author = ctx.message.author\r\n guild = ctx.guild\r\n target = string.capwords(target)\r\n targets = await self.thief.get_guild_targets(guild)\r\n\r\n if target not in targets:\r\n return await ctx.send(\"That target does not exist.\")\r\n\r\n keys = [x for x in targets[target]]\r\n keys.append(\"Name\")\r\n check = lambda m: m.content.title() in keys and m.author == author\r\n\r\n await ctx.send(\"Which property of {} would you like to edit?\\n\"\r\n \"{}\".format(target, \", \".join(keys)))\r\n\r\n response = await self.bot.wait_for('MESSAGE', timeout=15, check=check)\r\n\r\n if response is None:\r\n return await ctx.send(\"Canceling removal. You took too long.\")\r\n\r\n if response.content.title() == \"Name\":\r\n await ctx.send(\"What would you like to rename the target to?\\n*Cannot be a name \"\r\n \"currently in use.*\")\r\n check2 = lambda m: string.capwords(m.content) not in targets and m.author == author\r\n\r\n elif response.content.title() in [\"Vault\", \"Vault Max\"]:\r\n await ctx.send(\"What would you like to set the {} \"\r\n \"to?\".format(response.content.title()))\r\n check2 = lambda m: m.content.isdigit() and int(m.content) > 0 and m.author == author\r\n\r\n elif response.content.title() == \"Success\":\r\n await ctx.send(\"What would you like to change the success rate to?\")\r\n check2 = lambda m: m.content.isdigit() and 0 < int(m.content) <= 100 and m.author == author\r\n\r\n elif response.content.title() == \"Crew\":\r\n await ctx.send(\"What would you like to change the max crew size to?\\n Cannot be \"\r\n \"the same as another target and will be the maximum number of \"\r\n \"players for that target.\")\r\n crew_sizes = [subdict[\"Crew\"] for subdict in targets.values()]\r\n check2 = lambda m: m.content.isdigit() and int(m.content) not in crew_sizes and m.author == author\r\n\r\n choice = await self.bot.wait_for('MESSAGE', timeout=15, check=check2)\r\n\r\n if choice is None:\r\n return await ctx.send(\"Canceling removal. You took too long.\")\r\n\r\n if response.content.title() == \"Name\":\r\n new_name = string.capwords(choice.content)\r\n targets[new_name] = targets.pop(target)\r\n await self.thief.save_targets(guild, targets)\r\n await ctx.send(\"Changed {}'s {} to {}.\".format(target, response.content,\r\n choice.content))\r\n else:\r\n targets[target][response.content.title()] = int(choice.content)\r\n await self.thief.save_targets(guild, targets)\r\n await ctx.send(\"Changed {}'s {} to {}.\".format(target, response.content,\r\n choice.content))", "def choose_action(self, d, name):\n # some initializations\n current_state = d.state\n caution = False\n confidence = False\n buy_rules = [0,0,0,0]\n next_vec = self.transition_matrix[name][int(current_state)]\n num_undesirable_states = (self.trigger_states[name]+1)\n num_desirable_states = (next_vec.size-num_undesirable_states)\n if num_undesirable_states<5:\n left_basket_max = 2\n else:\n left_basket_max = num_undesirable_states//3\n if num_desirable_states<5:\n right_basket_min = next_vec.size-2\n else:\n right_basket_min = next_vec.size-num_undesirable_states//3\n # check if rules are satisfied\n # rule-1\n m1 = np.max(next_vec[:self.trigger_states[name]+1])\n m1_idx = np.argmax(next_vec[:self.trigger_states[name]+1])\n m2 = np.max(next_vec[self.trigger_states[name]+1:])\n m2_idx = np.argmax(next_vec[self.trigger_states[name]+1:])+\\\n next_vec[:self.trigger_states[name]+1].size\n if m2-m1>=0.1: # threshold\n #print('Rule #1 satisfied.')\n buy_rules[0]=1\n # rule-2\n if np.sum(next_vec[self.trigger_states[name]+1:])-\\\n np.sum(next_vec[:self.trigger_states[name]+1])>=0.25: # threshold\n #print('Rule #2 satisfied.')\n buy_rules[1]=1\n # rule-3 \n if m1_idx<left_basket_max: \n if buy_rules[0]!=1:\n caution=True\n #print('Predicted state is very undesirable.')\n # rule-3\n if m2_idx>=right_basket_min:\n if buy_rules[0]==1:\n confidence=True\n #print('Predicted state is very desirable.')\n if d.MACD>d.signal_line:\n #print('Rule #3 satisfied.')\n buy_rules[2] = True\n # sum of k most undesirable vs k most desirable\n temp_1 = np.sort(next_vec[self.trigger_states[name]+1:])\n temp_2 = np.sort(next_vec[:self.trigger_states[name]+1])\n size = 3\n if temp_1.size<size or temp_2.size<size:\n size = min(temp_1.size, temp_2.size)\n k1 = np.sum(temp_1[::-size])\n k2 = np.sum(temp_2[::-size])\n if k1-k2>0.25:\n #print('Rule #4 satisfied.')\n buy_rules[3] = True\n # finally, make a call using the rules\n if confidence or sum(buy_rules)>=3:\n return 'buy'\n elif caution or (buy_rules[0]==0 and sum(buy_rules)<=2 and m1-m2>0.05):\n return 'sell'\n else:\n return 'hold'", "def choose_luggage(self, *args, **kwargs):", "def target(self, assault_army: dict, arr_target: dict) -> None:\n raise NotImplementedError(\"target() method is undefined\")" ]
[ "0.6439152", "0.5870092", "0.5696178", "0.5516288", "0.53760064", "0.53070027", "0.52721477", "0.51333565", "0.51169074", "0.51167685", "0.50744283", "0.5036501", "0.50308836", "0.49818993", "0.49748594", "0.49631476", "0.49574703", "0.49414706", "0.49384534", "0.49269494", "0.49226442", "0.49226442", "0.49170947", "0.49148884", "0.48944974", "0.4879226", "0.48720896", "0.48517603", "0.4851324", "0.48365942" ]
0.70045894
0
Implementing set route class to create new target buy policy functionality
def set_route_class_to_create_new_target_buy_policy(self, route_class): self.single_selection_from_static_kendo_dropdown(self.route_class_single_kendo_dropdown_arrow_locator, route_class)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def __init__(self, source_node, source_gate_name, target_node, target_slot_name, weight=1):\n self.link(source_node, source_gate_name, target_node, target_slot_name, weight)", "def policy(agent):", "def apply_policy(self, policy):\n tenant_name = policy['tenant_name']\n fw_id = policy['fw_id']\n fw_name = policy['fw_name']\n LOG.debug(\"asa_apply_policy: tenant=%s fw_id=%s fw_name=%s\",\n tenant_name, fw_id, fw_name)\n cmds = [\"conf t\", \"changeto context \" + tenant_name]\n\n rule_dict = policy['rules']\n for rule_id in rule_dict:\n rule = rule_dict[rule_id]\n protocol = rule['protocol']\n name = rule['name']\n enabled = rule['enabled']\n dst_port = rule['destination_port']\n src_port = rule['source_port']\n\n if (rule['source_ip_address'] is not None):\n src_ip = IPNetwork(rule['source_ip_address'])\n else:\n src_ip = IPNetwork('0.0.0.0/0')\n\n if (rule['destination_ip_address'] is not None):\n dst_ip = IPNetwork(rule['destination_ip_address'])\n else:\n dst_ip = IPNetwork('0.0.0.0/0')\n\n if rule['action'] == 'allow':\n action = 'permit'\n else:\n action = 'deny'\n\n LOG.debug(\"rule[%s]: name=%s enabled=%s prot=%s dport=%s sport=%s \\\n dip=%s %s sip=%s %s action=%s\",\n rule_id, name, enabled, protocol, dst_port, src_port,\n dst_ip.network, dst_ip.netmask,\n src_ip.network, src_ip.netmask, action)\n\n acl = \"access-list \"\n acl = (acl + tenant_name + \" extended \" + action + \" \" +\n protocol + \" \")\n if (rule['source_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(src_ip.network) + \" \" + (\n str(src_ip.netmask) + \" \")\n if (src_port is not None):\n if (':' in src_port):\n range = src_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + src_port + \" \"\n if (rule['destination_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(dst_ip.network) + \" \" + \\\n str(dst_ip.netmask) + \" \"\n if (dst_port is not None):\n if (':' in dst_port):\n range = dst_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + dst_port + \" \"\n if (enabled is False):\n acl = acl + 'inactive'\n\n # remove the old ace for this rule\n if (rule_id in self.rule_tbl):\n cmds.append('no ' + self.rule_tbl[rule_id])\n\n self.rule_tbl[rule_id] = acl\n if tenant_name in self.tenant_rule:\n if rule_id not in self.tenant_rule[tenant_name]['rule_lst']:\n self.tenant_rule[tenant_name]['rule_lst'].append(rule_id)\n cmds.append(acl)\n cmds.append(\"access-group \" + tenant_name + \" global\")\n cmds.append(\"write memory\")\n\n LOG.debug(cmds)\n data = {\"commands\": cmds}\n return self.rest_send_cli(data)", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def pre_route_target_create(self, resource_dict):\n pass", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def route(self):\n pass", "def test_add_taxation_strategy_to_rate_plan(self):\n pass", "def do_turn(self, price: int) -> SalesmanAction:", "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def make_decision_with_policy(self, policy_type, *args):\n if policy_type == 1: # ADP\n assert len(args) == 2, 'args should be exactly 2'\n cur_K = -self.K_im_traj[-1]\n distance_2_tan, radian_at_tan = args\n self.dis_sum += distance_2_tan\n pwm_l_new, pwm_r_new = policy.adp(distance_2_tan, radian_at_tan, self.dis_sum, cur_K)\n elif policy_type == 2: # pure pursuit\n l_d, sin_alpha = args\n amp = 150\n pwm_l_new, pwm_r_new = policy.pure_pursuit(l_d, sin_alpha, amp)\n elif policy_type == 3: # Car following with ADP\n assert len(args) == 3, 'args should be exactly 3'\n cur_K = -self.K_im_traj[-1]\n distance_2_tan, radian_at_tan, estimated_dis = args\n self.dis_sum += distance_2_tan\n if self.is_recording and self.counter % 100 == 0:\n np.save('./.out/record', self.record)\n pwm_l_new, pwm_r_new = policy.car_following_with_adp(distance_2_tan, radian_at_tan, self.dis_sum, cur_K, estimated_dis, self.record)\n print(self.counter)\n self.counter += 1\n elif policy_type == 4:\n K = 0.5\n dis2car, = args\n pwm_l_new, pwm_r_new = policy.car_following(dis2car, K)\n elif policy_type == 5:\n d_arc, d_curve, theta = args\n pwm_l_new, pwm_r_new = policy.adp_coupled_car_following(d_arc, d_curve, theta, self.z, self.K_coupled)\n else:\n pwm_l_new, pwm_r_new = 0, 0\n print('Policy Not Found')\n self.motor.motor_set_new_speed(pwm_l_new, pwm_r_new)", "def create_order_con(self, target, price_list):\n\tprint()\n\tprint('OH - pl_create_order_con')\n\tprint(self)\n\tprint(target)\n\tprint(price_list)\n\n\t# Search Partner\n\tprint()\n\tprint('Search partner')\n\tpartner = self.env['res.partner'].search([\n\t\t\t\t\t\t\t\t\t\t\t\t('name', '=', self.patient.name),\n\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\tlimit=1,)\n\tprint(partner)\n\n\t# Search\n\tprint()\n\tprint('Search pricelist')\n\tpricelist = self.env['product.pricelist'].search([\n\t\t\t\t\t\t\t\t\t\t\t#('active', 'in', [True]),\n\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\t#order='x_serial_nr asc',\n\t\t\t\t\t\t\t\t\t\t\tlimit=1,\n\t\t\t\t\t\t\t\t\t\t)\n\tprint(pricelist)\n\n\n\t# Create Order\n\tprint()\n\tprint('Create order')\n\tprint(partner.id)\n\tprint(self.patient.id)\n\tprint(self.patient.x_id_doc)\n\tprint(self.patient.x_id_doc_type)\n\tprint(self.physician.id)\n\torder = self.env['sale.order'].create({\n\t\t\t\t\t\t\t\t\t\t\t'patient': self.patient.id,\n\t\t\t\t\t\t\t\t\t\t\t'x_id_doc': self.patient.x_id_doc,\n\t\t\t\t\t\t\t\t\t\t\t'x_id_doc_type': self.patient.x_id_doc_type,\n\t\t\t\t\t\t\t\t\t\t\t'x_doctor': self.physician.id,\n\t\t\t\t\t\t\t\t\t\t\t'state':'draft',\n\t\t\t\t\t\t\t\t\t\t\t'partner_id': partner.id,\n\t\t\t\t\t\t\t\t\t\t\t'x_family': 'consultation',\n\t\t\t\t\t\t\t\t\t\t\t'treatment': self.id,\n\n\t\t\t\t\t\t\t\t\t\t\t'pricelist_id': pricelist.id,\n\t\t\t\t\t\t\t\t\t\t})\n\tprint(order)\n\n\t# Init\n\t_dic_con = {\n\t\t\t\t\t'medical':\t\t'CONSULTA MEDICA',\n\t\t\t\t\t'gynecology':\t'CONSULTA GINECOLOGICA',\n\t\t\t\t\t'premium':\t\t'CONSULTA MEDICA DR. CHAVARRI',\n\t}\n\tname = _dic_con[target]\n\n\t# Search\n\tprint()\n\tprint('Search product')\n\tproduct = self.env['product.product'].search([\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t('name', 'in', [name]),\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t('pl_price_list', 'in', [price_list]),\n\t\t\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t#order='date_begin asc',\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t#limit=1,\n\t\t\t\t\t\t\t\t\t\t\t\t)\n\tprint(product)\n\tprint(product.name)\n\n\t# Create Order Line\n\tprint()\n\tprint('Create order line')\n\tol = order.order_line.create({\n\t\t\t\t\t\t\t\t\t'name': \t\t\tproduct.name,\n\t\t\t\t\t\t\t\t\t'product_id': \t\tproduct.id,\n\t\t\t\t\t\t\t\t\t'order_id': \t\torder.id,\n\t\t\t\t\t\t\t\t})\n\treturn order", "def test_create_hyperflex_proxy_setting_policy(self):\n pass", "def post_route_target_create(self, resource_dict):\n pass", "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def _create_new_route(self, tick):\n if self.target_node_id is None:\n self.source_node_id = random.choice(Network.nodes).getID()\n else:\n self.source_node_id = self.target_node_id # We start where we stopped\n # random target\n self.target_node_id = random.choice(Network.nodes).getID()\n self.current_route_id = self.id + \"-\" + str(self.rounds)\n self.current_router_result = CustomRouter.minimal_route(self.source_node_id, self.target_node_id)\n \n if len(self.current_router_result.edges) > 0:\n traci.route.add(self.current_route_id, self.current_router_result.edges)\n return self.current_route_id\n else:\n # try again\n return self._create_new_route(tick)", "def onTradeGenRoute(self, item):\n self.frame.mode.addTradeRoute(0, 0, 0, self.mySystemDict['id'],\n self.lstToSystem.getSelectedItem().data, 'GEN')", "def test_create_rate_plan(self):\n pass", "def click_create_new_target_buy_policy_button(self):\n self.click_element(self.create_new_target_buy_policy_button_locator, script_executor=True, error_message='create new target buy policy button locator not found before specified time out')", "def set_target_policy(self, policy):\n self.target_policy = policy", "def __init__(self, target, policy):\n self.target_name = target.lower().strip()\n self.policy = os.path.abspath(policy)\n\n if not os.path.exists(self.policy):\n raise ValueError(f'Cannot find file \"{self.policy}\"')\n\n self.target = self._get_target(self.target_name, policy)\n\n self.memory_map = self.target.memory_map\n self.register_map = self.target.register_map\n self.policy_parser = self.target.policy_parser\n self.policy_validator = self.target.policy_validator\n self.policy_filter = self.target.policy_filter\n\n # Validate policy file\n if not self.policy_validator.validate():\n logger.error('Policy validation failed')\n return", "def __init__(self, nav,\n waypoint=ll.LatLon(50.742810, 1.014469), # somewhere in the solent\n target_radius=2, waypoint_id=None,\n ):\n self.nav = nav\n self.waypoint = waypoint\n self.waypoint_id = waypoint_id\n x, y = self.nav.latlon_to_utm(waypoint.lat.decimal_degree, waypoint.lon.decimal_degree)\n self.waypoint_xy = Point(x, y)\n self.target_area = self.waypoint_xy.buffer(target_radius)", "def __createNewRoute(self, tick):\n # import here because python can not handle circular-dependencies\n if self.targetID is None:\n self.sourceID = random.choice(Network.nodes).getID()\n else:\n self.sourceID = self.targetID # We start where we stopped\n # random target\n self.targetID = random.choice(Network.nodes).getID()\n self.currentRouteID = self.id + \"-\" + str(self.rounds)\n self.currentRouterResult = CustomRouter.route(self.sourceID, self.targetID, tick, self)\n if len(self.currentRouterResult.route) > 0:\n traci.route.add(self.currentRouteID, self.currentRouterResult.route)\n # set color to red\n return self.currentRouteID\n else:\n # recursion aka. try again as this should work!\n return self.__createNewRoute(tick)", "def setup_rule(self, client):\n pass", "def setup_rule(self, client):\n pass", "def __init__(__self__, *,\n accelerator_id: Optional[pulumi.Input[str]] = None,\n address: Optional[pulumi.Input[str]] = None,\n custom_routing_endpoint_traffic_policy_id: Optional[pulumi.Input[str]] = None,\n endpoint_group_id: Optional[pulumi.Input[str]] = None,\n endpoint_id: Optional[pulumi.Input[str]] = None,\n listener_id: Optional[pulumi.Input[str]] = None,\n port_ranges: Optional[pulumi.Input[Sequence[pulumi.Input['CustomRoutingEndpointTrafficPolicyPortRangeArgs']]]] = None,\n status: Optional[pulumi.Input[str]] = None):\n if accelerator_id is not None:\n pulumi.set(__self__, \"accelerator_id\", accelerator_id)\n if address is not None:\n pulumi.set(__self__, \"address\", address)\n if custom_routing_endpoint_traffic_policy_id is not None:\n pulumi.set(__self__, \"custom_routing_endpoint_traffic_policy_id\", custom_routing_endpoint_traffic_policy_id)\n if endpoint_group_id is not None:\n pulumi.set(__self__, \"endpoint_group_id\", endpoint_group_id)\n if endpoint_id is not None:\n pulumi.set(__self__, \"endpoint_id\", endpoint_id)\n if listener_id is not None:\n pulumi.set(__self__, \"listener_id\", listener_id)\n if port_ranges is not None:\n pulumi.set(__self__, \"port_ranges\", port_ranges)\n if status is not None:\n pulumi.set(__self__, \"status\", status)", "def __init__(self, plan):\n self.plan = plan", "def gen_route_target_list(route_target):\n rt = RouteTargetList(route_target=route_target)\n return rt", "def __init__(self, params=None):\n\n rights = ghop_access.GHOPChecker(params)\n rights['show'] = ['allow']\n rights['create'] = [('checkSeeded', ['checkHasRoleForScope',\n host_logic.logic])]\n rights['edit'] = [('checkIsHostForProgram', [ghop_program_logic.logic])]\n rights['delete'] = ['checkIsDeveloper']\n rights['accepted_orgs'] = [('checkIsAfterEvent',\n ['student_signup_start',\n '__all__', ghop_program_logic.logic])]\n rights['list_participants'] = [('checkIsHostForProgram',\n [ghop_program_logic.logic])]\n rights['task_difficulty'] = [('checkIsHostForProgram',\n [ghop_program_logic.logic])]\n rights['task_type'] = [('checkIsHostForProgram',\n [ghop_program_logic.logic])]\n rights['difficulty_tag_edit'] = [('checkIsHostForProgram',\n [ghop_program_logic.logic])]\n rights['type_tag_edit'] = [('checkIsHostForProgram',\n [ghop_program_logic.logic])]\n\n new_params = {}\n new_params['logic'] = soc.modules.ghop.logic.models.program.logic\n new_params['rights'] = rights\n\n new_params['name'] = \"GHOP Program\"\n new_params['module_name'] = \"program\"\n new_params['sidebar_grouping'] = 'Programs'\n new_params['document_prefix'] = 'ghop_program'\n\n new_params['module_package'] = 'soc.modules.ghop.views.models'\n new_params['url_name'] = 'ghop/program'\n\n new_params['extra_dynaexclude'] = ['task_difficulties', 'task_types']\n\n patterns = []\n patterns += [\n (r'^%(url_name)s/(?P<access_type>assign_task_quotas)/%(key_fields)s$',\n '%(module_package)s.%(module_name)s.assign_task_quotas',\n 'Assign task quota limits'),\n (r'^%(url_name)s/(?P<access_type>task_difficulty)/%(key_fields)s$',\n '%(module_package)s.%(module_name)s.task_difficulty_edit',\n 'Edit Task Difficulty Tags'),\n (r'^%(url_name)s/(?P<access_type>task_type)/%(key_fields)s$',\n '%(module_package)s.%(module_name)s.task_type_edit',\n 'Edit Task Type Tags'),\n (r'^%(url_name)s/(?P<access_type>difficulty_tag_edit)$',\n '%(module_package)s.%(module_name)s.difficulty_tag_edit',\n 'Edit a Difficulty Tag'),\n (r'^%(url_name)s/(?P<access_type>type_tag_edit)$',\n '%(module_package)s.%(module_name)s.task_type_tag_edit',\n 'Edit a Task Type Tag'),\n ]\n\n new_params['public_field_keys'] = [\"name\", \"scope_path\"]\n new_params['public_field_names'] = [\"Program Name\", \"Program Owner\"]\n\n new_params['extra_django_patterns'] = patterns\n\n new_params['org_app_logic'] = org_app_logic\n new_params['org_app_prefix'] = 'ghop'\n\n # used to list the participants in this program\n new_params['participants_logic'] = [\n (ghop_org_admin_logic.logic, 'program'),\n (ghop_mentor_logic.logic, 'program'),\n (ghop_student_logic.logic, 'scope')]\n\n params = dicts.merge(params, new_params, sub_merge=True)\n\n super(View, self).__init__(params=params)\n\n dynafields = [\n {'name': 'overview_task_difficulties',\n 'base': forms.CharField,\n 'label': 'Task Difficulty Levels',\n 'group': 'Task Settings',\n 'widget': widgets.ReadOnlyInput(),\n 'required': False,\n 'help_text': ugettext('Lists all the difficulty levels that '\n 'can be assigned to a task. Edit them '\n 'from the Program menu on sidebar.'),\n },\n {'name': 'overview_task_types',\n 'base': forms.CharField,\n 'label': 'Task Type Tags',\n 'group': 'Task Settings',\n 'widget': widgets.ReadOnlyInput(),\n 'required': False,\n 'help_text': ugettext('Lists all the types a task can be in. '\n 'Edit them from the Program menu on sidebar.'),\n },\n ]\n\n dynaproperties = params_helper.getDynaFields(dynafields)\n\n edit_form = dynaform.extendDynaForm(\n dynaform=self._params['edit_form'],\n dynaproperties=dynaproperties)\n\n self._params['edit_form'] = edit_form", "def create_target(self):\n\n # I used a random number variable (rand_target) in order to randomize the target created each time this function\n # is called.\n stand = StandardTarget()\n strong = StrongTarget()\n safe = SafeTarget()\n bird = Bird()\n\n rand_target = random.randint(1, 4)\n if rand_target == 1:\n self.targets.append(stand)\n elif rand_target == 2:\n self.targets.append(strong)\n elif rand_target == 3:\n self.targets.append(safe)\n elif rand_target == 4:\n self.targets.append(bird)" ]
[ "0.57578945", "0.56873965", "0.5348911", "0.5346214", "0.5319583", "0.53184915", "0.53172463", "0.52937037", "0.52722543", "0.5266859", "0.52648765", "0.525317", "0.52372414", "0.52235234", "0.51925033", "0.51838106", "0.5182945", "0.5146547", "0.5137112", "0.51151454", "0.5055275", "0.50413054", "0.50037014", "0.4953614", "0.4953614", "0.4950223", "0.49470842", "0.49273017", "0.4925292", "0.49160904" ]
0.66891956
0
Implementing click include cli checkbox functionality
def click_include_cli_checkbox(self): self.click_element(self.include_cli_checkbox_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_guarantee_cli_checkbox(self):\n self.click_element(self.guarantee_cli_checkbox_locator)", "def click_automate_generation_checkbox(self):\n self.click_element(self.automate_generation_checkbox_locator)", "def CheckBoxClicked(self,chkb):\r\n\r\n print(\"{} Selecionado.\", format(chkb.text()))", "def flag(i, com, xi, yi):\r\n chk = Checkbutton(window1, variable=list_cb[i], command=com)\r\n chk.focus()\r\n chk.place(x=xi, y=yi)\r\n return", "def action_checkbox(self, obj):\n if self.check_concurrent_action:\n return helpers.checkbox.render(helpers.ACTION_CHECKBOX_NAME,\n force_str(\"%s,%s\" % (obj.pk, get_revision_of_object(obj))))\n else: # pragma: no cover\n return super().action_checkbox(obj)", "def htmlCheckbox(labelText, parName, args, labelAttr='', attr=''):\n snippet = htmlLabel(labelText,parName,labelAttr)\n checked = 'checked=\"checked\"' if parName in args else ''\n snippet += '<input type=\"checkbox\" name=\"%s\"%s%s/>\\n' % (parName,sep(checked),sep(attr))\n return snippet", "def on_CheckPunish_clicked(self):\n # TODO: not implemented yet\n raise NotImplementedError", "def action_checkbox(self):\n self.checkbox_online_var = not self.checkbox_online_var", "def on_action_clicked(self, content):\n checked = content['checked']\n self.set_guarded(checked=checked)\n self.clicked(checked)", "def clicked_checkbox_use_antimasks(self):\n # TODO\n self._get_selected_model().metadata[\"antimask_flag\"] \\\n = self.checkbox_use_antimasks.isChecked()\n return None", "def uiCheckboxChecked(checkbox):\n\n return clibui.uiCheckboxChecked(checkbox)", "def on_checkBox_kongtoukai_clicked(self, checked):\n # TODO: not implemented yet\n raise NotImplementedError", "def isSelected(*args):", "def isSelected(*args):", "def _create_boolean_widget(self,frame,name,widget_options):\n # CB: might be necessary to pass actions to command option of Checkbutton;\n # could be cause of test pattern boolean not working?\n return T.Checkbutton(frame,variable=self._tkvars[name],**widget_options)", "def _add_checkbox(self, text, state_changed, tooltip, checked=True,\n enabled=True, button_label=True):\n cbox = QtWidgets.QCheckBox('' if button_label else text, self)\n self.control.layout().addWidget(cbox)\n btn = None\n if button_label:\n btn = QtWidgets.QPushButton(text, self)\n self.control.layout().addWidget(btn)\n\n def cb(checked, cbox=cbox, state_changed=state_changed):\n state_changed(cbox.isChecked(), one_shot=True)\n\n btn.clicked.connect(cb)\n btn.setToolTip(tooltip)\n cbox.setChecked(checked)\n cbox.setEnabled(enabled)\n cbox.stateChanged.connect(state_changed)\n cbox.setToolTip(tooltip)\n self.control.layout().addItem(QtWidgets.QSpacerItem(20, 0))\n return cbox", "def symbolCheckBox(*args, annotation: Union[AnyStr, bool]=\"\", backgroundColor: Union[List[float,\n float, float], bool]=None, changeCommand: Script=None, defineTemplate:\n AnyStr=\"\", disableOffImage: Union[AnyStr, bool]=\"\", disableOnImage:\n Union[AnyStr, bool]=\"\", docTag: Union[AnyStr, bool]=\"\", dragCallback:\n Script=None, dropCallback: Script=None, enable: bool=True, enableBackground:\n bool=True, enableKeyboardFocus: bool=True, exists: bool=True, fullPathName:\n bool=True, height: Union[int, bool]=0, highlightColor: Union[List[float,\n float, float], bool]=None, image: Union[AnyStr, bool]=\"\", innerMargin:\n bool=True, isObscured: bool=True, ltVersion: Union[AnyStr, bool]=\"\", manage:\n bool=True, noBackground: bool=True, numberOfPopupMenus: bool=True,\n offCommand: Script=None, offImage: Union[AnyStr, bool]=\"\", onCommand:\n Script=None, onImage: Union[AnyStr, bool]=\"\", parent: Union[AnyStr, bool]=\"\",\n popupMenuArray: bool=True, preventOverride: bool=True, statusBarMessage:\n AnyStr=\"\", useTemplate: AnyStr=\"\", value: bool=True, version: Union[AnyStr,\n bool]=\"\", visible: bool=True, visibleChangeCommand: Union[Script, bool]=None,\n width: Union[int, bool]=0, q=True, query=True, e=True, edit=True,\n **kwargs)->Union[AnyStr, Any]:\n pass", "def checkBox(*args, align: Union[AnyStr, bool]=\"\", annotation: Union[AnyStr, bool]=\"\",\n backgroundColor: Union[List[float, float, float], bool]=None, changeCommand:\n Union[Script, bool]=None, defineTemplate: AnyStr=\"\", docTag: Union[AnyStr,\n bool]=\"\", dragCallback: Script=None, dropCallback: Script=None, editable:\n bool=True, enable: bool=True, enableBackground: bool=True, enableKeyboardFocus:\n bool=True, exists: bool=True, fullPathName: bool=True, height: Union[int, bool]=0,\n highlightColor: Union[List[float, float, float], bool]=None, isObscured: bool=True,\n label: Union[AnyStr, bool]=\"\", manage: bool=True, noBackground: bool=True,\n numberOfPopupMenus: bool=True, offCommand: Union[Script, bool]=None, onCommand:\n Union[Script, bool]=None, parent: Union[AnyStr, bool]=\"\", popupMenuArray:\n bool=True, preventOverride: bool=True, recomputeSize: bool=True, statusBarMessage:\n AnyStr=\"\", useTemplate: AnyStr=\"\", value: bool=True, visible: bool=True,\n visibleChangeCommand: Union[Script, bool]=None, width: Union[int, bool]=0, q=True,\n query=True, e=True, edit=True, **kwargs)->Union[AnyStr, Any]:\n pass", "def on_action_toggled(self, content):\n checked = content['checked']\n self.set_guarded(checked=checked)\n self.toggled(checked)", "def flagEnable(self, item):\n if item.checkState() == Qt.Checked:\n logging.debug('\"%s\" Checked' % item.text())\n self.flags[item.text()].enable()\n elif item.checkState() == Qt.Unchecked:\n logging.debug('\"%s\" Unchecked' % item.text())\n self.flags[item.text()].disable()\n else:\n logging.debug('\"%s\" Clicked' % item.text())", "def _clicked_yes_button(self):\n self.yes = True", "def recursive_clicked(self, checkbox):\n\n\t\tlogging.progress('recursive_clicked() called.')\n\n\t\tself.update_keyword_usage()\n\t\tself.update_notebook()", "def comb_box_click(self):\n\n if self.checkBox.isChecked():\n self.completer.setFilterMode(Qt.MatchStartsWith)\n else:\n self.completer.setFilterMode(Qt.MatchContains)", "def on_checkBox_duotoukai_clicked(self, checked):\n # TODO: not implemented yet\n raise NotImplementedError", "def __reference_frame_checkbox(self, c):\n if len(self.__robots) > 0:\n self.__robots[self.__selected_robot].set_reference_visibility(\n c.checked)", "def click_start_automatically_ssh_checkbox_and_enable_the_ssh_service(driver):\n assert wait_on_element(driver, 5, xpaths.services.title)\n time.sleep(1)\n assert wait_on_element(driver, 5, '//tr[contains(.,\"SSH\")]//mat-checkbox')\n value_exist = attribute_value_exist(driver, '//tr[contains(.,\"SSH\")]//mat-checkbox', 'class', 'mat-checkbox-checked')\n if not value_exist:\n driver.find_element_by_xpath('//tr[contains(.,\"SSH\")]//mat-checkbox').click()\n assert wait_on_element(driver, 5, '//tr[contains(.,\"SSH\")]//mat-slide-toggle/label', 'clickable')\n value_exist = attribute_value_exist(driver, xpaths.services.ssh_Service_Toggle, 'class', 'mat-checked')\n if not value_exist:\n driver.find_element_by_xpath('//tr[contains(.,\"SSH\")]//mat-slide-toggle/label').click()\n time.sleep(1)", "def check_box(self, grid: object, name: str, xposition: int, yposition: int,\n synchronize: bool = False, xspan: int = 1, yspan: int = 1) -> QtWidgets.QCheckBox:\n label = QtWidgets.QLabel()\n label.setText(TR().tr(name) + ':')\n grid.addWidget(label, yposition, xposition, 1, 1)\n\n input = QtWidgets.QCheckBox()\n input.setObjectName(name)\n if synchronize:\n self.synchronize(input)\n grid.addWidget(input, yposition, xposition + 1, yspan, xspan)\n input.stateChanged.connect(self.data_changed)\n\n return input", "def checkbox(self):\r\n return self._checkbox", "def sd_yes_clicked(self, widget, data=None):\n return True", "def segmentNeedle(self):\n #productive #event\n profprint()\n if self.fiducialButton.isEnabled():\n print \"new checked state: \",not self.fiducialButton.checked\n self.onStartStopGivingNeedleTipsToggled(not self.fiducialButton.checked)" ]
[ "0.7180335", "0.633742", "0.6186127", "0.61447966", "0.60904425", "0.6061627", "0.6051627", "0.597381", "0.59416974", "0.59316313", "0.5918221", "0.58853", "0.58538675", "0.58538675", "0.5849901", "0.5836241", "0.58296305", "0.58069646", "0.5802643", "0.57914114", "0.57344997", "0.5730153", "0.57152385", "0.5706675", "0.5677495", "0.5647356", "0.5595114", "0.55800736", "0.5568433", "0.5563624" ]
0.82027
0
Implementing click automate generation checkbox functionality
def click_automate_generation_checkbox(self): self.click_element(self.automate_generation_checkbox_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_guarantee_cli_checkbox(self):\n self.click_element(self.guarantee_cli_checkbox_locator)", "def on_checkBox_kongtoukai_clicked(self, checked):\n # TODO: not implemented yet\n raise NotImplementedError", "def _add_checkbox(self, text, state_changed, tooltip, checked=True,\n enabled=True, button_label=True):\n cbox = QtWidgets.QCheckBox('' if button_label else text, self)\n self.control.layout().addWidget(cbox)\n btn = None\n if button_label:\n btn = QtWidgets.QPushButton(text, self)\n self.control.layout().addWidget(btn)\n\n def cb(checked, cbox=cbox, state_changed=state_changed):\n state_changed(cbox.isChecked(), one_shot=True)\n\n btn.clicked.connect(cb)\n btn.setToolTip(tooltip)\n cbox.setChecked(checked)\n cbox.setEnabled(enabled)\n cbox.stateChanged.connect(state_changed)\n cbox.setToolTip(tooltip)\n self.control.layout().addItem(QtWidgets.QSpacerItem(20, 0))\n return cbox", "def click_include_cli_checkbox(self):\n self.click_element(self.include_cli_checkbox_locator)", "def create_checkboxes(self):\n self.create_y_crop_box()", "def htmlCheckbox(labelText, parName, args, labelAttr='', attr=''):\n snippet = htmlLabel(labelText,parName,labelAttr)\n checked = 'checked=\"checked\"' if parName in args else ''\n snippet += '<input type=\"checkbox\" name=\"%s\"%s%s/>\\n' % (parName,sep(checked),sep(attr))\n return snippet", "def on_checkBox_kongtouping_clicked(self, checked):\n # TODO: not implemented yet\n raise NotImplementedError", "def _create_boolean_widget(self,frame,name,widget_options):\n # CB: might be necessary to pass actions to command option of Checkbutton;\n # could be cause of test pattern boolean not working?\n return T.Checkbutton(frame,variable=self._tkvars[name],**widget_options)", "def on_checkBox_duotoukai_clicked(self, checked):\n # TODO: not implemented yet\n raise NotImplementedError", "def flag(i, com, xi, yi):\r\n chk = Checkbutton(window1, variable=list_cb[i], command=com)\r\n chk.focus()\r\n chk.place(x=xi, y=yi)\r\n return", "def CheckBoxClicked(self,chkb):\r\n\r\n print(\"{} Selecionado.\", format(chkb.text()))", "def action_checkbox(self, obj):\n if self.check_concurrent_action:\n return helpers.checkbox.render(helpers.ACTION_CHECKBOX_NAME,\n force_str(\"%s,%s\" % (obj.pk, get_revision_of_object(obj))))\n else: # pragma: no cover\n return super().action_checkbox(obj)", "def recursive_clicked(self, checkbox):\n\n\t\tlogging.progress('recursive_clicked() called.')\n\n\t\tself.update_keyword_usage()\n\t\tself.update_notebook()", "def _generateCellCheckedState(self, obj, **args):\n result = []\n if self._script.utilities.hasMeaningfulToggleAction(obj):\n oldRole = self._overrideRole(pyatspi.ROLE_CHECK_BOX, args)\n result.extend(self.generate(obj, **args))\n self._restoreRole(oldRole, args)\n\n return result", "def on_checkBox_duotouping_clicked(self, checked):\n # TODO: not implemented yet\n raise NotImplementedError", "def on_CheckPunish_clicked(self):\n # TODO: not implemented yet\n raise NotImplementedError", "def test_checkboxtextgroup(self):\r\n self.check_group('checkboxtextgroup', 'choice', 'checkbox')", "def DrawCheckBox(*args, **kwargs):\n return _gdi_.RendererNative_DrawCheckBox(*args, **kwargs)", "def symbolCheckBox(*args, annotation: Union[AnyStr, bool]=\"\", backgroundColor: Union[List[float,\n float, float], bool]=None, changeCommand: Script=None, defineTemplate:\n AnyStr=\"\", disableOffImage: Union[AnyStr, bool]=\"\", disableOnImage:\n Union[AnyStr, bool]=\"\", docTag: Union[AnyStr, bool]=\"\", dragCallback:\n Script=None, dropCallback: Script=None, enable: bool=True, enableBackground:\n bool=True, enableKeyboardFocus: bool=True, exists: bool=True, fullPathName:\n bool=True, height: Union[int, bool]=0, highlightColor: Union[List[float,\n float, float], bool]=None, image: Union[AnyStr, bool]=\"\", innerMargin:\n bool=True, isObscured: bool=True, ltVersion: Union[AnyStr, bool]=\"\", manage:\n bool=True, noBackground: bool=True, numberOfPopupMenus: bool=True,\n offCommand: Script=None, offImage: Union[AnyStr, bool]=\"\", onCommand:\n Script=None, onImage: Union[AnyStr, bool]=\"\", parent: Union[AnyStr, bool]=\"\",\n popupMenuArray: bool=True, preventOverride: bool=True, statusBarMessage:\n AnyStr=\"\", useTemplate: AnyStr=\"\", value: bool=True, version: Union[AnyStr,\n bool]=\"\", visible: bool=True, visibleChangeCommand: Union[Script, bool]=None,\n width: Union[int, bool]=0, q=True, query=True, e=True, edit=True,\n **kwargs)->Union[AnyStr, Any]:\n pass", "def check_box(self, grid: object, name: str, xposition: int, yposition: int,\n synchronize: bool = False, xspan: int = 1, yspan: int = 1) -> QtWidgets.QCheckBox:\n label = QtWidgets.QLabel()\n label.setText(TR().tr(name) + ':')\n grid.addWidget(label, yposition, xposition, 1, 1)\n\n input = QtWidgets.QCheckBox()\n input.setObjectName(name)\n if synchronize:\n self.synchronize(input)\n grid.addWidget(input, yposition, xposition + 1, yspan, xspan)\n input.stateChanged.connect(self.data_changed)\n\n return input", "def action_checkbox(self):\n self.checkbox_online_var = not self.checkbox_online_var", "def uiCheckboxChecked(checkbox):\n\n return clibui.uiCheckboxChecked(checkbox)", "def _create_checkboxes(self) -> widgets.VBox:\n checkboxes = []\n pgons_checkboxes = []\n graph_checkboxes = []\n\n graphs = [\n (name, \"graphs\", layer_subtype, graph)\n for name, graph in self.viewer.layer_dict[\"graphs\"].items()\n for layer_subtype in [\"graph\", \"pgons\"]\n ]\n maps = [\n (name, \"maps\", \"map\", map_layer[\"map\"])\n for name, map_layer in self.viewer.layer_dict[\"maps\"].items()\n ]\n\n # Add checkboxes for all maps and graphs (including habitats)\n for idx, (layer_name, layer_type, layer_subtype, layer_dict) in enumerate(\n maps + graphs\n ):\n\n layout = widgets.Layout(padding=\"0px 0px 0px 0px\")\n\n # Indent habitat checkboxes\n if layer_type == \"graphs\":\n if layer_dict[\"is_habitat\"]:\n layout = widgets.Layout(padding=\"0px 0px 0px 25px\")\n\n checkbox = widgets.Checkbox(\n value=True,\n description=\"{} ({})\".format(layer_name, layer_subtype),\n disabled=False,\n indent=False,\n layout=layout,\n )\n checkbox.add_traits(\n layer_type=traitlets.Unicode().tag(sync=True),\n layer_subtype=traitlets.Unicode().tag(sync=True),\n layer_name=traitlets.Unicode().tag(sync=True),\n )\n checkbox.layer_type = layer_type\n checkbox.layer_name = layer_name\n checkbox.layer_subtype = layer_subtype\n\n checkbox.observe(self._switch_layer_visibility)\n\n if idx == 0:\n checkboxes.append(widgets.HTML(\"<b>Map Data</b>\"))\n\n checkboxes.append(checkbox)\n\n if layer_subtype == \"graph\":\n graph_checkboxes.append(checkbox)\n elif layer_subtype == \"pgons\":\n pgons_checkboxes.append(checkbox)\n\n # Add habitats header if last part of main graph\n if (\n layer_type == \"graphs\"\n and layer_subtype == \"pgons\"\n and not layer_dict[\"is_habitat\"]\n ):\n checkboxes.append(\n widgets.HTML(\n \"<b>Habitats in {}</b>\".format(layer_name),\n layout=widgets.Layout(padding=\"0px 0px 0px 25px\"),\n )\n )\n\n # Add horizontal rule if last map to separate from graphs\n if idx == len(maps) - 1:\n checkboxes.append(widgets.HTML(\"<hr/>\"))\n checkboxes.append(widgets.HTML(\"<b>Graph Data</b>\"))\n\n # Create button to toggle all polygons at once\n hide_pgon_button = widgets.ToggleButton(description=\"Toggle all polygons\")\n\n def toggle_all_pgons(change):\n try:\n if change[\"name\"] == \"value\":\n for box in pgons_checkboxes:\n box.value = change[\"new\"]\n except: # pylint: disable=bare-except\n self.logger.exception(\"Exception in view button callback on click.\")\n\n hide_pgon_button.observe(toggle_all_pgons)\n\n # Create button to toggle all graphs at once\n hide_graph_button = widgets.ToggleButton(description=\"Toggle all graphs\")\n\n def toggle_all_graphs(change):\n try:\n if change[\"name\"] == \"value\":\n for box in graph_checkboxes:\n box.value = change[\"new\"]\n except: # pylint: disable=bare-except\n self.logger.exception(\"Exception in view button callback on click.\")\n\n hide_graph_button.observe(toggle_all_graphs)\n\n checkboxes.append(widgets.HTML(\"<hr/>\"))\n buttons = widgets.HBox([hide_pgon_button, hide_graph_button])\n checkboxes.append(buttons)\n\n return widgets.VBox(checkboxes)", "def add_CheckBox(self, checkbox_rating):\n for item in checkbox_rating:\n checkbox = MyCheckbox(item)\n self.layout.add_widget(checkbox)\n self.list_checkboxes.append(checkbox)", "def GridCheck(Parent,DefaultSelected,Row,Column):\r\n dummyvar = IntVar()\r\n C = Checkbutton(Parent,var=dummyvar)\r\n if DefaultSelected == 1:\r\n C.select()\r\n C.grid(row=Row,column=Column)\r\n C.isChecked = dummyvar\r\n return C", "def add_check_box(self, name, caption, value=False, label=None, add_indicator=None, location=(None,0)):\n widget=QtWidgets.QCheckBox(self)\n widget.setText(_translate(self.name,caption,None))\n widget.setObjectName(_fromUtf8(self.name+\"_\"+name))\n widget.setChecked(value)\n return self.add_simple_widget(name,widget,label=label,add_indicator=add_indicator,location=location)", "def on_action_clicked(self, content):\n checked = content['checked']\n self.set_guarded(checked=checked)\n self.clicked(checked)", "def clicked_checkbox_use_antimasks(self):\n # TODO\n self._get_selected_model().metadata[\"antimask_flag\"] \\\n = self.checkbox_use_antimasks.isChecked()\n return None", "def checkbox(self, label, initial=False, handler=None, **kwargs):\n handler = self._changed_handler(handler)\n cb = wx.CheckBox(self, label=label)\n #cb.span = 2\n cb.SetValue(initial)\n cb.Bind(wx.EVT_CHECKBOX, handler)\n self.pack(\"\", cb, **kwargs)\n return cb", "def add_checkbox(Frame, row, label, gained_gold, long_gold=None):\n if (long_gold != None):\n right_label = \"{:5.2f} ({:6.2f})\".format(gained_gold, long_gold)\n\n else:\n right_label = \"{:6.2f}\".format(gained_gold)\n\n num_letters = 35 - len(label) - len(right_label)\n label += \" \" * num_letters + right_label\n \n checkbox = tk.Checkbutton(Frame, text=label, bg='WHITE')\n checkbox.grid(row=row, sticky=tk.W)" ]
[ "0.67904496", "0.6535885", "0.65002245", "0.6481565", "0.6431148", "0.63907903", "0.63848615", "0.6366873", "0.6341021", "0.63073826", "0.63037425", "0.62903464", "0.62731963", "0.620092", "0.6196953", "0.6037659", "0.6002369", "0.591591", "0.5890406", "0.58776736", "0.58606", "0.5860545", "0.58600825", "0.5851673", "0.579197", "0.5788369", "0.5783467", "0.5775771", "0.57538855", "0.5751139" ]
0.7924865
0
Implementing click save target buy policy button functionality
def click_save_target_buy_policy_button(self): self.click_element(self.save_target_buy_policy_button_locator, error_message='save target buy policy button locator not found before specified time out') self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out') self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out') self.wait_for_ajax_spinner_load()
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_save_edited_target_buy_policy_button(self):\n self.click_element(self.save_edited_target_buy_policy_button_locator)\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def click_buy_and_sell_deal_save_button(self):\n self.click_element(self.save_vendor_profile_locator, True)\n try:\n self.wait().until(EC.visibility_of_element_located(self.confirmation_popup_locator), 'confirmation popup locator not found before specified time out')\n self.click_element(self.ok_button_locator, True)\n except:\n raise", "def click_cancel_edited_target_buy_policy_button(self):\n self.click_element(self.cancel_edited_target_buy_policy_button_locator)", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def click_create_new_target_buy_policy_button(self):\n self.click_element(self.create_new_target_buy_policy_button_locator, script_executor=True, error_message='create new target buy policy button locator not found before specified time out')", "def saveMenuHandler(self, action):\n button_text = action.text()\n\n if button_text == 'Save As':\n subsystem_controller = self.getCurrentSubsystemController()\n self.saveAsHandler(subsystem_controller)\n\n elif button_text == 'Save':\n subsystem_controller = self.getCurrentSubsystemController()\n self.saveHandler(subsystem_controller)\n\n elif button_text == 'Save As Scenario':\n self.saveScenarioHandler()", "def save_edit(self):\r\n self.driver.find_element_by_xpath('//*[@id=\"vnos\"]/span[1]/div/a').click()", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def add_option_save(self):\n logger.debug(\"Adding save option\")\n btnsave = ttk.Button(self.optsframe,\n image=get_images().icons[\"save\"],\n command=self.save_items)\n btnsave.pack(padx=2, side=tk.RIGHT)\n Tooltip(btnsave,\n text=\"Save {}(s) to file\".format(self.tabname),\n wraplength=200)", "def click_assign_adjustment_charge_save_button(self):\n self.click_element(self.assign_adjustment_charge_save_button_locator)\n try:\n self.wait().until(EC.visibility_of_element_located(self.error_log_viewer_locator), \"error log viewer locator not found before specified time out\")\n self.click_element(self.error_log_viewer_close_button_locator)\n self.click_element(self.assign_adjustment_charge_save_button_locator)\n except:\n pass\n try:\n self.wait().until(EC.visibility_of_element_located(self.success_message_locator), 'success message locator not found before specified time out')\n self.click_element(self.ok_button_locator)\n except:\n raise", "def click_the_save_button_which_should_be_returned_to_the_storage_page(driver):\n assert wait_on_element(driver, 5, '//button[contains(.,\"Save Access Control List\")]', 'clickable')\n driver.find_element_by_xpath('//button[contains(.,\"Save Access Control List\")]').click()\n time.sleep(1)\n assert wait_on_element_disappear(driver, 30, '//h6[contains(.,\"Please wait\")]')", "def tool_save_target_info_clicked(self, widget, data=None):\n self._save_target_info()", "def save(self, *args, **kwargs):\n wallet = self.wallet.withdraw(self.value)\n super(Payment, self).save(*args, **kwargs)", "def click_save_vendor_profile_button(self):\n save_vendor_profile_button_element = self.wait().until(EC.element_to_be_clickable(self.save_vendor_profile_button_locator), \"save vendor profile button locator not found before spcified time out\")\n save_vendor_profile_button_element.click()\n self.wait_for_ajax_spinner_load()\n try:\n self.wait().until(EC.presence_of_element_located(self.confirmation_popup_locator), 'confirmation popup locator not found before specified time')\n ok_button_element = self.wait().until(EC.element_to_be_clickable(self.ok_button_locator), 'ok button locator not found before specified time')\n ok_button_element.click()\n except:\n raise\n self.wait_for_ajax_spinner_load()", "def save(self):\n self.click(\".action-save\")\n self.page.wait_for_ajax()", "def on_save_button_clicked(self, widget):\n active_page = self.get_active_tab()\n active_page.save()", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def on_press_save(self):\n\n if self.dbChk.isChecked():\n self.processed_fields['db'] = self.dbPath.text()\n self.dbForm = DBFormWindow(self.processed_fields, self)\n self.dbForm.show()\n\n if self.savePathChk.isChecked():\n if self.savePath.text():\n shutil.copy(self.tempFile[1], os.path.join(self.savePath.text(), self.saveName.text()+'.wav'))\n os.close(self.tempFile[0])\n os.remove(self.tempFile[1])\n\n QMessageBox.information(self, 'Saved', f'Saved to: {os.path.join(self.savePath.text(), self.saveName.text()+\".wav\")}')\n\n self.saveBtn.setEnabled(False)\n self.deleteBtn.setEnabled(False)\n self.recordBtn.setEnabled(True)\n self.recordBtn.setIcon(QIcon(r'.\\assets\\record.png'))\n self.recordShortCut.setEnabled(True)\n self.inputDropDown.setEnabled(True)\n self.canvasStack.setCurrentWidget(self.microphoneCanvas)\n self.state = 'Waiting'", "def click_create_adjustment_charge_save_button(self):\n self.click_element(self.create_adjustment_charge_save_button_locator)\n try:\n self.wait().until(EC.visibility_of_element_located(self.success_message_locator), 'success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, True)\n except:\n raise\n self.wait_for_ajax_spinner_load()", "def saveButtonMethod(self):\n return AbstractAction(\"Save\")", "def on_ConfirmWalletOP_clicked(self):\n # TODO: not implemented yet\n raise NotImplementedError", "def click_download_button(self):\n self._basket.click_download_button()", "def _save(self):\n\n # Get option\n if self.button1.isChecked():\n option = 'Steunpunt'\n uncertainties = self.supportloc_unc_table\n elif self.button2.isChecked():\n # Check if the harboruncertainties are filled:\n if pd.isnull(self.harbor_unc_table.model._data).any().any():\n raise ValueError('Niet alle modelonzekerheden voor het havenmodel zijn ingevuld.')\n option = 'Havenmodel'\n uncertainties = self.harbor_unc_table\n elif self.button3.isChecked():\n if pd.isnull(self.combined_unc_table.model._data).any().any():\n raise ValueError('De gecombineerde modelonzekerheden zijn nog niet berekend.')\n option = 'Combinatie'\n uncertainties = self.combined_unc_table\n else:\n raise ValueError('Selecteer een optie voor de te gebruiken onzekerheid')\n\n self.parent.adjust_selection(uncertainties=uncertainties, option=option)\n\n # Toegevoegd Svasek 31/10/2018 - Sluit het onzekerheden input scherm als er op opslaan gedrukt wordt\n self.close()", "def awaiting_payment(self):", "def save(self, *args, **kwargs):\n\n # Call the \"real\" save() method.\n super(Product, self).save(*args, **kwargs)\n\n # If sold and is a Widget\n if self.sale and self.kind == self.WIDGET:\n # But has not stamp\n try:\n self.stamp\n except:\n s = Stamp(owned_by=self.sale.customer, obtained_with=self)\n s.save()", "def saveAsButtonMethod(self):\n return AbstractAction(\"Save As\")", "def onSaveNotesButton(self, button):\n pass", "def pricing_save(request, simulation):\n # Retrieve the formset from the POST data.\n formset = PolicyFormSet(request.POST)\n if formset.is_valid():\n # Save the formset (updated values and newly created objects).\n formset.save()\n simulation.has_changed = True\n simulation.save()\n else:\n # Redirect to a page with the errors.\n context = {\n 'simulation': simulation,\n 'demandsegment': demandsegment,\n 'form': formset,\n }\n return render(request, 'metro_app/errors.html', context)\n\n return HttpResponseRedirect(reverse(\n 'metro:pricing_edit', args=(simulation.id, demandsegment.id,)\n ))" ]
[ "0.7729644", "0.72661537", "0.6927698", "0.69045794", "0.6451391", "0.6321213", "0.6035302", "0.6026688", "0.5929195", "0.5800937", "0.5781004", "0.572054", "0.56928444", "0.56849253", "0.56754345", "0.5626255", "0.56152767", "0.5538557", "0.552185", "0.55135155", "0.5491271", "0.54869205", "0.5468869", "0.54612464", "0.54569143", "0.5446169", "0.54379123", "0.5428911", "0.54260796", "0.5381104" ]
0.7788845
0
Implementing is target buy list overrides screen loaded functionality
def is_target_buy_list_overrides_screen_loaded(self): return self.is_element_visible(self.target_buy_list_overrides_page_header_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def listctrl_switch_loadout(self, source):\n if source == self.active_loadout:\n return False\n\n self.gear_slots[self.active_loadout.name] = self.create_blrevive_weapon()\n\n if source.name in self.gear_slots:\n self.load_blrevive_weapon(self.gear_slots[source.name])\n\n else:\n self.clear_equipped_weapon()\n\n self.update_main_preview_image(wx.NullBitmap)\n self.active_loadout = source\n self.export_current_loadouts()\n return True", "def on_use(self):\n assert self.can_use, 'Used an unuseable item!'", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def activated(self):", "def on_activate(self):", "def will_activate(self):\n pass", "def targeted(self):\n\t\tpass", "def on_activate(self) -> None:", "def on_load(self):", "def _test_display_down_button(self):\n return (self.product_displays.top_index + self.limits.screen_products) < len(self.product_displays)", "def available(self, wl, *args):\n return True", "def is_target_buy_policies_tab_displayed(self):\n return self.is_element_visible(self.target_buy_policies_tab_locator)", "def is_expected_dropdown_list_control_visible_in_target_buy_overrides_screen(self, expected_list_control_name):\n list_control_locator = (By.XPATH, \"//select[@name='%s']\" %(expected_list_control_name))\n return self.is_element_present(list_control_locator)", "def _test_display_up_button(self):\n return (self.product_displays.top_index > 0)", "def is_target_buy_details_section_present_in_vendor_profile_page(self):\n return self.is_element_present(self.target_buy_details_section_locator)", "def onDrinkCreated(self, event):", "def on_resultList_itemSelectionChanged(self):\n self.__installButton.setEnabled(\n len(self.resultList.selectedItems()) > 0)\n self.__showDetailsButton.setEnabled(\n len(self.resultList.selectedItems()) == 1)", "def test_nres_targets_list(self):\n response = self.client.get(reverse('nres_calibrations:nres_home'))\n self.assertContains(response, self.target.id)", "def _after_plot_list_load(self):\r\n self.button_send_list.enable()\r\n self.label_list_status.set_text('Shown list does not match device list.')", "def pick_up(self):", "def __load(self, *args, **kwargs):\n self.__should_load = True\n return Menu.CONTINUE", "def setup_class(cls):\n super().setup_class()\n cls.add_item(\"skill\", str(cls.GENERIC_SELLER.public_id), local=False)", "def on_load(self):\n pass", "def on_load(self):\n pass", "def onShow(self):\n pass", "def is_available(self, product_url):\n\t\tpass", "def __handle_view_item(self, gamestate_component):", "def battle_screen_stage_3_action(position, screen,buttons, screen_status, button_status, card_database_filter, user):\n button_status.battle_screen_instruction_bar_text = \"Pick a target to attack with monster: \" + position\n button_status.battle_screen_instruction_bar_yes_display = False\n button_status.battle_screen_instruction_bar_yes_backend = False", "def onToSystemSelected(self, item):\n if not item:\n self.btnAddTradeRoute.disable()\n self.btnTradeAllRoute.disable()\n self.btnAddOneTimeTrade.disable()\n else:\n self.btnAddTradeRoute.enable()\n self.btnTradeAllRoute.enable()\n self.btnAddOneTimeTrade.enable()", "def init_shopping(self):\n all_items = self.db.get_items()\n scroll_parent = Window\n sw = self.ids.shopping_wrapper\n for t in all_items:\n item = ItemToBuy()\n state = self.string_to_bool(t[1])\n if t[1] == \"True\":\n item.bcg_clr = [0.5, 0.5, 0.5, 0.5]\n item.is_done.active = 1\n\n item.name = t[2]\n item.size_hint = [None, None]\n item.size = [scroll_parent.width / 1.35, dp(65)]\n\n sw.add_widget(item)" ]
[ "0.6001915", "0.5983745", "0.58766633", "0.5687606", "0.567117", "0.56279045", "0.5524678", "0.5523005", "0.54764086", "0.542335", "0.5407171", "0.53916734", "0.5346383", "0.531239", "0.5288447", "0.52460074", "0.524296", "0.523115", "0.5228215", "0.5227995", "0.5221563", "0.5202568", "0.5196564", "0.5196564", "0.51919705", "0.5176313", "0.5146909", "0.5138636", "0.5136021", "0.51318467" ]
0.72665113
0
Implementing is expected dropdown list control visible functionality
def is_expected_dropdown_list_control_visible_in_target_buy_overrides_screen(self, expected_list_control_name): list_control_locator = (By.XPATH, "//select[@name='%s']" %(expected_list_control_name)) return self.is_element_present(list_control_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def is_view_dropdown_visible(self):\n return self.is_element_visible(self.view_dropdown_locator)", "def is_dropdown_expanded(self):\n return self.dropdown.element(class_name=\"dropdown-focus\").exists", "def is_visible(self):", "def is_discrepancy_dropdown_visible(self):\n return self.is_element_visible(self.discrepancy_dropdown_locator)", "def is_status_dropdown_visible(self):\n return self.is_element_visible(self.status_dropdown_locator)", "def visible(self, show):", "def HasDropDown(self):\r\n\r\n return self.dropdown", "def dropdown(self):\n return self._root.element(class_name=\"multiselect-dropdown\")", "def IsDropDownClicked(self):\r\n\r\n return self.is_dropdown_clicked", "def test_options_menu_autoopen(self):\n self.login()\n page = Page.objects.create(owner=self.user)\n path = page.get_absolute_url()\n self.selenium.get('%s%s' % (self.live_server_url, path))\n dropdown_menu_btn = self._get_one('.page-options-dropdown .dropdown-toggle')\n show_opts_btn = self._get_one('.options-btn')\n options_menu = self._get_one('.pageOptions')\n self.assertTrue(dropdown_menu_btn.is_displayed())\n self.assertFalse(show_opts_btn.is_displayed())\n self.assertTrue(options_menu.is_displayed())", "def SetHasDropDown(self, b):\r\n\r\n self.dropdown = b", "def is_in_routing_flags_dropdown_visible(self):\n return self.is_element_visible(self.in_routing_flags_dropdown_locator)", "def dropdown(self):\n # defaults = DDConfig(0.6, 0.6, 0, 0)\n return DropDown(self.name, self.command, **self.config)", "def expand_dropdown(self):\n if not self.is_dropdown_expanded:\n self.dropdown.click()", "def isVisible(self):\n\t\treturn True", "def ShowWindowMenu(self):\r\n \r\n tabCtrl = self.GetActiveTabCtrl()\r\n idx = tabCtrl.GetArtProvider().ShowDropDown(tabCtrl, tabCtrl.GetPages(), tabCtrl.GetActivePage())\r\n\r\n if not self.GetEnabled(idx):\r\n return False\r\n\r\n if idx != -1:\r\n e = AuiNotebookEvent(wxEVT_COMMAND_AUINOTEBOOK_PAGE_CHANGING, tabCtrl.GetId())\r\n e.SetSelection(idx)\r\n e.SetOldSelection(tabCtrl.GetActivePage())\r\n e.SetEventObject(tabCtrl)\r\n self.GetEventHandler().ProcessEvent(e)\r\n\r\n return True\r\n \r\n else:\r\n \r\n return False", "def selectable(cls):\n return True", "def ShowDropDown(self, wnd, pages, active_idx):\r\n \r\n useImages = self.GetAGWFlags() & AUI_NB_USE_IMAGES_DROPDOWN\r\n menuPopup = wx.Menu()\r\n\r\n longest = 0\r\n for i, page in enumerate(pages):\r\n \r\n caption = page.caption\r\n\r\n # if there is no caption, make it a space. This will prevent\r\n # an assert in the menu code.\r\n if caption == \"\":\r\n caption = \" \"\r\n\r\n # Save longest caption width for calculating menu width with\r\n width = wnd.GetTextExtent(caption)[0]\r\n if width > longest:\r\n longest = width\r\n\r\n if useImages:\r\n menuItem = wx.MenuItem(menuPopup, 1000+i, caption)\r\n if page.bitmap:\r\n menuItem.SetBitmap(page.bitmap)\r\n\r\n menuPopup.AppendItem(menuItem)\r\n \r\n else:\r\n \r\n menuPopup.AppendCheckItem(1000+i, caption)\r\n \r\n menuPopup.Enable(1000+i, page.enabled)\r\n\r\n if active_idx != -1 and not useImages:\r\n \r\n menuPopup.Check(1000+active_idx, True)\r\n \r\n # find out the screen coordinate at the bottom of the tab ctrl\r\n cli_rect = wnd.GetClientRect()\r\n\r\n # Calculate the approximate size of the popupmenu for setting the\r\n # position of the menu when its shown.\r\n # Account for extra padding on left/right of text on mac menus\r\n if wx.Platform in ['__WXMAC__', '__WXMSW__']:\r\n longest += 32\r\n\r\n # Bitmap/Checkmark width + padding\r\n longest += 20\r\n\r\n if self.GetAGWFlags() & AUI_NB_CLOSE_BUTTON:\r\n longest += 16\r\n\r\n pt = wx.Point(cli_rect.x + cli_rect.GetWidth() - longest,\r\n cli_rect.y + cli_rect.height)\r\n\r\n cc = AuiCommandCapture()\r\n wnd.PushEventHandler(cc)\r\n wnd.PopupMenu(menuPopup, pt)\r\n command = cc.GetCommandId()\r\n wnd.PopEventHandler(True)\r\n\r\n if command >= 1000:\r\n return command - 1000\r\n\r\n return -1", "def IsShown(self):\r\n \r\n return not self.HasFlag(self.optionHidden)", "def test_list_options(self):\n pass", "def setVisible(*args):", "def setVisible(*args):", "def setVisible(*args):", "def setVisible(*args):", "def setVisible(*args):", "def setVisible(*args):", "def setVisible(*args):", "def setVisible(*args):", "def setVisible(*args):", "def setVisible(*args):" ]
[ "0.6721152", "0.62846535", "0.6274534", "0.62375015", "0.62283164", "0.62241983", "0.6187858", "0.60083485", "0.5904933", "0.58766943", "0.5875189", "0.57837415", "0.576537", "0.5752084", "0.57505965", "0.5628871", "0.56188196", "0.5590547", "0.5584492", "0.55437773", "0.5531586", "0.5531586", "0.5531586", "0.5531586", "0.5531586", "0.5531586", "0.5531586", "0.5531586", "0.5531586", "0.5531586" ]
0.699368
0
Implementing is destination input textbox visible functionality
def is_destination_input_textbox_visible(self): return self.is_element_visible(self.destination_input_textbox_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def is_destination_textbox_visible(self):\n return self.is_element_visible(self.reference_rates_destination_textbox_locator)", "def is_visible(self):", "def visible(self, show):", "def ev_textinput(self, event: TextInput) -> None:", "def show(self): \n InputItem.show(self)\n self.input.selectAll()", "def show(self): \n InputItem.show(self)\n self.input.selectAll()", "def _textChangedSlot(self, text):\n self._param = QtXml.QDomElement()\n if qt4:\n qs = str(self.sender().property(\"dom address\").toPyObject())\n else:\n qs = str(self.sender().property(\"dom address\"))\n\n ind = qs.rfind('/')\n ids = qs[ind:]\n\n self._param = self.qhash[qs].elem.firstChildElement(\"Activate\")\n while(self._param.isNull() is False):\n qs = str(self._param.text()).strip() + ids\n widget_visible = self.qhash[qs].elem.attribute(\"Visible\", \"Uknown\")\n if(text != \"\"):\n self.qhash[qs].widget.setEnabled(True)\n self.qhash[qs].widget.show()\n self.qhash[qs].label.show()\n else:\n self.qhash[qs].widget.setEnabled(False)\n if(widget_visible != \"Unknown\"):\n self.qhash[qs].label.hide()\n self.qhash[qs].widget.hide()\n self._param = self._param.nextSiblingElement(\"Activate\")", "def show(self): \n InputItem.show(self,*args)\n self.input.selectAll()", "def show(self): \n InputItem.show(self,*args)\n self.input.selectAll()", "def modal(self):\r\n self.win.move(0, 0)\r\n self.editing = True\r\n goxapi.start_thread(self.cursor_placement_thread, \"TextBox cursor placement\")\r\n self.value = self.box.edit(self.validator)\r\n self.editing = False\r\n return self.result", "def TransferFromWindow(self):\n return True", "def input(self, *fields, title=\"User input window\"):\n pass", "def XPIsWidgetVisible(inWidget):\n pass", "def TransferToWindow(self):\n return True", "def isVisible(self):\n\t\treturn True", "def AcceptsFocus(self):\n\n return self.IsShown() and self.IsEnabled()", "def isVisible(self, p_int): # real signature unknown; restored from __doc__\n return False", "def isVisible(self, p_int): # real signature unknown; restored from __doc__\n return False", "def check_target(event):\n target_text.set(target_text.get().lower())", "def is_shown(self, request):\n return True", "def test_visible_widgets(plugin_dialog):\n\n assert plugin_dialog.direct_entry_edit.isVisible()\n assert plugin_dialog.direct_entry_btn.isVisible()", "def input_text(self,loc,value,img_name):\r\n self.wait_ele_visible(loc,img_name)\r\n ele = self.get_element(loc,img_nameue)\r\n try:\r\n ele.send_keys(value)\r\n except:\r\n self.getScreenShot(img_name)\r\n logging.exception(\"********input text fail********\")\r\n raise", "def handle_keyup(self, key, string):\r\n if self.get_visible():\r\n for i in self.widgets:\r\n if i.get_visible():\r\n if i.handle_keyup(key, string):\r\n return True\r\n return False", "def test_constructor_visible_widgets(plugin_dialog_constructor):\n assert not plugin_dialog_constructor.direct_entry_edit.isVisible()\n assert not plugin_dialog_constructor.direct_entry_btn.isVisible()", "def onFocus(*args):", "def onFocus(*args):", "def onFocus(*args):", "def onFocus(*args):", "def buildEnter(self): \n \n ttk.Label(self, text='Enter accession number(s), gi(s), or FASTA sequence(s)', \n font=('Arial', '12', 'bold')).grid(row = self.ROW , column = 1, columnspan=4, sticky ='w')\n self.clear_button = tk.Button(self, text='Clear', font=('Arial', '9', 'underline'),command = \n (lambda view = self: self.controller.clear_query(view)))\n self.clear_button.grid(row = self.ROW, column =5, sticky = 'E')\n ttk.Label(self, text = 'Subrange:', font=('Arial', '12', 'bold', 'underline')\n ).grid(row = self.ROW, column = 6, columnspan = 2, sticky = 'E')\n self.ROW += 1\n \n self.query_box = scrolledtext.ScrolledText(self, width = 70, height = 7, wrap=tk.CHAR)\n self.query_box.grid(row = self.ROW, column = 1, rowspan = 6, columnspan = 5)\n self.model_vars['textbox'] = self.query_box\n #Event generated only refers to scrolledtext need a reference to load_query_button\n \n self.query_box.bind('<Key>', lambda event, view = self : self.controller.disable_upload_button(event, view))\n\n tk.Label(self, text = 'From:').grid(row = self.ROW, column = 6, sticky = 'E')\n\n self.query_from = ttk.Entry(self, textvariable = self.model_vars['from'], font=('Arial', 10), width = 15)\n self.query_from.grid(row = self.ROW, column = 7, columnspan = 2, sticky = 'W')\n \n self.ROW+=2\n \n tk.Label(self, text = 'To:').grid(row = self.ROW, column = 6, sticky = 'E')\n self.query_to = tk.Entry(self, textvariable = self.model_vars['to'], font=('Arial', 10), width = 15)\n self.query_to.grid(row = self.ROW, column = 7, columnspan =2 , sticky = 'W')\n \n self.ROW+=5\n #There are objects that inherit from this one that will need to know this value for genetic code widget\n self.upload_file_row = self.ROW\n \n ttk.Label(self, text ='Or, Upload File:', font=('Arial', 10, 'bold')).grid(row = self.ROW, column=1, sticky = 'E')\n \n self.load_query_button = ttk.Button(self, text='Choose File', command = \n (lambda view = self: self.controller.load_handler(view)))\n self.load_query_button.grid(row = self.ROW, column = 2)\n self.load_status = ttk.Label(self, text='No file chosen', font=('Arial', '10'))\n self.load_status.grid(row = self.ROW , column = 3, columnspan = 7, sticky = 'W')", "def XPGetWidgetWithFocus():\n pass" ]
[ "0.7082717", "0.6056446", "0.58619386", "0.5751972", "0.5742402", "0.5742402", "0.5723534", "0.5681438", "0.5681438", "0.5649161", "0.55968016", "0.5532368", "0.5513671", "0.547241", "0.5431309", "0.53958184", "0.53434575", "0.53434575", "0.5325499", "0.52807295", "0.5265454", "0.524183", "0.52406", "0.5206914", "0.5206451", "0.5206451", "0.5206451", "0.5206451", "0.51932776", "0.5192586" ]
0.788917
0
Implementing is expected grid column present functionality
def is_expected_grid_column_present(self, expected_column_name): grid_column_locator = (By.XPATH, "//th[@data-title='%s']" %(expected_column_name)) return self.is_element_present(grid_column_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def columns(self):\n \n pass", "def is_specific_column_present(self, grid_div_id, column_name):\n specific_column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-field='%s']\" % (grid_div_id, column_name))\n return self.is_element_present(specific_column_locator)", "def verify_rates_grid_columns(self):\n is_verified = True\n if(self.is_element_present(self.rates_grid_destination_column_locator) == False):\n is_verified = False\n if (is_verified != False):\n for column_name in self.compare_price_lists_list:\n rates_grid_new_column_locator = (By.XPATH, \"//div[@id='divComaprePriceListRate']/descendant::span[contains(text(),'%s')]\" % column_name)\n if (self.is_element_present(rates_grid_new_column_locator) == False):\n is_verified = False\n break\n return is_verified", "def test_overall_report_columns():\n assert (len(overall_data['columns']) == 31)", "def is_specific_column_present_in_workflow_tab(self, column_name):\n column_locator = (By.XPATH, \"//div[contains(@id, 'divOutboundWorkFlowGrid_')]/descendant::div[@class='k-grid-header']/descendant::th[@data-title='%s']\" % column_name)\n return self.is_element_present(column_locator)", "def assert_check_bounds_column(self):\n value = logic.check_bounds_column(config.NR_COLS-1)\n self.assertTrue(value)\n value = logic.check_bounds_column(config.NR_COLS)\n self.assertFalse(value)\n value = logic.check_bounds_column(config.NR_COLS+1)\n self.assertFalse(value)", "def checkColumns( self ):\n\n for x in list(range(0,3)):\n firstVal = self.__grid[x]\n secondVal = self.__grid[x+3]\n thirdVal = self.__grid[x+6]\n\n compiledVal = str(firstVal) + str(secondVal) + str(thirdVal)\n\n if 'xx' in compiledVal.lower():\n return ('X', compiledVal)\n\n elif 'oo' in compiledVal.lower():\n return ('O', compiledVal)\n\n elif compiledVal.lower() == 'x4x' or \\\n compiledVal.lower() == 'x5x' or \\\n compiledVal.lower() == 'x6x':\n\n return ('X', compiledVal) \n\n return None", "def test_num_columns(self):\n pass", "def is_target_buy_policies_grid_column_present(self, column_name):\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (self.target_buy_policies_grid_div_id, column_name))\n return self.is_element_present(column_locator)", "def test_get_col_index(self):\n self.dboard.set_col_names([\"A\", \"B\"])\n self.assertEqual(0, self.dboard._get_col_index(\"A\"))\n self.assertEqual(1, self.dboard._get_col_index(\"B\"))", "def is_specific_column_on_vendor_profile_grid_present(self, column_name):\n column_locator = (By.XPATH, \"//div[contains(@id, 'divCustomerDialedDigit')]/descendant::a[text()='%s']\" % column_name)\n return self.is_element_present(column_locator)", "def valid_column(self, col: int) -> bool:\n\n return self.check_bounds(0, col) and self.grid[0][col] == \" \"", "def test_002_range_columns(self):\n assert(len(\n self.range_transformer.fit_transform(\n self.data[self.range_col]\n ).columns\n ) == 1)", "def test_column_formats(self):\n attrs = '%(html_col_attr)s'\n col_tag = 'span'\n single_col_tag = ''\n col_data = '%(errors)s%(label)s %(field)s%(help_text)s'\n expected_col = self.form._html_tag(col_tag, col_data, attrs)\n expected_single = col_data if not single_col_tag else self.form._html_tag(single_col_tag, col_data, attrs)\n actual_col, actual_single = self.form.column_formats(None, col_tag, single_col_tag, '', col_data)\n self.assertEqual(expected_col, actual_col)\n self.assertEqual(expected_single, actual_single)", "def test_c2_check_bounds_column_change_config(self):\n config.NR_COLS = 9\n self.assert_check_bounds_column()", "def is_comments_column_present(self):\n is_present = False\n self.column_name_list = self.get_grid_column_names_by_order(self.qos_grid_div_id)\n for column_name in self.column_name_list:\n if column_name == \"Comments\":\n is_present = True\n break\n return is_present", "def test_get_column(self):\n expected = [8, 9, 5, 2, 4, 7, 1, 3, 6]\n self.assertEqual(expected, self.sudoku.get_column(6))", "def verify_dialed_digits_columns(self):\n is_verified = True\n if(self.is_element_present(self.dialed_digits_grid_destination_column_locator) == False or self.is_element_present(self.dialed_digits_grid_dialed_digit_column_locator) == False):\n is_verified = False\n if (is_verified != False):\n for column_name in self.compare_price_lists_list:\n dialed_digits_grid_new_column_locator = (By.XPATH, \"//div[@id='divComparePriceListDialedDigit']/descendant::span[contains(text(),'%s')]\" % column_name)\n if (self.is_element_present(dialed_digits_grid_new_column_locator) == False):\n is_verified = False\n break\n return is_verified", "def check_cols(self):\r\n for i in range(3):\r\n if self.grid[i][-1] != ' ' and self.grid[i][-1] == self.grid[i+3][-1] and self.grid[i+3][-1] == self.grid[i+6][-1]:\r\n return (i, (self.grid[i], self.grid[i+6]))\r\n return (-1, None)", "def becomes_column(self):\n row_type = self.get_type()\n # Note: \"begin repeat\" does become a column\n non_columns = ('begin group', 'end group', 'end repeat')\n return row_type not in non_columns", "def test_collected_columns_no_table_one_col_from_one(self):\n col_double, allow_colspan = False, False # as_type != 'table'\n col_args = ('span', False, {})\n name, multi_field_row = 'first', False\n names = [name]\n row = {name: self.form.fields[name]}\n col_count = 1\n expected = [self.form.collect_col_data(name, self.form.fields[name], *col_args) for name in names]\n for ea in expected:\n if multi_field_row:\n ea['css_classes'] = ' '.join(['nowrap', ea['css_classes']])\n ea['html_head_attr'] = ' class=\"nowrap\"'\n val = ea.pop('css_classes', '')\n val = ' class=\"%s\"' % val if val else ''\n ea['html_col_attr'] = val\n col_settings = (multi_field_row, col_count, col_double, allow_colspan)\n actual = self.form.collect_columns(row, col_settings, *col_args)\n\n self.assertEqual(len(expected), len(actual))\n for expect, got in zip(expected, actual):\n self.assertEqual(len(expect), len(got))\n self.assertListEqual(list(expect.keys()), list(got.keys()))\n self.assertListEqual(list(expect.values()), list(got.values()))\n self.assertEqual(expected, actual)", "def testGetColumnSolution(self): \n actionlist = [1,2,3,4,5]\n for action in actionlist:\n if action == 1:\n val = getColumnSelection(action)\n self.assertEqual(val,\"bookID\")\n if action == 2:\n val = getColumnSelection(action)\n self.assertEqual(val,\"bookAuthor\")\n if action == 3:\n val = getColumnSelection(action)\n self.assertEqual(val,\"ISBN\")\n if action == 4:\n val = getColumnSelection(action)\n self.assertEqual(val,\"numPurchased\")\n if action == 5:\n val = getColumnSelection(action)\n self.assertEqual(val,\"numCheckedOut\")\n if action == 6:\n val = getColumnSelection(action)\n self.assertEqual(val,\"bookTitle\")\n if action == 7:\n val = getColumnSelection(action)\n self.assertEqual(val,\"bookPrice\")", "def valid_col_tester(self, state):\n vert_state = self.cols(state)\n for line in vert_state:\n line_index = vert_state.index(line)\n vert_word = self.check_word(vert_state[line_index])\n if not(vert_word):\n return False\n return True", "def check_col(df):\n students = list(df[\"name\"])\n exp_display = True\n skill_display = True\n # check if gap, cap, exp, skill cols are in df\n if \"gap\" not in list(df.columns):\n # add gap col\n gap_list = [constants.GAP] * len(students)\n df[\"gap\"] = gap_list\n else:\n df.fillna(constants.GAP, inplace=True)\n if \"cap\" not in list(df.columns):\n # add cap col\n shift_cap_list = [constants.CAP] * len(students)\n df[\"cap\"] = shift_cap_list # add cap column to df\n else:\n df.fillna(constants.CAP, inplace=True)\n if \"experience\" not in list(df.columns):\n # add experience col\n exp_list = [constants.EXP] * len(students)\n df[\"experience\"] = exp_list # add experience column to df\n exp_display = False # if using default exp, don't make a display\n else:\n df.fillna(constants.EXP, inplace=True)\n if \"skill\" not in list(df.columns):\n skill_list = [constants.SKILL] * len(students)\n df[\"skill\"] = skill_list\n skill_display = False # if using default skill, don't make a display\n else:\n df.fillna(constants.SKILL, inplace=True)\n\n return [exp_display, skill_display]", "def is_in_the_grid(self, row: int, col: int) -> bool:\n return 0 <= row < self.n_row and 0 <= col < self.n_col", "def test_display_custom_table(self):\n\n column_field = MoveSearchForm.column\n columns = [value for (value, label) in column_field.kwargs['choices']]\n\n response = self.do_search(id=u'1', display='custom-table',\n column=columns)\n self.assert_(\n response.tmpl_context.results,\n \"\"\"Custom table columns don't crash\"\"\".format(value)\n )", "def _prepare_grid(self):\n raise NotImplementedError", "def check_columns(self):\n\t\ti=0\n\t\tfor i in range(len(self.board[i])):\n\t\t\tpts = 0\n\t\t\tfor j in range(len(self.board)):\n\t\t\t\tif self.board[j][i] == self.marker:\n\t\t\t\t\tpts+=1\n\t\t\tif pts == 3:\n\t\t\t\tprint('YOU WON')\n\t\t\t\treturn True", "def in_col(n: int, row: int, col: int, grid: List) -> bool:\n for x in range(9):\n if x!= row and n == grid[x][col]:\n return True\n return False", "def test_collected_columns_no_table_two_col_from_two(self):\n col_double, allow_colspan = False, False # as_type != 'table'\n col_args = ('span', False, {})\n names, multi_field_row = ('first', 'billing_address_1'), True\n row = {name: self.form.fields[name] for name in names}\n col_count = 2\n expected = [self.form.collect_col_data(name, self.form.fields[name], *col_args) for name in names]\n for ea in expected:\n if multi_field_row:\n ea['css_classes'] = ' '.join(['nowrap', ea['css_classes']])\n ea['html_head_attr'] = ' class=\"nowrap\"'\n val = ea.pop('css_classes', '')\n val = ' class=\"%s\"' % val if val else ''\n ea['html_col_attr'] = val\n col_settings = (multi_field_row, col_count, col_double, allow_colspan)\n actual = self.form.collect_columns(row, col_settings, *col_args)\n\n self.assertEqual(len(expected), len(actual))\n for expect, got in zip(expected, actual):\n self.assertEqual(len(expect), len(got))\n self.assertListEqual(list(expect.keys()), list(got.keys()))\n self.assertListEqual(list(expect.values()), list(got.values()))\n self.assertEqual(expected, actual)" ]
[ "0.6262402", "0.6219727", "0.6189594", "0.61687696", "0.6151684", "0.6127107", "0.6107462", "0.610121", "0.60943484", "0.6070248", "0.60317606", "0.5992439", "0.59751236", "0.5942852", "0.59316164", "0.5924672", "0.591986", "0.5913953", "0.58789116", "0.5877305", "0.5858077", "0.5842788", "0.5827345", "0.5822145", "0.58167803", "0.5811138", "0.5810126", "0.5799891", "0.578383", "0.578262" ]
0.711884
0
Implementing select target buy list overrides grid row checkbox functionality
def select_target_buy_list_overrides_grid_row_checkbox(self): for i in range(2): target_buy_list_overrides_grid_checkbox_locator = (By.XPATH, "(//div[@data-grid-name='%s']/descendant::input[@type='checkbox'])[%s]" % (self.target_buy_list_overrides_data_grid_name, str(i + 2))) self.click_element(target_buy_list_overrides_grid_checkbox_locator, True)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def select_vendor_price_list_grid_row_checkbox(self):\n for row in range(0,2):\n column_locator = (By.XPATH, \"//div[@id='%s']/descendant::th[@data-title='%s']\" % (self.vendor_price_list_grid_div_id, self.price_list_column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n column_value_locator = (By.XPATH, \"//div[@id='%s']/descendant::tbody/descendant::tr[%s]/td[%s]/span\" % (self.vendor_price_list_grid_div_id, (row + 1), str(column_index)))\n column_value_element = self.wait().until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n price_list_item = column_value_element.text\n self.compare_price_lists_list.append(price_list_item)\n self.select_grid_row_checkbox(self.customer_price_list_grid_div_id, price_list_item, self.select_grid_row_checkbox_column_number)", "def select_buy_and_sell_deal_management_grid_first_row_checkbox(self):\n self.select_grid_row_checkbox(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_checkbox_column_number, True)", "def select_add_disputes_grid_row_checkbox(self, list_of_audit_account):\n self.select_grid_row_checkbox(self.add_disputes_grid_div_id, list_of_audit_account, self.add_disputes_grid_checkbox_column_number)\n self.wait_for_ajax_spinner_load()", "def DoCheck(self,event):\r\n index = event.GetSelection()\r\n item = self.items[index]\r\n if self.list.IsChecked(index):\r\n self.data.check(item)\r\n else:\r\n self.data.uncheck(item)\r\n #self.list.SetSelection(index)\r", "def GridCheck(Parent,DefaultSelected,Row,Column):\r\n dummyvar = IntVar()\r\n C = Checkbutton(Parent,var=dummyvar)\r\n if DefaultSelected == 1:\r\n C.select()\r\n C.grid(row=Row,column=Column)\r\n C.isChecked = dummyvar\r\n return C", "def sgnDoubleClickInList(self):\n\n self.uiSwitchSelectedCheckStateInList()", "def select_charges_grid_row_checkbox(self, charge_name):\n self.select_grid_row_checkbox(self.charges_grid_div_id, charge_name, self.charges_grid_checkbox_column_number)", "def select_non_traffic_charges_grid_row_checkbox(self, row_identifier_text):\n self.select_grid_row_checkbox(self.non_traffic_charges_grid_div_id, row_identifier_text, self.non_traffic_charges_grid_checkbox_column_number, first_row=True)", "def on_row_selected(self, obj_name):\n if obj_name == 'none':\n for act in self.app.ui.menuobjects.actions():\n act.setChecked(False)\n return\n\n # get the name of the selected objects and add them to a list\n name_list = []\n for obj in self.get_selected():\n name_list.append(obj.options['name'])\n\n # set all actions as unchecked but the ones selected make them checked\n for act in self.app.ui.menuobjects.actions():\n act.setChecked(False)\n if act.text() in name_list:\n act.setChecked(True)", "def _generateCellCheckedState(self, obj, **args):\n result = []\n if self._script.utilities.hasMeaningfulToggleAction(obj):\n oldRole = self._overrideRole(pyatspi.ROLE_CHECK_BOX, args)\n result.extend(self.generate(obj, **args))\n self._restoreRole(oldRole, args)\n\n return result", "def populate_filter(self, src, col, tbl):\r\n sub_cats = set([c for l in kit.SQL_pull(col, tbl) for c in l])\r\n select = {g: tk.BooleanVar() for g in sub_cats}\r\n for key in select:\r\n select[key].set(True)\r\n \r\n line = 1\r\n for g in sub_cats:\r\n line += 1\r\n tk.Checkbutton(src,\r\n text = g,\r\n variable = select[g],\r\n command = lambda x = tbl,\r\n y = False,\r\n z = select: self.checked(x, y, z),\r\n **jt.filter_style).grid(row = line, \r\n sticky = tk.W)\r\n \r\n return select", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def on_source_checked(self, source_item):\n # don't bother for non-checked events\n if not source_item.has_checked_changed():\n return\n # leave checked for as long as search is running\n if self.is_searching():\n source_item.setCheckState(Qt.Checked)\n return\n # keep to the same if exporting\n if self.is_exporting():\n if source_item.is_checked:\n source_item.setCheckState(Qt.Checked)\n else:\n source_item.setCheckState(Qt.Unchecked)\n return\n is_checked = source_item.current_state()\n for key, geometry in source_item.geometries.iteritems():\n if is_checked:\n geometry.enable_source(source_item.title)\n else:\n geometry.disable_source(source_item.title)\n for key, type_entry in source_item.type_entries.iteritems():\n if is_checked:\n type_entry.enable_source(source_item.title, self.geometries)\n else:\n type_entry.disable_source(source_item.title, self.geometries)\n source_item.update_checked()", "def setSelected(*args):", "def setSelected(*args):", "def select_received_charges_grid_row_checkbox(self, row_identifier_text):\n self.select_grid_row_checkbox(self.received_charges_grid_div_id, row_identifier_text, self.received_charges_grid_checkbox_column_number)", "def enable_selected(self, window, values, branch_log_dict, key_event):\n utils.convert_to_numeric(values)\n if(values[key_event] in branch_log_dict[key_event]):#if there is branching for the chosen option\n for element_key in branch_log_dict[key_event][values[key_event]]:\n #values the element can take\n if not isinstance(window[element_key], sg.Text):\n window[element_key].update(disabled = False)\n window[element_key].metadata = True\n window[element_key+\"_label\"].update(text_color = \"#FFFFFF\")#every non-text field has a label\n window[element_key].update(visible = True)", "def select_transactions_grid_row_checkbox(self, row_identifier_text):\n self.select_grid_row_checkbox(self.transactions_grid_div_id, row_identifier_text, self.transactions_grid_checkbox_column_number)\n self.click_element(self.page_header_locator)", "def cbTriggered(self, value):\n global selectedCheckboxes\n selectedCheckboxes=value", "def isSelected(*args):", "def isSelected(*args):", "def onTradeSelected(self, item):\n if self.lstTrade.getMultiSelectedItems() == []:\n self.btnCancelTrade.disable()\n else:\n self.btnCancelTrade.enable()", "def _on_selection_type(self, event):\n def check_item_and_children(control, check_value=True):\n self.tree_ctrl.CheckItem(data_ctrl, check_value)\n if data_ctrl.HasChildren():\n if check_value and not control.IsExpanded():\n # Only select children if control is expanded\n # Always deselect children, regardless (see ticket #259)\n return\n for child_ctrl in data_ctrl.GetChildren():\n self.tree_ctrl.CheckItem(child_ctrl, check_value)\n\n option = self.selection_cbox.GetValue()\n\n pos = self.selection_cbox.GetSelection()\n if pos == wx.NOT_FOUND:\n return\n option = self.selection_cbox.GetString(pos)\n for item in self.list_cb_data.values():\n data_ctrl, _, _, _, _, _, _, _ = item\n _, data_class, _ = self.tree_ctrl.GetItemPyData(data_ctrl)\n if option == 'Select all Data':\n check_item_and_children(data_ctrl, check_value=True)\n elif option == 'Unselect all Data':\n check_item_and_children(data_ctrl, check_value=False)\n elif option == 'Select all Data 1D':\n if data_class == 'Data1D':\n check_item_and_children(data_ctrl, check_value=True)\n elif option == 'Unselect all Data 1D':\n if data_class == 'Data1D':\n check_item_and_children(data_ctrl, check_value=False)\n elif option == 'Select all Data 2D':\n if data_class == 'Data2D':\n check_item_and_children(data_ctrl, check_value=True)\n elif option == 'Unselect all Data 2D':\n if data_class == 'Data2D':\n check_item_and_children(data_ctrl, check_value=False)\n self.enable_append()\n self.enable_freeze()\n self.enable_plot()\n self.enable_import()\n self.enable_remove()", "def change_entries(self, grid_sizer, options, border):\n\n if (self.diag_type is None): return grid_sizer # no data selected, don't change anything\n\n # setup the grid of possible values\n header0 = wx.StaticText(self, -1, \"Add/Remove\")\n header1 = wx.StaticText(self, -1, \"Quantity Code\")\n header2 = wx.StaticText(self, -1, \"Name\")\n header3 = wx.StaticText(self, -1, \"LaTeX Formula\")\n grid_sizer.Add(header0, pos=(0,0), flag=options, border=border)\n grid_sizer.Add(header1, pos=(0,1), flag=options, border=border)\n grid_sizer.Add(header2, pos=(0,2), flag=options, border=border)\n grid_sizer.Add(header3, pos=(0,3), flag=options, border=border)\n grid_sizer.Add(wx.StaticLine(self), pos=(1,0), span=(1,4),\n flag=wx.ALL|wx.EXPAND|wx.GROW, border=border)\n\n self.selected_values = [] # keep track of selected quantities\n\n quantities = self.output_quantities.diagnostic_types[self.diag_type]\n\n # choose a good height/width for formulas\n if (self.diag_type in [\"Linear_Forces\", \"Angular_Momentum\", \"Energy_Flux\",\n \"Induction\", \"Inertial_Forces\", \"Lorentz_Forces\",\n \"Poynting_Flux\", \"TurbKE_Budget\"]):\n width = 100\n elif (self.diag_type == \"Thermal_Equation\"):\n width = 150\n else:\n width = 30\n height = 20\n\n row = 2\n iquant = 0\n for Q in quantities:\n but = wx.ToggleButton(self, Q.code, \"Add\") # build button and place it in second column\n but.Bind(wx.EVT_TOGGLEBUTTON, self.OnToggle)\n grid_sizer.Add(but, pos=(row,0), flag=options, border=border)\n\n q_code = wx.StaticText(self, -1, str(Q.code)) # build other column entries\n q_name = wx.StaticText(self, -1, Q.name) # name\n\n formula = self.RenderTeX(Q, size=(width,height))\n\n # place column entries\n grid_sizer.Add(q_code, pos=(row,1), flag=options, border=border)\n grid_sizer.Add(q_name, pos=(row,2), flag=options, border=border)\n grid_sizer.Add(formula, pos=(row,3), flag=options, border=border)\n\n iquant += 1\n\n # add horizontal line every 5 quantities\n if (iquant % 5 == 0):\n grid_sizer.Add(wx.StaticLine(self), pos=(row+1,0), span=(1,4),\n flag=wx.ALL|wx.EXPAND|wx.GROW, border=border)\n row_inc = 2\n else:\n row_inc = 1\n\n row += row_inc\n\n grid_sizer.AddGrowableCol(2,1) # make the name/formula columns \"1\" growable, i.e., grows as necessary\n grid_sizer.AddGrowableCol(3,1)\n\n return grid_sizer", "def toggleSelection(self):\n for item in self.getItemsToModify():\n checked = item.checkState() == Qt.Checked\n item.setCheckState(Qt.Unchecked if checked else Qt.Checked)", "def toggleSelection(self):\n for item in self.getItemsToModify():\n checked = item.checkState() == Qt.Checked\n item.setCheckState(Qt.Unchecked if checked else Qt.Checked)", "def select_disputes_grid_row_checkbox(self, statement_number):\n self.select_grid_row_checkbox(self.disputes_grid_div_id, statement_number, self.disputes_grid_checkbox_column_number)\n self.wait_for_ajax_spinner_load()", "def onIndustrySelected(self, item):\n if self.lstIndustry.getMultiSelectedItems() == []:\n self.btnRemoveIndustry.disable()\n self.btnUpgradeIndustry.disable()\n else:\n self.enableButtons(self.lstIndustry, [self.btnRemoveIndustry, self.btnUpgradeIndustry])", "def accept(self):\n # self.selectedoptions = [\"ID\", \"Sub district\"]\n self.selectedoptions = [\"ID\", \"District\"]\n model = self.lstLayers.model()\n for i in range(model.rowCount()):\n item = model.item(i)\n if item.checkState() == Qt.Checked:\n self.selectedoptions.append(item.text())\n QDialog.accept(self)", "def toggle_select(self):\r\n if not len(self.items):\r\n return\r\n item = self.items[self.item_sel]\r\n if item in self.selected:\r\n self.selected.remove(item)\r\n else:\r\n self.selected.append(item)\r\n self.do_paint()" ]
[ "0.6595565", "0.6579554", "0.64243436", "0.6204201", "0.61319244", "0.59698147", "0.5958927", "0.595838", "0.59310436", "0.59297544", "0.5927377", "0.58760035", "0.5861238", "0.58487624", "0.58487624", "0.58440185", "0.58437294", "0.5766345", "0.5711271", "0.56679034", "0.56679034", "0.56521577", "0.5646735", "0.56105965", "0.56082714", "0.56082714", "0.55731446", "0.55380154", "0.5525007", "0.55123305" ]
0.7345537
0
Implementing is target buy details section present in vendor profile page functionality
def is_target_buy_details_section_present_in_vendor_profile_page(self): return self.is_element_present(self.target_buy_details_section_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def is_specific_field_on_target_buy_details_section_present(self, field_name):\n field_locator = (By.XPATH, \"//div[contains(@id, 'divcustomerProfileDetails_')]/descendant::label[text()='%s']\" % field_name)\n return self.is_element_present(field_locator)", "def is_target_buy_policies_tab_displayed(self):\n return self.is_element_visible(self.target_buy_policies_tab_locator)", "def is_target_buy_list_overrides_screen_loaded(self):\n return self.is_element_visible(self.target_buy_list_overrides_page_header_locator)", "def test_get_pay_in_details(self):\n pass", "def get_vendor_price_lists_details(self):\n try:\n self.vendor_price_lists_dict = self.get_grid_row_details(self.customer_price_list_grid_div_id, self.vendor_price_lists_dict)\n return True\n except:\n return False", "def test_get_small_and_light_eligibility_by_seller_sku(self):\n pass", "def test_get_order_buyer_info(self):\n pass", "def is_buy_and_sell_deal_details_screen_status_dropdown_visible(self):\n return self.is_element_present(self.buy_and_sell_deal_details_screen_status_dropdown_locator)", "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def is_view_price_list_detail_present(self):\n return self.is_element_present(self.view_price_list_detail_locator)", "def input_payment_details(self):\n pass", "def test_get_order_items_buyer_info(self):\n pass", "def get_details(self):", "def enable_details(self):\n self.cs.detail = True", "def test_verifyvendordetails(self):\n self.tc_id = \"Ts_003\"\n self.tc_desc = \"Verify the admin is able to login and verify the vendor\"\n self.tc_step = \"TC Start\"\n\n login = adminLoginVerify(self.driver)\n\n self.tc_step = \"Launch the url\"\n login.launchUrl(self.url)\n\n self.tc_step = \"Enter the login details\"\n login.clickVendorLogin()\n login.enterloginDetails(self.username, self.password)\n login.clickLogin()\n login.clickvendorTab()\n login.entervendornameSearch(self.entervendornameSearch)\n login.clickVendor()\n login.verifyVendor()\n login.enterstaffDetails(self.staffname, self.staffdesc)\n login.confirmVerification()", "def is_vendor_profile_present(self):\n return self.is_element_present(self.vendor_profile_locator)", "def purchase(self, item_type):", "def is_available(self, product_url):\n\t\tpass", "def is_specific_tab_on_vendor_profile_page_present(self, tab_name):\n tab_locator = (By.XPATH, \"//div[contains(@id, 'SourceProfileTabStrip')]/descendant::a[text()='%s']\" % tab_name)\n return self.is_element_present(tab_locator)", "async def should_handle(self):\n return self.main.base_amount > 4 and self.main.can_build_unique(UnitTypeId.INFESTATIONPIT, self.main.pits)", "def click_buy_and_sell_deal_management_link(self):\n self.select_static_tab(self.buy_and_sell_deal_management_locator, message=\"buy and sell deal management locator not found before specified time out\")", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def test_confirm_customization_details(self):\n pass", "def test_get_small_and_light_enrollment_by_seller_sku(self):\n pass", "def has_details(self):\n return hasattr(self, 'details')", "def _test_display_up_button(self):\n return (self.product_displays.top_index > 0)", "def test_buy_now(self):\n catalog_page = CatalogPage(self.driver)\n product_page = ProductPage(self.driver)\n payment_page = PaymentPage(self.driver)\n payment_review_page = PaymentReviewPage(self.driver)\n payment_info_page = PaymentInfoPage(self.driver)\n success_page = SuccessPage(self.driver)\n # buy the new product\n navigate_to(self.driver, ProductPage.URL(self.new_product['product']['title']))\n product_page.add_to_cart.click()\n # by an old product\n catalog_page.catalog.click()\n # Sort products to move the newly created to last page\n catalog_page.sorting_order.select_by_visible_text(\"Date, old to new\")\n catalog_page.image.random_click()\n product = product_page.product.get_text()\n product_page.add_to_cart.click()\n catalog_page.catalog.click()\n catalog_page.cart.click()\n payment_dic = {\n 'address' : f'{randint(1, 99999)} {random_name(5, 8)}',\n 'city' : \"San Francisco\",\n 'email_or_mobile_phone_number_input' : random_name(8) + \"@gmail.com\",\n 'last_name' : random_name(3, 12),\n 'zip_code' : '94107',\n }\n if randint(0, 1):\n payment_dic['first_name'] = random_name(4, 16)\n if randint(0, 1):\n payment_dic['address2'] = random_name(5)\n for _ in payment_dic:\n exec(f\"payment_page.{_}.enter(payment_dic['{_}'])\")\n payment_page.continue_to_shipping.click()\n payment_review_page.continue_to_payment.click()\n payment_info_page.full_address.get_text()\n # validate address\n for _ in ['address', 'city', 'zip_code']:\n assert_and_log(payment_dic[_] in payment_info_page.full_address.get_text(),\n f\"{_} in full address\")\n payment_info_page.enter_bogus_payment(1)\n assert_and_log(success_page.thank_you.find_visible_element(),\n \"'Thank you' appeared as a sign of successful transaction\",\n continue_on_error=False)\n validate(success_page.basic_validation_list)", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def select_vendor_price_list_detail_reference_rates_tab(self):\n self.click_element(self.vendor_price_list_details_reference_rates_tab_locator, True)" ]
[ "0.64984775", "0.59864783", "0.5873134", "0.5831746", "0.57233775", "0.56565744", "0.56486034", "0.56228536", "0.5586765", "0.5521409", "0.546854", "0.54677045", "0.5465543", "0.5402681", "0.53863895", "0.53619283", "0.53444225", "0.52923024", "0.52921027", "0.5267506", "0.52491593", "0.5244132", "0.5225149", "0.52245116", "0.51811576", "0.51537395", "0.5146594", "0.51357687", "0.5126538", "0.51214004" ]
0.76688296
0
Implementing is specific field on target buy details section present functionality
def is_specific_field_on_target_buy_details_section_present(self, field_name): field_locator = (By.XPATH, "//div[contains(@id, 'divcustomerProfileDetails_')]/descendant::label[text()='%s']" % field_name) return self.is_element_present(field_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def is_target_buy_details_section_present_in_vendor_profile_page(self):\n return self.is_element_present(self.target_buy_details_section_locator)", "def get_details(self):", "def enable_details(self):\n self.cs.detail = True", "def purchase(self, item_type):", "def input_payment_details(self):\n pass", "def test_get_pay_in_details(self):\n pass", "def test_get_additional_seller_inputs(self):\n pass", "def has_details(self):\n return hasattr(self, 'details')", "def test_get_order_buyer_info(self):\n pass", "def is_view_price_list_detail_present(self):\n return self.is_element_present(self.view_price_list_detail_locator)", "def qualified_item(self):\n # Price check with base item\n self.log.info(\"Price checking Qual 1 via PLU\")\n pos.click(\"Price Check\")\n pos.enter_keypad(\"030\", after='enter')\n if self.selection_list_visible():\n pos.select_list_item(\"Qual 1 ($5.00)\")\n pos.click(\"enter\")\n else:\n tc_fail(\"Selection list didn't appear.\")\n\n # Confirm the right item, at the right price\n self.read_price_check(\"Qual 1\", \"$5.00\")\n # Add the item\n pos.click(\"Sell Item\")\n\n # Price check with qualifier\n self.log.info(\"Price checking Qual 1 via PLU\")\n pos.click(\"Price Check\")\n pos.enter_keypad(\"030\", after='enter')\n if self.selection_list_visible():\n pos.select_list_item(\"Test Type ($10.00)\")\n pos.click(\"enter\")\n else:\n tc_fail(\"Selection list didn't appear.\")\n\n # Confirm the right item, at the right price\n self.read_price_check(\"Qualifier 1\", \"$10.00\")\n # Add the item\n pos.click(\"Sell Item\")\n\n # Confirm we are in a transaction\n if not self.in_transaction():\n self.tc_fail(\"POS did not start a transaction; can not confirm item was added\")\n else:\n self.log.info(\"Confirmed we are in a transaction\")\n \n # Confirm we added the item\n ret1 = self.confirm_line(-2, \"Qual 1\", \"$5.00\")\n if ret1:\n self.log.info(\"Confirmed Qual 1 item added\")\n else:\n self.tc_fail(ret1)\n \n # Confirm we added the linked item\n ret2 = self.confirm_line(-1, \"Qualifier 1\", \"$10.00\")\n if ret2:\n self.log.info(\"Confirmed Qualifier 1 item added\")\n else:\n self.tc_fail(ret2)\n \n # Setup for next test\n self.recover()", "def get_display_price_data(self, source, commitment):\n overage, included = self.get_price_data(source, commitment)\n if self.name == settings.BILLING_DEFAULT_PLAN_NAME:\n included = OFFICIAL_BUILDER_LIMITS[source]\n return overage, included", "def details(self):\n pass", "def test_get_order_items_buyer_info(self):\n pass", "def test_get_small_and_light_fee_preview(self):\n pass", "def test_get_small_and_light_eligibility_by_seller_sku(self):\n pass", "def _set_details_of_charges_71A(self, val):\n self.swift_obj.DetailsOfCharges = val\n self.swift_obj.DetailsOfCharges.swiftTag = '71A'", "def test_get_details7(self):\n pass", "def get_vendor_price_lists_details(self):\n try:\n self.vendor_price_lists_dict = self.get_grid_row_details(self.customer_price_list_grid_div_id, self.vendor_price_lists_dict)\n return True\n except:\n return False", "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def test_get_additional_seller_inputs_old(self):\n pass", "def _set_additional_fields(self, data):\n # Remove the non required rules data.\n if 'rules' in data:\n del data['rules']\n # Change description into proper string.\n data['description'] = re.sub(\"[\\'\\\"]\", \"\", data['description'])\n # Calculate and update the premium field.\n premium = str(data.get('premium', \"false\")).lower() == 'true'\n data['pvtVuln'] = premium\n return data", "def on_market_info(self):\n pass", "def handle_list(self, detail, *args, **kwargs):\n for product_type in models.ProductType.objects.all():\n print(product_type.name)\n if detail:\n for coverage_type in product_type.allowed_coverage_types.all():\n print(\"\\t%s\" % coverage_type.name)", "def additional_first_vax_info(self):\n myElem = WebDriverWait(self.driver, self.delay)\\\n .until(EC.presence_of_element_located((By.ID, 'first_vaccine_brand')))\n for click_field in (\"first_vaccine_brand\"):\n click_value = self.user_info[click_field]\n click_field_final = f\"{click_field}_{click_value}\"\n checkbox = self.driver.find_element_by_id(click_field_final)\n checkbox.click()\n for fill_field in (\"patient_patient_question_answers_attributes_1_additional_info\"):\n fill_value = self.user_info[fill_field]\n fill_web_loc = self.driver.find_element_by_id(fill_field)\n fill_web_loc.send_keys(fill_value)", "def getDetail(self):\n\t\t\n\t\treturn (super().setParameters(0,self.getDefense(),0))\n\t\t\n\t\t#return \"\\n#########################################################\\n\"+\"\\nItem of Defense, Name of item:\"+self.getName()+\"\\nCapacity of defense:\"+str(self.getDefense())+\"\\nCapacity of attack:0 \\n Capacity of heal:0 \\n\"+\"#########################################################\\n\"", "async def should_handle(self):\n return self.main.base_amount > 4 and self.main.can_build_unique(UnitTypeId.INFESTATIONPIT, self.main.pits)", "def action_show_details(self):\n self.ensure_one()\n if self.is_subcontract:\n rounding = self.product_uom.rounding\n production = self.move_orig_ids.production_id\n if self._has_tracked_subcontract_components() and\\\n float_compare(production.qty_produced, production.product_uom_qty, precision_rounding=rounding) < 0 and\\\n float_compare(self.quantity_done, self.product_uom_qty, precision_rounding=rounding) < 0:\n return self._action_record_components()\n action = super(StockMove, self).action_show_details()\n if self.is_subcontract and self._has_tracked_subcontract_components():\n action['views'] = [(self.env.ref('stock.view_stock_move_operations').id, 'form')]\n action['context'].update({\n 'show_lots_m2o': self.has_tracking != 'none',\n 'show_lots_text': False,\n })\n return action", "def getTitle(self):\n return \"ERP5BankingStopPayment\"", "def print_report(self):\n assert len(self) == 1, 'This option should only be used for a single id at a time.'\n datas = {\n 'form': \n {\n 'company_id': self.company_id and [self.company_id.id] or [],\n 'warehouse_ids': [y.id for y in self.warehouse_ids],\n 'start_date': self.start_date,\n 'end_date': self.end_date,\n 'include_zero': self.include_zero,\n 'sort_order': self.sort_order,\n 'value': self.value,\n 'id': self.id,\n }\n }\n\n if [y.id for y in self.warehouse_ids] and (not self.company_id):\n self.warehouse_ids = []\n raise Warning(_('Please select company of those warehouses to get correct view.\\nYou should remove all warehouses first from selection field.'))\n return self.env.ref(\n 'most_selling_product.action_ir_most_selling_product'\n ).report_action(self, data=datas)" ]
[ "0.6218292", "0.58803385", "0.57047385", "0.55989105", "0.5592674", "0.5399362", "0.5347198", "0.531534", "0.5296534", "0.5213632", "0.5194612", "0.5192922", "0.5177318", "0.51669794", "0.5142594", "0.5140994", "0.5122961", "0.5118239", "0.5090551", "0.5051972", "0.50412834", "0.5009991", "0.49997178", "0.49672276", "0.49540892", "0.4950815", "0.49497515", "0.49494627", "0.49379268", "0.4919366" ]
0.6910262
0
Implementing set target buy distribution link code functionality
def set_target_buy_distribution_link_code(self): self.set_value_into_input_field(self.target_buy_distribution_link_code_textbox_locator, self.random_string_generator(6)) self.wait_for_ajax_spinner_load()
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _set_link(\n meta: Dict,\n link: Optional[Union[type(None), str, bool, KEChainPages]] = None,\n link_value: Optional[CardWidgetLinkValue] = None,\n link_target: Optional[Union[str, LinkTargets]] = LinkTargets.SAME_TAB,\n **kwargs,\n) -> Dict:\n meta[\"linkTarget\"] = check_enum(link_target, LinkTargets, \"link_target\")\n\n from pykechain.models import Activity\n\n if isinstance(link, Activity):\n if link.activity_type == ActivityType.TASK:\n default_link_value = CardWidgetLinkValue.TASK_LINK\n else:\n default_link_value = CardWidgetLinkValue.TREE_VIEW\n\n meta.update(\n {\n MetaWidget.CUSTOM_LINK: link.id,\n MetaWidget.SHOW_LINK_VALUE: default_link_value,\n }\n )\n elif isinstance(link, str) and is_uuid(link):\n meta.update(\n {\n MetaWidget.CUSTOM_LINK: link,\n MetaWidget.SHOW_LINK_VALUE: CardWidgetLinkValue.TASK_LINK,\n }\n )\n elif link is None or link is False:\n meta.update(\n {\n MetaWidget.CUSTOM_LINK: None,\n MetaWidget.SHOW_LINK_VALUE: CardWidgetLinkValue.NO_LINK,\n }\n )\n elif link in KEChainPages.values():\n meta.update(\n {\n MetaWidget.CUSTOM_LINK: \"\",\n MetaWidget.SHOW_LINK_VALUE: CardWidgetKEChainPageLink[link],\n }\n )\n else:\n meta.update(\n {\n MetaWidget.CUSTOM_LINK: link,\n MetaWidget.SHOW_LINK_VALUE: CardWidgetLinkValue.EXTERNAL_LINK,\n }\n )\n\n if link_value is not None:\n meta.update(\n {\n MetaWidget.SHOW_LINK_VALUE: check_enum(\n link_value, CardWidgetLinkValue, \"link_value\"\n ),\n }\n )\n\n return meta", "def execute(self, target_link: str) -> ShortenedLink:\n\n # Here is a place to add some business features\n # For example check for banned domains or something\n\n link_key = self.key_generation_service.generate(target_link)\n\n shortened_link = ShortenedLink(target_link=target_link, key=link_key)\n\n self.shortened_link_dao.upsert(shortened_link)\n\n return shortened_link", "def target(self):", "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def __init__(self, source_node, source_gate_name, target_node, target_slot_name, weight=1):\n self.link(source_node, source_gate_name, target_node, target_slot_name, weight)", "def target(self, value):\n self._target = value", "def target(self, target) :\n\t\ttry :\n\t\t\tself._target = target\n\t\texcept Exception as e:\n\t\t\traise e", "def target(self, target) :\n\t\ttry :\n\t\t\tself._target = target\n\t\texcept Exception as e:\n\t\t\traise e", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def target_product(self, target_product):\n\n self._target_product = target_product", "def target(self, assault_army: dict, arr_target: dict) -> None:\n raise NotImplementedError(\"target() method is undefined\")", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def target(self, target):\n self.__target = float(target)", "def use(target, name):\n return \"You use the \" + name + \" to pay \" + target.name", "def repo_link_set(self, repo_id, link_type, target):\n self.send(repo_id, 'repo_link_set', link_type, target)", "def target(self, target):\n\n self._target = target", "def calculate_appropriate_target(self):\n pass", "def calculate_appropriate_target(self):\n pass", "def initialise_target(self, c, key):\n return 0", "def set_target_buy_policy_abbreviation(self, prefix):\n self.set_value_into_input_field(self.target_buy_policy_abbreviation_input_field_locator, value=(prefix + self.random_string_generator(size=4)))", "def test_update_pci_link(self):\n pass", "def use(target, name):", "def set_target_buy_distribution_receiving_member(self, member_prefix):\n member_name = member_prefix + self.random_string_generator(size=4)\n self.set_value_into_input_field(self.target_buy_distribution_receiving_member_textbox_locator, member_name)\n self.wait_for_ajax_spinner_load()", "def create_deb_for_target(ctx, target=target):\n pass", "def transfer(self,source,target):\n self.__transfer_dict[tuple([source,target])]()", "def handle_all_link_nak(self, target, cmd1, cmd2, user_data, hops_left):", "def update_target(self):\n pass", "def set_target_buy_list_number_plan(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_number_plan_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_number_plan_kendo_dropdown_arrow_locator, item_name)", "def menu_active_python_distribution(self, event=None):\n self.link('http://www.activestate.com')" ]
[ "0.57581264", "0.57360834", "0.5704888", "0.57009506", "0.56327444", "0.5612569", "0.5577209", "0.5577209", "0.55430907", "0.5542682", "0.5497914", "0.5482014", "0.5472685", "0.5469825", "0.54515964", "0.54339606", "0.54190713", "0.5416186", "0.5416186", "0.54143345", "0.5406166", "0.54038846", "0.5378979", "0.5368973", "0.5367905", "0.53668016", "0.5354917", "0.5338118", "0.5336816", "0.5282774" ]
0.7068939
0
Implementing set target buy distribution receiving member functionality
def set_target_buy_distribution_receiving_member(self, member_prefix): member_name = member_prefix + self.random_string_generator(size=4) self.set_value_into_input_field(self.target_buy_distribution_receiving_member_textbox_locator, member_name) self.wait_for_ajax_spinner_load()
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def target_sell_price(self):\n return super(Player, self).target_sell_price", "async def _vis_buy(self, ctx, *args):\n if has_post_permission(ctx.guild.id, ctx.channel.id):\n number, item = ch.parse_number_and_name(args)\n if item:\n await ctx.send(vis_helpers.shop_buy(ctx.user_object, item, number))", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def purchase(self, item_type):", "def target(self):", "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def _buy(self, units=1):\n self.quantity -= units", "def calculate_appropriate_target(self):\n pass", "def calculate_appropriate_target(self):\n pass", "def __call__(self, auctioneer):\n possible_bid = self.bid_increase_perc * auctioneer.get_highest_bid()\n if possible_bid < self.budget and random.random() <= self.bid_probability:\n self.highest_bid = possible_bid\n auctioneer.accept_bid(possible_bid, self)", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "async def buy(self, ctx, *args):\n if has_post_permission(ctx.guild.id, ctx.channel.id):\n number, item = ch.parse_number_and_name(args)\n if number and item:\n out = item_helpers.buy(ctx.author.id, item, number=number)\n await ctx.send(out)", "def transfer(self,source,target):\n self.__transfer_dict[tuple([source,target])]()", "def passive(self,target):\r\n target.dmg_increase(self.__dmg_increase,buff_type = \"permanent\")", "def ReflectingBuyer(Buyer):\n increase_step = 0.01\n\n if Buyer.has_bought == True:\n Buyer.like_buy *= (1-increase_step)\n elif Buyer.like_buy * (1+increase_step) >= Buyer.max_value and Buyer.has_bought == False:\n Buyer.like_buy = Buyer.max_value\n else:\n Buyer.like_buy *= (1+increase_step)\n Buyer.has_bought = False #return to normal state", "def transferfunds(self):", "def target(self, target):\n self.__target = float(target)", "def _borrow(self):\n\t\tif self.debt_link == None:\n\t\t\tchoice_list = [a for s in self.site.neighbors for a in s.agents_in_site if a.stock >= (self.consumption_demanded - self.consumed)]\n\t\t\tif len(choice_list) > 0: \n\t\t\t\tchoosed = numpy.random.choice(choice_list)\n\t\t\t\tloan_value = choosed._lend(self, self.consumption_demanded - self.consumed)\n\t\t\t\tself.consumed += loan_value\n\t\t\t\tself.consumption_deficit -= loan_value", "def add_for_target(self, target, products):\n self._products_by_target[target].update(products)", "async def buybait(self, ctx:commands.Context, quantity:int, *bait_type:str):\r\n\r\n if not await self.IsSpecialized(ctx.guild, ctx.channel.id, SHOP_CHANNEL):\r\n await ctx.send('Cannot buy bait here\\nUse `add shop` to turn this channel into a shop')\r\n return\r\n\r\n bait_type = ' '.join(bait_type)\r\n if not bait_type in fishing_bait:\r\n await ctx.send(f'{bait_type} is not a valid form of bait')\r\n\r\n bulk_mod = await self.GetModifier(ctx.guild, 'bulk_purchase_mod')\r\n bulk_requirement = await self.GetSetting(ctx.guild, 'bulk_minimum')\r\n total = int(bait_prices[bait_type] * quantity * await self.GetModifier(ctx.guild, \"bait_price\") * (1 if quantity < bulk_requirement else 1 - bulk_mod))\r\n\r\n if not bank.can_spend(ctx.message.author, total):\r\n await ctx.send(f'You don\\'t have enough {await bank.get_currency_name(ctx.guild)}')\r\n return\r\n\r\n msg = await ctx.send(f'Are you sure you want to buy {bait_type} x{quantity} ({total} {await bank.get_currency_name(ctx.guild)})'\r\n + (f'\\n*-{100 * bulk_mod}% for buying in bulk*' if quantity >= bulk_requirement else ''))\r\n\r\n start_adding_reactions(msg, ReactionPredicate.YES_OR_NO_EMOJIS)\r\n pred = ReactionPredicate.yes_or_no(msg, ctx.author)\r\n try:\r\n await ctx.bot.wait_for(\"reaction_add\", check=pred, timeout=15)\r\n except asyncio.TimeoutError:\r\n await msg.clear_reactions()\r\n return\r\n\r\n if pred.result is True:\r\n member_bait = await self.config.member(ctx.message.author).bait()\r\n member_bait[bait_type] += quantity\r\n await self.config.member(ctx.message.author).bait.set(member_bait)\r\n await msg.edit(content=f'{quantity} {bait_type} bought for {total} {await bank.get_currency_name(ctx.guild)}')\r\n await bank.withdraw_credits(ctx.message.author, total)\r\n else:\r\n await msg.edit(content='Sale cancelled')\r\n\r\n await msg.clear_reactions()", "def use(target, name):\n return \"You use the \" + name + \" to pay \" + target.name", "def test_share_distributor(self):\n\n # =================================================================\n # test: complete test, share distribution int\n # =================================================================\n\n sell_list = [StockOrderWrapper(self.order_2)]\n sell = sell_list[0]\n buy_orders = [StockOrderWrapper(self.order_3), StockOrderWrapper(self.order_4), StockOrderWrapper(self.order_7),\n StockOrderWrapper(self.order_8), StockOrderWrapper(self.order_11),\n StockOrderWrapper(self.order_12)]\n\n result = self.bidding_round_manager.share_distributor(sell_orders=sell_list, buy_orders=buy_orders)\n\n order_real_3 = result[0]\n order_real_4 = result[1]\n order_real_7 = result[2]\n order_real_8 = result[3]\n order_real_11 = result[4]\n order_real_12 = result[5]\n\n sell_order_dict_1 = {sell: 1}\n sell_order_dict_2 = {sell: 2}\n\n order_exp_3 = StockOrderWrapper(self.order_3)\n order_exp_4 = StockOrderWrapper(self.order_4)\n order_exp_7 = StockOrderWrapper(self.order_7)\n order_exp_8 = StockOrderWrapper(self.order_8)\n order_exp_11 = StockOrderWrapper(self.order_11)\n order_exp_12 = StockOrderWrapper(self.order_12)\n\n order_exp_3.transaction_dict = sell_order_dict_1\n order_exp_4.transaction_dict = sell_order_dict_2\n order_exp_7.transaction_dict = sell_order_dict_2\n order_exp_8.transaction_dict = sell_order_dict_2\n order_exp_11.transaction_dict = sell_order_dict_2\n order_exp_12.transaction_dict = sell_order_dict_1\n\n order_exp_3.shares_left = 4\n order_exp_4.shares_left = 8\n order_exp_7.shares_left = 8\n order_exp_8.shares_left = 8\n order_exp_11.shares_left = 8\n order_exp_12.shares_left = 4\n\n self.is_equal_order_wrapper(order_real_3, order_exp_3)\n self.is_equal_order_wrapper(order_real_4, order_exp_4)\n self.is_equal_order_wrapper(order_real_7, order_exp_7)\n self.is_equal_order_wrapper(order_real_8, order_exp_8)\n self.is_equal_order_wrapper(order_real_11, order_exp_11)\n self.is_equal_order_wrapper(order_real_12, order_exp_12)\n\n # =================================================================\n # test: share distribution float easy\n # =================================================================\n\n sell_list = [StockOrderWrapper(self.order_2)]\n sell = sell_list[0]\n\n buy_order_3 = StockOrderWrapper(self.order_3)\n buy_order_4 = StockOrderWrapper(self.order_4)\n buy_order_7 = StockOrderWrapper(self.order_7)\n buy_order_8 = StockOrderWrapper(self.order_8)\n buy_order_11 = StockOrderWrapper(self.order_11)\n buy_order_12 = StockOrderWrapper(self.order_12)\n\n sell.shares_left = 6\n buy_order_3.shares_left = 8 # 2.4 -> 3\n buy_order_4.shares_left = 4 # 1.2 -> 1\n buy_order_7.shares_left = 4 # 1.2 -> 1\n buy_order_8.shares_left = 4 # 1.2 -> 1\n buy_order_11.shares_left = 0 # 0.0 -> 0\n buy_order_12.shares_left = 0 # 0.0 -> 0\n\n buy_orders = [buy_order_3, buy_order_4, buy_order_7, buy_order_8, buy_order_11, buy_order_12]\n\n result = self.bidding_round_manager.share_distributor(sell_orders=sell_list, buy_orders=buy_orders)\n\n result_dict = {}\n\n for res in result:\n result_dict[res.stock_order.order_id] = res\n\n try:\n # noinspection PyUnusedLocal\n dummy = result[4]\n raise AssertionError(\"IndexError exception expected\")\n except IndexError:\n pass\n\n sell_order_dict_1 = {sell: 1}\n sell_order_dict_3 = {sell: 3}\n\n order_exp_3.transaction_dict = sell_order_dict_3\n order_exp_4.transaction_dict = sell_order_dict_1\n order_exp_7.transaction_dict = sell_order_dict_1\n order_exp_8.transaction_dict = sell_order_dict_1\n\n order_exp_3.shares_left = 5\n order_exp_4.shares_left = 3\n order_exp_7.shares_left = 3\n order_exp_8.shares_left = 3\n\n self.is_equal_order_wrapper(result_dict.get(3), order_exp_3)\n self.is_equal_order_wrapper(result_dict.get(4), order_exp_4)\n self.is_equal_order_wrapper(result_dict.get(7), order_exp_7)\n self.is_equal_order_wrapper(result_dict.get(8), order_exp_8)\n\n # =================================================================\n # test: share distribution float complex (multiple float iterations)\n # =================================================================\n\n sell_list = [StockOrderWrapper(self.order_2)]\n sell = sell_list[0]\n\n buy_order_3 = StockOrderWrapper(self.order_3)\n buy_order_4 = StockOrderWrapper(self.order_4)\n buy_order_7 = StockOrderWrapper(self.order_7)\n buy_order_8 = StockOrderWrapper(self.order_8)\n buy_order_11 = StockOrderWrapper(self.order_11)\n buy_order_12 = StockOrderWrapper(self.order_12)\n\n sell.shares_left = 18\n sell.stock_order.order_status = DEFINITIVE\n\n self.order_3.order_status = DEFINITIVE\n self.order_4.order_status = DEFINITIVE\n self.order_7.order_status = DEFINITIVE\n self.order_8.order_status = DEFINITIVE\n self.order_11.order_status = DEFINITIVE\n self.order_12.order_status = DEFINITIVE\n\n buy_order_3.shares_left = 6 # 3.6 -> 4\n buy_order_4.shares_left = 4 # 2.4 -> 2\n buy_order_7.shares_left = 2 # 1.2 -> 1\n buy_order_8.shares_left = 2 # 1.2 -> 1\n buy_order_11.shares_left = 8 # 4.8 -> 5\n buy_order_12.shares_left = 8 # 4.8 -> 5\n\n buy_orders = [buy_order_3, buy_order_4, buy_order_7, buy_order_8, buy_order_11, buy_order_12]\n\n result = self.bidding_round_manager.share_distributor(sell_orders=sell_list, buy_orders=buy_orders)\n\n result_dict = {}\n\n for res in result:\n result_dict[res.stock_order.order_id] = res\n\n try:\n # noinspection PyUnusedLocal\n dummy = result[6]\n raise AssertionError(\"IndexError exception expected\")\n except IndexError:\n pass\n\n sell_order_dict_1 = {sell: 1}\n sell_order_dict_2 = {sell: 2}\n sell_order_dict_4 = {sell: 4}\n sell_order_dict_5 = {sell: 5}\n\n order_exp_3.transaction_dict = sell_order_dict_4\n order_exp_4.transaction_dict = sell_order_dict_2\n order_exp_7.transaction_dict = sell_order_dict_1\n order_exp_8.transaction_dict = sell_order_dict_1\n order_exp_11.transaction_dict = sell_order_dict_5\n order_exp_12.transaction_dict = sell_order_dict_5\n\n order_exp_3.shares_left = 2\n order_exp_4.shares_left = 2\n order_exp_7.shares_left = 1\n order_exp_8.shares_left = 1\n order_exp_11.shares_left = 3\n order_exp_12.shares_left = 3\n\n self.is_equal_order_wrapper(result_dict.get(3), order_exp_3)\n self.is_equal_order_wrapper(result_dict.get(4), order_exp_4)\n self.is_equal_order_wrapper(result_dict.get(7), order_exp_7)\n self.is_equal_order_wrapper(result_dict.get(8), order_exp_8)\n self.is_equal_order_wrapper(result_dict.get(11), order_exp_11)\n self.is_equal_order_wrapper(result_dict.get(12), order_exp_12)\n\n # =================================================================\n # test: share distribution date (two stocks left after one float iteration)\n # =================================================================\n\n sell_list = [StockOrderWrapper(self.order_2)]\n sell = sell_list[0]\n\n buy_order_3 = StockOrderWrapper(self.order_3)\n buy_order_4 = StockOrderWrapper(self.order_4)\n buy_order_7 = StockOrderWrapper(self.order_7)\n buy_order_8 = StockOrderWrapper(self.order_8)\n buy_order_11 = StockOrderWrapper(self.order_11)\n buy_order_12 = StockOrderWrapper(self.order_12)\n\n sell.shares_left = 19\n\n buy_order_3.shares_left = 4 # 2.4 -> 2 -> 3\n buy_order_4.shares_left = 4 # 2.4 -> 2 -> 3\n buy_order_7.shares_left = 4 # 2.4 -> 2 -> 2\n buy_order_8.shares_left = 2 # 1.2 -> 1 -> 1\n buy_order_11.shares_left = 8 # 4.8 -> 5 -> 5\n buy_order_12.shares_left = 8 # 4.8 -> 5 -> 5\n\n buy_orders = [buy_order_3, buy_order_4, buy_order_7, buy_order_8, buy_order_11, buy_order_12]\n\n result = self.bidding_round_manager.share_distributor(sell_orders=sell_list, buy_orders=buy_orders)\n\n result_dict = {}\n\n for res in result:\n result_dict[res.stock_order.order_id] = res\n\n try:\n # noinspection PyUnusedLocal\n dummy = result[6]\n raise AssertionError(\"IndexError exception expected\")\n except IndexError:\n pass\n\n sell_order_dict_1 = {sell: 1}\n sell_order_dict_2 = {sell: 2}\n sell_order_dict_3 = {sell: 3}\n sell_order_dict_5 = {sell: 5}\n\n order_exp_3.transaction_dict = sell_order_dict_3\n order_exp_4.transaction_dict = sell_order_dict_3\n order_exp_7.transaction_dict = sell_order_dict_2\n order_exp_8.transaction_dict = sell_order_dict_1\n order_exp_11.transaction_dict = sell_order_dict_5\n order_exp_12.transaction_dict = sell_order_dict_5\n\n order_exp_3.shares_left = 1\n order_exp_4.shares_left = 1\n order_exp_7.shares_left = 2\n order_exp_8.shares_left = 1\n order_exp_11.shares_left = 3\n order_exp_12.shares_left = 3\n\n order_exp_11.order_status = DEFINITIVE\n order_exp_12.order_status = DEFINITIVE\n\n self.is_equal_order_wrapper(result_dict.get(3), order_exp_3)\n self.is_equal_order_wrapper(result_dict.get(4), order_exp_4)\n self.is_equal_order_wrapper(result_dict.get(7), order_exp_7)\n self.is_equal_order_wrapper(result_dict.get(8), order_exp_8)\n self.is_equal_order_wrapper(result_dict.get(11), order_exp_11)\n self.is_equal_order_wrapper(result_dict.get(12), order_exp_12)\n\n # =================================================================\n # test: share distribution int sell shares left\n # =================================================================\n\n sell_list = [StockOrderWrapper(self.order_2)]\n sell = sell_list[0]\n\n sell.shares_left = 100\n\n buy_order_3 = StockOrderWrapper(self.order_3)\n buy_order_4 = StockOrderWrapper(self.order_4)\n buy_order_7 = StockOrderWrapper(self.order_7)\n buy_order_8 = StockOrderWrapper(self.order_8)\n buy_order_11 = StockOrderWrapper(self.order_11)\n buy_order_12 = StockOrderWrapper(self.order_12)\n\n buy_order_3.shares_left = 4\n buy_order_4.shares_left = 7\n buy_order_7.shares_left = 9\n buy_order_8.shares_left = 4\n buy_order_11.shares_left = 8\n buy_order_12.shares_left = 3\n\n buy_orders = [buy_order_3, buy_order_4, buy_order_7, buy_order_8, buy_order_11, buy_order_12]\n\n result = self.bidding_round_manager.share_distributor(sell_orders=sell_list, buy_orders=buy_orders)\n\n order_exp_2 = StockOrderWrapper(self.order_2)\n order_exp_2.shares_left = 65\n\n self.is_equal_order_wrapper(result[0], order_exp_2)\n\n sell_order_dict_3 = {sell: 3}\n sell_order_dict_4 = {sell: 4}\n sell_order_dict_7 = {sell: 7}\n sell_order_dict_8 = {sell: 8}\n sell_order_dict_9 = {sell: 9}\n\n order_exp_3.transaction_dict = sell_order_dict_4\n order_exp_4.transaction_dict = sell_order_dict_7\n order_exp_7.transaction_dict = sell_order_dict_9\n order_exp_8.transaction_dict = sell_order_dict_4\n order_exp_11.transaction_dict = sell_order_dict_8\n order_exp_12.transaction_dict = sell_order_dict_3\n\n order_exp_3.shares_left = 0\n order_exp_4.shares_left = 0\n order_exp_7.shares_left = 0\n order_exp_8.shares_left = 0\n order_exp_11.shares_left = 0\n order_exp_12.shares_left = 0\n\n order_exp_3.order_status = PROCESSED\n order_exp_4.order_status = PROCESSED\n order_exp_7.order_status = PROCESSED\n order_exp_8.order_status = PROCESSED\n order_exp_11.order_status = PROCESSED\n order_exp_12.order_status = PROCESSED\n\n self.is_equal_order_wrapper(buy_order_3, order_exp_3)\n self.is_equal_order_wrapper(buy_order_4, order_exp_4)\n self.is_equal_order_wrapper(buy_order_7, order_exp_7)\n self.is_equal_order_wrapper(buy_order_8, order_exp_8)\n self.is_equal_order_wrapper(buy_order_11, order_exp_11)\n self.is_equal_order_wrapper(buy_order_12, order_exp_12)", "def _process_buy(self, base, close_arr, sample, pred_buy):\n for bt in range(len(self.btl)):\n if pred_buy >= self.btl[bt]:\n for st in range(len(self.stl)):\n if self.perf[bt, st, self.BUY_PRICE] == 0:\n # first buy of a possible sequence of multiple buys before sell\n self.perf[bt, st, self.BUY_PRICE] = close_arr[sample]\n self.perf[bt, st, self.BUY_IX] = sample", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def buyTradedVal(self, buyTradedVal):\n\n self._buyTradedVal = buyTradedVal", "async def trade(self, ctx, *args):\n if has_post_permission(ctx.guild.id, ctx.channel.id):\n if len(args) < 4:\n await ctx.send('Arguments missing. '\n 'Syntax is `~trade [name] [number] [item] [offer]`.')\n return\n\n try:\n trade = {'user1': ctx.author.id,\n 'user2': args[0],\n 'amount1': args[1],\n 'amount2': args[-1],\n 'item1': ' '.join(args[2:-1]),\n 'item2': 'coins'}\n ctx.bot.trade_manager.add_trade(ctx, trade)\n except TradeError as e:\n await ctx.send(e.msg)\n return\n\n name = args[0]\n for member in ctx.guild.members:\n if name.lower() in member.name.lower():\n name_member = member\n break\n\n offer = users.parse_int(args[-1])\n number = users.parse_int(args[1])\n itemid = items.find_by_name(' '.join(args[2:-1]))\n name = get_display_name(ctx.author)\n offer_formatted = '{:,}'.format(offer)\n out = (f'{items.SHOP_HEADER}{name.title()} wants to sell {name_member.mention} '\n f'{items.add_plural(number, itemid)} for {offer_formatted} coins. '\n f'To accept this offer, reply to this post with a :thumbsup:. '\n f'Otherwise, this offer will expire in one minute.')\n msg = await ctx.send(out)\n\n if await self.confirm(ctx, msg, out, timeout=60):\n price = {\"0\": offer}\n users.update_inventory(name_member.id, price, remove=True)\n users.update_inventory(ctx.author.id, price)\n loot = {itemid: number}\n users.update_inventory(ctx.author.id, loot, remove=True)\n users.update_inventory(name_member.id, loot)\n\n buyer_name = get_display_name(name_member)\n await ctx.send(f'{items.SHOP_HEADER}{name.title()} successfully sold '\n f'{items.add_plural(number, itemid)} to {buyer_name} for '\n f'{offer_formatted} coins!')\n ctx.bot.trade_manager.reset_trade(trade, ctx.author.id, name_member.id)", "async def _targetadd_heist(self, ctx):\r\n\r\n author = ctx.message.author\r\n guild = ctx.guild\r\n cancel = ctx.prefix + \"cancel\"\r\n check = lambda m: m.author == author and (m.content.isdigit() and int(m.content) > 0 or m.content == cancel)\r\n start = (\"This will walk-through the target creation process. You may cancel this process \"\r\n \"at anytime by typing {}cancel. Let's begin with the first question.\\nWhat is the \"\r\n \"name of this target?\".format(ctx.prefix))\r\n\r\n await ctx.send(start)\r\n name = await self.bot.wait_for('MESSAGE', timeout=35, check=lambda x: x.author == author)\r\n\r\n if name is None:\r\n await ctx.send(\"You took too long. canceling target creation.\")\r\n return\r\n\r\n if name.content == cancel:\r\n await ctx.send(\"Target creation cancelled.\")\r\n return\r\n\r\n targets = await self.thief.get_guild_targets(guild)\r\n if string.capwords(name.content) in targets:\r\n await ctx.send(\"A target with that name already exists. canceling target \"\r\n \"creation.\")\r\n return\r\n\r\n await ctx.send(\"What is the max crew size for this target? Cannot be the same as \"\r\n \"other targets.\\n*Crews over this size will go to the next \"\r\n \"largest bank.*\")\r\n crew = await self.bot.wait_for('MESSAGE', timeout=35, check=check)\r\n\r\n if crew is None:\r\n await ctx.send(\"You took too long. canceling target creation.\")\r\n return\r\n\r\n if crew.content == cancel:\r\n await ctx.send(\"Target creation cancelled.\")\r\n return\r\n\r\n if int(crew.content) in [subdict[\"Crew\"] for subdict in targets.values()]:\r\n await ctx.send(\"Group size conflicts with another target. Canceling target creation.\")\r\n return\r\n\r\n await ctx.send(\"How many starting credits does this target have?\")\r\n vault = await self.bot.wait_for('MESSAGE', timeout=35, check=check)\r\n\r\n if vault is None:\r\n await ctx.send(\"You took too long. canceling target creation.\")\r\n return\r\n\r\n if vault.content == cancel:\r\n await ctx.send(\"Target creation cancelled.\")\r\n return\r\n\r\n await ctx.send(\"What is the maximum number of credits this target can hold?\")\r\n vault_max = await self.bot.wait_for('MESSAGE', timeout=35, check=check)\r\n\r\n if vault_max is None:\r\n await ctx.send(\"You took too long. canceling target creation.\")\r\n return\r\n\r\n if vault_max.content == cancel:\r\n await ctx.send(\"Target creation cancelled.\")\r\n return\r\n\r\n await ctx.send(\"What is the individual chance of success for this target? 1-100\")\r\n check = lambda m: m.content.isdigit() and 0 < int(m.content) <= 100 or m.content == cancel\r\n success = await self.bot.wait_for('MESSAGE', timeout=35, check=check)\r\n\r\n if success is None:\r\n await ctx.send(\"You took too long. canceling target creation.\")\r\n return\r\n\r\n if success.content == cancel:\r\n await ctx.send(\"Target creation cancelled.\")\r\n return\r\n else:\r\n msg = (\"Target Created.\\n```Name: {}\\nGroup: {}\\nVault: {}\\nVault Max: \"\r\n \" {}\\nSuccess: {}%```\".format(string.capwords(name.content), crew.content,\r\n vault.content, vault_max.content,\r\n success.content)\r\n )\r\n target_fmt = {\"Crew\": int(crew.content), \"Vault\": int(vault.content),\r\n \"Vault Max\": int(vault_max.content), \"Success\": int(success.content)}\r\n targets[string.capwords(name.content)] = target_fmt\r\n await self.thief.save_targets(guild, targets)\r\n await ctx.send(msg)", "def set_target_buy_distribution_link_code(self):\n self.set_value_into_input_field(self.target_buy_distribution_link_code_textbox_locator, self.random_string_generator(6))\n self.wait_for_ajax_spinner_load()", "def update_target(self):\n pass", "def Trading(Seller,Buyer):\n if Seller.has_sold == False:\n if Buyer.like_buy >= Seller.like_sell:\n Seller.has_sold = True\n Buyer.has_bought = True\n Seller.sold_objects += 1\n Buyer.bought_objects += 1\n print('A trade has been made')\n else:\n Buyer.has_bought = False\n Seller.has_sold = False\n print('There was no deal')\n else:\n Buyer.has_bought = False" ]
[ "0.5813365", "0.57899123", "0.5756613", "0.57438093", "0.5740053", "0.5614278", "0.56005937", "0.5541609", "0.5541609", "0.55083424", "0.550767", "0.54989284", "0.54829466", "0.5476783", "0.5433824", "0.5425738", "0.54128516", "0.54126567", "0.5404454", "0.53718203", "0.53610206", "0.53357506", "0.52940077", "0.52872527", "0.52773154", "0.5275863", "0.5272356", "0.52680737", "0.5260119", "0.5255194" ]
0.68895245
0
Implementing set target buy list functionality
def set_target_buy_list(self, item_name, is_first_item): if is_first_item is True: self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True) else: self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_target_stocks_list(self, list_of_stocks):\n self.target_stocks = list_of_stocks", "def setBuyList(self, buyList):\n parsedBuyList = []\n for bought in buyList:\n if hasattr(bought, \"unitType\"):\n parsedBuyList.append(bought)\n elif isinstance(bought, dict) and u'unitType' in bought and u'territory' in bought:\n parsedBuyList.append(createBoughtUnitFromDict(bought, self.board.territories))\n else:\n raise Exception(\"Invalid buy list\", buyList)\n\n sumCost = self.costOfUnits(parsedBuyList)\n\n if sumCost <= self.board.currentCountry.money:\n self.board.buyList = parsedBuyList[:] # copy in buyList\n return True\n else:\n return False", "def set_target_buy_list_number_plan(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_number_plan_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_number_plan_kendo_dropdown_arrow_locator, item_name)", "def target_prices(self, target_prices: List[float]):\n\n self._target_prices = target_prices", "def set_buy_sell_deal_account(self, account_list):\n self.multiple_items_selection_from_kendo_dropdown(self.buy_sell_deal_account_dropdown_locator, account_list)\n self.wait_for_ajax_spinner_load()", "def add_for_target(self, target, products):\n self._products_by_target[target].update(products)", "def give(self, objlst):\n\t\tfor neuron in self.linkto:\n\t\t\tprint(str(self.id) + \" giving val to \" + str(neuron))\n\t\t\tobjlst[neuron].take(self.id, self.activation)", "def buys(self, buys):\n\n self._buys = buys", "def extra_products(self, target):\r\n return []", "def select_target_buy_list_overrides_grid_row_checkbox(self):\n for i in range(2):\n target_buy_list_overrides_grid_checkbox_locator = (By.XPATH, \"(//div[@data-grid-name='%s']/descendant::input[@type='checkbox'])[%s]\" % (self.target_buy_list_overrides_data_grid_name, str(i + 2)))\n self.click_element(target_buy_list_overrides_grid_checkbox_locator, True)", "def set_target(self, target):\n # parse target objects\n res = []\n targets = target.split(',')\n for item in targets:\n res.append(item)\n self.target = res\n \n # create conversion table for new index\n self.conversion = {}\n for i, cat in enumerate(self.target):\n self.conversion[cat] = f'{i}'", "def set_targets(self, targets: List[float]):\n self.targets = targets", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def update_target(self):\n pass", "def buy2(budget, items):\n\n # Initialize the treatment table\n TREATMENT.append([(0, 0)] * (budget+1))\n\n for i in xrange(len(items)):\n TREATMENT.append([])\n if items[i].lim == 0:\n limit = int(math.floor(budget/items[i].cost)) + 1\n else:\n limit = items[i].lim + 1\n\n for j in xrange(budget + 1):\n temp = [treat(items[i], n) +\n TREATMENT[i][j-n*items[i].cost][1]\n for n in xrange(\n min(limit, int(math.floor(j/items[i].cost)+1)))\n ]\n\n value = max(temp)\n # the index in list 'temp' is the number of item to buy\n num = temp.index(value)\n TREATMENT[i+1].append((num, value))", "def target_product(self, target_product):\n\n self._target_product = target_product", "def pick_up(self):", "def target_list(obj, product_name, slo_id):\n client = get_client(obj)\n\n product = client.product_list(name=product_name)\n if not product:\n fatal_error('Product {} does not exist'.format(product_name))\n\n product = product[0]\n\n slo = client.slo_list(product, id=slo_id)\n if not slo:\n fatal_error('SLO {} does not exist'.format(slo_id))\n\n slo = slo[0]\n\n res = client.target_list(slo)\n\n print(json.dumps(res, indent=4))", "def _process_buy(self, base, close_arr, sample, pred_buy):\n for bt in range(len(self.btl)):\n if pred_buy >= self.btl[bt]:\n for st in range(len(self.stl)):\n if self.perf[bt, st, self.BUY_PRICE] == 0:\n # first buy of a possible sequence of multiple buys before sell\n self.perf[bt, st, self.BUY_PRICE] = close_arr[sample]\n self.perf[bt, st, self.BUY_IX] = sample", "def target(self):", "def targets_placeholder(self):", "def traverseBuyBand(bandsg, high, low, day, amount, targetprofit, stoploss):\n buyprice = 0\n history = 0 # nothing, buy = 1, sell = -1\n buyindex = 0\n previous_day = 0\n ybandsg = np.empty(bandsg.size)\n ybandsg.fill(np.nan)\n for i in range(bandsg.size):\n if day[i] != previous_day: # a new day reset everything\n if history == 1:\n # the previous batch o/f signals will be saved with Nan don't want to train with that\n ybandsg[buyindex] = np.nan\n buyprice = 0\n history = 0 # nothing, buy = 1, sell = -1\n buyindex = 0\n previous_day = day[i]\n if int(bandsg[i]) == 1:\n if history == 0:\n buyprice = high[i]\n buyindex = i\n ybandsg[i] = 1 # save this buy\n else: # another buy in sequence -> cancel the first (\n # the previous batch of signals will be saved with this class (hold)\n ybandsg[buyindex] = 0 # reclassify the previous buy as hold\n # new buy signal\n buyprice = high[i]\n buyindex = i\n #print('y: ', 0)\n history=1\n # net mode\n elif int(bandsg[i]) == -1: # a sell, cancel the first buy\n ybandsg[buyindex] = 0 # reclassify the previous buy as hold\n #print('y: ', 0)\n history=0\n elif history == 1:\n profit = (low[i]-buyprice)*amount # current profit\n #print('profit: ', profit)\n if profit >= targetprofit:\n # ybandsg[buyindex] = 1 # a real (buy) index class nothing to do\n history = 0\n #print('y: ', 1)\n elif profit <= (-stoploss): # reclassify the previous buy as hold\n ybandsg[buyindex] = 0 # not a good deal, was better not to entry\n history = 0\n #print('y: ', 0)\n # reached the end of data but did not close one buy previouly open\n if history == 1: # don't know about the future cannot train with this guy\n ybandsg[buyindex] = np.nan # set it to be ignored\n return ybandsg # will only have 0 (false positive) or 1's", "def buyTradedVal(self, buyTradedVal):\n\n self._buyTradedVal = buyTradedVal", "def buy(self, buyDate, sellDate, buyLimit, unitLimit, repeat):\n\t\tstartDate, startPrice = self.getOpeningPrice()\n\t\tmaxAmount = int(buyLimit)/int(startPrice)\n\t\t\t\t\n\t\t# If you've selected a one-off stock purchase\n\t\tif repeat == 'never' and maxAmount > 0:\n\t\t\tself.purchased = (maxAmount * startPrice)\n\t\t\tself.units = maxAmount\n\t\t\tself.purchasedList.append(maxAmount * startPrice)\n\t\t\tself.unitsList.append(maxAmount)\n\t\t\tself.dateList = [startDate]\t\n\t\t\t\n\t\t# Otherwise there's some iteration required\n\t\telse:\n\t\t\tcurrentDate = startDate\n\t\t\tendDate, endPrice = self.getClosingPrice()\n\t\t\t\n\t\t\t# Iterate on the date value\n\t\t\twhile currentDate <= endDate:\n\t\t\t\tdate, price = self.getPrice(currentDate)\n\n\t\t\t\tif date is not None and price is not None:\n\t\t\t\t\tmaxAmount = int(buyLimit)/int(price)\n\t\t\t\t\t\n\t\t\t\t\t# Check there's no unit limit\t\t\t\t\n\t\t\t\t\tif maxAmount > unitLimit and unitLimit > 0:\n\t\t\t\t\t\tmaxAmount = unitLimit\n\t\t\t\t\t\n\t\t\t\t\t# If you can afford it, but the stock, otherwise skip it\n\t\t\t\t\tif maxAmount > 0:\n\t\t\t\t\t\tself.purchased += (maxAmount * price)\n\t\t\t\t\t\tself.units += maxAmount\n\t\t\t\t\t\tself.purchasedList.append(maxAmount * price)\n\t\t\t\t\t\tself.unitsList.append(maxAmount)\n\t\t\t\t\t\tself.dateList.append(date)\t\n\t\t\t\t\n\t\t\t\t# If there's no later stock data remaining break the loop\n\t\t\t\tcurrentDate = self.getNextDate(currentDate, startDate, repeat)\n\t\t\t\tif currentDate is None:\n\t\t\t\t\tbreak", "def click_target_buy_list_overrides_bulk_edit_button(self):\n self.click_element(self.target_buy_list_overrides_bulk_edit_button_locator, error_message='target buy list overrides bulk edit button locator not found before specified time out')", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def targeted(self):\n\t\tpass", "def targetFromSignals(obars, nbands=3, amount=1, targetprofit=15., stoploss=45.):\n # bandsg, yband, ask, bid, day, amount, targetprofit, stoploss\n bars = obars.copy()\n for j in range(nbands): # for each band traverse it\n ibandsg = bars.columns.get_loc('bandsg'+str(j))\n # being pessimistic ... right\n ybandsell = traverseSellBand(bars.iloc[:, ibandsg].values.astype(int),\n bars.H.values, bars.L.values, bars.date.values,\n amount, targetprofit, stoploss)\n ybandbuy = traverseBuyBand(bars.iloc[:, ibandsg].values.astype(int),\n bars.H.values, bars.L.values, bars.date.values,\n amount, targetprofit, stoploss)\n bars['y'+str(j)] = mergebandsignals(ybandsell, ybandbuy)\n\n return bars", "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def _borrow(self):\n\t\tif self.debt_link == None:\n\t\t\tchoice_list = [a for s in self.site.neighbors for a in s.agents_in_site if a.stock >= (self.consumption_demanded - self.consumed)]\n\t\t\tif len(choice_list) > 0: \n\t\t\t\tchoosed = numpy.random.choice(choice_list)\n\t\t\t\tloan_value = choosed._lend(self, self.consumption_demanded - self.consumed)\n\t\t\t\tself.consumed += loan_value\n\t\t\t\tself.consumption_deficit -= loan_value" ]
[ "0.6495771", "0.62618667", "0.60916185", "0.5946068", "0.5715191", "0.567597", "0.56602305", "0.5574808", "0.5572295", "0.5452745", "0.5394016", "0.5347834", "0.5309349", "0.53008556", "0.5293746", "0.5289541", "0.527092", "0.5261056", "0.5248171", "0.5235941", "0.51916426", "0.51317805", "0.5123512", "0.5116645", "0.51159585", "0.5112233", "0.5107676", "0.5090837", "0.50878596", "0.5078379" ]
0.66756034
0
Implementing set target buy list number plan functionality
def set_target_buy_list_number_plan(self, item_name, is_first_item): if is_first_item is True: self.single_selection_from_static_kendo_dropdown(self.target_buy_list_number_plan_kendo_dropdown_arrow_locator, first_item=True) else: self.single_selection_from_static_kendo_dropdown(self.target_buy_list_number_plan_kendo_dropdown_arrow_locator, item_name)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def plans():", "def set_target_buy_list(self, item_name, is_first_item):\n if is_first_item is True:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, first_item=True)\n else:\n self.single_selection_from_static_kendo_dropdown(self.target_buy_list_kendo_dropdown_arrow_locator, item_name)", "def set_target_buy_distribution_link_code(self):\n self.set_value_into_input_field(self.target_buy_distribution_link_code_textbox_locator, self.random_string_generator(6))\n self.wait_for_ajax_spinner_load()", "def set_target_stocks_list(self, list_of_stocks):\n self.target_stocks = list_of_stocks", "def setBuyList(self, buyList):\n parsedBuyList = []\n for bought in buyList:\n if hasattr(bought, \"unitType\"):\n parsedBuyList.append(bought)\n elif isinstance(bought, dict) and u'unitType' in bought and u'territory' in bought:\n parsedBuyList.append(createBoughtUnitFromDict(bought, self.board.territories))\n else:\n raise Exception(\"Invalid buy list\", buyList)\n\n sumCost = self.costOfUnits(parsedBuyList)\n\n if sumCost <= self.board.currentCountry.money:\n self.board.buyList = parsedBuyList[:] # copy in buyList\n return True\n else:\n return False", "def target_prices(self, target_prices: List[float]):\n\n self._target_prices = target_prices", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def set_buy_sell_deal_account(self, account_list):\n self.multiple_items_selection_from_kendo_dropdown(self.buy_sell_deal_account_dropdown_locator, account_list)\n self.wait_for_ajax_spinner_load()", "def buy2(budget, items):\n\n # Initialize the treatment table\n TREATMENT.append([(0, 0)] * (budget+1))\n\n for i in xrange(len(items)):\n TREATMENT.append([])\n if items[i].lim == 0:\n limit = int(math.floor(budget/items[i].cost)) + 1\n else:\n limit = items[i].lim + 1\n\n for j in xrange(budget + 1):\n temp = [treat(items[i], n) +\n TREATMENT[i][j-n*items[i].cost][1]\n for n in xrange(\n min(limit, int(math.floor(j/items[i].cost)+1)))\n ]\n\n value = max(temp)\n # the index in list 'temp' is the number of item to buy\n num = temp.index(value)\n TREATMENT[i+1].append((num, value))", "def _borrow(self):\n\t\tif self.debt_link == None:\n\t\t\tchoice_list = [a for s in self.site.neighbors for a in s.agents_in_site if a.stock >= (self.consumption_demanded - self.consumed)]\n\t\t\tif len(choice_list) > 0: \n\t\t\t\tchoosed = numpy.random.choice(choice_list)\n\t\t\t\tloan_value = choosed._lend(self, self.consumption_demanded - self.consumed)\n\t\t\t\tself.consumed += loan_value\n\t\t\t\tself.consumption_deficit -= loan_value", "def set_target(self, target):\n # parse target objects\n res = []\n targets = target.split(',')\n for item in targets:\n res.append(item)\n self.target = res\n \n # create conversion table for new index\n self.conversion = {}\n for i, cat in enumerate(self.target):\n self.conversion[cat] = f'{i}'", "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def create_order_con(self, target, price_list):\n\tprint()\n\tprint('OH - pl_create_order_con')\n\tprint(self)\n\tprint(target)\n\tprint(price_list)\n\n\t# Search Partner\n\tprint()\n\tprint('Search partner')\n\tpartner = self.env['res.partner'].search([\n\t\t\t\t\t\t\t\t\t\t\t\t('name', '=', self.patient.name),\n\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\tlimit=1,)\n\tprint(partner)\n\n\t# Search\n\tprint()\n\tprint('Search pricelist')\n\tpricelist = self.env['product.pricelist'].search([\n\t\t\t\t\t\t\t\t\t\t\t#('active', 'in', [True]),\n\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\t#order='x_serial_nr asc',\n\t\t\t\t\t\t\t\t\t\t\tlimit=1,\n\t\t\t\t\t\t\t\t\t\t)\n\tprint(pricelist)\n\n\n\t# Create Order\n\tprint()\n\tprint('Create order')\n\tprint(partner.id)\n\tprint(self.patient.id)\n\tprint(self.patient.x_id_doc)\n\tprint(self.patient.x_id_doc_type)\n\tprint(self.physician.id)\n\torder = self.env['sale.order'].create({\n\t\t\t\t\t\t\t\t\t\t\t'patient': self.patient.id,\n\t\t\t\t\t\t\t\t\t\t\t'x_id_doc': self.patient.x_id_doc,\n\t\t\t\t\t\t\t\t\t\t\t'x_id_doc_type': self.patient.x_id_doc_type,\n\t\t\t\t\t\t\t\t\t\t\t'x_doctor': self.physician.id,\n\t\t\t\t\t\t\t\t\t\t\t'state':'draft',\n\t\t\t\t\t\t\t\t\t\t\t'partner_id': partner.id,\n\t\t\t\t\t\t\t\t\t\t\t'x_family': 'consultation',\n\t\t\t\t\t\t\t\t\t\t\t'treatment': self.id,\n\n\t\t\t\t\t\t\t\t\t\t\t'pricelist_id': pricelist.id,\n\t\t\t\t\t\t\t\t\t\t})\n\tprint(order)\n\n\t# Init\n\t_dic_con = {\n\t\t\t\t\t'medical':\t\t'CONSULTA MEDICA',\n\t\t\t\t\t'gynecology':\t'CONSULTA GINECOLOGICA',\n\t\t\t\t\t'premium':\t\t'CONSULTA MEDICA DR. CHAVARRI',\n\t}\n\tname = _dic_con[target]\n\n\t# Search\n\tprint()\n\tprint('Search product')\n\tproduct = self.env['product.product'].search([\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t('name', 'in', [name]),\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t('pl_price_list', 'in', [price_list]),\n\t\t\t\t\t\t\t\t\t\t\t\t\t],\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t#order='date_begin asc',\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t#limit=1,\n\t\t\t\t\t\t\t\t\t\t\t\t)\n\tprint(product)\n\tprint(product.name)\n\n\t# Create Order Line\n\tprint()\n\tprint('Create order line')\n\tol = order.order_line.create({\n\t\t\t\t\t\t\t\t\t'name': \t\t\tproduct.name,\n\t\t\t\t\t\t\t\t\t'product_id': \t\tproduct.id,\n\t\t\t\t\t\t\t\t\t'order_id': \t\torder.id,\n\t\t\t\t\t\t\t\t})\n\treturn order", "def build_plan(self):\n assert False, \"Not implemented.\"", "def calculate_appropriate_target(self):\n pass", "def calculate_appropriate_target(self):\n pass", "def test_create_rate_plan(self):\n pass", "def addTargetNumber(target, number) -> str:\n targetPropeties = target.split(\"|\")\n targetPropeties[1] = str(number)\n return \" | \".join(targetPropeties)", "def set_target_buy_distribution_receiving_member(self, member_prefix):\n member_name = member_prefix + self.random_string_generator(size=4)\n self.set_value_into_input_field(self.target_buy_distribution_receiving_member_textbox_locator, member_name)\n self.wait_for_ajax_spinner_load()", "def select_target_buy_list_overrides_grid_row_checkbox(self):\n for i in range(2):\n target_buy_list_overrides_grid_checkbox_locator = (By.XPATH, \"(//div[@data-grid-name='%s']/descendant::input[@type='checkbox'])[%s]\" % (self.target_buy_list_overrides_data_grid_name, str(i + 2)))\n self.click_element(target_buy_list_overrides_grid_checkbox_locator, True)", "def __init__(self, plan):\n self.plan = plan", "def set_bill_account(self, bill_account_list):\n self.multiple_items_selection_from_kendo_dropdown(self.bill_account_kendo_dropdown_locator, bill_account_list)\n page_header_element = self.wait().until(EC.element_to_be_clickable(self.page_header_locator), 'page header locator not found before specified time out')\n page_header_element.click()", "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def test_set_plan_renewal(self):\n up = baker.make(\"UserPlan\")\n o = baker.make(\"Order\", amount=10)\n up.set_plan_renewal(order=o, card_masked_number=\"1234\")\n self.assertEqual(up.recurring.amount, 10)\n self.assertEqual(up.recurring.card_masked_number, \"1234\")\n old_id = up.recurring.id\n\n # test setting new values\n up.set_plan_renewal(order=o)\n self.assertEqual(up.recurring.amount, 10)\n self.assertEqual(up.recurring.card_masked_number, None)\n self.assertEqual(up.recurring.id, old_id)", "def get_target_market():\n return choice(GraphParameters.TARGET_MARKETS)", "def plan_targets(target, site, darkness, date, plan_range):\n targets = iop_targets.resolve_target_list(target)\n options = parse_options(site, darkness, date, plan_range)\n summarize_options(options, targets)\n\n # Setup the nights to plan and calculate sun rise/set times\n planned_nights = iop_nights.setup_nights(\n options[\"date\"], options[\"site\"], options[\"darkness\"], options[\"range\"]\n )\n\n # calculate 1 night per thread TODO: Threadding\n for night in planned_nights:\n for target in targets:\n night.plan_target(target)\n\n sched = Schedule(planned_nights)\n\n # allocate targets", "def target_product(self, target_product):\n\n self._target_product = target_product", "def setNumPurchased(self, val):\n self.numberPurchased = val", "def action_number(self, *args):\n # obj_ret = self.browse()\n if self.type == 'out_invoice':\n self._cr.execute(\n 'SELECT id, number '\n 'FROM account_wh_src '\n 'WHERE id IN (' + ','.join([str(item) for item in self.ids]) + ')')\n\n for (aws_id, number) in self._cr.fetchall():\n if not number:\n number = self.env['ir.sequence'].get(\n 'account.wh.src.%s' % self.type)\n self._cr.execute('UPDATE account_wh_src SET number=%s '\n 'WHERE id=%s', (number, aws_id))\n\n return True", "def plan(self):\n raise NotImplementedError('You must implement the plan() method '\n 'yourself!')" ]
[ "0.57999843", "0.5784943", "0.5556903", "0.55306053", "0.53336537", "0.533109", "0.5294177", "0.5224904", "0.5156525", "0.51557237", "0.5153142", "0.514415", "0.5124864", "0.51236206", "0.51054424", "0.51054424", "0.51011217", "0.50561494", "0.5020014", "0.5019802", "0.50181764", "0.50180817", "0.50160015", "0.50034314", "0.499574", "0.49928406", "0.4966621", "0.49559307", "0.49531", "0.49485675" ]
0.69643253
0
Implementing set target buy policy type functionality
def set_target_buy_policy_type(self, type): self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_target_buy_policy_name(self, prefix):\n global target_buy_policy\n target_buy_policy = (prefix + self.random_string_generator(size=4))\n self.set_value_into_input_field(self.target_buy_policy_input_field_locator, value=target_buy_policy)", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def purchase(self, item_type):", "def set_target_policy(self, policy):\n self.target_policy = policy", "def policy(agent):", "def target_type(self):", "def _setup_policy_object(policy_type,\n policy_entitlement_type,\n service_target_type,\n policy_membership_type,\n container_object,\n name: str,\n priority: int,\n description: Optional[str] = None,\n keywords: Optional[str] = None,\n caption: Optional[str] = None,\n available_to_subunits: Optional[bool] = None,\n enabled: Optional[bool] = None,\n membership_type: Optional[str] = None,\n membership_role_dns: Optional[List[str]] = None,\n entitlements: List[Dict] = []):\n\n policy_object = policy_type()\n\n if description is not None:\n policy_object['description'] = description\n policy_object['name'] = name\n\n if keywords is not None:\n policy_object['keywords'] = keywords\n\n if caption is not None:\n policy_object['caption'] = caption\n\n entitlement_list = []\n\n # Iterate through the entitlements argument and add each one to the request\n for entitlement in entitlements:\n entitlement_object = policy_entitlement_type()\n service_target_object = service_target_type()\n\n # Set type 0 for a service type (specify the name of the service profile in the name. MAKE SURE IT IS EXACT-\n # IT IS CASE_SENSITIVE).\n # Set type 1 for a specific service (specify it's DN in the name).\n # Set type 2 for all services (specify * as the name).\n # Set type 3 for a service selection policy (specify the name of the service profile in the name. MAKE SURE IT\n # IS EXACT- IT IS CASE_SENSITIVE). The service selection policy will be automatically selected based on the\n # service profile selected.\n\n if entitlement['target_type'] is not None:\n if entitlement['target_type'] == 'all':\n service_target_object['name'] = '*'\n service_target_object['type'] = '2'\n elif entitlement['target_type'] == 'type':\n service_target_object['name'] = entitlement['service_type']\n service_target_object['type'] = '0'\n elif entitlement['target_type'] == 'policy':\n service_target_object['name'] = entitlement['service_type']\n service_target_object['type'] = '3'\n elif entitlement['target_type'] == 'specific':\n service_target_object['name'] = entitlement['service_dn']\n service_target_object['type'] = '1'\n else:\n raise ValueError(\"Invalid target_type value in entitlement. Valid values are 'all', 'type', 'policy', \"\n \"or 'specific'.\")\n\n entitlement_object['serviceTarget'] = service_target_object\n\n if entitlement['automatic'] is not None:\n # The type value should be set to 0 for manual provisioning, or 1 for automatic provisioning\n if entitlement['automatic']:\n entitlement_object['type'] = 1\n else:\n entitlement_object['type'] = 0\n\n if entitlement['workflow_dn'] is not None:\n entitlement_object['processDN'] = str(entitlement['workflow_dn'])\n\n if entitlement['ownership_type'] is not None:\n if entitlement['ownership_type'].lower() == 'all':\n entitlement_object['ownershipType'] = '*'\n elif entitlement['ownership_type'].lower() == 'device':\n entitlement_object['ownershipType'] = 'Device'\n elif entitlement['ownership_type'].lower() == 'individual':\n entitlement_object['ownershipType'] = 'Individual'\n elif entitlement['ownership_type'].lower() == 'system':\n entitlement_object['ownershipType'] = 'System'\n elif entitlement['ownership_type'].lower() == 'vendor':\n entitlement_object['ownershipType'] = 'Vendor'\n else:\n raise ValueError(\"Invalid value for entitlement ownership_type. Valid values are 'all', 'device', \"\n \"'individual', 'system', or 'vendor'.\")\n\n entitlement_list.append(entitlement_object)\n\n policy_object['entitlements'] = {'item': entitlement_list}\n\n # Add membership information to the request\n membership_list = []\n membership_object = policy_membership_type()\n\n if membership_type is not None:\n # Set type 2 for all users in the organization. Specify '*' as the name.\n # Set type 3 to specify a specific role. Specify the role DN as the name. Create more membership objects for\n # more roles.\n # Set type 4 for all other users who are not granted to the entitlement(s) defined by this provisioning policy\n # via other policies. Specify '*' as the name.\n if membership_type == 'all':\n membership_object['name'] = '*'\n membership_object['type'] = '2'\n membership_list.append(membership_object)\n elif membership_type == 'other':\n membership_object['name'] = '*'\n membership_object['type'] = '4'\n membership_list.append(membership_object)\n elif membership_type == 'roles':\n for role in membership_role_dns:\n membership_object = policy_membership_type()\n membership_object['name'] = str(role)\n membership_object['type'] = '3'\n membership_list.append(membership_object)\n else:\n raise ValueError(\"Invalid value for membership_type. Valid values are 'all', 'other', or 'roles'.\")\n\n policy_object['membership'] = {'item': membership_list}\n\n if priority is not None:\n if priority < 1:\n raise ValueError(\"Invalid priority value. Priority must be an integer greater than 0.\")\n policy_object['priority'] = priority\n\n if available_to_subunits is not None:\n # Scope should be set to 1 for 'this business unit only' and 2 for 'this business unit and its subunits'\n if available_to_subunits:\n policy_object['scope'] = 2\n else:\n policy_object['scope'] = 1\n\n if container_object is not None:\n policy_object['organizationalContainer'] = container_object\n\n if enabled is not None:\n policy_object['enabled'] = enabled\n\n return policy_object", "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def make_decision_with_policy(self, policy_type, *args):\n if policy_type == 1: # ADP\n assert len(args) == 2, 'args should be exactly 2'\n cur_K = -self.K_im_traj[-1]\n distance_2_tan, radian_at_tan = args\n self.dis_sum += distance_2_tan\n pwm_l_new, pwm_r_new = policy.adp(distance_2_tan, radian_at_tan, self.dis_sum, cur_K)\n elif policy_type == 2: # pure pursuit\n l_d, sin_alpha = args\n amp = 150\n pwm_l_new, pwm_r_new = policy.pure_pursuit(l_d, sin_alpha, amp)\n elif policy_type == 3: # Car following with ADP\n assert len(args) == 3, 'args should be exactly 3'\n cur_K = -self.K_im_traj[-1]\n distance_2_tan, radian_at_tan, estimated_dis = args\n self.dis_sum += distance_2_tan\n if self.is_recording and self.counter % 100 == 0:\n np.save('./.out/record', self.record)\n pwm_l_new, pwm_r_new = policy.car_following_with_adp(distance_2_tan, radian_at_tan, self.dis_sum, cur_K, estimated_dis, self.record)\n print(self.counter)\n self.counter += 1\n elif policy_type == 4:\n K = 0.5\n dis2car, = args\n pwm_l_new, pwm_r_new = policy.car_following(dis2car, K)\n elif policy_type == 5:\n d_arc, d_curve, theta = args\n pwm_l_new, pwm_r_new = policy.adp_coupled_car_following(d_arc, d_curve, theta, self.z, self.K_coupled)\n else:\n pwm_l_new, pwm_r_new = 0, 0\n print('Policy Not Found')\n self.motor.motor_set_new_speed(pwm_l_new, pwm_r_new)", "def set_target_buy_policy_abbreviation(self, prefix):\n self.set_value_into_input_field(self.target_buy_policy_abbreviation_input_field_locator, value=(prefix + self.random_string_generator(size=4)))", "def set_route_class_to_create_new_target_buy_policy(self, route_class):\n self.single_selection_from_static_kendo_dropdown(self.route_class_single_kendo_dropdown_arrow_locator, route_class)", "def test_add_taxation_strategy_to_rate_plan(self):\n pass", "def __get_settlement_type():\n return 'Beneficial Ownership'", "def prepare_trade(self, input_, prediction):\n if prediction > input_:\n # Price will go up, so we should buy\n # amount = self.amount\n amount = self.amount\n allowed, amount_ret = self.check_risk('buy', amount)\n assert amount == amount_ret or amount == 'max', \"Mistake in check_risk function\"\n if allowed:\n return 'buy', amount_ret\n else:\n return False, amount_ret\n elif prediction < input_:\n # Sell, short or hold?\n amount = -1 * self.amount\n allowed, amount_ret = self.check_risk('buy', amount)\n assert amount == amount_ret, \"Mistake in check_risk function\"\n if allowed:\n return 'sell', amount_ret\n else:\n return False, amount_ret", "def get_target_market():\n return choice(GraphParameters.TARGET_MARKETS)", "def policy(self, s):\r\n if s.dealer_sum >= 16:\r\n return Action.STICK\r\n else:\r\n return Action.HIT", "def __init__(__self__, *,\n type: Optional[pulumi.Input['PlacementPolicyType']] = None):\n if type is not None:\n pulumi.set(__self__, \"type\", type)", "def test_create_hyperflex_proxy_setting_policy(self):\n pass", "def click_save_target_buy_policy_button(self):\n self.click_element(self.save_target_buy_policy_button_locator, error_message='save target buy policy button locator not found before specified time out')\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def update_policy(self, *args, **kwargs):\r\n pass", "def __init__(__self__, *,\n object_type: pulumi.Input[str]):\n pulumi.set(__self__, \"object_type\", 'CopyOnExpiryOption')", "def update_policy(self):\n pass", "def getPolicy(self, state):\n \"\"\"Description:\n Find all of q-values of current state, and choose the action \n with the hight q-value as optimal policy\n \"\"\"\n \"\"\" YOUR CODE HERE \"\"\"\n legalActions = self.getLegalActions(state)\n action = None\n policy = util.Counter() # use counter to store action and its q-value\n \n if len(legalActions) == 0:\n return action\n \n for a in legalActions:\n policy[a] = self.getQValue(state, a)\n action = policy.argMax()\n return action\n\n \"\"\" END CODE \"\"\"", "def trade(self, action=None):\n #print(\"Trading {}\".format(action))\n # Buy\n if action > 0.2 : self.posture = 1\n # Hold\n if action < 0.2 and action > -0.2: self.posture = 0\n # Sell\n if action < -0.2: self.posture = -1\n \n # Evaluate posture and calculare actual cost of trade\n #print(\"Posture: {}\".format(self.posture))\n if self.posture == 1:\n _amt = self.amt_buy\n _base = (_amt * self.state['price'] \\\n + (_amt * self.commission)) * -1\n \n elif self.posture == -1:\n _amt = self.amt_sell\n _base = _amt * self.state['price'] \\\n + (_amt * self.commission) \\\n + (_amt * self.gap)\n _amt = _amt * -1 \n\n # Set posture to 0 if no balance available\n if (self.posture == 1 and self.balance < abs(_base)) \\\n or (self.posture == -1 and self.bag < abs(_amt)):\n print(\"NOT enough amount!!\")\n self.stop=True\n self.posture = 0\n\n if self.posture == 0:\n _amt = 0\n _base = 0\n\n # Modify balances\n self.transaction = _base\n self.amt = _amt\n self.balance = self.balance + _base\n self.bag = self.bag + _amt\n self.value = self.calculate_value()\n #print(\"Posture : {} // Transaction: {}\".format(self.posture, self.transaction))\n\n return self", "def test_get_hyperflex_proxy_setting_policy_by_moid(self):\n pass", "def click_create_new_target_buy_policy_button(self):\n self.click_element(self.create_new_target_buy_policy_button_locator, script_executor=True, error_message='create new target buy policy button locator not found before specified time out')", "def apply_policy(self, policy):\n tenant_name = policy['tenant_name']\n fw_id = policy['fw_id']\n fw_name = policy['fw_name']\n LOG.debug(\"asa_apply_policy: tenant=%s fw_id=%s fw_name=%s\",\n tenant_name, fw_id, fw_name)\n cmds = [\"conf t\", \"changeto context \" + tenant_name]\n\n rule_dict = policy['rules']\n for rule_id in rule_dict:\n rule = rule_dict[rule_id]\n protocol = rule['protocol']\n name = rule['name']\n enabled = rule['enabled']\n dst_port = rule['destination_port']\n src_port = rule['source_port']\n\n if (rule['source_ip_address'] is not None):\n src_ip = IPNetwork(rule['source_ip_address'])\n else:\n src_ip = IPNetwork('0.0.0.0/0')\n\n if (rule['destination_ip_address'] is not None):\n dst_ip = IPNetwork(rule['destination_ip_address'])\n else:\n dst_ip = IPNetwork('0.0.0.0/0')\n\n if rule['action'] == 'allow':\n action = 'permit'\n else:\n action = 'deny'\n\n LOG.debug(\"rule[%s]: name=%s enabled=%s prot=%s dport=%s sport=%s \\\n dip=%s %s sip=%s %s action=%s\",\n rule_id, name, enabled, protocol, dst_port, src_port,\n dst_ip.network, dst_ip.netmask,\n src_ip.network, src_ip.netmask, action)\n\n acl = \"access-list \"\n acl = (acl + tenant_name + \" extended \" + action + \" \" +\n protocol + \" \")\n if (rule['source_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(src_ip.network) + \" \" + (\n str(src_ip.netmask) + \" \")\n if (src_port is not None):\n if (':' in src_port):\n range = src_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + src_port + \" \"\n if (rule['destination_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(dst_ip.network) + \" \" + \\\n str(dst_ip.netmask) + \" \"\n if (dst_port is not None):\n if (':' in dst_port):\n range = dst_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + dst_port + \" \"\n if (enabled is False):\n acl = acl + 'inactive'\n\n # remove the old ace for this rule\n if (rule_id in self.rule_tbl):\n cmds.append('no ' + self.rule_tbl[rule_id])\n\n self.rule_tbl[rule_id] = acl\n if tenant_name in self.tenant_rule:\n if rule_id not in self.tenant_rule[tenant_name]['rule_lst']:\n self.tenant_rule[tenant_name]['rule_lst'].append(rule_id)\n cmds.append(acl)\n cmds.append(\"access-group \" + tenant_name + \" global\")\n cmds.append(\"write memory\")\n\n LOG.debug(cmds)\n data = {\"commands\": cmds}\n return self.rest_send_cli(data)", "def set_order_conditions(self, df_lt_spm: pd.DataFrame, df_lt_repair: pd.DataFrame, procurement_mode: int = 1):\n\n # Set procurement types based on any match found\n self.is_spm = any([material.is_spm for material in self.materials])\n self.is_repairable = any([material.is_repairable for material in self.materials])\n self.is_buyable = any([material.is_buyable for material in self.materials])\n self.is_dismountable = any([material.is_dismountable for material in self.materials])\n\n # If no procurement type set as buyable\n self.has_procurement_type = self.is_dismountable or self.is_buyable or self.is_spm or self.is_repairable\n if not self.has_procurement_type:\n self.is_buyable = True\n\n # --------------\n\n # Set unique values (and override if needed)\n\n # If CORE VOZ, set dismountable instead of repairable\n if procurement_mode == 0:\n self.procurement_type = 'Buyable'\n self.leadtime = max([material.leadtime for material in self.materials if material.is_buyable])\n self.leadtime_sd = max([material.leadtime_sd for material in self.materials if material.is_buyable])\n return\n elif procurement_mode == 1:\n if self.domain == 'CORE VOZ' and not self.is_spm and not self.is_buyable\\\n and self.is_dismountable and self.is_repairable:\n self.procurement_type = 'Dismountable'\n self.leadtime = 90\n self.leadtime_sd = 0\n return\n\n if self.is_spm:\n self.procurement_type = 'SPM'\n\n # Override\n if (self.domain, self.brand) in df_lt_spm.index:\n try:\n new_leadtime = df_lt_spm.loc[(self.domain, self.brand)]['leadtime_spm']\n new_leadtime = float(new_leadtime)\n\n self.leadtime_override = True\n self.leadtime = new_leadtime\n self.leadtime_sd = 0\n return\n except:\n self.leadtime = 2\n self.leadtime_sd = 0\n return\n else:\n self.leadtime = 2\n self.leadtime_sd = 0\n return\n\n if self.is_repairable:\n self.procurement_type = 'Repairable'\n\n # Override\n if self.domain in df_lt_repair.index:\n try:\n new_leadtime = df_lt_repair.loc[self.domain]['leadtime_reparable']\n new_leadtime = float(new_leadtime)\n\n self.leadtime_override = True\n self.leadtime = new_leadtime\n self.leadtime_sd = 0\n return\n except:\n self.leadtime = 30\n self.leadtime_sd = 0\n return\n else:\n self.leadtime = 30\n self.leadtime_sd = 0\n return\n\n if self.is_buyable:\n self.procurement_type = 'Buyable'\n try:\n self.leadtime = max([material.leadtime for material in self.materials if material.is_buyable])\n self.leadtime_sd = max([material.leadtime_sd for material in self.materials if material.is_buyable])\n return\n except:\n self.leadtime = 90\n self.leadtime_sd = 0\n\n # Else, return Dismountable\n self.procurement_type = 'Dismountable'\n self.leadtime = 90\n self.leadtime_sd = 0\n return\n else:\n raise Exception(\"procurement_mode not valid.\")", "def set_type(self,typ):\n self._typ = typ\n if typ == 'Sine':\n self._type = 7\n elif typ == 'Sawtooth up':\n self._type = 0\n elif typ == 'Sawtooth down':\n self._type = 1\n elif typ == 'Square':\n self._type = 2\n elif typ == 'Triangle':\n self._type = 3\n elif typ == 'Pulse':\n self._type = 4\n elif typ == 'Bipolar pulse':\n self._type = 5\n elif typ == 'Sample and hold':\n self._type = 6\n else:\n print \"Unrecognized type keyword!\"\n print \"Please use only the following keywords:\"\n print \"Choices are :\"\n print \" 0. Saw up\"\n print \" 1. Saw down\"\n print \" 2. Square\"\n print \" 3. Triangle\"\n print \" 4. Pulse\"\n print \" 5. Bipolar pulse\"\n print \" 6. Sample and hold\"\n print \" 7. Modulated Sine\"\n self._type = 7\n super(self.__class__, self).setType(self, self._type):" ]
[ "0.6160238", "0.6142805", "0.5944419", "0.5795669", "0.577346", "0.5734563", "0.56611264", "0.5657722", "0.5645191", "0.5593953", "0.5516085", "0.5394442", "0.53024334", "0.5253677", "0.51763785", "0.5144148", "0.51351786", "0.51129127", "0.50925", "0.5092124", "0.5087189", "0.5056498", "0.50552475", "0.50445133", "0.5032238", "0.50256044", "0.5007574", "0.50051546", "0.4984992", "0.4982321" ]
0.8053538
0
Implementing get target buy policy grid row details functionality
def get_target_buy_policy_grid_row_details(self): self.grid_row_data.clear() self.grid_row_data = {"Status":"", "Target Buy Policy Type":"", "Call Type":"", "Route Class":""} return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_outbound_statement_grid_row_data(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary = {\"Bill Account\": \"\", \"Begin Date\": \"\", \"End Date\": \"\", \"Amount ( EUR )\": \"\"}\n self.outbound_statement_grid_row_data = self.get_grid_row_details(self.outbound_statements_grid_div_id, self.grid_row_details_dictionary)\n return self.outbound_statement_grid_row_data", "def get_transactions_grid_row_details(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary.update({\"Transaction Code\": \"\", \"Credit/Debit\": \"\", \"Transaction Details\": \"\", \"Currency\": \"\", \"Amount\": \"\"})\n return self.get_grid_row_details(self.transactions_grid_div_id, self.grid_row_details_dictionary, self.transactions_data_grid_name)", "def click_target_buy_policy_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, column_number=2, first_row=True)\n self.wait_for_ajax_spinner_load()", "def get_all_rows_of_rate_increase_violation_column_from_vendor_price_list_detail_grid(self, column_name):\n self.all_row_data = []\n self.wait_for_ajax_spinner_load()\n row_count = len(self.wait().until(EC.presence_of_all_elements_located(self.vendor_price_list_detail_row_locator), 'vendor price list details row element locator not found before specified time'))\n next_page = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (self.vendor_price_list_detail_rates_grid_div_id, column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n row = 1\n try:\n while row < row_count + 1:\n column_value_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tbody[@role='rowgroup']/tr[%s]/td[%s]\" % (self.vendor_price_list_detail_rates_grid_div_id, row, str(column_index)))\n column_value_element = self.wait(10).until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n column_value = str(column_value_element.get_attribute(\"innerHTML\"))\n if column_value not in self.all_row_data:\n self.all_row_data.append(column_value)\n else:\n row_count += 1\n if row % 50 == 0:\n next_page = True\n next_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-arrow-e']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(next_page_button_locator, True)\n row = 0\n row_count -= 50\n row += 1\n except WebDriverException:\n pass\n if next_page is True:\n first_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-seek-w']\" % self.vendor_price_list_detail_rates_grid_div_id)\n self.click_element(first_page_button_locator, True)\n return self.all_row_data", "def get_received_charges_grid_row_details(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary.update({\"Tariff Type\": \"\", \"Line Item Type\": \"\", \"Service Group\": \"\", \"Product\": \"\", \"Time Band\": \"\", \"Begin Date\": \"\", \"End Date\": \"\", \"Rating Unit\": \"\", \"Minutes/Quantity\": \"\", \"Rate\": \"\"})\n return self.get_grid_row_details(self.received_charges_grid_div_id, self.grid_row_details_dictionary, self.received_charges_data_grid_name)", "def get_disputes_grid_row_details(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary.update({\"Statement Begin Date\": \"\", \"Statement End Date\": \"\", \"Statement Number\": \"\", \"Dispute Type\": \"\", \"Dispute Amount\": \"\", \"Status\": \"\"})\n return self.get_grid_row_details(self.disputes_grid_div_id, self.grid_row_details_dictionary)", "def get_vendor_profile_page_grid_row_details(self, div_id, row_data):\n for key in row_data.keys():\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (div_id, str(key)))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\"))+1\n column_value_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tr/td[%s]\" % (div_id, str(column_index)))\n column_value_element = self.wait().until(EC.presence_of_element_located(column_value_locator), 'column value not found before specified time out')\n column_value = column_value_element.text\n row_data[key] = str(column_value)\n return row_data", "def click_target_buy_policies_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, self.target_buy_policies_grid_inline_action_column_number, first_row=True)", "def get_details(self):", "def get_charges_grid_row_details(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary.update({\"Charge Begin Date\": \"\", \"Charge End Date\": \"\", \"Statement Number\": \"\", \"Rate Type\": \"\", \"Rating Quantity\": \"\", \"Adjustment Statement\": \"\"})\n return self.get_grid_row_details(self.charges_grid_div_id, self.grid_row_details_dictionary, self.charges_grid_data_grid_name)", "def rosterRowData(self):", "def abc_create_row(self, row):\n # Lifted from action_assign on stock.move\n product = self.env['product.product'].browse(row['product_id'])\n location = self.location_id\n main_domain = [('reservation_id', '=', False), ('qty', '>', 0)]\n quants = self.env['stock.quant'].quants_get_prefered_domain(\n location,\n product,\n row['quantity'] or 1.0,\n domain=main_domain,\n prefered_domain_list=[])\n # Find suggested location. Could be more than one given quantity > 0.\n # All that stuff remains to be solved.\n for quant in quants:\n if quant[0]:\n location = quant[0].location_id\n row.update({\n '_name': 'stock.transfer_detailsitems',\n 'product_id': self.abc_make_records(product, ['display_name'])[0],\n 'is_offer' : self.abc_make_records(product, ['is_offer'])[0],\n 'destinationloc_id': self.abc_make_records(self.location_dest_id)[0],\n 'sourceloc_id': self.abc_make_records(location)[0],\n 'product_uom_id': self.abc_make_records(product.uom_id)[0],\n })\n return row", "def get_non_traffic_charges_grid_row_details(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary.update({\"Accounting Code\": \"\", \"Start Date\": \"\", \"End Date\": \"\", \"Charge Name\": \"\", \"Charge Category\": \"\", \"Currency\": \"\", \"Amount\": \"\", \"Tax Template\": \"\"})\n non_traffic_charges_grid_row_details = self.get_grid_row_details(self.non_traffic_charges_grid_div_id, self.grid_row_details_dictionary)\n if \",\" in non_traffic_charges_grid_row_details[\"Amount\"]:\n non_traffic_charges_grid_row_details[\"Amount\"] = non_traffic_charges_grid_row_details[\"Amount\"].replace(\",\", \"\")\n return non_traffic_charges_grid_row_details", "def get_target_buy_policy(self):\n global target_buy_policy\n return target_buy_policy", "def get_inbound_statement_grid_row_details(self, unique_identifier):\n self.inbound_statement_grid_row_data.clear()\n self.inbound_statement_grid_row_data = {\"Audit Account\": \"\", \"Statement Number\": \"\", \"Begin Date\": \"\", \"End Date\": \"\", \"Amount\": \"\", \"Final Payment Due Date\": \"\"}\n grid_row_data = self.get_grid_row_details_with_unique_identifier(self.inbound_statement_grid_div_id, self.inbound_statement_grid_row_data, unique_identifier)\n amount_from_grid = str(grid_row_data[\"Amount\"]).replace(\",\", \"\")\n grid_row_data[\"Amount\"] = amount_from_grid\n return grid_row_data", "def get_display_price_data(self, source, commitment):\n overage, included = self.get_price_data(source, commitment)\n if self.name == settings.BILLING_DEFAULT_PLAN_NAME:\n included = OFFICIAL_BUILDER_LIMITS[source]\n return overage, included", "def test_get_rate_plan_by_product(self):\n pass", "def get_rows(self):\n rows = []\n common_fields = self.get_common_data()\n\n p_identities = self.data['participantIdentities']\n p_data = self.data['participants']\n winning_team = self.get_winning_team()\n items_purchased = self.get_items_purchased()\n team_champions = self.get_team_champions()\n teams = set(team_champions.keys())\n gold_per_player = self.get_gold_per_player()\n xp_per_player = self.get_xp_per_player()\n gold_diff = self.get_gold_diff()\n xp_diff = self.get_xp_diff()\n dmg_taken_diff = self.get_dmg_taken_diff()\n\n for p in p_identities:\n p_id = int(p['participantId'])\n p_idx = p_id - 1\n team_id = p_data[p_idx]['teamId']\n opposing_team_id = (teams - {team_id}).pop()\n player_purchases = items_purchased[p_idx]\n purchase_list = [item_pair[0] for item_pair in player_purchases]\n items_10min = {\n item_pair[0] for item_pair in player_purchases\n if item_pair[1] < 1000*60*10\n }\n cur_row = {\n 'summonerId': p['player']['summonerId'],\n\n 'allyChampions': team_champions[team_id],\n 'championId': p_data[p_idx]['championId'],\n 'gold5': gold_per_player[p_idx].get(5, None),\n 'gold10': gold_per_player[p_idx].get(10, None),\n 'gold15': gold_per_player[p_idx].get(15, None),\n 'gold20': gold_per_player[p_idx].get(20, None),\n 'xp5': xp_per_player[p_idx].get(5, None),\n 'xp10': xp_per_player[p_idx].get(10, None),\n 'xp15': xp_per_player[p_idx].get(15, None),\n 'xp20': xp_per_player[p_idx].get(20, None),\n 'items10': list(items_10min),\n 'opponentChampions': team_champions[opposing_team_id],\n 'purchases': purchase_list,\n 'lane': p_data[p_idx]['timeline']['lane'],\n 'role': p_data[p_idx]['timeline']['role'],\n 'xpDiff10': xp_diff[p_idx],\n 'damageTakenDiff10': dmg_taken_diff[p_idx],\n 'gdPerMin10': gold_diff[p_idx],\n 'winner': (team_id == winning_team),\n }\n\n cur_row.update(common_fields)\n rows.append(cur_row)\n\n return rows", "def get_all_rows_of_specific_column_only_from_vendor_price_list_detail_grid(self, div_id, column_name):\n row_count = len(self.wait().until(EC.presence_of_all_elements_located(self.vendor_price_list_detail_row_locator), 'vendor price list details row element locator not found before specified time'))\n all_row_data = []\n next_page = False\n column_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::th[@data-title='%s']\" % (div_id, column_name))\n column_element = self.wait().until(EC.presence_of_element_located(column_locator), 'column locator not found before specified time out')\n column_index = int(column_element.get_attribute(\"data-index\")) + 1\n row = 1\n try:\n while row < row_count+1:\n column_value_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::div[@class='k-grid-content']/descendant::tbody[@role='rowgroup']/tr[%s]/td[%s]/descendant::td\" % (div_id, row, str(column_index)))\n column_value_element = self.wait(10).until(EC.presence_of_element_located(column_value_locator), 'column value locator not found before specified time out')\n column_value = str(column_value_element.get_attribute(\"innerHTML\"))\n if column_value not in all_row_data:\n all_row_data.append(column_value)\n else:\n row_count += 1\n if row % 50 == 0:\n next_page = True\n next_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-arrow-e']\" % div_id)\n self.click_element(next_page_button_locator, True)\n row = 0\n row_count -= 50\n row += 1\n except WebDriverException:\n pass\n if next_page is True:\n first_page_button_locator = (By.XPATH, \"//div[contains(@id, '%s')]/descendant::span[@class='k-icon k-i-seek-w']\" % div_id)\n self.click_element(first_page_button_locator, True)\n return all_row_data", "def select_target_buy_policies_tab(self):\n self.select_static_tab(self.target_buy_policies_tab_locator, 'target buy policies tab locator not found before specified time out')", "def get_add_disputes_grid_row_details(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary.update({\"Statement Begin Date\": \"\", \"Statement End Date\": \"\", \"Statement Number\": \"\", \"Dispute Type\": \"\", \"Dispute Amount\": \"\", \"Status\": \"\"})\n return self.get_grid_row_details(self.add_disputes_grid_div_id, self.grid_row_details_dictionary)", "def select_target_buy_list_overrides_grid_row_checkbox(self):\n for i in range(2):\n target_buy_list_overrides_grid_checkbox_locator = (By.XPATH, \"(//div[@data-grid-name='%s']/descendant::input[@type='checkbox'])[%s]\" % (self.target_buy_list_overrides_data_grid_name, str(i + 2)))\n self.click_element(target_buy_list_overrides_grid_checkbox_locator, True)", "def item_view_bought(request):\n\n r = {}\n u = request.user\n\n p = Product.objects.get_by_sku(request.POST['sku'])\n if p is not None:\n #r = p.details(u)\n\n if u.experiment.id in [1,3]:\n purchases = TransactionLineItem.objects.filter(product=p).exclude(transaction__party=u)\n r['people'] = [pu.transaction.party.get_json() for pu in purchases]\n else:\n purchases = TransactionLineItem.objects.filter(product=p, transaction__party__in=u.friends()).exclude(transaction__party=u)\n r['people'] = [pu.transaction.party.get_json(level=1) for pu in purchases]\n\n #reviews = Review.objects.filter(product=p)\n #r['reviews'] = {'count': str(reviews.count()),\n # 'reviews': [rev.get_json(me=u) for rev in reviews]}\n else:\n r['result'] = '0'\n\n return JSONHttpResponse(r)", "def test_get_rate_plan_by_product_and_rate_plan(self):\n pass", "def getRows(self, context, obj, data, aggData):\n raise NotImplementedError()", "def execQ4():\n frame = pan.DataFrame(data, columns=['Product', 'Price', 'Period'])\n tuna = frame[dFrame.Series_title_1 == \"Tuna - canned (supermarket only), 185g\"]\n expensive = tuna.sort_values(by=\"Price\", ascending=False).head(1)\n return expensive", "def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)", "def test_get_small_and_light_eligibility_by_seller_sku(self):\n pass", "def get_resolutions_grid_row_data(self):\n self.grid_row_details_dictionary.clear()\n self.grid_row_details_dictionary = {\"Resolution Type\": \"\", \"Discrepancy Source\": \"\", \"Resolution Amount\": \"\", \"Description\": \"\"}\n resolutions_grid_row_data = self.get_grid_row_details(self.resolution_grid_div_id, self.grid_row_details_dictionary)\n resolutions_grid_row_data[\"Resolution Amount\"] = resolutions_grid_row_data[\"Resolution Amount\"].replace(\",\", \"\")\n return resolutions_grid_row_data", "def get_details_of_item(self, item):\n row = {}\n try:\n with Transaction().start(DBNAME, 1):\n product = self.Product.search([('name', '=', item),\n ('description', '=', 'Stock'), ('type', '=', 'goods')])[-1]\n row['code'] = product.code\n row['category'] = product.template.category.name\n row['units'] = product.template.default_uom.name\n row['rate'] = product.template.list_price.to_eng()\n suppliers = product.template.product_suppliers\n if suppliers:\n row['supplier'] = suppliers[0].party.name\n return row\n except Exception:\n if settings.level == 10:\n logger.exception('raised exception')\n return row" ]
[ "0.575327", "0.56671774", "0.54983485", "0.5497504", "0.54278195", "0.53913313", "0.5331615", "0.5325904", "0.5314808", "0.53105533", "0.5297771", "0.52571106", "0.52271354", "0.52199143", "0.519113", "0.5143383", "0.5140523", "0.5123071", "0.51150763", "0.50886685", "0.50832075", "0.5015577", "0.5013142", "0.5009531", "0.49823964", "0.49637672", "0.48990068", "0.4893585", "0.48865873", "0.48730484" ]
0.8331571
0
Implementing click target buy policy grid first row inline action button functionality
def click_target_buy_policy_grid_first_row_inline_action_button(self): self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, column_number=2, first_row=True) self.wait_for_ajax_spinner_load()
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_target_buy_policies_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, self.target_buy_policies_grid_inline_action_column_number, first_row=True)", "def click_buy_and_sell_deal_management_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_inline_action_column_number, True)", "def click_vendor_price_list_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.vendor_price_list_grid_div_id, None, self.view_price_list_column_number, True)", "def click_adjustment_charges_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.adjustment_charges_grid_div_id, None, self.adjustment_charges_grid_inline_action_column_number, True)", "def click_outbound_statement_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, None, self.outbound_statements_grid_inline_action_column_number, True)", "def click_inbound_statement_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.inbound_statement_grid_div_id, None, self.inbound_statement_grid_inline_action_column_number, True)", "def click_vendors_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.vendors_div_id, None, self.grid_column_number, True)", "def click_outbound_statements_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, None, self.outbound_statements_grid_inline_action_column_number, True)", "def click_disputes_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.statement_details_dispute_grid_div_id, None, self.disputes_grid_inline_action_column_number, True)", "def click_adjustments_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.adjustments_grid_div_id, None, self.adjustments_grid_inline_action_column_number, True)", "def click_line_item_disputes_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.line_item_disputes_grid_div_id, None, self.line_item_disputes_grid_inline_action_column_number, True)", "def click_re_analysis_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.re_analysis_grid_div_id, None, self.re_analysis_grid_inline_action_column_number, True)", "def click_statement_validation_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.statement_validation_grid_div_id, None, self.statement_validation_grid_inline_action_column_number, True)", "def click_view_price_list_detail_first_row_inline_action_button(self):\n self.click_inline_action_button(self.view_price_list_div_id, None, self.view_price_list_column_number, True)\n self.wait_for_ajax_spinner_load()", "def click_resolution_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.resolution_grid_div_id, None, self.resolution_grid_inline_column_number, True)", "def click_inbound_statement_grid_inline_action_button(self, row_identifier_text):\n self.click_inline_action_button(self.inbound_statement_grid_div_id, row_identifier_text, self.inbound_statement_grid_inline_action_column_number)", "def click_buy_page_inline_action_button(self, vendor):\n self.click_inline_action_button(self.vendors_div_id, vendor, self.grid_column_number)", "def click_outbound_statements_grid_inline_action_button(self, row_identifier_text):\n self.click_inline_action_button(self.outbound_statements_grid_div_id, row_identifier_text, self.outbound_statements_grid_inline_action_column_number)", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def select_buy_and_sell_deal_management_grid_first_row_checkbox(self):\n self.select_grid_row_checkbox(self.buy_and_sell_management_grid_div_id, None, self.buy_and_sell_management_grid_checkbox_column_number, True)", "def click(self):\r\n pass", "def set_shortfall_grid_actions(self, action_item):\n self.single_selection_from_kendo_in_grid(self.shortfall_grid_div_id, self.actions_column_name, action_item)", "def click_statement_fields_grid_first_row_checkbox(self):\n self.click_element(self.statement_fields_grid_first_row_checkbox_locator, True)", "def get_target_buy_policy_grid_row_details(self):\n self.grid_row_data.clear()\n self.grid_row_data = {\"Status\":\"\", \"Target Buy Policy Type\":\"\", \"Call Type\":\"\", \"Route Class\":\"\"}\n return self.get_grid_row_details(self.target_buy_policies_grid_div_id, self.grid_row_data, data_grid_name=self.target_buy_policy_data_grid_name)", "def click_view_price_list_detail_page_inline_action_button(self, price_list_item):\n self.click_inline_action_button(self.view_price_list_div_id, price_list_item, self.view_price_list_column_number)\n self.wait_for_ajax_spinner_load()", "def select_adjustment_charge_grid_first_row_checkbox(self):\n self.click_element(self.adjustment_charges_grid_first_row_checkbox_locator, True, True)", "def _ClickPrimaryActionButton(self):\n self._ExecuteOobeApi('Oobe.clickGaiaPrimaryButtonForTesting')", "def select_bill_and_audit_module_grid_inline_item(self, item_name):\n self.select_inline_action_item(item_name, module_name = self.bill_audit_module_name)\n self.wait_for_ajax_spinner_load()", "def OnLeftUp_ClickButton(self, event):\r\n \r\n self._hover_button = None\r\n\r\n if self._action_part:\r\n self.RefreshButton(self._action_part)\r\n\r\n # make sure we're still over the item that was originally clicked\r\n if self._action_part == self.HitTest(*event.GetPosition()):\r\n \r\n # fire button-click event\r\n e = AuiManagerEvent(wxEVT_AUI_PANE_BUTTON)\r\n e.SetManager(self)\r\n e.SetPane(self._action_part.pane)\r\n e.SetButton(self._action_part.button.button_id)\r\n self.ProcessMgrEvent(e)" ]
[ "0.8202686", "0.80335456", "0.74357104", "0.74149716", "0.72845346", "0.72829384", "0.72605", "0.72499174", "0.7216867", "0.7202817", "0.70367014", "0.69755983", "0.68936044", "0.68760645", "0.6781373", "0.6636045", "0.6578328", "0.6401699", "0.636651", "0.6061604", "0.60248244", "0.58237", "0.5816428", "0.5729777", "0.5721168", "0.56961125", "0.5690402", "0.5673926", "0.56108594", "0.55957127" ]
0.8149763
1
Implementing click edit target buy policy button functionality
def click_edit_target_buy_policy_button(self): self.click_element(self.edit_target_buy_policy_button_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def click_cancel_edited_target_buy_policy_button(self):\n self.click_element(self.cancel_edited_target_buy_policy_button_locator)", "def click_save_edited_target_buy_policy_button(self):\n self.click_element(self.save_edited_target_buy_policy_button_locator)\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def click_save_target_buy_policy_button(self):\n self.click_element(self.save_target_buy_policy_button_locator, error_message='save target buy policy button locator not found before specified time out')\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def click_target_buy_list_overrides_bulk_edit_button(self):\n self.click_element(self.target_buy_list_overrides_bulk_edit_button_locator, error_message='target buy list overrides bulk edit button locator not found before specified time out')", "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def save_edit(self):\r\n self.driver.find_element_by_xpath('//*[@id=\"vnos\"]/span[1]/div/a').click()", "def click_bulk_edit_button(self):\n self.click_element(self.bulk_edit_button_locator, True)", "def edit_tools(self, e):\n #GETTING SELECTION\n\n self.selected_item = self.user_inventory.selection()\n self.select_name = self.user_inventory.item([i for i in self.selected_item], \"values\")[0]\n self.select_entdate = self.user_inventory.item([i for i in self.selected_item], \"values\")[3]\n\n self.df_same_name = self.df_user.query(\"title == @self.select_name\")\n #this is the selected one for sure\n self.df_the_selected_item = self.df_same_name.loc[self.df_same_name[\"entry date\"] == self.select_entdate]\n\n #GETTING THE INDEX NUMBER OF THE SELECTION IN .CSV FILE\n self.index_select = self.df_the_selected_item.index\n self.index_select_number = self.index_select.tolist()\n\n #bottom buttons appear:\n self.changing_item_label.config(text=\"Now editing \"+self.select_name+\" that added on \"+self.select_entdate+\":\")\n\n self.delete_but = Button (self.bottom_frame, text=\"DELETE\", command=self.delete_button)\n self.delete_but.place(relx=0.1, rely=0.7, relwidth=0.28, anchor=\"w\")\n\n self.servings_list = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]\n self.serv_drop = Combobox(self.bottom_frame, value=self.servings_list, state=\"readonly\")\n self.serv_drop.place(relx=0.5, rely=0.7, relwidth=0.2, anchor=CENTER)\n\n \n self.serv_but = Button(self.bottom_frame, text=\"CHANGE AMOUNT\", command=self.change_amount_button, state=\"disabled\")\n self.serv_but.place(relx=0.9, rely=0.7, relwidth=0.28, anchor=\"e\")\n\n self.serv_drop.bind(\"<<ComboboxSelected>>\", self.activate_button)", "def click_the_edit_button_that_appears(driver):\n driver.find_element_by_xpath(xpaths.users.eric_Edit_Button).click()", "def click_statement_fields_bulk_edit_button(self):\n self.click_element(self.statement_fields_bulk_edit_button_locator, True)", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def click_bulk_edit_submit_button(self):\n self.click_element(self.bulk_edit_submit_button_locator)", "def click_bulk_edit_submit_button(self):\n self.click_element(self.bulk_edit_submit_button_locator, True)", "def edit(self, **kwargs):\n ...", "def edit(self):\n\n pass", "def edit_event(self):\n WebDriverWait(self.driver, 20).until(expected_conditions.presence_of_element_located((By.XPATH, '/html/body/div[4]/div[3]/div/div[2]/button'))).click()\n\n #elem = self.driver.find_elements(By.XPATH, \"/html/body/div[1]/div/div[3]/div/div[2]/div/div[2]/div[2]/div/div/div/div/div/div[11]/div/a/div/div[4]/div[2]/div/a[3]\")\n elem = self.driver.find_element_by_xpath(\"/html/body/div[1]/div/div[3]/div/div[2]/div/div[2]/div[2]/div/div/div/div/div/div[11]/div/a/div/div[4]/div[2]/div/a[3]\")\n actions = ActionChains(self.driver)\n actions.move_to_element(elem).perform()\n sleep(5)\n WebDriverWait(self.driver, 20).until(expected_conditions.presence_of_element_located((By.XPATH, \"/html/body/div[1]/div/div[3]/div/div[2]/div/div[2]/div[2]/div/div/div/div/div/div[11]/div/a/div/div[4]/div[2]/div/a[3]\"))).click()", "def edit_button_clicked(self, obj):\n handle = self.get_selected()\n if handle:\n note = self.dbstate.db.get_note_from_handle(handle)\n try:\n from .. import EditNote\n EditNote(self.dbstate, self.uistate, self.track, note,\n callertitle = self.callertitle,\n extratype = [self.notetype] )\n except WindowActiveError:\n pass", "def edit():", "def update_policy(self, *args, **kwargs):\r\n pass", "def pressS(self):\n self.myParent.mode.modifyIndustry(self.myParent.mySystemDict['id'], self.currentValue, self.myIndustryData.id)\n self.disableButton('S')", "def click_create_new_target_buy_policy_button(self):\n self.click_element(self.create_new_target_buy_policy_button_locator, script_executor=True, error_message='create new target buy policy button locator not found before specified time out')", "def click_request_new_deal_button(self):\n self.click_element(self.request_new_deal_button_locator)", "def change_amount_button(self):\n self.pop_up_amount = Toplevel(master)\n self.pop_up_amount.geometry(\"500x50\")\n\n self.select_amo = self.user_inventory.item([i for i in self.selected_item], \"values\")[2]\n\n self.del_label = Label(self.pop_up_amount, text=\"Are you sure you want to change servings amount from \"+self.select_amo+\" to \"+self.serv_drop.get()+\"?\", font=\"roboto 12\")\n self.del_label.place(relx=0.5, rely=0.01, anchor=\"n\")\n\n self.change_button = Button(self.pop_up_amount, text=\"OK\", command=self.change_amount_incsv)\n self.change_button.place(relx=0.4, rely=0.5, anchor=\"n\")\n\n self.nochange_button = Button(self.pop_up_amount, text=\"CANCEL\", command=self.close_2)\n self.nochange_button.place(relx=0.6, rely=0.5, anchor=\"n\")", "def click_target_buy_policy_grid_first_row_inline_action_button(self):\n self.click_inline_action_button(self.target_buy_policies_grid_div_id, None, column_number=2, first_row=True)\n self.wait_for_ajax_spinner_load()", "def click_target_buy_overrides_button(self):\n self.click_element(self.target_buy_overrides_button_locator, script_executor=True, error_message='target buy overrides button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def editChange(self,editBtn):\n if self.edit ==True:\n self.updateDetails(\"normal\") #update details column\n self.edit = False #switch boolean\n self.editBtn.setText(\"Edit\") #update button text\n else:\n self.updateDetails(\"edit\") #update details column\n self.edit= True #switch boolean\n self.editBtn.setText(\"Stop Editing\") #update button text", "def click_buy_page_inline_action_button(self, vendor):\n self.click_inline_action_button(self.vendors_div_id, vendor, self.grid_column_number)", "def click_buy_and_sell_deal_save_button(self):\n self.click_element(self.save_vendor_profile_locator, True)\n try:\n self.wait().until(EC.visibility_of_element_located(self.confirmation_popup_locator), 'confirmation popup locator not found before specified time out')\n self.click_element(self.ok_button_locator, True)\n except:\n raise", "def click_bulk_edit_cancel_button(self):\n self.click_element(self.bulk_edit_cancel_button_locator, True)" ]
[ "0.72049874", "0.71058124", "0.6989213", "0.6368108", "0.631195", "0.6140118", "0.61324686", "0.60960543", "0.6026939", "0.58555514", "0.585286", "0.57779235", "0.5776125", "0.5776095", "0.5768035", "0.57594347", "0.55886054", "0.55702305", "0.5557164", "0.55496216", "0.55267096", "0.5498815", "0.54957706", "0.5482461", "0.54742956", "0.54660165", "0.54615057", "0.54261005", "0.54222685", "0.5419801" ]
0.8397314
0
Implementing click save edited target buy policy button functionality
def click_save_edited_target_buy_policy_button(self): self.click_element(self.save_edited_target_buy_policy_button_locator) self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out') self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out') self.wait_for_ajax_spinner_load()
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def click_save_target_buy_policy_button(self):\n self.click_element(self.save_target_buy_policy_button_locator, error_message='save target buy policy button locator not found before specified time out')\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def click_cancel_edited_target_buy_policy_button(self):\n self.click_element(self.cancel_edited_target_buy_policy_button_locator)", "def click_buy_and_sell_deal_bulk_edit_button(self):\n self.click_element(self.buy_and_sell_deal_bulk_edit_button_lcoator, True)", "def click_buy_and_sell_deal_create_button(self):\n self.click_element(self.save_vendor_profile_locator)", "def click_buy_and_sell_deal_save_button(self):\n self.click_element(self.save_vendor_profile_locator, True)\n try:\n self.wait().until(EC.visibility_of_element_located(self.confirmation_popup_locator), 'confirmation popup locator not found before specified time out')\n self.click_element(self.ok_button_locator, True)\n except:\n raise", "def save_edit(self):\r\n self.driver.find_element_by_xpath('//*[@id=\"vnos\"]/span[1]/div/a').click()", "def saveMenuHandler(self, action):\n button_text = action.text()\n\n if button_text == 'Save As':\n subsystem_controller = self.getCurrentSubsystemController()\n self.saveAsHandler(subsystem_controller)\n\n elif button_text == 'Save':\n subsystem_controller = self.getCurrentSubsystemController()\n self.saveHandler(subsystem_controller)\n\n elif button_text == 'Save As Scenario':\n self.saveScenarioHandler()", "def save_change(self):\n try:\n income = abs(float(self.listOfIncomeSEK.item(1).text()))\n fixed_expenses = {\n \"rent\": abs(float(self.listOfExpensesSEK.item(4).text())),\n \"subscription\": abs(float(self.listOfExpensesSEK.item(2).text())),\n \"insurance\": abs(float(self.listOfExpensesSEK.item(3).text())),\n \"others\": abs(float(self.listOfExpensesSEK.item(5).text()))\n }\n variable_expenses = {\n \"food\": abs(float(self.listOfExpensesSEK.item(11).text())),\n \"bills\": abs(float(self.listOfExpensesSEK.item(12).text())),\n \"transportation\": abs(float(self.listOfExpensesSEK.item(13).text())),\n \"hygien\": abs(float(self.listOfExpensesSEK.item(14).text())),\n \"clothes\": abs(float(self.listOfExpensesSEK.item(15).text())),\n \"entertainment\": abs(float(self.listOfExpensesSEK.item(16).text())),\n \"others\": abs(float(self.listOfExpensesSEK.item(17).text()))\n }\n self.customer.budget.set_budget(income, variable_expenses, fixed_expenses)\n DB.set_variable_expenses(self.customer.email, variable_expenses)\n DB.set_fixed_expenses(self.customer.email, fixed_expenses)\n DB.not_new_customer(self.customer.email)\n DB.update_income(income, self.customer.email)\n\n self.displayUi = MenuScreen()\n self.hide()\n self.displayUi.show()\n except Exception:\n self.popUp.exec_()", "def save(self, *args, **kwargs):\n wallet = self.wallet.withdraw(self.value)\n super(Payment, self).save(*args, **kwargs)", "def click_assign_adjustment_charge_save_button(self):\n self.click_element(self.assign_adjustment_charge_save_button_locator)\n try:\n self.wait().until(EC.visibility_of_element_located(self.error_log_viewer_locator), \"error log viewer locator not found before specified time out\")\n self.click_element(self.error_log_viewer_close_button_locator)\n self.click_element(self.assign_adjustment_charge_save_button_locator)\n except:\n pass\n try:\n self.wait().until(EC.visibility_of_element_located(self.success_message_locator), 'success message locator not found before specified time out')\n self.click_element(self.ok_button_locator)\n except:\n raise", "def click_create_new_target_buy_policy_button(self):\n self.click_element(self.create_new_target_buy_policy_button_locator, script_executor=True, error_message='create new target buy policy button locator not found before specified time out')", "def _save(self):\n\n # Get option\n if self.button1.isChecked():\n option = 'Steunpunt'\n uncertainties = self.supportloc_unc_table\n elif self.button2.isChecked():\n # Check if the harboruncertainties are filled:\n if pd.isnull(self.harbor_unc_table.model._data).any().any():\n raise ValueError('Niet alle modelonzekerheden voor het havenmodel zijn ingevuld.')\n option = 'Havenmodel'\n uncertainties = self.harbor_unc_table\n elif self.button3.isChecked():\n if pd.isnull(self.combined_unc_table.model._data).any().any():\n raise ValueError('De gecombineerde modelonzekerheden zijn nog niet berekend.')\n option = 'Combinatie'\n uncertainties = self.combined_unc_table\n else:\n raise ValueError('Selecteer een optie voor de te gebruiken onzekerheid')\n\n self.parent.adjust_selection(uncertainties=uncertainties, option=option)\n\n # Toegevoegd Svasek 31/10/2018 - Sluit het onzekerheden input scherm als er op opslaan gedrukt wordt\n self.close()", "def save(self):\n self.click(\".action-save\")\n self.page.wait_for_ajax()", "def on_save_button_clicked(self, widget):\n active_page = self.get_active_tab()\n active_page.save()", "def tool_save_target_info_clicked(self, widget, data=None):\n self._save_target_info()", "def set_target_buy_policy_type(self, type):\n self.single_selection_from_kendo_dropdown(self.target_buy_policy_type_kendo_dropdown_locator, type)", "def add_option_save(self):\n logger.debug(\"Adding save option\")\n btnsave = ttk.Button(self.optsframe,\n image=get_images().icons[\"save\"],\n command=self.save_items)\n btnsave.pack(padx=2, side=tk.RIGHT)\n Tooltip(btnsave,\n text=\"Save {}(s) to file\".format(self.tabname),\n wraplength=200)", "def click_the_save_button_which_should_be_returned_to_the_storage_page(driver):\n assert wait_on_element(driver, 5, '//button[contains(.,\"Save Access Control List\")]', 'clickable')\n driver.find_element_by_xpath('//button[contains(.,\"Save Access Control List\")]').click()\n time.sleep(1)\n assert wait_on_element_disappear(driver, 30, '//h6[contains(.,\"Please wait\")]')", "def pricing_save(request, simulation):\n # Retrieve the formset from the POST data.\n formset = PolicyFormSet(request.POST)\n if formset.is_valid():\n # Save the formset (updated values and newly created objects).\n formset.save()\n simulation.has_changed = True\n simulation.save()\n else:\n # Redirect to a page with the errors.\n context = {\n 'simulation': simulation,\n 'demandsegment': demandsegment,\n 'form': formset,\n }\n return render(request, 'metro_app/errors.html', context)\n\n return HttpResponseRedirect(reverse(\n 'metro:pricing_edit', args=(simulation.id, demandsegment.id,)\n ))", "def onSaveNotesButton(self, button):\n pass", "def click_target_buy_list_overrides_bulk_edit_button(self):\n self.click_element(self.target_buy_list_overrides_bulk_edit_button_locator, error_message='target buy list overrides bulk edit button locator not found before specified time out')", "def saveButtonMethod(self):\n return AbstractAction(\"Save\")", "def click_outbound_statements_grid_save_changes_button(self):\n self.click_element(self.outbound_statements_grid_save_changes_button_locator)", "def update_policy(self, *args, **kwargs):\r\n pass", "def _save_clicked(self, info):\n\n assignment = self._validate(info)\n if assignment is None:\n return\n\n # Update the data in the database.\n try:\n get_permissions_manager().policy_manager.policy_storage.set_assignment(assignment.user_name, [r.name for r in assignment.roles])\n\n info.ui.dispose()\n except PolicyStorageError, e:\n self._ps_error(e)", "def save(self, *args, **kwargs):\n\n # Call the \"real\" save() method.\n super(Product, self).save(*args, **kwargs)\n\n # If sold and is a Widget\n if self.sale and self.kind == self.WIDGET:\n # But has not stamp\n try:\n self.stamp\n except:\n s = Stamp(owned_by=self.sale.customer, obtained_with=self)\n s.save()", "def save_model(self, request, obj, form, change):\n if request.user.has_perm(\"easypublisher.can_approve_for_publication\"):\n obj.save()\n else:\n reversion.revision.add_meta(EasyPublisherMetaData, status='draft', language=request.LANGUAGE_CODE)\n reversion.revision.comment = \"Draft\"\n\n if not change and hasattr(obj, 'published'):\n obj.published = False\n obj.save()\n obj.published = True\n \n reversion.revision.post_save_receiver(obj, 0)", "def save(self, *args, **kwargs):\n return", "def pressS(self):\n self.myParent.mode.modifyIndustry(self.myParent.mySystemDict['id'], self.currentValue, self.myIndustryData.id)\n self.disableButton('S')" ]
[ "0.7632212", "0.7429682", "0.6681149", "0.6559746", "0.6546259", "0.6481419", "0.6439012", "0.5897788", "0.5761528", "0.5743864", "0.5709328", "0.5687823", "0.5661773", "0.5654063", "0.56287414", "0.5625229", "0.5610592", "0.5592306", "0.5590449", "0.55888796", "0.5561194", "0.55611193", "0.551927", "0.5518546", "0.54882586", "0.5468552", "0.5454214", "0.545384", "0.545352", "0.5451796" ]
0.78032136
0
Implementing click cancel edited target buy policy button functionality
def click_cancel_edited_target_buy_policy_button(self): self.click_element(self.cancel_edited_target_buy_policy_button_locator)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def click_edit_target_buy_policy_button(self):\n self.click_element(self.edit_target_buy_policy_button_locator)", "def click_cancel(self):\n self.click_element(self.cancel_button_selector)", "def click_save_edited_target_buy_policy_button(self):\n self.click_element(self.save_edited_target_buy_policy_button_locator)\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def onCancelButtonClick(self, event):\n self.EndModal(wx.ID_CANCEL)\n event.Skip()", "def action_cancel(self):\n self.state = 'canceled'", "def click_bulk_edit_cancel_button(self):\n self.click_element(self.bulk_edit_cancel_button_locator, True)", "def click_save_target_buy_policy_button(self):\n self.click_element(self.save_target_buy_policy_button_locator, error_message='save target buy policy button locator not found before specified time out')\n self.click_element(self.confirmation_popup_locator, error_message='success message locator not found before specified time out')\n self.click_element(self.ok_button_locator, error_message='ok button locator not found before specified time out')\n self.wait_for_ajax_spinner_load()", "def buttonCancel_Clicked( self, event ):\n\t\tself.EndModal(wx.ID_CANCEL)", "def cancel(self):", "def cancel(self):", "def cancel(self):", "def cancel_on_pos(self):\n # Start a transaction\n self.log.info(\"Starting a transaction...\")\n pos.click('generic item')\n\n # Click customer id and then cancel\n self.log.info(\"Clicking the customer ID button...\")\n if pos.is_element_present(self.customer_id_button, timeout = self.wait_time):\n pos.click('customer id')\n else:\n tc_fail(\"Customer ID button did not appear.\")\n\n self.log.info(\"Clicking cancel...\")\n if pos.is_element_present(self.manual_button, timeout = self.wait_time):\n pos.click('cancel')\n else:\n tc_fail(\"Did not change to the customer ID screen.\")\n\n msg = pos.read_message_box(timeout = self.wait_time)\n if not msg:\n tc_fail(\"No popup appeared.\")\n elif not \"cancel\" in msg.lower():\n tc_fail(\"Did not display the correct popup message after cancelling.\")\n\n pos.click('ok')\n\n # Make sure we returned to the right screen after cancelling\n if pos.is_element_present(self.customer_id_button, timeout = self.wait_time):\n self.log.info(\"Successfully cancelled input of customer ID!\")\n else:\n tc_fail(\"Did not return from customer ID screen.\")", "def unlink(self):\n context = self._context or {}\n for src_brw in self.browse():\n if src_brw.state != 'cancel':\n raise exceptions.except_orm(\n _(\"Invalid Procedure!!\"),\n _(\"The withholding document needs to be in cancel state to\"\n \" be deleted.\"))\n else:\n super(AccountWhSrc, self).unlink(\n )\n return True", "def onCancelOrder(self, item):\n self.frame.mode.cancelIndustryOrder(self.lstOrders.getMultiSelectedItems(), self.mySystemDict['id'])", "def cancel(self): #$NON-NLS-1$\r", "def do_cancel(order):\r\n self.gox.cancel(order.oid)", "def _do_cancel(self):\r\n\r\n def do_cancel(order):\r\n \"\"\"cancel a single order\"\"\"\r\n self.gox.cancel(order.oid)\r\n\r\n if not len(self.items):\r\n return\r\n if not len(self.selected):\r\n order = self.items[self.item_sel]\r\n do_cancel(order)\r\n else:\r\n for order in self.selected:\r\n do_cancel(order)", "def od_cancel_clicked(self, widget, data=None):\n self.open_chooser.hide()", "def _onButtonCancelClick(self, widget):\n self.delete()", "def _cancel(self, __button):\r\n\r\n self.destroy()", "def onCancelOrder(self, item):\n self.frame.mode.cancelMarketOrder(self.lstOrders.getMultiSelectedItems(), self.mySystemDict['id'])", "def OnCancel(self, event):\n pass", "def OnCancel(self, event):\n pass", "def cancel(self, uuid):\n return self.__call__('market', 'tradecancel',\n {'orderId': uuid})", "def OnButtonRateCancelButton(self, event):\r\n\t\tself.Hide()", "def cancel():", "def click_statement_entry_cancel_button(self):\n self.click_element(self.statement_entry_cancel_button_locator, False, True)\n try:\n self.dismiss_alert_pop_up()\n except:\n pass", "def onCancelTrade(self, item):\n self.frame.mode.cancelTradeRoute(self.lstTrade.getMultiSelectedItems(), self.mySystemDict['id'])", "def _cancel(self, __button=None):\r\n\r\n self.destroy()", "def test_cancel_pending_payment(self):\n pass" ]
[ "0.7026351", "0.68475145", "0.66155964", "0.6442423", "0.6418752", "0.638268", "0.63307065", "0.63213784", "0.6320915", "0.6320915", "0.6320915", "0.6234889", "0.623244", "0.62140983", "0.6192419", "0.61831707", "0.61803746", "0.6167173", "0.61546427", "0.61361396", "0.61344296", "0.61112756", "0.61112756", "0.61073965", "0.6021574", "0.60203147", "0.60177386", "0.6017065", "0.600579", "0.5996722" ]
0.8608011
0
Try to update null parameters from environment_params_path, or DEFAULT_PARAMS
def update_custom_environment_params(self): allowed_parameter_keys = [ k for k, v in signature(Environment).parameters.items() if v.kind == v.KEYWORD_ONLY ] for k in allowed_parameter_keys: if getattr(self, k) is None: setattr(self, k, self.DEFAULT_PARAMS.get(k, None))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def overwrite_hyperparams(self):\n try:\n default_hyperparams = self.hyperparams\n for key in default_hyperparams:\n try:\n flag = self.FLAGS[key]\n param_value = flag.value\n if param_value is not None:\n self.hyperparams[key] = param_value\n except:\n pass\n except:\n pass", "def resetParam(key, value):\n \n if globalParams == {}:\n warning(\"WARNING: runtime parameters not yet initialized\")\n LoadParams(\"_defaults\")\n \n if key in globalParams.keys():\n globalParams[key] = value\n else:\n raise ValueError()", "def test_reset_params(self):\n\t\t\n\t\tparams = DEFAULT_PARAMS.copy()\n\t\tparams['fit']=PL\n\t\tvalid = self.watcher.valid_params(params)\n\t\tself.assertTrue(valid)\n\t\tparams = self.watcher.normalize_params(params)\n\t\tself.assertEqual(params['fit'], POWER_LAW)\n\t\t\n\t\tparams = DEFAULT_PARAMS.copy()\n\t\tparams['fit']=TPL\n\t\tvalid = self.watcher.valid_params(params)\n\t\tself.assertFalse(valid)\n\t\t\n\t\tparams = DEFAULT_PARAMS.copy()\n\t\tparams['fit']=TPL\n\t\tparams[PL_PACKAGE]=POWERLAW_PACKAGE\n\t\tparams[XMAX]=XMAX_FORCE\n\n\t\tvalid = self.watcher.valid_params(params)\n\t\tself.assertTrue(valid)\n\t\tparams = self.watcher.normalize_params(params)\n\t\tself.assertEqual(params['fit'], TRUNCATED_POWER_LAW)", "def _transform_env(self) -> None:\n self.env = None if self.env == {} else self.env", "def reset(self):\n\n for value in self.__dict__.itervalues():\n if isinstance(value, EnvParm):\n value._value = 'use_env'\n getattr(value, 'value')", "def update_environ():\n\n # Environment variables to set.\n BASE = os.getcwd()\n PLUGINS = os.path.join(BASE, 'lib')\n RESOURCES = os.path.join(BASE, 'res')\n MODELS = os.path.join(RESOURCES, 'models')\n\n # Set the vaue to '' to set the var to ''.\n # Anything else will be added to current var value.\n minimapper_env = {\n 'GAZEBO_RESOURCE_PATH': RESOURCES,\n 'GAZEBO_MODEL_PATH': MODELS,\n 'GAZEBO_PLUGIN_PATH': PLUGINS,\n 'GAZEBO_MODEL_DATABASE_URI': None\n }\n\n # Conditionally set environment variables.\n env = os.environ.copy()\n for key, val in minimapper_env.items():\n if val is None:\n env[key] = ''\n elif key not in env:\n env[key] = val\n elif key in env and val not in env[key]:\n env[key] = val + ':' + env[key]\n\n return env", "def _use_existing_params(self):\n sh = shelve.open(os.path.expanduser('~/.config/scheduler/params'))\n self.params = sh['params']\n sh.close()", "def resetparams(self, parameters):\n try:\n utils.update_dictionary_items(self.params,parameters)\n except AttributeError:\n # Variable self.params does not exist, so not updated\n # Create an empty set of params for future reference\n self.params = {}", "def _initialize_defaults(self):\n for key, value in defaults.items():\n if key not in self.source_params:\n self.source_params[key] = value", "def _initialize_defaults(self):\n for key, value in defaults.items():\n if key not in self.source_params:\n self.source_params[key] = value", "def _set_params(self, params, defaults):\n new_params = OrderedDict(\n zip(params, [x if isinstance(x, Parameter) else Parameter() for x in defaults])\n )\n for key, value in self._src.items():\n if key in new_params:\n new_params[key] = value\n\n self._src = new_params", "def _validate_params(self):\n assert set(self.required_params) - set(self._params) == set()\n for par, val in self.optional_params.items():\n if par not in self._params:\n self._params[par] = val", "def update_workflow_params(**kwargs) -> None:\n workflow_params_file = upsearch(WORKFLOW_PARAMS_FILENAME)\n workflow_params = load_workflow_params()\n for k, v in kwargs.items():\n if k in workflow_params:\n workflow_params[k] = v\n\n with workflow_params_file.open(\"w\") as f:\n f.write(json.dumps(workflow_params, indent=4))", "def update_settings(self, param):\n if param.name() == '':\n pass", "def test_required(self, missing_param):\n with mock.patch.dict('os.environ', {\n **REQUIRED_SETTINGS,\n missing_param: '',\n }, clear=True), self.assertRaises(ImproperlyConfigured):\n self.reload_settings()", "def runtime_updatable_params(self) -> Optional[pulumi.Input['RuntimeUpdatableParamsArgs']]:\n return pulumi.get(self, \"runtime_updatable_params\")", "def _update(self):\n self.all_params = {}\n self._update_experiment_params()\n self._update_preprocessing_params()\n self._update_model_params()", "def _update_params(self):\n pass", "def resetparams(self, parameters):\n self.weights = None\n try:\n self.params = parameters\n except AttributeError:\n # Variable self.params does not exist, so not updated\n # Create an empty set of params for future reference\n self.params = {}", "def _setup_params(self,**params):\n ### a parameter might be passed in for one of the extra_pos;\n ### if a key in the params dict is not a *parameter* of this\n ### PO, then try it on the extra_pos\n for n,p in params.items():\n if n not in self.params():\n self.set_parameter_value(n,p)\n del params[n]\n\n Parameterized._setup_params(self,**params)", "def update_params(argv: list, prm: dict):\n\n\tfor a in argv[1:]:\n\t\ttoks = a.split('=',1)\n\t\tif len(toks)<2: continue\n\t\tk,v = toks[:2]\n\t\tif k not in prm: continue\n\t\tprm[k] = v", "def _update_params(self):\n _load = not self.san_interface.runmode\n params={}\n if ('iosched' in self._updatedattr or _load) and self.iosched<>IoSchedType.default:\n params['iosched']=str(self.iosched)\n if ('readahead' in self._updatedattr or _load) and self.readahead :\n params['readahead']=self.readahead\n if params:\n for pt in self.paths():\n pt.provider.set_dev_params(pt,params)", "def check_raw_parameters(self):\n # exclude some irrelevant or mandatory parameters\n excluded_keys = (\"cmd\", \"client\", \"resource_group_name\", \"name\")\n # check whether the remaining parameters are set\n # the default value None or False (and other empty values, like empty string) will be considered as not set\n is_changed = any(v for k, v in self.context.raw_param.items() if k not in excluded_keys)\n\n # special cases\n # some parameters support the use of empty string or dictionary to update/remove previously set values\n is_default = (\n self.context.get_cluster_autoscaler_profile() is None and\n self.context.get_api_server_authorized_ip_ranges() is None and\n self.context.get_nodepool_labels() is None and\n self.context.get_nodepool_taints() is None\n )\n\n if not is_changed and is_default:\n reconcilePrompt = 'no argument specified to update would you like to reconcile to current settings?'\n if not prompt_y_n(reconcilePrompt, default=\"n\"):\n # Note: Uncomment the followings to automatically generate the error message.\n option_names = [\n '\"{}\"'.format(format_parameter_name_to_option_name(x))\n for x in self.context.raw_param.keys()\n if x not in excluded_keys\n ]\n error_msg = \"Please specify one or more of {}.\".format(\n \" or \".join(option_names)\n )\n raise RequiredArgumentMissingError(error_msg)", "def _overwrite_parameters_from_input(self, new_params):\n params_overwrite = self.inputs.params_overwrite.get_dict()\n for key, val in params_overwrite.items():\n if key in new_params:\n self.report('ATTENTION: overwriting parameter from \"params_overwrite\" input')\n self.report(f'key: {key}')\n self.report(f'old value: {new_params[key]}')\n self.report(f'overwritten value: {val}')\n new_params[key] = val", "def override_environment(**kwargs):\n overridden = {}\n for key, value in kwargs.items():\n overridden[key] = os.environ.get(key)\n if value is None:\n os.environ.pop(key, None)\n else:\n os.environ[key] = value\n\n yield\n\n for key, value in overridden.items():\n if value is None:\n os.environ.pop(key, None)\n else:\n os.environ[key] = value", "def _set_from_env(name, context, default):\n if default is _DEFAULT_ARG and name not in os.environ:\n return\n\n context[name] = os.environ.get(name, default)", "def _validatePypetParameters(self, runParams):\n\n # fix rng seed, which is saved as a string if None\n if \"seed\" in runParams:\n if runParams[\"seed\"] == \"None\":\n runParams[\"seed\"] = None\n return runParams", "def _apply_params(self):\n config = self.get_startup_config()\n # Pass true to _set_params so we know these are startup values\n self._set_params(config, True)", "def get_parm_value(parameters, name, env_name, default_value):\n value = parameters.get(name, '')\n return os.environ.get(env_name, default=default_value) if not value else value", "def set_params_BS(self):\n params = self.ctx.input_params_KKR\n\n # maybe overwrite some inputs\n if 'params_kkr_overwrite' in self.inputs:\n self.report(f'found params_kkr_overwrite: {self.inputs.params_kkr_overwrite.get_dict()}')\n updatenode = self.inputs.params_kkr_overwrite\n updatenode.label = 'params overwrite'\n params = update_params_wf(params, updatenode)\n\n input_dict = params.get_dict()\n para_check = kkrparams()\n try:\n for key, val in input_dict.items():\n para_check.set_value(key, val, silent=True)\n except:\n return self.exit_codes.ERROR_CALC_PARAMETERS_INVALID # pylint: disable=no-member\n label = ''\n descr = f'(pk - {self.inputs.remote_data.pk}, and uuid - {self.inputs.remote_data.uuid})'\n\n missing_list = para_check.get_missing_keys(use_aiida=True)\n\n if missing_list != []:\n kkrdefaults = kkrparams.get_KKRcalc_parameter_defaults()[0]\n kkrdefaults_updated = []\n for key_default, val_default in list(kkrdefaults.items()):\n if key_default in missing_list:\n para_check.set_value(key_default, val_default)\n kkrdefaults_updated.append(key_default)\n missing_list.remove(key_default)\n if len(missing_list) > 0:\n self.report(f'ERROR: calc_parameters misses keys: {missing_list}')\n return self.exit_codes.ERROR_CALC_PARAMETERS_INCOMPLETE # pylint: disable=no-member\n else:\n self.report(f'updated KKR parameter node with default values: {kkrdefaults_updated}')\n label = 'add_defaults_'\n descr = 'added missing default keys, '\n ##+++ Starts to add the NTP2, EMAX and EMIN from the\n econt_new = self.ctx.BS_params_dict\n if self.ctx.struc_is_alloy:\n if econt_new.get('kmesh', None) is None:\n econt_new['kmesh'] = [1, 1, 1] # overwrite kmesh since the kpoints are used from the input\n kkr_calc = self.inputs.remote_data.get_incoming(node_class=KkrCalculation).first().node\n ef = kkr_calc.outputs.output_parameters.get_dict()['fermi_energy'] # unit in Ry\n self.ctx.fermi_energy = ef ## in Ry unit\n\n # Set BS params\n try:\n para_check = set_energy_params(econt_new, ef, para_check)\n except:\n return self.exit_codes.ERROR_CALC_PARAMETERS_INVALID # pylint: disable=no-member\n\n para_check.set_multiple_values(\n NPT1=0,\n NPT3=0,\n NPOL=0,\n )\n\n updatenode = Dict(para_check.get_dict())\n updatenode.label = label + 'KKRparam_BS'\n updatenode.description = 'KKR parameter node extracted from remote_folder' + descr + ' as well as wf_parameter input node.'\n\n paranode_BS = update_params_wf(self.ctx.input_params_KKR, updatenode)\n self.ctx.BS_kkrparams = paranode_BS" ]
[ "0.6267459", "0.60361874", "0.6005112", "0.59332246", "0.5914079", "0.58198893", "0.578207", "0.57676196", "0.57331485", "0.57331485", "0.56856215", "0.5666313", "0.5595226", "0.5546338", "0.55456305", "0.5544343", "0.55403894", "0.55393416", "0.5523232", "0.5497067", "0.54677385", "0.5466806", "0.54613966", "0.5446688", "0.5444374", "0.543951", "0.5435035", "0.54328704", "0.54327106", "0.54233664" ]
0.6903751
0
Generate a key to describe the current Environment's crossexperiment parameters
def generate_cross_experiment_key(self): parameters = dict( metrics_params=self.metrics_params, cv_params=self.cv_params, target_column=self.target_column, id_column=self.id_column, do_predict_proba=self.do_predict_proba, prediction_formatter=self.prediction_formatter, train_dataset=self.train_dataset, test_dataset=self.test_dataset, holdout_dataset=self.holdout_dataset, cross_experiment_params=self.cross_experiment_params.copy(), to_csv_params=self.to_csv_params, ) #################### Revert Aliases for Compatibility #################### # If any aliases were used during call to `Environment.__init__`, replace the default names # in `parameters` with the alias used. This ensures compatibility with Environment keys # made in earlier versions aliases_used = getattr(self, "__hh_aliases_used", {}) # noinspection PyUnusedLocal def _visit(path, key, value): if key in aliases_used: key = aliases_used.pop(key) return (key, value) if aliases_used: parameters = remap(parameters, visit=_visit) #################### Make `cross_experiment_key` #################### self.cross_experiment_key = CrossExperimentKeyMaker(parameters)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _course_key(self):\r\n return \"slashes:{org}+{number}+{run}\".format(**self._course_dict)", "def _key(\n service=None, # type: Optional[str]\n env=None, # type: Optional[str]\n ):\n # type: (...) -> str\n service = service or \"\"\n env = env or \"\"\n return \"service:\" + service + \",env:\" + env", "def _key(self):\n key_args = [self.__class__.__name__] + [str(a) for a in self.args]\n return (\":\".join(key_args))", "def _create_key(chip):\n try:\n suffix = chip['version'][0]\n except IndexError:\n suffix = ''\n\n if chip['classification'] == 'secret':\n classification = 'z'\n else:\n classification = chip['classification'][0]\n\n return '%s-%s%s%s' % (chip['game'], classification, chip['indice'], suffix)", "def generate_prototype_key(self):\n return str(uuid.uuid5(UUID_XYZ_NAMESPACE, str((self.X, self.Y, self.Z))))", "def environment_workflow(self):\n self.update_custom_environment_params()\n self.validate_parameters()\n self.format_result_paths()\n self.generate_cross_experiment_key()\n G.log(\"Cross-Experiment Key: '{!s}'\".format(self.cross_experiment_key))", "def get_config(self):\n return 'c2Key', self.key", "def generate_prototype_key(self, *args):\n return str(uuid.uuid5(UUID_XYZ_NAMESPACE, str((self.X, self.Y, self.Z, *args))))", "def describe(self) -> str:\r\n env = str(self.env)\r\n observation_space = str(self.observations)\r\n action_space = str(self.actions)\r\n policy = str(self.policy)\r\n gamma = f\"Gamma({self.gamma})\"\r\n\r\n return env, observation_space, action_space, policy, gamma", "def keygen(cls, device, **kwargs):\n key = '%s/%s' % (cls.__name__, device)\n for v in kwargs.values():\n key += '/' + str(v)\n return key", "def get_key(self) -> str:\n return f'{self.address}_{self.port}'", "def create_key ():", "def key_name(self) -> str:\n return pulumi.get(self, \"key_name\")", "def infocalypse_genkey(ui_, **opts):\n params, dummy = get_config_info(ui_, opts)\n execute_genkey(ui_, params)", "def get_env_key(obj, key=None):\n return str.join('_', [obj.__module__.replace('.','_').upper(),\n key.upper()])", "def generate_key():\r\n\t\treturn ''.join(random.SystemRandom().choice(string.ascii_lowercase) for _ in range(123))", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")", "def key(self) -> str:\n return pulumi.get(self, \"key\")" ]
[ "0.6446606", "0.61374927", "0.60621893", "0.5975565", "0.5934612", "0.5891562", "0.5889384", "0.5880724", "0.58526766", "0.5810926", "0.57407105", "0.5730642", "0.5688418", "0.56631047", "0.56109715", "0.55618995", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926", "0.5547926" ]
0.70411134
0
Get a `DatasetSentinel` representing an Experiment's `fold_train_input` Returns
def train_input(self): return DatasetSentinel("train_input", **self._dataset_sentinel_helper())
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def validation_input(self):\n return DatasetSentinel(\"validation_input\", **self._dataset_sentinel_helper())", "def train_target(self):\n return DatasetSentinel(\"train_target\", **self._dataset_sentinel_helper())", "def validation_target(self):\n return DatasetSentinel(\"validation_target\", **self._dataset_sentinel_helper())", "def get_train(self, preprocess=False):\n return self._dataset('train', self._directory, 'sharded_train_0shifted_affnist.tfrecords')", "def holdout_input(self):\n return DatasetSentinel(\n \"holdout_input\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def get_train_input(self, prev, i):\n return self.get_eval_input(prev, i)", "def get_train_input(self, prev, i):\n return self.get_eval_input(prev, i)", "def get_input_fn_training(Xtrain_ul, Xtrain_l, Xtest, ytrain_ul, ytrain_l, ytest, batch_size, num_labeled):\n dataset = input_data.Data(Xtrain_ul,\n Xtrain_l,\n Xtest,\n ytrain_ul,\n ytrain_l,\n ytest,\n num_labeled, \n batch_size, \n shuffle=True)\n return dataset.next_batch()", "def get_train_inputs(self, example):\n return example", "def load_input(io_config):\n\n fold = io_config['fold']\n\n train_data_fnames = io_config['file']['train_data']\n\n train_fname = [t[1] for t in train_data_fnames\n if t[0] == fold][0]\n\n return train_fname", "def get_train_input(self, prev, i):\n pass", "def get_train_input(self, prev, i):\n if i == 0:\n return self._zero_label\n else:\n # TODO(gorban): update to gradually introduce gt labels.\n return self._labels_one_hot[:, i - 1, :]", "def holdout_target(self):\n return DatasetSentinel(\n \"holdout_target\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def get_eval_input(self, prev, i):\n del prev, i\n # The attention_decoder will fetch image features from the net, no need for\n # extra inputs.\n return self._zero_label", "def train_input_fn():\n # When choosing shuffle buffer sizes, larger sizes result in better\n # randomness, while smaller sizes use less memory. MNIST is a small\n # enough dataset that we can easily shuffle the full epoch.\n ds = dtrain(flags_obj.data_dir)\n ds = ds.cache().shuffle(buffer_size=50000).batch(flags_obj.batch_size)\n\n # Iterate through the dataset a set number (`epochs_between_evals`) of times\n # during each training session.\n ds = ds.repeat()\n return ds", "def _get_training_dataset(self):\n\n return tf.data.Dataset.zip((self.conditioned_dataset, self.dataset))", "def get_input(self, prev, i):\n if self.is_training():\n return self.get_train_input(prev, i)\n else:\n return self.get_eval_input(prev, i)", "def get_train_input(self, prev, i):\n if i == 0:\n prev = self._zero_label\n else:\n prev = self._labels_one_hot[:, i - 1, :]\n image_feature = self.get_image_feature(i)\n return tf.concat([image_feature, prev], 1)", "def get_train_index():\n data_size = (NUM_CLASS - 1) * NUM_DATA_PER_CLASS\n return np.array([i for i in range(0, data_size)])", "def get_train_iterator(self) -> tf.contrib.data.Iterator:\n return self.train.make_initializable_iterator()", "def train(self) -> tf.contrib.data.Dataset:\n return self.__train_dataset", "def get_train_dataset(self):\n return SubsetDataset(self.ds, self.train_ids)", "def _dataset_sentinel_helper(self):\n params = self.cross_experiment_key.parameters\n return dict(\n dataset_hash=params[\"train_dataset\"],\n cv_type=params[\"cross_experiment_params\"].get(\n \"cv_type\", params[\"cross_experiment_params\"].get(\"cross_validation_type\", None)\n ),\n global_random_seed=params[\"cross_experiment_params\"][\"global_random_seed\"],\n random_seeds=params[\"cross_experiment_params\"][\"random_seeds\"],\n )", "def get_infer_fn(infer_step: InferStepCallable, batch_size: int,\n train_state_axes: train_state_lib.TrainState,\n partitioner: partitioning.BasePartitioner):\n infer_step = partitioner.partition(\n infer_step,\n in_axis_resources=(train_state_axes.params, PartitionSpec('data',)),\n out_axis_resources=PartitionSpec('data',))\n\n data_layout = partitioner.get_data_layout(batch_size)\n shard_id = data_layout.shard_id\n num_shards = data_layout.num_shards\n\n per_shard_batch_size = batch_size // num_shards\n\n def infer_fn(ds: tf.data.Dataset, train_state: train_state_lib.TrainState):\n ds_shapes = jax.tree_map(lambda x: jnp.array(x.shape), ds.element_spec)\n multihost_utils.assert_same(\n ds_shapes, 'Dataset element shapes do not agree across hosts. '\n 'This could be an indication that the dataset is nondeterministic.')\n try:\n dataset_remainder = len(ds) % batch_size # pytype:disable=wrong-arg-types\n logging.info('length of dataset = %s', len(ds))\n except TypeError as e:\n if str(e) == 'dataset length is unknown.':\n logging.warning(\n 'The following error is likely due to the use of TensorFlow v1 in '\n 'your dataset pipeline. Verify you are not importing from '\n '`tf.compat.v1` as part of your pipeline.')\n raise e\n\n if dataset_remainder:\n dataset_pad_amt = batch_size - dataset_remainder\n logging.info(\n 'Padding infer dataset with %d examples for even per-replica shards.',\n dataset_pad_amt)\n # Pad with the first example using an index of -1 so seqio will ignore.\n pad_ds = ds.take(1).map(lambda i, x: (np.int64(-1), x)).repeat(\n dataset_pad_amt)\n ds = ds.concatenate(pad_ds)\n\n # Shard the infer dataset across replica sets.\n sharded_ds = ds.shard(num_shards, shard_id).batch(\n per_shard_batch_size, drop_remainder=True)\n multihost_utils.assert_same(\n jnp.array(len(sharded_ds)),\n 'Dataset lengths do not agree across hosts.')\n\n logging.info(\n 'The infer dataset is sharded into %d shards with per-shard '\n 'batch size of %d', num_shards, per_shard_batch_size)\n\n # Run inference for each replica set.\n batched_results, all_indices = [], []\n for index, infer_batch in sharded_ds.as_numpy_iterator():\n # Run fast inference on batch.\n # [B, ...] -> [B, ...]\n batch_result = infer_step(train_state.params, infer_batch)\n logging.info('Inference of batch %s done.', index)\n # Issue asynchronous copy request which serves as prefetching to the host.\n # The result value is synchronized with host_allgather in the loop below.\n try:\n jax.tree_map(lambda x: x.copy_to_host_async(), batch_result)\n except AttributeError:\n # Similar to jax.device_get, we skip transfers for non DeviceArrays.\n pass\n batched_results.append(batch_result)\n all_indices.append(index)\n logging.info('Inference of all batches done.')\n all_inferences = []\n for batch_result in batched_results:\n # [B, ...] -> [H, B, ...]\n batch_result = multihost_utils.host_allgather(\n batch_result, num_shards, shard_id,\n data_layout.is_first_host_in_replica_set)\n all_inferences.append(batch_result)\n\n # List[H, B, ...] -> List[B, H, ...]\n all_inferences = jax.tree_map(lambda x: np.moveaxis(x, 0, 1),\n all_inferences)\n\n # List[B, H, ...] -> [B * batch_count, H, ...]\n all_inferences = jax.tree_multimap(lambda *args: np.concatenate(args),\n *all_inferences)\n # List[B] -> [B * batch_count]\n all_indices = np.concatenate(all_indices)\n # Collect all batches from across hosts.\n # [B * batch_count] -> [H, B * batch_count]\n all_indices = multihost_utils.host_allgather(\n all_indices, num_shards, shard_id,\n data_layout.is_first_host_in_replica_set)\n # [H, B * batch_count] -> [B * batch_count, H]\n all_indices = np.transpose(all_indices)\n all_inferences, all_indices = _deshard_and_remove_padding(\n all_inferences, all_indices)\n\n # Translate [B, ...] -> List[...] by flattening inferences making sure to\n # preserve structure of individual elements (inferences are not assumed to\n # be simple np.array). Finally, zip inferences with corresponding indices\n # and convert leaf np.arrays into lists.\n all_inferences, struct = jax.tree_flatten(all_inferences)\n all_inferences = map(\n functools.partial(jax.tree_unflatten, struct), zip(*all_inferences))\n indices_and_outputs = list(zip(all_indices, all_inferences))\n indices_and_outputs = jax.tree_map(lambda x: np.array(x).tolist(),\n indices_and_outputs)\n return indices_and_outputs\n\n return infer_fn", "def get_init_fn():\n checkpoint_exclude_scopes = [\"InceptionV1/Logits\", \"InceptionV1/AuxLogits\"]\n\n exclusions = [scope.strip() for scope in checkpoint_exclude_scopes]\n\n variables_to_restore = []\n for var in slim.get_model_variables():\n excluded = False\n for exclusion in exclusions:\n if var.op.name.startswith(exclusion):\n excluded = True\n break\n if not excluded:\n variables_to_restore.append(var)\n\n return slim.assign_from_checkpoint_fn(\n os.path.join(checkpoints_dir, 'inception_v1.ckpt'),\n variables_to_restore)", "def get_inputs_train():\n x = tf.constant(extract_pandas_data(x_train))\n y = tf.constant(y_train.values)\n return x, y", "def get_train(self, preprocess=False):\n return self._dataset(self._directory, 'images_background_small1', preprocess)", "def train_input_fn(self) -> types.FeatureAndLabelTensors:\n return self._input_fn_from_file(self._train_path)", "def get_train_examples(self):\n raise NotImplementedError()", "def preprocess_train_dataset(dataset):\n return (dataset\n # Shuffle according to the largest client dataset\n .shuffle(buffer_size=MAX_CLIENT_DATASET_SIZE)\n # Repeat to do multiple local epochs\n .repeat(CLIENT_EPOCHS_PER_ROUND)\n # Batch to a fixed client batch size\n .batch(CLIENT_BATCH_SIZE, drop_remainder=False)\n # Preprocessing step\n .map(reshape_emnist_element))" ]
[ "0.7197259", "0.7058317", "0.6581115", "0.63265425", "0.62734795", "0.61224383", "0.61224383", "0.609924", "0.59627235", "0.5961343", "0.58399856", "0.5539047", "0.55349284", "0.5525194", "0.54774237", "0.54644793", "0.5406239", "0.540372", "0.5370397", "0.5369286", "0.53544694", "0.5328077", "0.5297147", "0.520762", "0.51923513", "0.5188809", "0.5186567", "0.5182556", "0.5180823", "0.5177131" ]
0.81839544
0
Get a `DatasetSentinel` representing an Experiment's `fold_train_target` Returns
def train_target(self): return DatasetSentinel("train_target", **self._dataset_sentinel_helper())
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def validation_target(self):\n return DatasetSentinel(\"validation_target\", **self._dataset_sentinel_helper())", "def train_input(self):\n return DatasetSentinel(\"train_input\", **self._dataset_sentinel_helper())", "def holdout_target(self):\n return DatasetSentinel(\n \"holdout_target\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def validation_input(self):\n return DatasetSentinel(\"validation_input\", **self._dataset_sentinel_helper())", "def get_train(self, preprocess=False):\n return self._dataset('train', self._directory, 'sharded_train_0shifted_affnist.tfrecords')", "def holdout_input(self):\n return DatasetSentinel(\n \"holdout_input\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def sample_train_data(dataset ,target,data_len, resp = True ):\r\n np.random.seed(222)\r\n ixes = np.random.choice(dataset.index, data_len, replace = False)\r\n print(ixes)\r\n under_df = dataset.iloc[ixes]\r\n if resp==True:\r\n under_target = target.iloc[ixes]\r\n return under_df, under_target\r\n else:\r\n return under_df", "def get_input_fn_training(Xtrain_ul, Xtrain_l, Xtest, ytrain_ul, ytrain_l, ytest, batch_size, num_labeled):\n dataset = input_data.Data(Xtrain_ul,\n Xtrain_l,\n Xtest,\n ytrain_ul,\n ytrain_l,\n ytest,\n num_labeled, \n batch_size, \n shuffle=True)\n return dataset.next_batch()", "def _get_training_dataset(self):\n\n return tf.data.Dataset.zip((self.conditioned_dataset, self.dataset))", "def training_target(training_data):\n return training_data.status == \"DEFAULT\"", "def get_train_input(self, prev, i):\n return self.get_eval_input(prev, i)", "def get_train_input(self, prev, i):\n return self.get_eval_input(prev, i)", "def _dataset_sentinel_helper(self):\n params = self.cross_experiment_key.parameters\n return dict(\n dataset_hash=params[\"train_dataset\"],\n cv_type=params[\"cross_experiment_params\"].get(\n \"cv_type\", params[\"cross_experiment_params\"].get(\"cross_validation_type\", None)\n ),\n global_random_seed=params[\"cross_experiment_params\"][\"global_random_seed\"],\n random_seeds=params[\"cross_experiment_params\"][\"random_seeds\"],\n )", "def get_dataset(self):\n\n trainset = datasets.MNIST('datasets/MNIST/train/', train=True, transform=self.train_transforms,\n target_transform=None, download=True)\n valset = datasets.MNIST('datasets/MNIST/test/', train=False, transform=self.val_transforms,\n target_transform=None, download=True)\n\n return trainset, valset", "def get_train_inputs(self, example):\n return example", "def train(self) -> tf.contrib.data.Dataset:\n return self.__train_dataset", "def get_train_input(self, prev, i):\n pass", "def get_train_index():\n data_size = (NUM_CLASS - 1) * NUM_DATA_PER_CLASS\n return np.array([i for i in range(0, data_size)])", "def get_dataset(self):\n trainset = datasets.KMNIST('datasets/KMNIST/train/', train=True, transform=self.train_transforms,\n target_transform=None, download=True)\n valset = datasets.KMNIST('datasets/KMNIST/test/', train=False, transform=self.val_transforms,\n target_transform=None, download=True)\n\n return trainset, valset", "def get_test(self, preprocess=False):\n return self._dataset('test', self._directory, 'sharded_test_0shifted_affnist.tfrecords')", "def get_train_input(self, prev, i):\n if i == 0:\n return self._zero_label\n else:\n # TODO(gorban): update to gradually introduce gt labels.\n return self._labels_one_hot[:, i - 1, :]", "def load_input(io_config):\n\n fold = io_config['fold']\n\n train_data_fnames = io_config['file']['train_data']\n\n train_fname = [t[1] for t in train_data_fnames\n if t[0] == fold][0]\n\n return train_fname", "def get_train_dataset(self):\n return SubsetDataset(self.ds, self.train_ids)", "def _next_train_batch(self, batch_size=128, replace=False):\n mask = np.random.choice(self.train_data.shape[0], batch_size, replace=replace)\n return self.train_data[mask], self.train_label[mask]", "def get_dataset(self):\n\n trainset = datasets.FashionMNIST('datasets/FashionMNIST/train/', train=True, transform=self.train_transforms,\n target_transform=None, download=True)\n valset = datasets.FashionMNIST('datasets/FashionMNIST/test/', train=False, transform=self.val_transforms,\n target_transform=None, download=True)\n\n return trainset, valset", "def get_train_full_examples(self, data_dir):\n raise NotImplementedError()", "def example(self):\n result = getattr(self, '_example', None)\n if result is None:\n # No example batch was found, so get one from the `.train` dataset\n result = next(iter(self.train))\n # And cache it for next time\n self._example = result\n return result", "def get_train(self, preprocess=False):\n return self._dataset(self._directory, 'images_background_small1', preprocess)", "def get_iris_data():\n iris = datasets.load_iris()\n data = iris[\"data\"]\n target = iris[\"target\"]\n\n # Prepend the column of 1s for bias\n N, M = data.shape\n all_X = np.ones((N, M + 1))\n all_X[:, 1:] = data\n\n # Convert into one-hot vectors\n num_labels = len(np.unique(target))\n all_Y = np.eye(num_labels)[target]\n return train_test_split(all_X, all_Y, test_size=0.33, random_state=RANDOM_SEED)", "def get_target_encoder(self, train: NumpyOrPandas) -> Optional[type]:\n target_encoder = None\n if train.folds is not None:\n if train.task.name in [\"binary\", \"reg\"]:\n target_encoder = TargetEncoder\n else:\n n_classes = train.target.max() + 1\n if n_classes <= self.multiclass_te_co:\n target_encoder = MultiClassTargetEncoder\n\n return target_encoder" ]
[ "0.72998613", "0.713837", "0.66336054", "0.6162712", "0.5902305", "0.57828236", "0.5423811", "0.5414221", "0.53984654", "0.5350862", "0.52556276", "0.52556276", "0.5214808", "0.5211418", "0.51941895", "0.51866794", "0.5174885", "0.5166052", "0.51568496", "0.5156731", "0.5136225", "0.5072823", "0.5066894", "0.50632143", "0.50560087", "0.50398403", "0.5035402", "0.5030994", "0.5027986", "0.5027658" ]
0.81183374
0
Get a `DatasetSentinel` representing an Experiment's `fold_validation_input` Returns
def validation_input(self): return DatasetSentinel("validation_input", **self._dataset_sentinel_helper())
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def train_input(self):\n return DatasetSentinel(\"train_input\", **self._dataset_sentinel_helper())", "def validation_target(self):\n return DatasetSentinel(\"validation_target\", **self._dataset_sentinel_helper())", "def holdout_input(self):\n return DatasetSentinel(\n \"holdout_input\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def train_target(self):\n return DatasetSentinel(\"train_target\", **self._dataset_sentinel_helper())", "def get_train_input(self, prev, i):\n return self.get_eval_input(prev, i)", "def get_train_input(self, prev, i):\n return self.get_eval_input(prev, i)", "def get_input_fn_training(Xtrain_ul, Xtrain_l, Xtest, ytrain_ul, ytrain_l, ytest, batch_size, num_labeled):\n dataset = input_data.Data(Xtrain_ul,\n Xtrain_l,\n Xtest,\n ytrain_ul,\n ytrain_l,\n ytest,\n num_labeled, \n batch_size, \n shuffle=True)\n return dataset.next_batch()", "def load_input(io_config):\n\n fold = io_config['fold']\n\n train_data_fnames = io_config['file']['train_data']\n\n train_fname = [t[1] for t in train_data_fnames\n if t[0] == fold][0]\n\n return train_fname", "def get_eval_input(self, prev, i):\n del prev, i\n # The attention_decoder will fetch image features from the net, no need for\n # extra inputs.\n return self._zero_label", "def get_input(self, prev, i):\n if self.is_training():\n return self.get_train_input(prev, i)\n else:\n return self.get_eval_input(prev, i)", "def get_train_inputs(self, example):\n return example", "def get_init_fn():\n checkpoint_exclude_scopes = [\"InceptionV1/Logits\", \"InceptionV1/AuxLogits\"]\n\n exclusions = [scope.strip() for scope in checkpoint_exclude_scopes]\n\n variables_to_restore = []\n for var in slim.get_model_variables():\n excluded = False\n for exclusion in exclusions:\n if var.op.name.startswith(exclusion):\n excluded = True\n break\n if not excluded:\n variables_to_restore.append(var)\n\n return slim.assign_from_checkpoint_fn(\n os.path.join(checkpoints_dir, 'inception_v1.ckpt'),\n variables_to_restore)", "def holdout_target(self):\n return DatasetSentinel(\n \"holdout_target\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def _dataset_sentinel_helper(self):\n params = self.cross_experiment_key.parameters\n return dict(\n dataset_hash=params[\"train_dataset\"],\n cv_type=params[\"cross_experiment_params\"].get(\n \"cv_type\", params[\"cross_experiment_params\"].get(\"cross_validation_type\", None)\n ),\n global_random_seed=params[\"cross_experiment_params\"][\"global_random_seed\"],\n random_seeds=params[\"cross_experiment_params\"][\"random_seeds\"],\n )", "def get_train_input(self, prev, i):\n pass", "def validation(self) -> tf.contrib.data.Dataset:\n return self.__validation_dataset", "def split_validation_training_index(allind, splitsize, do_offset, offset_steps):\n i = offset_steps\n lval = splitsize\n if not do_offset:\n i_val = allind[:lval]\n i_train = allind[lval:]\n else:\n i_val = allind[i * lval:(i + 1) * lval]\n i_train = np.concatenate([allind[0:i * lval], allind[(i + 1) * lval:]], axis=0)\n if len(i_val) <= 0:\n print(\"Warning: #Validation data is 0, take 1 training sample instead\")\n i_val = i_train[:1]\n\n return i_train, i_val", "def get_train(self, preprocess=False):\n return self._dataset('train', self._directory, 'sharded_train_0shifted_affnist.tfrecords')", "def _input_fn(input_pipeline_context=None):\n return _create_dataset(options, is_training, input_pipeline_context)", "def get_eval_input(self, prev, i):\n del prev\n return self.get_image_feature(i)", "def _partition_elms(\n self, max_elms: int = None, fold: int = None\n ) -> Tuple[np.ndarray, np.ndarray, np.ndarray]:\n # get ELM indices from datafile\n elm_index, _ = self._read_file()\n\n # limit the data according to the max number of events passed\n if max_elms is not None and max_elms != -1:\n LOGGER.info(f\"Limiting data read to {max_elms} events.\")\n n_elms = max_elms\n else:\n n_elms = len(elm_index)\n\n # split the data into train, validation and test sets\n training_elms, test_elms = model_selection.train_test_split(\n elm_index[:n_elms],\n test_size=self.fraction_test,\n shuffle=True,\n random_state=config.seed,\n )\n\n # kfold cross validation\n if self.kfold and fold is None:\n raise Exception(\n f\"K-fold cross validation is passed but fold index in range [0, {config.folds}) is not specified.\"\n )\n\n if self.kfold:\n LOGGER.info(\"Using K-fold cross validation\")\n self._kfold_cross_val(training_elms)\n training_elms = self.df[self.df[\"fold\"] != fold][\"elm_events\"]\n validation_elms = self.df[self.df[\"fold\"] == fold][\"elm_events\"]\n else:\n LOGGER.info(\n \"Creating training and validation datasets by simple splitting\"\n )\n training_elms, validation_elms = model_selection.train_test_split(\n training_elms, test_size=self.fraction_validate\n )\n LOGGER.info(f\"Number of training ELM events: {training_elms.size}\")\n LOGGER.info(f\"Number of validation ELM events: {validation_elms.size}\")\n LOGGER.info(f\"Number of test ELM events: {test_elms.size}\")\n\n return training_elms, validation_elms, test_elms", "def _kfold_cross_val(self, training_elms: np.ndarray) -> None:\n kf = model_selection.KFold(\n n_splits=config.folds, shuffle=True, random_state=config.seed\n )\n self.df[\"elm_events\"] = training_elms\n self.df[\"fold\"] = -1\n for f_, (_, valid_idx) in enumerate(kf.split(X=training_elms)):\n self.df.loc[valid_idx, \"fold\"] = f_", "def get_train_input(self, prev, i):\n if i == 0:\n return self._zero_label\n else:\n # TODO(gorban): update to gradually introduce gt labels.\n return self._labels_one_hot[:, i - 1, :]", "def get_init_fn():\n\n variables_to_restore = []\n for var in slim.get_model_variables():\n variables_to_restore.append(var)\n\n checkpoint_path = tf.train.latest_checkpoint(\"./base_checkpoint\")\n\n tf.logging.info('Fine-tuning from %s' % checkpoint_path)\n\n return slim.assign_from_checkpoint_fn(\n checkpoint_path,\n variables_to_restore,\n ignore_missing_vars=False)", "def _get_init_fn():\n exclusions = []\n if FLAGS.checkpoint_exclude_scopes:\n exclusions = [scope.strip()\n for scope in FLAGS.checkpoint_exclude_scopes.split(',')]\n\n variables_to_restore = []\n for var in slim.get_model_variables():\n excluded = False\n for exclusion in exclusions:\n if var.op.name.startswith(exclusion):\n excluded = True\n break\n if not excluded:\n variables_to_restore.append(var)\n\n if tf.gfile.IsDirectory(FLAGS.checkpoint_path):\n checkpoint_path = tf.train.latest_checkpoint(FLAGS.checkpoint_path)\n else:\n checkpoint_path = FLAGS.checkpoint_path\n\n tf.logging.info('Fine-tuning from {}'.format(checkpoint_path))\n\n return slim.assign_from_checkpoint_fn(checkpoint_path, variables_to_restore)", "def get_validation_batch(self, batch_size):\n if batch_size > len(self.val_indices):\n return self.data_handler.slice_data(self.val_indices)\n else:\n return self.data_handler.slice_data(list(np.random.choice(self.val_indices, size=batch_size)))", "def get_eval_input(self, prev, i):\n raise AssertionError('Not implemented')", "def test_early_stopping_no_val_step(tmpdir):\n\n class CurrentModel(EvalModelTemplate):\n def training_step(self, *args, **kwargs):\n output = super().training_step(*args, **kwargs)\n output.update({'my_train_metric': output['loss']}) # could be anything else\n return output\n\n model = CurrentModel()\n model.validation_step = None\n model.val_dataloader = None\n\n stopping = EarlyStopping(monitor='my_train_metric', min_delta=0.1)\n trainer = Trainer(\n default_root_dir=tmpdir,\n early_stop_callback=stopping,\n overfit_pct=0.20,\n max_epochs=5,\n )\n result = trainer.fit(model)\n\n assert result == 1, 'training failed to complete'\n assert trainer.current_epoch < trainer.max_epochs", "def get_input_fn(is_train):\n d = DataInfo(ddir,evalddir)\n hparams = d.generate()\n params = utils.Params(**hparams)\n\n if is_train:\n input_fn = data.get_input_fn(dataset_fn=data.get_train_dataset, mode=TRAIN, params=params, shuffle_queue=10000, repeat=False)\n \n else:\n input_fn = data.get_input_fn(dataset_fn=data.get_eval_dataset, mode=EVAL, params=params, shuffle_queue=10000, repeat=False)\n \n return input_fn, params", "def get_eval_input(self, prev, i):\n if i == 0:\n prev = self._zero_label\n else:\n logit = self.char_logit(prev, char_index=i - 1)\n prev = self.char_one_hot(logit)\n image_feature = self.get_image_feature(char_index=i)\n return tf.concat([image_feature, prev], 1)" ]
[ "0.7417075", "0.7251253", "0.6297536", "0.6207659", "0.5739387", "0.5739387", "0.57109046", "0.5591265", "0.55240226", "0.54016316", "0.53963184", "0.5390611", "0.5354384", "0.5352668", "0.52997166", "0.50982046", "0.509454", "0.5057065", "0.50273144", "0.49092582", "0.48859358", "0.48842114", "0.48813173", "0.48661986", "0.4842402", "0.48323932", "0.4805341", "0.47921005", "0.47847784", "0.47741088" ]
0.79688627
0
Get a `DatasetSentinel` representing an Experiment's `fold_validation_target` Returns
def validation_target(self): return DatasetSentinel("validation_target", **self._dataset_sentinel_helper())
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def train_target(self):\n return DatasetSentinel(\"train_target\", **self._dataset_sentinel_helper())", "def validation_input(self):\n return DatasetSentinel(\"validation_input\", **self._dataset_sentinel_helper())", "def train_input(self):\n return DatasetSentinel(\"train_input\", **self._dataset_sentinel_helper())", "def holdout_target(self):\n return DatasetSentinel(\n \"holdout_target\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def holdout_input(self):\n return DatasetSentinel(\n \"holdout_input\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def validation(self) -> tf.contrib.data.Dataset:\n return self.__validation_dataset", "def training_target(training_data):\n return training_data.status == \"DEFAULT\"", "def __get_x_y_from_training_validation(\n logger, training, validation, predictors, target):\n if training is not None:\n training_X, training_Y = __get_x_y_from_data(logger, training, predictors, target)\n if validation is not None:\n validation_X, validation_Y = __get_x_y_from_data(logger, validation, predictors, target)\n return training_X, training_Y, validation_X, validation_Y", "def _dataset_sentinel_helper(self):\n params = self.cross_experiment_key.parameters\n return dict(\n dataset_hash=params[\"train_dataset\"],\n cv_type=params[\"cross_experiment_params\"].get(\n \"cv_type\", params[\"cross_experiment_params\"].get(\"cross_validation_type\", None)\n ),\n global_random_seed=params[\"cross_experiment_params\"][\"global_random_seed\"],\n random_seeds=params[\"cross_experiment_params\"][\"random_seeds\"],\n )", "def get_input_fn_training(Xtrain_ul, Xtrain_l, Xtest, ytrain_ul, ytrain_l, ytest, batch_size, num_labeled):\n dataset = input_data.Data(Xtrain_ul,\n Xtrain_l,\n Xtest,\n ytrain_ul,\n ytrain_l,\n ytest,\n num_labeled, \n batch_size, \n shuffle=True)\n return dataset.next_batch()", "def split_validation_training_index(allind, splitsize, do_offset, offset_steps):\n i = offset_steps\n lval = splitsize\n if not do_offset:\n i_val = allind[:lval]\n i_train = allind[lval:]\n else:\n i_val = allind[i * lval:(i + 1) * lval]\n i_train = np.concatenate([allind[0:i * lval], allind[(i + 1) * lval:]], axis=0)\n if len(i_val) <= 0:\n print(\"Warning: #Validation data is 0, take 1 training sample instead\")\n i_val = i_train[:1]\n\n return i_train, i_val", "def test_early_stopping_no_val_step(tmpdir):\n\n class CurrentModel(EvalModelTemplate):\n def training_step(self, *args, **kwargs):\n output = super().training_step(*args, **kwargs)\n output.update({'my_train_metric': output['loss']}) # could be anything else\n return output\n\n model = CurrentModel()\n model.validation_step = None\n model.val_dataloader = None\n\n stopping = EarlyStopping(monitor='my_train_metric', min_delta=0.1)\n trainer = Trainer(\n default_root_dir=tmpdir,\n early_stop_callback=stopping,\n overfit_pct=0.20,\n max_epochs=5,\n )\n result = trainer.fit(model)\n\n assert result == 1, 'training failed to complete'\n assert trainer.current_epoch < trainer.max_epochs", "def get_train_input(self, prev, i):\n return self.get_eval_input(prev, i)", "def get_train_input(self, prev, i):\n return self.get_eval_input(prev, i)", "def get_init_fn():\n checkpoint_exclude_scopes = [\"InceptionV1/Logits\", \"InceptionV1/AuxLogits\"]\n\n exclusions = [scope.strip() for scope in checkpoint_exclude_scopes]\n\n variables_to_restore = []\n for var in slim.get_model_variables():\n excluded = False\n for exclusion in exclusions:\n if var.op.name.startswith(exclusion):\n excluded = True\n break\n if not excluded:\n variables_to_restore.append(var)\n\n return slim.assign_from_checkpoint_fn(\n os.path.join(checkpoints_dir, 'inception_v1.ckpt'),\n variables_to_restore)", "def _generate_validation_fold(self):\n\n for offset in range(self.nb_folds):\n # Load all the data from cache (do this to save memory)\n with open(self.data_cache_path_str + \"data_cache.pkl\", \"rb\") as f:\n data_df, target_df = pickle.load(f)\n\n # Generate train and test sets\n data_dates_lst = data_df[\"date\"].drop_duplicates().sort_values().tolist()\n train_start_day = len(data_dates_lst) - ((self.nb_folds - offset) * self.test_nb_days + self.train_nb_days)\n train_end_day = train_start_day + self.train_nb_days\n test_start_day = train_end_day\n test_end_day = test_start_day + self.test_nb_days\n\n train_dates_lst = data_dates_lst[train_start_day:train_end_day]\n test_dates_lst = data_dates_lst[test_start_day:test_end_day]\n\n # Generate train and test labels\n training_set_df = data_df.loc[data_df[\"date\"].isin(train_dates_lst)].reset_index(drop = True)\n testing_set_df = data_df.loc[data_df[\"date\"].isin(test_dates_lst)].reset_index(drop = True)\n new_target_df = target_df.loc[data_df[\"date\"].isin(train_dates_lst)].reset_index(drop = True)\n truth_df = target_df.loc[data_df[\"date\"].isin(test_dates_lst)].reset_index(drop = True)\n\n # Reduce memory usage\n del data_df, target_df\n gc.collect()\n\n # Return result\n yield (training_set_df, testing_set_df, new_target_df, truth_df)", "def test_train_test_split_uni_exo(load_uni_exo_data_target):\n data, target = load_uni_exo_data_target\n\n ####################################\n #### Continuous fh without Gaps ####\n ####################################\n\n #### Integer fh ----\n exp = TSForecastingExperiment()\n fh = 12\n exp.setup(data=data, target=target, fh=fh, seasonal_period=4, session_id=42)\n assert np.all(exp.dataset.index == data.index)\n assert np.all(exp.train.index == data.iloc[: (len(data) - fh)].index)\n assert np.all(exp.test.index == data.iloc[-fh:].index)\n assert np.all(exp.X.index == data.index)\n assert np.all(exp.y.index == data.index)\n assert np.all(exp.X_train.index == data.iloc[: (len(data) - fh)].index)\n assert np.all(exp.X_test.index == data.iloc[-fh:].index)\n assert np.all(exp.y_train.index == data.iloc[: (len(data) - fh)].index)\n assert np.all(exp.y_test.index == data.iloc[-fh:].index)\n assert np.all(exp.dataset_transformed.index == data.index)\n assert np.all(exp.train_transformed.index == data.iloc[: (len(data) - fh)].index)\n assert np.all(exp.test_transformed.index == data.iloc[-fh:].index)\n assert np.all(exp.X_transformed.index == data.index)\n assert np.all(exp.y_transformed.index == data.index)\n assert np.all(exp.X_train_transformed.index == data.iloc[: (len(data) - fh)].index)\n assert np.all(exp.X_test_transformed.index == data.iloc[-fh:].index)\n assert np.all(exp.y_train_transformed.index == data.iloc[: (len(data) - fh)].index)\n assert np.all(exp.y_test_transformed.index == data.iloc[-fh:].index)\n\n #### Numpy fh ----\n exp = TSForecastingExperiment()\n fh = np.arange(1, 10) # 9 values\n exp.setup(data=data, target=target, fh=fh, seasonal_period=4, session_id=42)\n assert np.all(exp.dataset.index == data.index)\n assert np.all(exp.train.index == data.iloc[: (len(data) - max(fh))].index)\n assert np.all(exp.test.index == data.iloc[-len(fh) :].index)\n assert np.all(exp.X.index == data.index)\n assert np.all(exp.y.index == data.index)\n assert np.all(exp.X_train.index == data.iloc[: (len(data) - max(fh))].index)\n assert np.all(exp.X_test.index == data.iloc[-len(fh) :].index)\n assert np.all(exp.y_train.index == data.iloc[: (len(data) - max(fh))].index)\n assert np.all(exp.y_test.index == data.iloc[-len(fh) :].index)\n assert np.all(exp.dataset_transformed.index == data.index)\n assert np.all(\n exp.train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert np.all(exp.test_transformed.index == data.iloc[-len(fh) :].index)\n assert np.all(exp.X_transformed.index == data.index)\n assert np.all(exp.y_transformed.index == data.index)\n assert np.all(\n exp.X_train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert np.all(exp.X_test_transformed.index == data.iloc[-len(fh) :].index)\n assert np.all(\n exp.y_train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert np.all(exp.y_test_transformed.index == data.iloc[-len(fh) :].index)\n\n #### List fh ----\n exp = TSForecastingExperiment()\n fh = [1, 2, 3, 4, 5, 6]\n exp.setup(data=data, target=target, fh=fh, seasonal_period=4, session_id=42)\n assert np.all(exp.dataset.index == data.index)\n assert np.all(exp.train.index == data.iloc[: (len(data) - max(fh))].index)\n assert np.all(exp.test.index == data.iloc[-len(fh) :].index)\n assert np.all(exp.X.index == data.index)\n assert np.all(exp.y.index == data.index)\n assert np.all(exp.X_train.index == data.iloc[: (len(data) - max(fh))].index)\n assert np.all(exp.X_test.index == data.iloc[-len(fh) :].index)\n assert np.all(exp.y_train.index == data.iloc[: (len(data) - max(fh))].index)\n assert np.all(exp.y_test.index == data.iloc[-len(fh) :].index)\n assert np.all(exp.dataset_transformed.index == data.index)\n assert np.all(\n exp.train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert np.all(exp.test_transformed.index == data.iloc[-len(fh) :].index)\n assert np.all(exp.X_transformed.index == data.index)\n assert np.all(exp.y_transformed.index == data.index)\n assert np.all(\n exp.X_train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert np.all(exp.X_test_transformed.index == data.iloc[-len(fh) :].index)\n assert np.all(\n exp.y_train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert np.all(exp.y_test_transformed.index == data.iloc[-len(fh) :].index)\n\n #################################\n #### Continuous fh with Gaps ####\n #################################\n\n #### Numpy fh ----\n exp = TSForecastingExperiment()\n fh = np.arange(7, 13) # 6 values\n exp.setup(data=data, target=target, fh=fh, seasonal_period=4, session_id=42)\n assert np.all(exp.dataset.index == data.index)\n assert np.all(exp.train.index == data.iloc[: (len(data) - max(fh))].index)\n # `test`` call still refers to y_test indices and not X_test indices\n assert len(exp.test) == len(fh)\n assert np.all(exp.X.index == data.index)\n assert np.all(exp.y.index == data.index)\n assert np.all(exp.X_train.index == data.iloc[: (len(data) - max(fh))].index)\n # Exogenous variables will not have any gaps (only target has gaps)\n assert np.all(exp.X_test.index == data.iloc[-max(fh) :].index)\n assert np.all(exp.y_train.index == data.iloc[: (len(data) - max(fh))].index)\n assert len(exp.y_test) == len(fh)\n assert np.all(exp.dataset_transformed.index == data.index)\n assert np.all(\n exp.train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert len(exp.test_transformed) == len(fh)\n assert np.all(exp.X_transformed.index == data.index)\n assert np.all(exp.y_transformed.index == data.index)\n\n #### List fh ----\n exp = TSForecastingExperiment()\n fh = [4, 5, 6]\n exp.setup(data=data, target=target, fh=fh, seasonal_period=4, session_id=42)\n assert np.all(exp.dataset.index == data.index)\n assert np.all(exp.train.index == data.iloc[: (len(data) - max(fh))].index)\n # `test`` call still refers to y_test indices and not X_test indices\n assert len(exp.test) == len(fh)\n assert np.all(exp.X.index == data.index)\n assert np.all(exp.y.index == data.index)\n assert np.all(exp.X_train.index == data.iloc[: (len(data) - max(fh))].index)\n # Exogenous variables will not have any gaps (only target has gaps)\n assert np.all(exp.X_test.index == data.iloc[-max(fh) :].index)\n assert np.all(exp.y_train.index == data.iloc[: (len(data) - max(fh))].index)\n assert len(exp.y_test) == len(fh)\n assert np.all(exp.dataset_transformed.index == data.index)\n assert np.all(\n exp.train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert len(exp.test_transformed) == len(fh)\n assert np.all(exp.X_transformed.index == data.index)\n assert np.all(exp.y_transformed.index == data.index)\n\n ####################################\n #### Discontinuous fh with Gaps ####\n ####################################\n\n #### Numpy fh ----\n exp = TSForecastingExperiment()\n fh = np.array([4, 5, 6, 10, 11, 12]) # 6 values\n exp.setup(data=data, target=target, fh=fh, seasonal_period=4, session_id=42)\n assert np.all(exp.dataset.index == data.index)\n assert np.all(exp.train.index == data.iloc[: (len(data) - max(fh))].index)\n # `test`` call still refers to y_test indices and not X_test indices\n assert len(exp.test) == len(fh)\n assert np.all(exp.X.index == data.index)\n assert np.all(exp.y.index == data.index)\n assert np.all(exp.X_train.index == data.iloc[: (len(data) - max(fh))].index)\n # Exogenous variables will not have any gaps (only target has gaps)\n assert np.all(exp.X_test.index == data.iloc[-max(fh) :].index)\n assert np.all(exp.y_train.index == data.iloc[: (len(data) - max(fh))].index)\n assert len(exp.y_test) == len(fh)\n assert np.all(exp.dataset_transformed.index == data.index)\n assert np.all(\n exp.train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert len(exp.test_transformed) == len(fh)\n assert np.all(exp.X_transformed.index == data.index)\n assert np.all(exp.y_transformed.index == data.index)\n assert np.all(\n exp.X_train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n # Exogenous variables will not have any gaps (only target has gaps)\n assert np.all(exp.X_test_transformed.index == data.iloc[-max(fh) :].index)\n assert np.all(\n exp.y_train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert len(exp.y_test_transformed) == len(fh)\n\n #### List fh ----\n exp = TSForecastingExperiment()\n fh = [4, 5, 6, 10, 11, 12]\n exp.setup(data=data, target=target, fh=fh, seasonal_period=4, session_id=42)\n assert np.all(exp.dataset.index == data.index)\n assert np.all(exp.train.index == data.iloc[: (len(data) - max(fh))].index)\n # `test`` call still refers to y_test indices and not X_test indices\n assert len(exp.test) == len(fh)\n assert np.all(exp.X.index == data.index)\n assert np.all(exp.y.index == data.index)\n assert np.all(exp.X_train.index == data.iloc[: (len(data) - max(fh))].index)\n # Exogenous variables will not have any gaps (only target has gaps)\n assert np.all(exp.X_test.index == data.iloc[-max(fh) :].index)\n assert np.all(exp.y_train.index == data.iloc[: (len(data) - max(fh))].index)\n assert len(exp.y_test) == len(fh)\n assert np.all(exp.dataset_transformed.index == data.index)\n assert np.all(\n exp.train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert len(exp.test_transformed) == len(fh)\n assert np.all(exp.X_transformed.index == data.index)\n assert np.all(exp.y_transformed.index == data.index)\n assert np.all(\n exp.X_train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n # Exogenous variables will not have any gaps (only target has gaps)\n assert np.all(exp.X_test_transformed.index == data.iloc[-max(fh) :].index)\n assert np.all(\n exp.y_train_transformed.index == data.iloc[: (len(data) - max(fh))].index\n )\n assert len(exp.y_test_transformed) == len(fh)", "def _kfold_cross_val(self, training_elms: np.ndarray) -> None:\n kf = model_selection.KFold(\n n_splits=config.folds, shuffle=True, random_state=config.seed\n )\n self.df[\"elm_events\"] = training_elms\n self.df[\"fold\"] = -1\n for f_, (_, valid_idx) in enumerate(kf.split(X=training_elms)):\n self.df.loc[valid_idx, \"fold\"] = f_", "def get_eval_input(self, prev, i):\n del prev, i\n # The attention_decoder will fetch image features from the net, no need for\n # extra inputs.\n return self._zero_label", "def get_validation_batch(self, batch_size):\n if batch_size > len(self.val_indices):\n return self.data_handler.slice_data(self.val_indices)\n else:\n return self.data_handler.slice_data(list(np.random.choice(self.val_indices, size=batch_size)))", "def sample_train_data(dataset ,target,data_len, resp = True ):\r\n np.random.seed(222)\r\n ixes = np.random.choice(dataset.index, data_len, replace = False)\r\n print(ixes)\r\n under_df = dataset.iloc[ixes]\r\n if resp==True:\r\n under_target = target.iloc[ixes]\r\n return under_df, under_target\r\n else:\r\n return under_df", "def get_train(self, preprocess=False):\n return self._dataset('train', self._directory, 'sharded_train_0shifted_affnist.tfrecords')", "def load_input(io_config):\n\n fold = io_config['fold']\n\n train_data_fnames = io_config['file']['train_data']\n\n train_fname = [t[1] for t in train_data_fnames\n if t[0] == fold][0]\n\n return train_fname", "def cross_validation_datasets(self, fold):\n if fold > len(self): fold = len(self) / 2\n stratified = self.stratified_bunches(fold)\n datasets = []\n for index in range(len(stratified)):\n gold = GoldInstances(training_as_gold(stratified[index]))\n rest = flatten(stratified[:index]) + flatten(stratified[index + 1:])\n training = TrainingInstances(rest)\n datasets.append((training, gold))\n return datasets", "def _partition_elms(\n self, max_elms: int = None, fold: int = None\n ) -> Tuple[np.ndarray, np.ndarray, np.ndarray]:\n # get ELM indices from datafile\n elm_index, _ = self._read_file()\n\n # limit the data according to the max number of events passed\n if max_elms is not None and max_elms != -1:\n LOGGER.info(f\"Limiting data read to {max_elms} events.\")\n n_elms = max_elms\n else:\n n_elms = len(elm_index)\n\n # split the data into train, validation and test sets\n training_elms, test_elms = model_selection.train_test_split(\n elm_index[:n_elms],\n test_size=self.fraction_test,\n shuffle=True,\n random_state=config.seed,\n )\n\n # kfold cross validation\n if self.kfold and fold is None:\n raise Exception(\n f\"K-fold cross validation is passed but fold index in range [0, {config.folds}) is not specified.\"\n )\n\n if self.kfold:\n LOGGER.info(\"Using K-fold cross validation\")\n self._kfold_cross_val(training_elms)\n training_elms = self.df[self.df[\"fold\"] != fold][\"elm_events\"]\n validation_elms = self.df[self.df[\"fold\"] == fold][\"elm_events\"]\n else:\n LOGGER.info(\n \"Creating training and validation datasets by simple splitting\"\n )\n training_elms, validation_elms = model_selection.train_test_split(\n training_elms, test_size=self.fraction_validate\n )\n LOGGER.info(f\"Number of training ELM events: {training_elms.size}\")\n LOGGER.info(f\"Number of validation ELM events: {validation_elms.size}\")\n LOGGER.info(f\"Number of test ELM events: {test_elms.size}\")\n\n return training_elms, validation_elms, test_elms", "def _get_training_dataset(self):\n\n return tf.data.Dataset.zip((self.conditioned_dataset, self.dataset))", "def example(self):\n result = getattr(self, '_example', None)\n if result is None:\n # No example batch was found, so get one from the `.train` dataset\n result = next(iter(self.train))\n # And cache it for next time\n self._example = result\n return result", "def _get_init_fn():\n exclusions = []\n if FLAGS.checkpoint_exclude_scopes:\n exclusions = [scope.strip()\n for scope in FLAGS.checkpoint_exclude_scopes.split(',')]\n\n variables_to_restore = []\n for var in slim.get_model_variables():\n excluded = False\n for exclusion in exclusions:\n if var.op.name.startswith(exclusion):\n excluded = True\n break\n if not excluded:\n variables_to_restore.append(var)\n\n if tf.gfile.IsDirectory(FLAGS.checkpoint_path):\n checkpoint_path = tf.train.latest_checkpoint(FLAGS.checkpoint_path)\n else:\n checkpoint_path = FLAGS.checkpoint_path\n\n tf.logging.info('Fine-tuning from {}'.format(checkpoint_path))\n\n return slim.assign_from_checkpoint_fn(checkpoint_path, variables_to_restore)", "def keep_cross_validation_fold_assignment(self):\n return self._parms.get(\"keep_cross_validation_fold_assignment\")", "def get_target_encoder(self, train: NumpyOrPandas) -> Optional[type]:\n target_encoder = None\n if train.folds is not None:\n if train.task.name in [\"binary\", \"reg\"]:\n target_encoder = TargetEncoder\n else:\n n_classes = train.target.max() + 1\n if n_classes <= self.multiclass_te_co:\n target_encoder = MultiClassTargetEncoder\n\n return target_encoder" ]
[ "0.7272191", "0.7040783", "0.65106875", "0.6453621", "0.58669484", "0.52230227", "0.52210194", "0.5189654", "0.5189253", "0.5166722", "0.50729924", "0.5023478", "0.4998633", "0.4998633", "0.49692518", "0.4958584", "0.4885863", "0.4881906", "0.48708826", "0.4870653", "0.48660624", "0.48495704", "0.48440486", "0.48257494", "0.480864", "0.47701162", "0.47636676", "0.47615838", "0.47511327", "0.47497708" ]
0.80776197
0
Get a `DatasetSentinel` representing an Experiment's `holdout_input_data` Returns
def holdout_input(self): return DatasetSentinel( "holdout_input", self.cross_experiment_key.parameters["holdout_dataset"] )
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def holdout_target(self):\n return DatasetSentinel(\n \"holdout_target\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def train_input(self):\n return DatasetSentinel(\"train_input\", **self._dataset_sentinel_helper())", "def validation_input(self):\n return DatasetSentinel(\"validation_input\", **self._dataset_sentinel_helper())", "def train_target(self):\n return DatasetSentinel(\"train_target\", **self._dataset_sentinel_helper())", "def validation_target(self):\n return DatasetSentinel(\"validation_target\", **self._dataset_sentinel_helper())", "def _dataset_sentinel_helper(self):\n params = self.cross_experiment_key.parameters\n return dict(\n dataset_hash=params[\"train_dataset\"],\n cv_type=params[\"cross_experiment_params\"].get(\n \"cv_type\", params[\"cross_experiment_params\"].get(\"cross_validation_type\", None)\n ),\n global_random_seed=params[\"cross_experiment_params\"][\"global_random_seed\"],\n random_seeds=params[\"cross_experiment_params\"][\"random_seeds\"],\n )", "def test_additional_data(self):\n task = SentinelHubInputTask(\n bands_feature=(FeatureType.DATA, 'BANDS'),\n bands=['B01', 'B02', 'B05'],\n additional_data=[\n (FeatureType.MASK, 'dataMask', 'IS_DATA'),\n (FeatureType.MASK, 'CLM'),\n (FeatureType.MASK, 'SCL'),\n (FeatureType.MASK, 'SNW'),\n (FeatureType.MASK, 'CLD'),\n (FeatureType.DATA, 'CLP'),\n (FeatureType.DATA, 'viewAzimuthMean', 'view_azimuth_mean'),\n (FeatureType.DATA, 'sunAzimuthAngles'),\n (FeatureType.DATA, 'sunZenithAngles')\n ],\n size=self.size,\n maxcc=self.maxcc,\n time_difference=self.time_difference,\n data_collection=DataCollection.SENTINEL2_L2A,\n max_threads=self.max_threads\n )\n\n eopatch = task.execute(bbox=self.bbox, time_interval=self.time_interval)\n\n bands = eopatch[(FeatureType.DATA, 'BANDS')]\n is_data = eopatch[(FeatureType.MASK, 'IS_DATA')]\n clm = eopatch[(FeatureType.MASK, 'CLM')]\n scl = eopatch[(FeatureType.MASK, 'SCL')]\n snw = eopatch[(FeatureType.MASK, 'SNW')]\n cld = eopatch[(FeatureType.MASK, 'CLD')]\n clp = eopatch[(FeatureType.DATA, 'CLP')]\n view_azimuth_mean = eopatch[(FeatureType.DATA, 'view_azimuth_mean')]\n sun_azimuth_angles = eopatch[(FeatureType.DATA, 'sunAzimuthAngles')]\n sun_zenith_angles = eopatch[(FeatureType.DATA, 'sunZenithAngles')]\n\n self.assertTrue(np.allclose(array_stats(bands), [0.027, 0.0243, 0.0162]))\n\n width, height = self.size\n self.assertTrue(bands.shape == (4, height, width, 3))\n self.assertTrue(is_data.shape == (4, height, width, 1))\n self.assertTrue(is_data.dtype == bool)\n self.assertTrue(clm.shape == (4, height, width, 1))\n self.assertTrue(clm.dtype == np.uint8)\n self.assertTrue(scl.shape == (4, height, width, 1))\n self.assertTrue(snw.shape == (4, height, width, 1))\n self.assertTrue(cld.shape == (4, height, width, 1))\n self.assertTrue(clp.shape == (4, height, width, 1))\n self.assertTrue(view_azimuth_mean.shape == (4, height, width, 1))\n self.assertTrue(sun_azimuth_angles.shape == (4, height, width, 1))\n self.assertTrue(sun_zenith_angles.shape == (4, height, width, 1))\n self.assertTrue(len(eopatch.timestamp) == 4)", "def get_dataset_sampler(self):\n return None", "def get_interest_variable(\n in_dataset, sensor_var, date_col, hr_col, numeric_var, target_sensor=\"A620\"\n):\n dataset_pproc = in_dataset.loc[\n in_dataset[sensor_var] == target_sensor, [date_col, hr_col] + [numeric_var]\n ]\n hrs_str = dataset_pproc[hr_col].to_string()\n dates_str = dataset_pproc[date_col]\n\n dataset_pproc[date_col] = pd.to_datetime(dataset_pproc[date_col])\n dataset_pproc.set_index([date_col, hr_col], inplace=True)\n dataset_pproc.fillna(method=\"ffill\", inplace=True)\n dataset_pproc.interpolate(method=\"linear\", axis=0)\n\n return dataset_pproc", "def _create_dataset(source=''):\n return ExperimentalDataset()", "def get_train_inputs(self, example):\n return example", "def get_data_for_semi_supervised(self,percent_dropout,num_dropout_corruptions_per_point):\n\n\t\t\"\"\"Select the corrupted data if applicable, and otherwise the original training data\"\"\"\n\t\treturn self.__get_data__(percent_dropout,num_dropout_corruptions_per_point,bool_targetted_dropout=True)", "def input(self):\n try:\n return self.inputs[-1]\n except IndexError:\n pass\n raise ValueError(\"The sample method has not been called\")", "def get_input_data(sample):\n with checkpoints.query_portal.get(sample=sample).output[0].open() as f:\n data = json.read(f)\n return data", "def get_dataset(dataset_name):\n if dataset_name == \"Iris\":\n data = datasets.load_iris()\n elif dataset_name == \"Breast Cancer\":\n data = datasets.load_breast_cancer()\n else:\n data = datasets.load_wine()\n\n X = data.data\n y = data.target\n return X, y", "def get_input_fn_training(Xtrain_ul, Xtrain_l, Xtest, ytrain_ul, ytrain_l, ytest, batch_size, num_labeled):\n dataset = input_data.Data(Xtrain_ul,\n Xtrain_l,\n Xtest,\n ytrain_ul,\n ytrain_l,\n ytest,\n num_labeled, \n batch_size, \n shuffle=True)\n return dataset.next_batch()", "def _input_fn(input_pipeline_context=None):\n return _create_dataset(options, is_training, input_pipeline_context)", "def get_dataset(args):\n\n if args['experiment']['dataset'] == Dataset.mindsets:\n xs, ys, cs = make_mindsets(mindset_sizes=args['dataset']['mindset_sizes'],\n nb_questions=args['dataset']['nb_questions'],\n nb_useless=args['dataset']['nb_useless'],\n noise=args['dataset']['noise'],\n seed=args['experiment']['seed'])\n\n return Data(xs=xs, ys=ys, cs=cs)\n\n if args['experiment']['dataset'] == Dataset.questionnaire_likert:\n xs, ys, cs = make_likert_questionnaire(nb_samples=args['dataset']['nb_samples'],\n nb_features=args['dataset']['nb_features'],\n nb_mindsets=args['dataset']['nb_mindsets'],\n centers=args['dataset']['centers'],\n range_answers=args['dataset']['range_answers'],\n seed=args['experiment']['seed'])\n\n return Data(xs=xs, ys=ys, cs=cs)\n\n if args['experiment']['dataset'] == Dataset.retinal:\n xs, ys = load_RETINAL(root_path=args['root_dir'],\n nb_bins=args['dataset']['nb_bins'],\n max_idx=args['dataset']['max_idx'])\n\n return Data(xs=xs, ys=ys)\n\n if args['experiment']['dataset'] == Dataset.moons:\n xs, ys = make_moons(n_samples=args['dataset']['n_samples'],\n noise=args['dataset']['noise'],\n random_state=args['experiment']['seed'])\n\n return Data(xs=xs, ys=ys)\n\n if args['experiment']['dataset'] == Dataset.breast_cancer_wisconsin:\n xs, ys = load_CANCER(args['dataset']['nb_bins'])\n\n return Data(xs=xs, ys=ys)\n\n if args['experiment']['dataset'] == Dataset.SBM:\n A, ys, G = load_SBM(block_sizes=args['dataset']['block_sizes'],\n p_in=args['dataset']['p'],\n p_out=args['dataset']['q'],\n seed=args['experiment']['seed'])\n\n return Data(ys=ys, A=A, G=G)\n\n if args['experiment']['dataset'] == Dataset.gaussian_mixture:\n xs, ys = make_blobs(n_samples=args['dataset']['blob_sizes'],\n centers=args['dataset']['blob_centers'],\n n_features=args['dataset']['blob_centers'],\n cluster_std=args['dataset']['blob_variances'],\n random_state=args['experiment']['seed'])\n\n return Data(xs=xs, ys=ys)\n\n if args['experiment']['dataset'] == Dataset.LFR:\n A, ys, G = load_LFR(nb_nodes=args['dataset']['nb_nodes'],\n tau1=args['dataset']['tau1'],\n tau2=args['dataset']['tau2'],\n mu=args['dataset']['mu'],\n average_degree=args['dataset']['average_degree'],\n min_community=args['dataset']['min_community'],\n seed=args['experiment']['seed'])\n\n return Data(ys=ys, A=A, G=G)\n\n if args['experiment']['dataset'] == Dataset.wave:\n df = pd.read_csv('datasets/waveform.csv')\n xs = df[df.columns[:-1]].to_numpy()\n ys = df[df.columns[-1]].to_numpy()\n\n return Data(xs=xs, ys=ys)\n\n raise ValueError('Wrong name for a dataset')", "def fixture_input_data(wrapper_arguments):\n n_features, classes = wrapper_arguments\n num_data = 50\n data = np.random.rand(num_data, n_features)\n labels = np.random.choice(classes, size=num_data)\n\n return data, labels", "def readMNISTData():\n mnist = input_data.read_data_sets(\"MNIST_data\",one_hot=True) \n return mnist", "def get_seeds_null(samples,signal):\n theta_seeds={}\n if selected is None:\n SRlist = range(self.N_SR)\n theta_samples = samples[:,0,self.N_SR:].T\n else:\n SRlist = [selected] \n theta_samples = samples[:,0,1].T\n for i,x in zip(SRlist,theta_samples):\n theta_MLE = x\n theta_seeds['theta_{0}'.format(i)] = theta_MLE\n return theta_seeds", "def tempfeeder_exp():\n global _tempfeeder_exp\n if _tempfeeder_exp is None:\n _tempfeeder_exp = UserLoads_Experiment(DATA_WITHOUT_DUPES_PATH)\n return _tempfeeder_exp", "def iris():\n return IrisDataset()", "def find_input_dataset(stage, rconfig, data_type=None):\n\n # Use the stage-to-data mapping to find the data_type if none was handed in\n if data_type is None:\n data_type = DOCUMENT_PROCESSING_IO[stage]['in']\n # Get all data sets D for input name\n dirname = os.path.join(rconfig.target_path, 'data', data_type)\n datasets1 = [ds for ds in os.listdir(dirname) if ds.isdigit()]\n datasets2 = [DataSet(stage, data_type, rconfig, ds) for ds in datasets1]\n # Filer the datasets making sure that d.trace + d.head matches\n # rconfig.pipeline(txt).trace\n datasets3 = [ds for ds in datasets2 if ds.input_matches_global_config()]\n # If there is one result, return it, otherwise write a warning and exit\n if len(datasets3) == 1:\n return datasets3[0]\n elif len(datasets3) > 1:\n print \"WARNING, more than one approriate training set:\"\n for ds in datasets3:\n print ' ', ds\n sys.exit(\"Exiting...\")\n elif len(datasets3) == 0:\n print \"WARNING: no datasets available to meet input requirements\"\n sys.exit(\"Exiting...\")", "def sample_train_data(dataset ,target,data_len, resp = True ):\r\n np.random.seed(222)\r\n ixes = np.random.choice(dataset.index, data_len, replace = False)\r\n print(ixes)\r\n under_df = dataset.iloc[ixes]\r\n if resp==True:\r\n under_target = target.iloc[ixes]\r\n return under_df, under_target\r\n else:\r\n return under_df", "def default_dataset(self):\n return self.get_by_index(self._default_index)", "def get_data_only_random_dropout(self,percent_dropout,num_dropout_corruptions_per_point):\n\n\t\t\"\"\"Select the corrupted data if applicable, and otherwise the original training data\"\"\"\n\t\treturn self.__get_data__(percent_dropout,num_dropout_corruptions_per_point,bool_targetted_dropout=False)", "def get_datasetID(self):\n\t\treturn self.prDoc['inputs']['data'][0]['datasetID']", "def GetLabelInput(self) -> \"itkImageSS3 const *\":\n return _itkLabelStatisticsImageFilterPython.itkLabelStatisticsImageFilterIUS3ISS3_GetLabelInput(self)", "def GetInput(self) -> \"itkHistogramD const *\":\n return _itkHistogramThresholdCalculatorPython.itkHistogramThresholdCalculatorHDF_GetInput(self)" ]
[ "0.7605484", "0.70850253", "0.65761745", "0.62563837", "0.58847857", "0.5119277", "0.49333444", "0.48865077", "0.48654017", "0.48275244", "0.48057696", "0.47976065", "0.47965127", "0.4744823", "0.47400033", "0.47386026", "0.47028476", "0.4699324", "0.46962723", "0.4694681", "0.46845272", "0.46641162", "0.46598667", "0.46492717", "0.46456861", "0.4645556", "0.46335387", "0.46245515", "0.46165693", "0.461459" ]
0.8473368
0
Get a `DatasetSentinel` representing an Experiment's `holdout_target_data` Returns
def holdout_target(self): return DatasetSentinel( "holdout_target", self.cross_experiment_key.parameters["holdout_dataset"] )
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def holdout_input(self):\n return DatasetSentinel(\n \"holdout_input\", self.cross_experiment_key.parameters[\"holdout_dataset\"]\n )", "def train_target(self):\n return DatasetSentinel(\"train_target\", **self._dataset_sentinel_helper())", "def validation_target(self):\n return DatasetSentinel(\"validation_target\", **self._dataset_sentinel_helper())", "def train_input(self):\n return DatasetSentinel(\"train_input\", **self._dataset_sentinel_helper())", "def validation_input(self):\n return DatasetSentinel(\"validation_input\", **self._dataset_sentinel_helper())", "def sample_train_data(dataset ,target,data_len, resp = True ):\r\n np.random.seed(222)\r\n ixes = np.random.choice(dataset.index, data_len, replace = False)\r\n print(ixes)\r\n under_df = dataset.iloc[ixes]\r\n if resp==True:\r\n under_target = target.iloc[ixes]\r\n return under_df, under_target\r\n else:\r\n return under_df", "def get_dataset_sampler(self):\n return None", "def load_own_target():\n data = Data()\n target = data.get_label_col()\n return target.values", "def holdout(self, X, y):\n #X,y = self.define_dataset(self.df, self.col_list, self.target_var)\n X_train, X_test, y_train, y_test = train_test_split(\n X, y, test_size = 0.3, random_state = 0)\n\n return X_train, X_test, y_train, y_test", "def _dataset_sentinel_helper(self):\n params = self.cross_experiment_key.parameters\n return dict(\n dataset_hash=params[\"train_dataset\"],\n cv_type=params[\"cross_experiment_params\"].get(\n \"cv_type\", params[\"cross_experiment_params\"].get(\"cross_validation_type\", None)\n ),\n global_random_seed=params[\"cross_experiment_params\"][\"global_random_seed\"],\n random_seeds=params[\"cross_experiment_params\"][\"random_seeds\"],\n )", "def _create_dataset(source=''):\n return ExperimentalDataset()", "def get_data_only_random_dropout(self,percent_dropout,num_dropout_corruptions_per_point):\n\n\t\t\"\"\"Select the corrupted data if applicable, and otherwise the original training data\"\"\"\n\t\treturn self.__get_data__(percent_dropout,num_dropout_corruptions_per_point,bool_targetted_dropout=False)", "def get_dataset(dataset_name):\n if dataset_name == \"Iris\":\n data = datasets.load_iris()\n elif dataset_name == \"Breast Cancer\":\n data = datasets.load_breast_cancer()\n else:\n data = datasets.load_wine()\n\n X = data.data\n y = data.target\n return X, y", "def discard_none_targets(dataset):\r\n indices = []\r\n for (ii,sample) in enumerate(dataset):\r\n target = sample[1]\r\n if target is not None:\r\n indices.append(ii)\r\n\r\n return Subset(dataset,indices)", "def get_data_for_semi_supervised(self,percent_dropout,num_dropout_corruptions_per_point):\n\n\t\t\"\"\"Select the corrupted data if applicable, and otherwise the original training data\"\"\"\n\t\treturn self.__get_data__(percent_dropout,num_dropout_corruptions_per_point,bool_targetted_dropout=True)", "def target(self):\n return self.data.target", "def target(self, example):\n return example[self.target_index]", "def default_dataset(self):\n return self.get_by_index(self._default_index)", "def get_expo_data(\n data: Union[pd.Series, pd.DataFrame],\n target_column: str = \"close\",\n trend: str = \"A\",\n seasonal: str = \"A\",\n seasonal_periods: int = 7,\n dampen: str = \"F\",\n n_predict: int = 5,\n start_window: float = 0.85,\n forecast_horizon: int = 5,\n metric: str = \"mape\",\n) -> Tuple[\n List[TimeSeries],\n List[TimeSeries],\n List[TimeSeries],\n Optional[Union[float, ndarray]],\n ExponentialSmoothing,\n]:\n\n use_scalers = False\n _, ticker_series = helpers.get_series(data, target_column, is_scaler=use_scalers)\n\n if trend == \"M\":\n trend_model = ModelMode.MULTIPLICATIVE\n elif trend == \"N\":\n trend_model = ModelMode.NONE\n else: # Default\n trend_model = ModelMode.ADDITIVE\n\n if seasonal == \"M\":\n seasonal_model = SeasonalityMode.MULTIPLICATIVE\n elif seasonal == \"N\":\n seasonal_model = SeasonalityMode.NONE\n else: # Default\n seasonal_model = SeasonalityMode.ADDITIVE\n\n damped = True\n if dampen == \"F\":\n damped = False\n\n # Model Init\n model_es = ExponentialSmoothing(\n trend=trend_model,\n seasonal=seasonal_model,\n seasonal_periods=int(seasonal_periods),\n damped=damped,\n random_state=42,\n )\n\n try:\n # Historical backtesting\n historical_fcast_es = model_es.historical_forecasts(\n ticker_series, # backtest on entire ts\n start=float(start_window),\n forecast_horizon=int(forecast_horizon),\n verbose=True,\n )\n except Exception as e: # noqa\n error = str(e)\n # lets translate this to something everyone understands\n if \"with`overlap_end` set to `False`.\" in error:\n console.print(\n \"[red]Dataset too small.[/red]\"\n \"[red] Please increase size to at least 100 data points.[/red]\"\n )\n else:\n console.print(f\"[red]{error}[/red]\")\n return [], [], [], None, None\n\n # train new model on entire timeseries to provide best current forecast\n best_model = ExponentialSmoothing(\n trend=trend_model,\n seasonal=seasonal_model,\n seasonal_periods=int(seasonal_periods),\n damped=damped,\n random_state=42,\n )\n\n # we have the historical fcast, now lets train on entire set and predict.\n best_model.fit(ticker_series)\n probabilistic_forecast = best_model.predict(int(n_predict), num_samples=500)\n\n if metric == \"rmse\":\n precision = rmse(actual_series=ticker_series, pred_series=historical_fcast_es)\n elif metric == \"mse\":\n precision = mse(actual_series=ticker_series, pred_series=historical_fcast_es)\n elif metric == \"mape\":\n precision = mape(actual_series=ticker_series, pred_series=historical_fcast_es)\n elif metric == \"smape\":\n precision = smape(actual_series=ticker_series, pred_series=historical_fcast_es)\n\n console.print(\n f\"Exponential smoothing obtains {metric.upper()}: {precision:.2f}% \\n\"\n )\n\n return (\n ticker_series,\n historical_fcast_es,\n probabilistic_forecast,\n precision,\n best_model,\n )", "def define_holdout_set(\n train_set: pd.DataFrame,\n holdout_set: Union[pd.DataFrame, callable, str, None],\n target_column: Union[str, List[str]],\n) -> Tuple[pd.DataFrame, Optional[pd.DataFrame]]:\n #################### Update `holdout_set` ####################\n if callable(holdout_set):\n train_set, holdout_set = holdout_set(train_set, target_column)\n elif isinstance(holdout_set, str):\n holdout_set = pd.read_csv(holdout_set)\n #################### Validate `holdout_set` ####################\n try:\n if holdout_set is None or np.array_equal(train_set.columns, holdout_set.columns):\n return train_set, holdout_set\n except AttributeError:\n raise TypeError(f\"holdout_set must be None, DataFrame, callable, or str, not {holdout_set}\")\n raise ValueError(f\"Mismatched columns\\n{train_set.columns}\\n!=\\n{holdout_set.columns}\")", "def get_interest_variable(\n in_dataset, sensor_var, date_col, hr_col, numeric_var, target_sensor=\"A620\"\n):\n dataset_pproc = in_dataset.loc[\n in_dataset[sensor_var] == target_sensor, [date_col, hr_col] + [numeric_var]\n ]\n hrs_str = dataset_pproc[hr_col].to_string()\n dates_str = dataset_pproc[date_col]\n\n dataset_pproc[date_col] = pd.to_datetime(dataset_pproc[date_col])\n dataset_pproc.set_index([date_col, hr_col], inplace=True)\n dataset_pproc.fillna(method=\"ffill\", inplace=True)\n dataset_pproc.interpolate(method=\"linear\", axis=0)\n\n return dataset_pproc", "def get_dataset(self):\n return", "def backup_dataset(outfile=None):\n return backup_es(Dataset, outfile=outfile)", "def target_dummy(config: Configuration, seed: int) -> int:\n return seed", "def test_additional_data(self):\n task = SentinelHubInputTask(\n bands_feature=(FeatureType.DATA, 'BANDS'),\n bands=['B01', 'B02', 'B05'],\n additional_data=[\n (FeatureType.MASK, 'dataMask', 'IS_DATA'),\n (FeatureType.MASK, 'CLM'),\n (FeatureType.MASK, 'SCL'),\n (FeatureType.MASK, 'SNW'),\n (FeatureType.MASK, 'CLD'),\n (FeatureType.DATA, 'CLP'),\n (FeatureType.DATA, 'viewAzimuthMean', 'view_azimuth_mean'),\n (FeatureType.DATA, 'sunAzimuthAngles'),\n (FeatureType.DATA, 'sunZenithAngles')\n ],\n size=self.size,\n maxcc=self.maxcc,\n time_difference=self.time_difference,\n data_collection=DataCollection.SENTINEL2_L2A,\n max_threads=self.max_threads\n )\n\n eopatch = task.execute(bbox=self.bbox, time_interval=self.time_interval)\n\n bands = eopatch[(FeatureType.DATA, 'BANDS')]\n is_data = eopatch[(FeatureType.MASK, 'IS_DATA')]\n clm = eopatch[(FeatureType.MASK, 'CLM')]\n scl = eopatch[(FeatureType.MASK, 'SCL')]\n snw = eopatch[(FeatureType.MASK, 'SNW')]\n cld = eopatch[(FeatureType.MASK, 'CLD')]\n clp = eopatch[(FeatureType.DATA, 'CLP')]\n view_azimuth_mean = eopatch[(FeatureType.DATA, 'view_azimuth_mean')]\n sun_azimuth_angles = eopatch[(FeatureType.DATA, 'sunAzimuthAngles')]\n sun_zenith_angles = eopatch[(FeatureType.DATA, 'sunZenithAngles')]\n\n self.assertTrue(np.allclose(array_stats(bands), [0.027, 0.0243, 0.0162]))\n\n width, height = self.size\n self.assertTrue(bands.shape == (4, height, width, 3))\n self.assertTrue(is_data.shape == (4, height, width, 1))\n self.assertTrue(is_data.dtype == bool)\n self.assertTrue(clm.shape == (4, height, width, 1))\n self.assertTrue(clm.dtype == np.uint8)\n self.assertTrue(scl.shape == (4, height, width, 1))\n self.assertTrue(snw.shape == (4, height, width, 1))\n self.assertTrue(cld.shape == (4, height, width, 1))\n self.assertTrue(clp.shape == (4, height, width, 1))\n self.assertTrue(view_azimuth_mean.shape == (4, height, width, 1))\n self.assertTrue(sun_azimuth_angles.shape == (4, height, width, 1))\n self.assertTrue(sun_zenith_angles.shape == (4, height, width, 1))\n self.assertTrue(len(eopatch.timestamp) == 4)", "def target_temperature(self):\n return self._boiler.setpoint", "def eye_during_period(self, time_period, alias):\n\t\twith pd.get_store(self.input_object) as h5_file:\n\t\t\tperiod_block_nr = self.sample_in_block(sample = time_period[0], block_table = h5_file['%s/blocks'%alias])\n\t\t\teye = h5_file['%s/blocks'%alias]['eye_recorded'][period_block_nr]\n\t\treturn eye", "def get_dataset(args):\n\n if args['experiment']['dataset'] == Dataset.mindsets:\n xs, ys, cs = make_mindsets(mindset_sizes=args['dataset']['mindset_sizes'],\n nb_questions=args['dataset']['nb_questions'],\n nb_useless=args['dataset']['nb_useless'],\n noise=args['dataset']['noise'],\n seed=args['experiment']['seed'])\n\n return Data(xs=xs, ys=ys, cs=cs)\n\n if args['experiment']['dataset'] == Dataset.questionnaire_likert:\n xs, ys, cs = make_likert_questionnaire(nb_samples=args['dataset']['nb_samples'],\n nb_features=args['dataset']['nb_features'],\n nb_mindsets=args['dataset']['nb_mindsets'],\n centers=args['dataset']['centers'],\n range_answers=args['dataset']['range_answers'],\n seed=args['experiment']['seed'])\n\n return Data(xs=xs, ys=ys, cs=cs)\n\n if args['experiment']['dataset'] == Dataset.retinal:\n xs, ys = load_RETINAL(root_path=args['root_dir'],\n nb_bins=args['dataset']['nb_bins'],\n max_idx=args['dataset']['max_idx'])\n\n return Data(xs=xs, ys=ys)\n\n if args['experiment']['dataset'] == Dataset.moons:\n xs, ys = make_moons(n_samples=args['dataset']['n_samples'],\n noise=args['dataset']['noise'],\n random_state=args['experiment']['seed'])\n\n return Data(xs=xs, ys=ys)\n\n if args['experiment']['dataset'] == Dataset.breast_cancer_wisconsin:\n xs, ys = load_CANCER(args['dataset']['nb_bins'])\n\n return Data(xs=xs, ys=ys)\n\n if args['experiment']['dataset'] == Dataset.SBM:\n A, ys, G = load_SBM(block_sizes=args['dataset']['block_sizes'],\n p_in=args['dataset']['p'],\n p_out=args['dataset']['q'],\n seed=args['experiment']['seed'])\n\n return Data(ys=ys, A=A, G=G)\n\n if args['experiment']['dataset'] == Dataset.gaussian_mixture:\n xs, ys = make_blobs(n_samples=args['dataset']['blob_sizes'],\n centers=args['dataset']['blob_centers'],\n n_features=args['dataset']['blob_centers'],\n cluster_std=args['dataset']['blob_variances'],\n random_state=args['experiment']['seed'])\n\n return Data(xs=xs, ys=ys)\n\n if args['experiment']['dataset'] == Dataset.LFR:\n A, ys, G = load_LFR(nb_nodes=args['dataset']['nb_nodes'],\n tau1=args['dataset']['tau1'],\n tau2=args['dataset']['tau2'],\n mu=args['dataset']['mu'],\n average_degree=args['dataset']['average_degree'],\n min_community=args['dataset']['min_community'],\n seed=args['experiment']['seed'])\n\n return Data(ys=ys, A=A, G=G)\n\n if args['experiment']['dataset'] == Dataset.wave:\n df = pd.read_csv('datasets/waveform.csv')\n xs = df[df.columns[:-1]].to_numpy()\n ys = df[df.columns[-1]].to_numpy()\n\n return Data(xs=xs, ys=ys)\n\n raise ValueError('Wrong name for a dataset')", "def get_target_market():\n return choice(GraphParameters.TARGET_MARKETS)", "def _get__target(self):\n if self._saved_target is None:\n self._saved_target = self._metadata.get_callable()()\n \n return self._saved_target" ]
[ "0.7770308", "0.7121358", "0.64517236", "0.6015566", "0.54749423", "0.5369748", "0.5230045", "0.52188426", "0.5049947", "0.49974164", "0.49770722", "0.49149367", "0.48960745", "0.48660335", "0.4858429", "0.47984567", "0.47667906", "0.47544086", "0.47335964", "0.47325793", "0.46791846", "0.4668791", "0.46662125", "0.46469977", "0.46438655", "0.4628731", "0.46123934", "0.460944", "0.4598078", "0.45970562" ]
0.84920114
0
Create `holdout_set` (if necessary) by loading a DataFrame from a .csv file, or by separating `train_set`, and return the updated (`train_set`, `holdout_set`) pair
def define_holdout_set( train_set: pd.DataFrame, holdout_set: Union[pd.DataFrame, callable, str, None], target_column: Union[str, List[str]], ) -> Tuple[pd.DataFrame, Optional[pd.DataFrame]]: #################### Update `holdout_set` #################### if callable(holdout_set): train_set, holdout_set = holdout_set(train_set, target_column) elif isinstance(holdout_set, str): holdout_set = pd.read_csv(holdout_set) #################### Validate `holdout_set` #################### try: if holdout_set is None or np.array_equal(train_set.columns, holdout_set.columns): return train_set, holdout_set except AttributeError: raise TypeError(f"holdout_set must be None, DataFrame, callable, or str, not {holdout_set}") raise ValueError(f"Mismatched columns\n{train_set.columns}\n!=\n{holdout_set.columns}")
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_train_set(addition_to_filename = ''):\n df = f.create_df()\n print('original df size is ', df.shape)\n print('original df columns ', df.columns)\n\n df = f.calculate_missing_prices_for_train_set(df)\n print('df size after averaging price ', df.shape)\n df = f.downcast_dtypes(df)\n df = f.add_lag(all_data_df = df, df_to_add_lag= df, number_of_months=3)\n df = f.add_days_stat(df)\n\n print(df.columns)\n\n # df = h.add_holidays(df)\n # print('df size with holidays ', df.shape)\n\n timestr = time.strftime(\"%Y%m%d-%H%M%S\")\n pickle.dump(df, open(f\"{timestr}_{addition_to_filename}_train.pickle.dat\", \"wb\"))\n\n # save feature names for further use\n features_list = create_feature_names_list(df)\n pickle.dump(features_list, open(f\"{timestr}_{addition_to_filename}_features.pickle.dat\", \"wb\"))", "def read_random_data_from_csv(\n file_name, training_set_size, unlabeled_set_size, holdout_set_size, validation_set_size):\n data = samp_file_to_arr(\n file_name, training_set_size + unlabeled_set_size + holdout_set_size + validation_set_size)\n y_raw = np.array([x[0] for x in data])\n x_all = np.array([x[1:] for x in data])\n # Now transform so that the lower label is -1, always. \n uq = np.unique(y_raw) # Assumed to be only two unique labels!\n y_all = np.zeros(len(y_raw))\n y_all[np.where(y_raw == uq[0])[0]] = -1\n y_all[np.where(y_raw == uq[1])[0]] = 1\n xtrhoval, x_unl, ytrhoval, y_unl = sklearn.model_selection.train_test_split(\n x_all, y_all, test_size=unlabeled_set_size)\n x_trho, x_validate, y_trte, y_validate = sklearn.model_selection.train_test_split(\n xtrhoval, ytrhoval, test_size=validation_set_size)\n x_train, x_out, y_train, y_out = sklearn.model_selection.train_test_split(\n x_trho, y_trte, test_size=holdout_set_size)\n return (x_train, y_train, x_unl, y_unl, x_out, y_out, x_validate, y_validate)", "def get_train_data():\n # train set\n train = pd.read_csv(\"train.csv\")\n\n return train", "def load_data():\n train = pd.read_csv(\"../input/train.csv\", dtype={\"Age\": np.float64}, )\n test = pd.read_csv(\"../input/test.csv\", dtype={\"Age\": np.float64}, )\n\n train = train.set_index('PassengerId')\n test = test.set_index('PassengerId')\n\n train = train.apply(preprocess, axis=1)\n test = test.apply(preprocess, axis=1)\n\n x_train = train.drop(['Survived'], axis=1)\n y_train = train['Survived']\n x_test = test\n return {'train': {'x': x_train, 'y': y_train},\n 'test': {'x': x_test},\n 'full_features': pd.concat([x_train, x_test])}", "def learning_set_builders(self):\n # Transform in a dataframe:\n original_train_x = pd.DataFrame(self.original_train_x, columns=self.original_x_header)\n original_train_y = pd.DataFrame(self.original_train_y, columns=self.original_y_header)\n original_test_x = pd.DataFrame(self.original_test_x, columns=self.original_x_header)\n original_test_y = pd.DataFrame(self.original_test_y, columns=self.original_y_header)\n original_validation_x = pd.DataFrame(self.original_validation_x, columns=self.original_x_header)\n original_validation_y = pd.DataFrame(self.original_validation_y, columns=self.original_y_header)\n\n # Training set\n x, y = self.convertor(original_train_x, original_train_y)\n self.pairs_train_x = x.to_numpy()\n self.pairs_train_y = y.to_numpy()\n\n # Headers\n self.pairs_x_header = x.columns\n self.pairs_y_header = y.columns\n\n # Testing set\n x, y = self.convertor(original_test_x, original_test_y)\n self.pairs_test_x = x.to_numpy()\n self.pairs_test_y = y.to_numpy()\n\n # Validation set\n x, y = self.convertor(original_validation_x, original_validation_y)\n self.pairs_validation_x = x.to_numpy()\n self.pairs_validation_y = y.to_numpy()", "def train_set(self):\n if self._trainset is None: # loads the data to memory once and when requested.\n trainset_raw = self.read_dataset(self._trainset_path)\n trainset_spacy = self.read_spacy_pickle(self._trainset_spacy_path)\n if trainset_raw is None and trainset_spacy is None:\n # This is for languages we never see (French)\n self._trainset = None\n else:\n self._trainset = pd.concat([trainset_raw, trainset_spacy], axis=1)\n\n self._trainset['language'] = self._language\n self._trainset['dataset_name'] = self._dataset_name\n\n return self._trainset", "def createTrainTestSets():\n tweets = open(noDuplicatesFilename, 'r').read().splitlines()\n name_mapping = loadNameMapping()\n holdoutLocations = [u'Frederiksberg, Danmark', u'T\\xe5rnby, Danmark', u'Kolding, Danmark', u'T\\xe4by, Sverige', u'Kungsbacka, Sverige', u'Kristianstad, Sverige', u'Bod\\xf8, Norge', u'Kvinnherad, Norge', u'Ullensaker, Norge']\n testSetLocation = []\n rest = []\n for tweet in tweets:\n if stringToTweet(tweet).getFullName() in holdoutLocations:\n testSetLocation.append(tweet)\n else:\n rest.append(tweet)\n tweets = rest\n testIndex = int(round(len(tweets) * (1 - test_set_ratio)))\n random.seed(1)\n random.shuffle(tweets)\n trainSet = tweets[:testIndex]\n testSet = tweets[testIndex:]\n open(trainSetFilename, 'w').write('\\n'.join(trainSet))\n open(testSetNormalFilename, 'w').write('\\n'.join(testSet))\n open(testSetLocationFilename, 'w').write('\\n'.join(testSetLocation))\n print \"Wrote %d tweets to train set\" % len(trainSet)\n print \"Wrote %d tweets to normal test set\" % len(testSet)\n print \"Wrote %d tweets to location test set\" % len(testSetLocation)", "def load_testset(filename):\n pickle_name = filename + \".pickle\"\n try:\n print(\"trying to load \" + filename + \" from pickle\")\n dataset = pickle.load(open(pickle_name, \"rb\"))\n except:\n with open(filename, 'r') as csv_file:\n print(\"no pickle exists. parsing file \" + filename)\n dataset = [DataPoint(item[0:], \" \")\n for item\n in csv.reader(csv_file, delimiter=',')]\n pickle.dump(dataset, open(pickle_name, \"wb\"))\n print(\"loaded \" + filename)\n return dataset", "def load_data(self):\n with open('data/fordTrain.csv') as f:\n data = csv.reader(f, delimiter=',')\n train = [x for i, x in enumerate(data) if i > 0] \n # Extract features and target variable separately\n trainx = [x[3:] for x in train]\n trainy = [x[2] for x in train]\n\n with open('data/fordTest.csv') as f:\n data = csv.reader(f, delimiter=',')\n testx = [x[3:] for i, x in enumerate(data) if i > 0] \n\n with open('data/Solution.csv') as f:\n data = csv.reader(f, delimiter=',')\n testy = [x[2] for i, x in enumerate(data) if i > 0] \n\n # Extract features and target variable, convert to numpy array\n trainx = np.asarray(trainx, dtype=np.float32)\n trainy = np.asarray(trainy, dtype=np.int8)\n testx = np.asarray(testx, dtype=np.float32)\n testy = np.asarray(testy, dtype=np.int8)\n\n # Return training and test sets\n trainSet = Dataset(trainx, trainy)\n testSet = Dataset(testx, testy)\n return trainSet, testSet", "def load_data():\n data = gzip.open(\"mnist.pkl.gz\", \"rb\")\n train_set, valid_set, test_set = cPickle.load(data)\n data.close()\n\n # Combine validation and train folds to recreate the master 60k set.\n new_images = numpy.concatenate((train_set[0], valid_set[0]))\n new_labels = numpy.concatenate((train_set[1], valid_set[1]))\n\n train_set = (new_images, new_labels)\n \n return (train_set, test_set)", "def load_train_test() -> Tuple[pd.DataFrame, pd.DataFrame]:\n df = load_data() \n #Sample\n train_df = df.sample(frac=0.8, random_state=42)\n test_df = df[~df.index.isin(train_df.index)]\n #Rescale\n train_mean = train_df.iloc[:, :-1].mean()\n train_std = train_df.iloc[:, :-1].std()\n train_df.iloc[:, :-1] = (train_df.iloc[:, :-1] - train_mean) / train_std\n test_df.iloc[:, :-1] = (test_df.iloc[:, :-1] - train_mean) / train_std\n return train_df, test_df", "def get_dataset(filename, target_name, training_fraction):\n csv_reader = csv.DictReader(open(filename))\n attributes_names = [attribute_name for attribute_name in csv_reader.fieldnames\n if attribute_name != target_name]\n train_dataset = Dataset(attributes_names, target_name)\n test_dataset = Dataset(attributes_names, target_name)\n instances = []\n for row in csv_reader:\n instances.append(([row[attribute_name] for attribute_name in attributes_names],\n row[target_name]))\n all_instances_indexes = set(range(0, len(instances)))\n number_of_training_instances = int(len(all_instances_indexes) * training_fraction)\n training_instances_indexes = set(sample(all_instances_indexes, number_of_training_instances))\n test_instances_indexes = all_instances_indexes.difference(training_instances_indexes)\n for instance_index in training_instances_indexes:\n train_dataset.add_instance(instances[instance_index][0], instances[instance_index][1])\n for instance_index in test_instances_indexes:\n test_dataset.add_instance(instances[instance_index][0], instances[instance_index][1])\n return train_dataset, test_dataset", "def create_sets(test, data, test_size=0.2, write=False):\n y_test = test['y_old']\n X_test = test.drop('y_old', 1)\n y_data = data['y_old']\n X_data = data.drop('y_old', 1)\n X_train, X_val, y_train, y_val = train_test_split(X_data, y_data, test_size=test_size, random_state=123)\n if write:\n pickle.dump((X_train, X_val, y_train, y_val), open(obj_save_path+'train_val_df.p', 'wb'))\n #X_train, X_val, y_train, y_val = pickle.load(open(obj_save_path+'train_val_df.p', 'rb'))\n return X_train, y_train, X_val, y_val, X_test, y_test", "def initSets(self):\n data_frame = pd.read_csv(self.train_file, header=None)\n data_frame = data_frame.drop(columns=self.drop_cols)\n features = data_frame.iloc[:, :-1].values\n labels = data_frame.iloc[:, -1].values\n if self.test_file is None:\n self.train_feat, self.test_feat, self.train_labl, self.test_labl = train_test_split(features, labels, test_size=self.test_size)\n else:\n data_frame = pd.read_csv(self.test_file, header=None)\n data_frame = data_frame.drop(columns=self.drop_cols)\n self.train_feat, _, self.train_labl, _ = train_test_split(features, labels, test_size=self.test_size)\n features = data_frame.iloc[:, :-1].values\n labels = data_frame.iloc[:, -1].values\n _, self.test_feat, _, self.test_labl = train_test_split(features, labels, test_size=self.test_size)\n # kfold = KFold(n_splits=3)\n # self.train_index, self.test_index = kfold.split(features,labels)", "def time_split_dataset(df, train_start_date, train_end_date, holdout_end_date, date_col):\n\n train_set = df.copy()[\n (df[date_col] >= train_start_date) & (df[date_col] <= train_end_date)]\n\n test_set = df.copy()[\n (df[date_col] > train_end_date) & (df[date_col] <= holdout_end_date)]\n\n return train_set, test_set", "def load_dataset_train():\n df_train = load_csv_file(\"31_train.csv\")\n df_train_target = load_csv_file(\"31_target_train.csv\")\n\n return df_train.values, df_train_target.values", "def prepare_data(args):\n logger.info('Loading dataframe from %s' % args.newspath)\n df = pd.read_csv(args.newspath, encoding='gb18030')\n logger.info('Dataframe size: %d observations %d features after loaded' % (df.shape[0], df.shape[1]))\n\n # exclude rows with column source == NaN\n logger.info('Data cleansing...')\n df = df[~pd.isna(df['source'])]\n logger.info('Dataframe size: %d observations %d features after data cleansing' % (df.shape[0], df.shape[1]))\n\n # split the dataframe into training set and test set\n logger.info('Making training set & test set...')\n train_set, test_set = split_data(df)\n logger.info('Traning set size: %d' % train_set.shape[0])\n logger.info('Test set size: %d' % test_set.shape[0])\n\n # save the train set and test set to picke files\n logger.info('Save dataframes to files...')\n train_set.to_pickle(args.trainpath)\n test_set.to_pickle(args.testpath)", "def create_train_valid_set(self):\n\n if not self.eq_train:\n X_train_high_level, X_valid_high_level, X_train_low_level, X_valid_low_level, train_w, valid_w, y_train, y_valid = train_test_split(self.X_train_high_level, self.X_train_low_level, self.train_weights, self.y_train,\n train_size=0.7, test_size=0.3\n )\n else:\n X_train_high_level, X_valid_high_level, X_train_low_level, X_valid_low_level, train_w, valid_w, w_train_eq, w_valid_eq, y_train, y_valid = train_test_split(self.X_train_high_level, self.X_train_low_level,\n self.train_weights, self.train_weights_eq, self.y_train,\n train_size=0.7, test_size=0.3\n )\n self.train_weights_eq = w_train_eq\n\n #NOTE: might need to re-equalise weights in each folds as sumW_sig != sumW_bkg anymroe!\n self.train_weights = train_w\n self.valid_weights = valid_w #validation weights should never be equalised weights!\n\n print 'creating validation dataset'\n self.X_train_high_level = X_train_high_level\n self.X_train_low_level = self.join_objects(X_train_low_level)\n\n self.X_valid_high_level = X_valid_high_level\n self.X_valid_low_level = self.join_objects(X_valid_low_level)\n print 'finished creating validation dataset'\n\n self.y_train = y_train\n self.y_valid = y_valid", "def holdout(self, X, y):\n #X,y = self.define_dataset(self.df, self.col_list, self.target_var)\n X_train, X_test, y_train, y_test = train_test_split(\n X, y, test_size = 0.3, random_state = 0)\n\n return X_train, X_test, y_train, y_test", "def load_susy(trainsize=500, testsize=1000):\n filename = 'datasets/susysubset.csv'\n dataset = loadcsv(filename)\n trainset, testset = splitdataset(dataset, trainsize, testsize)\n return trainset, testset", "def read(train_path, test_path, label_name):\n train_dataset = pd.read_csv(train_path)\n test_dataset = pd.read_csv(test_path)\n\n train_labels = train_dataset.pop(label_name)\n\n imputer = DataFrameImputer().fit(train_dataset)\n train_dataset = imputer.transform(train_dataset)\n test_dataset = imputer.transform(test_dataset)\n\n train_dataset = pd.get_dummies(train_dataset)\n test_dataset = pd.get_dummies(test_dataset)\n\n train_dataset = train_dataset.drop(train_dataset.columns.difference(test_dataset.columns), axis=1)\n test_dataset = test_dataset.drop(test_dataset.columns.difference(train_dataset.columns), axis=1)\n\n scaler = StandardScaler().fit(train_dataset)\n train_dataset = scaler.transform(train_dataset)\n test_dataset = scaler.transform(test_dataset)\n\n return train_dataset, train_labels, test_dataset", "def create_sets():\n global train_x, train_y, val_x, val_y\n\n print('Creating sets')\n\n dataframe = pd.read_csv('LoggerBot.log', names=NAMES).sample(frac=1)\n inputs = dataframe.values[:,:-1].astype(np.float32)\n outputs = dataframe.values[:,-1].astype(np.int32)\n\n train_set_size = int(len(dataframe) * 0.7)\n train_x, train_y = inputs[:train_set_size], outputs[:train_set_size]\n val_x, val_y = inputs[train_set_size:], outputs[train_set_size:]", "def import_data(seed: object = 42) -> object:\n\n # Read input data\n df = pd.read_csv(\"x_train_gr_smpl.csv\").astype(int)\n\n # label data-frame rows based on sample data\n for x in range(10):\n index = ~pd.read_csv(\"y_train_smpl_%s.csv\" % x, squeeze=True).astype(bool) # reversed flags (~)\n df.loc[index, 'label'] = str(x)\n\n input_data_ordered = df.iloc[:, 0:2304].to_numpy()\n output_data_ordered = df.iloc[:, 2304].to_numpy()\n\n # Randomise instance order (forcing the same result each time)\n np.random.seed(seed)\n permutation = np.random.permutation(df.shape[0])\n\n # Create base input and output arrays\n input_data = input_data_ordered[permutation]\n output_data = output_data_ordered[permutation]\n\n return input_data, output_data, df, input_data_ordered, output_data_ordered", "def GetDataSet(filename: str):\n # Read csv files\n # Assign our data set name attribute to simplify our code\n df_train_set = pd.read_csv(filename + \"/train.csv\")\n df_train_set.name = \"train\"\n df_test_set = pd.read_csv(filename + \"/test.csv\")\n df_test_set.name = \"test\"\n # Combine train set and test set\n df_all = ConcatDF(df_train_set, df_test_set)\n df_all.name = \"all\"\n # Remember to return the train set and test set for the next step\n return df_train_set, df_test_set, df_all", "def create_test_and_train_sets(input_file):\n \n input_df = pd.read_csv(input_file, sep=',', header=0)\n df_items = pd.DataFrame({'productId': input_df.productId.unique()})\n df_sorted_items = df_items.sort_values('productId').reset_index()\n pds_items = df_sorted_items.productId\n \n df_user_items = input_df.groupby(['userId', 'productId']).agg({'Expense': 'sum'})\n \n # create a list of (userId, productId, Expense) ratings, where userId and productId are 0-indexed\n current_u = -1\n ux = -1\n pv_ratings = []\n user_ux = []\n \n for timeonpg in df_user_items.itertuples():\n user = timeonpg[0][0]\n item = timeonpg[0][1]\n if user != current_u:\n user_ux.append(user)\n ux += 1\n current_u = user\n ix = pds_items.searchsorted(item)[0]\n pv_ratings.append((ux, ix, timeonpg[1]))\n\n # convert ratings list and user map to np array\n pv_ratings = np.asarray(pv_ratings)\n user_ux = np.asarray(user_ux)\n \n # create train and test coos matrixes\n tr_sparse, test_sparse = _create_sparse_train_and_test(pv_ratings, ux + 1, df_items.size)\n \n return user_ux, pds_items.as_matrix(), tr_sparse, test_sparse", "def open_MRI_data(csv_path, train_set = 0.8, n_followups=5, normalize=True):\n\n data_df = pd.read_csv(csv_path)\n\n mri_col = data_df.columns.str.contains(\"SV_UCSFFSX_11_02_15_UCSFFSX51_08_01_16\")\n mri_col = data_df.columns[mri_col].values\n\n data_df = data_df.dropna(axis=0, subset=mri_col)\n\n # Select only the subjects with nfollowups\n # Code to only select 5 first appearances of each PTID\n ptid_list = np.unique(data_df[\"PTID\"])\n\n idx_to_drop = []\n for ptid in ptid_list:\n i_list = data_df.index[data_df['PTID'] == ptid].tolist()\n if len(i_list) < 5:\n idx_to_drop = idx_to_drop + i_list\n elif len(i_list) > 5:\n idx_to_drop = idx_to_drop + i_list[5:]\n\n data_final = data_df.drop(idx_to_drop)\n\n print(data_final.shape)\n\n # Normalize only features\n data_final.loc[:,mri_col] = data_final.loc[:,mri_col].apply(lambda x: (x-x.mean())/ x.std(), axis=0)\n\n # Divide between test and train\n from sklearn.model_selection import GroupShuffleSplit\n gss = GroupShuffleSplit(n_splits=1, test_size=1.0-train_set)\n train_dataset, test_dataset = next(gss.split(X=data_final, y=data_final.DX_bl.values, groups=data_final.PTID.values))\n\n df_train = data_final.iloc[train_dataset]\n df_test = data_final.iloc[test_dataset]\n\n df_train = df_train.reset_index(drop=True)\n df_test = df_test.reset_index(drop=True)\n\n # Return the features in the correct shape (Nsamples, timesteps, nfeatures)\n X_train = pandas_to_data_timeseries(df_train, mri_col)\n X_test = pandas_to_data_timeseries(df_test, mri_col)\n\n return X_train, X_test", "def dataset()-> pd.DataFrame:\n old= pd.read_csv(DIRECT_OLD)\n new= pd.read_csv(DIRECT_NEW)\n return pd.concat([old, new])", "def prep_data(df):\n y = df.target\n X = df.drop([\"target\"], axis=1)\n X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2)\n return X_train, X_test, y_train, y_test", "def data_prep(data, y, dropna=False):\n\n rand_state = 10 # Setting random state for later cv\n df = pd.read_pickle(data) # Reading in data\n if dropna is True:\n df.dropna(axis=0, inplace = True)\n else:\n pass\n X = df.drop(y, axis=1) # Assigning the feature space to X\n y = df[y] # Class labels to predict\n\n return X, y, rand_state", "def get_train_csv(self):\n try:\n self.train_article = pd.read_csv(constants.DATA_DIR / 'knn_article_tags.csv')\n except FileNotFoundError:\n train = pd.Series([])\n for csv_file in os.listdir(constants.CLEAN_DIR):\n if csv_file in self.article_feat_csvs:\n df = pd.read_csv(constants.CLEAN_DIR / csv_file)\n feat = csv_file[8:-4]\n g = df.dropna(axis=0).groupby(\"id\")[self.tag_ref[feat]]\n if train.empty:\n train = g.apply(lambda x: list(x.astype(str).str.lower()))\n else:\n g = g.apply(lambda x: list(x.astype(str).str.lower()))\n train = train.combine(g, lambda x1, x2: list(set(x1+x2)), fill_value=[])\n\n train = pd.DataFrame({'id':train.index, 'tags':train.values})\n train.to_csv(constants.DATA_DIR / 'knn_article_tags.csv', header=True)\n self.train_article = train\n\n try:\n self.train_image = pd.read_csv(constants.DATA_DIR / 'knn_image_tags.csv')\n except FileNotFoundError:\n train = pd.Series([])\n for csv_file in os.listdir(constants.CLEAN_DIR):\n if csv_file in self.image_feat_csvs:\n df = pd.read_csv(constants.CLEAN_DIR / csv_file)\n feat = csv_file[6:-4]\n g = df.dropna(axis=0).groupby(\"id\")[self.tag_ref[feat]]\n if train.empty:\n train = g.apply(lambda x: list(x.astype(str).str.lower()))\n else:\n g = g.apply(lambda x: list(x.astype(str).str.lower()))\n train = train.combine(g, lambda x1, x2: list(set(x1+x2)), fill_value=[])\n\n train = pd.DataFrame({'id':train.index, 'tags':train.values})\n train.to_csv(constants.DATA_DIR / 'knn_image_tags.csv', header=True)\n self.train_image = train" ]
[ "0.6308423", "0.6136544", "0.60832155", "0.60303485", "0.59481156", "0.5926183", "0.58851606", "0.5858881", "0.5792204", "0.57012284", "0.56984967", "0.569597", "0.56708777", "0.5638815", "0.56349504", "0.56267476", "0.56263214", "0.5608293", "0.55986565", "0.55914485", "0.55911386", "0.5584236", "0.55594903", "0.5529645", "0.55275583", "0.54748285", "0.5469353", "0.5464139", "0.54296494", "0.54276204" ]
0.6827967
0
Validate contents of blacklist. For most values, the corresponding file is saved upon completion of the experiment. See the "Notes" section below for details on some special cases
def validate_file_blacklist(blacklist): valid_values = [ # 'checkpoint', "description", "heartbeat", "predictions_holdout", "predictions_in_fold", "predictions_oof", "predictions_test", "script_backup", "tested_keys", "current_heartbeat", ] if blacklist == "ALL": G.warn('WARNING: Received `blacklist`="ALL". Nothing will be saved') return blacklist if not blacklist: return [] elif not isinstance(blacklist, list): raise TypeError("Expected blacklist to be a list, not: {}".format(blacklist)) elif not all([isinstance(_, str) for _ in blacklist]): invalid_files = [(type(_).__name__, _) for _ in blacklist if not isinstance(_, str)] raise TypeError("Expected blacklist contents to be strings, not: {}".format(invalid_files)) for a_file in blacklist: if a_file not in valid_values: raise ValueError(f"Invalid blacklist value: {a_file}.\nExpected one of: {valid_values}") if a_file in ["description", "tested_keys"]: G.warn(f"Including {a_file!r} in blacklist will severely impede library functionality") # Blacklist experiment-specific heartbeat if general (current) heartbeat is blacklisted if ("current_heartbeat" in blacklist) and ("heartbeat" not in blacklist): blacklist.append("heartbeat") return blacklist
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def load_blacklist(experiment):\n blacklist = np.loadtxt('../Slip_Property_Data/%s_blacklist.txt'%experiment)\n return blacklist", "def load_blacklist(experiment):\n blacklist = np.loadtxt('../Slip_Property_Data/%s_blacklist.txt'%experiment)\n return blacklist", "def save_blacklist(blacklist, output_dir):\n output_filepath = os.path.join(output_dir, 'blacklist.txt')\n print('[INFO] Writing blacklist to {}...'.format(output_filepath))\n with open(output_filepath, 'w') as f:\n for ent in blacklist:\n f.write('{}\\t{}\\n'.format(ent[0], ent[1]))", "def read_blacklist(self, _blacklist_fname):\n try:\n blacklist_f = codecs.open(_blacklist_fname, mode='r', encoding='utf-8')\n line_idx = 0\n for fline in blacklist_f:\n line_idx = line_idx + 1\n line = fline.strip()\n if ((len(line) > 0) and (line[0] != '#')):\n # non null and not started # line ... add to the set\n if (line in self.__black_list_set):\n print u'duplication found [' + line + u'] at ' + str(line_idx) + \\\n u' ignored'\n else:\n self.__black_list_set.add(line)\n\n print u'read blacklist_file [' + _blacklist_fname + \\\n u'], number of entries: ' + str(len(self.__black_list_set))\n except IOError as e:\n print \"I/O error({0}): {1}\".format(e.errno, e.strerror)\n print \"Can not open a blacklist file {0}\".format(_blacklist_fname)\n print \"Please create blacklist file (an empty file is also fine.)\"\n sys.exit(1)\n except:\n print \"Unexpected error:\", sys.exc_info()[0]\n raise", "def getBlackList(filename):\n #filename = \"filelist/blacklist_%s.txt\"%dataset.lstrip('/').replace('/','__')\n blacklist = [ ]\n if os.path.exists(filename):\n with open(filename,'r') as file:\n for line in file:\n line = line.rstrip('\\n')\n if line and '#' not in line:\n blacklist.append(line)\n return blacklist", "def open_blacklist(filepath):\n with open(filepath, 'r') as f:\n blacklist = [tuple(line.strip().split('\\t')) for line in f.readlines()]\n return blacklist", "def blacklist_file(self, fkey):\n self.blacklist.update([fkey])", "def test_blacklist(self):\n ts = self.ts_db\n fc1 = self.field_change1\n fc2 = self.field_change2\n fc3 = self.field_change3\n fc4 = self.field_change4\n\n valid = blacklist.filter_by_benchmark_name(ts, fc1)\n self.assertTrue(valid, \"Expect this to not be filtered.\")\n valid = blacklist.filter_by_benchmark_name(ts, fc2)\n self.assertTrue(valid, \"Expect this to not be filtered.\")\n bad = blacklist.filter_by_benchmark_name(ts, fc3)\n self.assertFalse(bad, \"Expect this to be filtered by regex.\")\n bad = blacklist.filter_by_benchmark_name(ts, fc4)\n self.assertFalse(bad, \"Expect this to be filtered by blacklist.\")", "def write_valids():\n har_path = os.path.join(PLT_SRC, 'data/replay/*')\n valid_path = '../data/filtered_stats/valids.txt'\n\n har_files = [f for f in glob(har_path)] # Include pc files?\n urls = \\\n [urlsafe_b64decode(f.split('/')[-1].split('.')[0]) for f in har_files]\n with open(valid_path, 'w') as f:\n for url, url_har_path in zip(urls, har_files):\n f.write('{0} {1}\\n'.format(url, url_har_path))", "def _parse_blacklist(path):\n if path is None:\n return []\n with open(path, 'rt') as f:\n return [line.strip() for line in f]", "def ReadBlackListFile(BlackListFile):\n blacklist = []\n if os.path.isfile(BlackListFile):\n with open(BlackListFile, 'r') as filecontent:\n for line in filecontent:\n #(chrom1, start1, chrom2, start2) = line.rstrip().split(\"\\t\")\n blacklist.append(line)\n return(blacklist)", "def apply_tempest_blacklist(self, black_list):\n LOGGER.debug(\"Applying tempest blacklist...\")\n if os.path.exists(self.raw_list):\n os.remove(self.raw_list)\n os.rename(self.list, self.raw_list)\n cases_file = self.read_file(self.raw_list)\n with open(self.list, 'w', encoding='utf-8') as result_file:\n black_tests = []\n try:\n deploy_scenario = env.get('DEPLOY_SCENARIO')\n if bool(deploy_scenario):\n # if DEPLOY_SCENARIO is set we read the file\n with open(black_list, encoding='utf-8') as black_list_file:\n black_list_yaml = yaml.safe_load(black_list_file)\n black_list_file.close()\n for item in black_list_yaml:\n scenarios = item['scenarios']\n in_it = rally.RallyBase.in_iterable_re\n if in_it(deploy_scenario, scenarios):\n tests = item['tests']\n black_tests.extend(tests)\n except Exception: # pylint: disable=broad-except\n black_tests = []\n LOGGER.debug(\"Tempest blacklist file does not exist.\")\n\n for cases_line in cases_file:\n for black_tests_line in black_tests:\n if re.search(black_tests_line, cases_line):\n break\n else:\n result_file.write(str(cases_line) + '\\n')", "def remove_blacklisted(blacklist, ssc, output_dir):\n print('[INFO] Writing blacklisted corpus to {}...'.format(output_dir))\n # assuming there is only 1 SSC, so take index 0\n ssc_filepaths = list(get_filepaths(ssc))[0]\n # for faster lookup\n blacklist = set(blacklist)\n for filepath in ssc_filepaths:\n with open(filepath, 'r') as f:\n # remove blacklisted entities\n lines = f.readlines()\n for i in range(1, len(lines) - 1):\n previous_tag = 'O' if lines[i-1] == '\\n' else lines[i-1].strip().split('\\t')[1]\n next_tag = 'O' if lines[i+1] == '\\n' else lines[i+1].strip().split('\\t')[1]\n single_token_entity = (previous_tag != 'I-' and next_tag != 'I-')\n blacklisted = tuple(lines[i].strip().split('\\t')) in blacklist\n if single_token_entity and blacklisted:\n lines[i] = '{}\\tO\\n'.format(lines[i].split('\\t')[0])\n # write blacklisted copy to disk\n corpus_name = os.path.basename(ssc) + '_blacklisted'\n output_directory = os.path.join(output_dir, corpus_name)\n make_dir(output_directory)\n output_filepath = os.path.join(output_directory, os.path.basename(filepath))\n with open(output_filepath, 'w') as f:\n for line in lines:\n f.write(line)", "def blacklist(self) -> List[str]:\n return self.raw_config.get(\"blacklist\", [])", "def blacklist_add():\n db = unitdata.kv()\n blacklist = db.get(BLACKLIST_KEY, [])\n for device in get_devices():\n if not os.path.exists(device):\n raise Error('{}: No such file or directory.'.format(device))\n if device not in blacklist:\n blacklist.append(device)\n db.set(BLACKLIST_KEY, blacklist)\n db.flush()", "def load_blocked_groups(self):\n print(\" ->[*] Loading group blacklist...\")\n blacklist = set()\n if os.access(\"blocked_groups\", os.F_OK):\n with codecs.open(\"blocked_groups\", \"r\", encoding=\"utf-8\") as groups:\n blocked_groups = groups.readlines()\n for group in blocked_groups:\n blacklist.add(group)\n return blacklist", "def is_blacklisted(fname):\n return is_dot(fname) or is_excluded_filetype(fname)", "def is_blacklisted(self):\r\n \r\n in_blacklist = False \r\n if self.chrompos in parser.blacklist:\r\n in_blacklist = True\r\n \r\n return in_blacklist", "def check_utxo_blacklist(commitment, persist=False):\n #TODO format error checking?\n fname = \"blacklist\"\n if jm_single().config.get(\"BLOCKCHAIN\", \"blockchain_source\") == 'regtest':\n fname += \"_\" + jm_single().nickname\n with jm_single().blacklist_file_lock:\n if os.path.isfile(fname):\n with open(fname, \"rb\") as f:\n blacklisted_commitments = [x.strip() for x in f.readlines()]\n else:\n blacklisted_commitments = []\n if commitment in blacklisted_commitments:\n return False\n elif persist:\n blacklisted_commitments += [commitment]\n with open(fname, \"wb\") as f:\n f.write('\\n'.join(blacklisted_commitments))\n f.flush()\n #If the commitment is new and we are *not* persisting, nothing to do\n #(we only add it to the list on sending io_auth, which represents actual\n #usage).\n return True", "def Validate(self, relative_file, contents):\n pass", "def handler(state, _):\n if state[0] == 'u':\n fname = player.playlist[player.playlist_pos]['filename']\n fkey = get_file_key(fname)\n col.blacklist_file(fkey)\n player.playlist_remove()\n os.remove(fname)\n print('Blacklisted: {}'.format(fname))", "def test_validation_negative(self):\n self.assertFalse(self.cut._validate_iban(\"FR14 2004 1010 0505 0001 3\"))\n self.assertFalse(self.cut._validate_iban(\"XX00 1234 5678 9012 3456 7890 1234 5678 90\"))\n self.assertFalse(self.cut._validate_iban(\"YY00123456789012345678901234567890\"))\n self.assertFalse(self.cut._validate_iban(\"XX22YYY1234567890123\"))\n self.assertFalse(self.cut._validate_iban(\"[email protected]\"))", "def validate(self):\n with open(os.path.join(settings.MEDIA_ROOT, self.file.name)) as file:\n lines = file.readlines()\n validators = ['os.', 'from os', 'io.', 'from io', 'open(', 'system(']\n for line in lines:\n for validator in validators:\n if validator in line:\n return False\n return True", "def test_remove_from_blacklist(self):\n\n self.feature_test.add_to_blacklist(3)\n self.feature_test.remove_from_blacklist(3)\n self.assertFalse(3 in Feature(\"testing\").blacklist)", "def CheckIfItIsBlacklisted(chr1, start1, chr2, start2, blacklist, extention):\n blacklistTag = None\n chr1 = str(chr1)\n chr2 = str(chr2)\n start1 = int(start1)\n start2 = int(start2)\n if chr1 == chr2:\n start = start1\n end = start2\n for entries in blacklist:\n (bchr1, bstart1, bchr2, bstart2) = entries.rstrip().split(\"\\t\")\n bstart1 = int(bstart1)\n bstart2 = int(bstart2)\n if(bchr1 == bchr2):\n if(chr1 == bchr1):\n if(((start <= int(bstart1 - extention)) and (start >= int(bstart1 + extention)))\n and ((end <= int(bstart2 - extention)) and (end <= int(bstart2 + extention)))):\n blacklistTag = True\n else:\n blacklistTag = False\n if(((end <= int(bstart1 - extention)) and (end >= int(bstart1 + extention)))\n and ((start <= int(bstart2 - extention)) and (start <= int(bstart2 + extention)))):\n blacklistTag = True\n else:\n blacklistTag = False\n else:\n continue\n else:\n continue\n\n else:\n for entries in blacklist:\n (bchr1, bstart1, bchr2, bstart2) = entries.rstrip().split(\"\\t\")\n bstart1 = int(bstart1)\n bstart2 = int(bstart2)\n if(((chr1 == bchr1) and (chr2 == bchr2)) or ((chr1 == bchr2) and (chr2 == bchr1))):\n if(((start1 <= int(bstart1 - extention)) and (start1 >= int(bstart1 + extention)))\n and ((start2 <= int(bstart2 - extention)) and (start2 <= int(bstart2 + extention)))):\n blacklistTag = True\n else:\n blacklistTag = False\n if(((start2 <= int(bstart1 - extention)) and (start2 >= int(bstart1 + extention)))\n and ((start1 <= int(bstart2 - extention)) and (start1 <= int(bstart2 + extention)))):\n blacklistTag = True\n else:\n blacklistTag = False\n else:\n continue\n\n return(blacklistTag)", "def test_copy_required_exclude_list(self):\n exclude_list = ['*.yar', 'skip/these/file*']\n self.assertTrue(clone_rules._copy_required('base.yara', [], exclude_list))\n self.assertTrue(clone_rules._copy_required('path/to/file.yara', [], exclude_list))\n self.assertFalse(clone_rules._copy_required('file.yar', [], exclude_list))\n self.assertFalse(clone_rules._copy_required('skip/these/file.yara', [], exclude_list))", "def write_data(self, blacklist=('normE', 'normEsquared', 'genRate')):\n\n start = time.time()\n self.data.write_data(blacklist=blacklist)\n end = time.time()\n self.log.info('Write time: %.2f seconds', end - start)", "def saveFileListLocal(dataset,filelist,blacklist=[ ],tag=\"\"):\n if '/pnfs/' in dataset:\n tag += \"_pnfs\"\n dataset = '__'.join(dataset.split('/')[-3:])\n filename = \"filelist/filelist_%s%s.txt\"%(dataset.replace('/','__'),tag)\n with open(filename,'w+') as file:\n for line in filelist:\n if line not in blacklist:\n file.write(line+'\\n')\n return filename", "def test_does_not_validate_invalid_files(self):\n bad_files = (\n 'newstest2019-defr-src-ts.de.sgm',\n 'newstest2019-defr-src-ts.de.xml',\n )\n for bad_file in bad_files:\n bad_path = join(getcwd(), 'testdata', bad_file)\n with self.assertRaises(ValueError):\n _ = valitest.ValidatableTestSet(bad_path)", "def test_add_to_blacklist(self):\n\n self.feature_test.add_to_blacklist(3)\n self.assertTrue(3 in Feature(\"testing\").blacklist)" ]
[ "0.6533181", "0.6533181", "0.6494651", "0.64306563", "0.6377956", "0.61511934", "0.6121788", "0.5941797", "0.5909957", "0.5882721", "0.5851798", "0.5822298", "0.56683594", "0.5648593", "0.5628515", "0.55158484", "0.54681826", "0.54313976", "0.5406495", "0.5349667", "0.5342094", "0.53107697", "0.5302546", "0.527642", "0.52742714", "0.5259476", "0.5247808", "0.52374256", "0.5211466", "0.52091473" ]
0.804985
0
Get the scsi data, and vtd mapping data and output in the fomat viosname,lunid,disk,vtdname,partitionid
def getVIOSmappings(vios,csvFile=''): """ We need the cec this vios is running to be able to map the lpar names """ cec = nim_attribute(vios,"mgmt_profile").split()[2] """ ssh comms all ok?? """ isOK = ssh_output(vios,"echo OK") if isOK == None or isOK == False: error_message("Unable to communicate with %s, host not responding." % vios ) return None """ scsi lun ids mapped to the disks """ scsidata = {} vhostMapping = {} vtdMapping = {} vhostMapping = getvhostmapping(vios) vtdMapping = getvhostmapping_with_vtd(vios) scsidata = get_EMC_CLARRiiON_inq(vios) logdata( lparMapping ) for disk in vhostMapping: lunid=scsidata.get(disk) lparName=lparMapping.get( cec + "-" + str(vtdMapping.get( vhostMapping[disk] ) ) ) if csvFile: csvFile.write( "\"%s\",\"%s\",\"%s\",\"%s\",\"%s\",\"%s\"\n" % ( vios, lunid, disk, vtdMapping.get(disk), vhostMapping[disk], lparName, ) ) else: print( "%-16s %34s %-9s %16s %-9s %-4s" % ( vios, lunid, disk, vtdMapping.get(disk), vhostMapping[disk], lparName, ) )
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def print_vscsi_attributes(self,objects):\n print(\"\\n\")\n print((\"LocalPartitionID\".ljust(35),\":\",objects.LocalPartitionID.value()))\n print((\"VirtualSlotNumber\".ljust(35),\":\",objects.VirtualSlotNumber.value()))\n print((\"RequiredAdapter\".ljust(35),\":\",objects.RequiredAdapter.value()))\n print((\"RemoteLogicalPartitionID\".ljust(35),\":\",objects.RemoteLogicalPartitionID.value()))\n print((\"RemoteSlotNumber\".ljust(35),\":\",objects.RemoteSlotNumber.value()))", "def get_device_file_dict():\n cmd = 'lshw -class disk'\n desc = \"description\"\n log_name = \"logical name\"\n serial = \"serial\"\n\n dev = []\n dev_list = []\n\n ret, output, err = run_gluster_command(cmd)\n output = output.decode('ASCII')\n dev_info = output.split('\\n')\n for line in dev_info:\n if re.search(desc, line):\n if dev:\n dev_list.append(dev)\n\n dev = []\n if re.search(log_name, line) or re.search(serial, line):\n temp = line.split(':')\n temp[1] = temp[1].strip(' ')\n dev.append(temp[1])\n dev_list.append(dev)\n for line in dev_list:\n print(line)", "def get_devices_lsscsi(self):\n\n try:\n message = \"Find SCSI Devices\"\n if self._include_enclosures:\n command = \"lsscsi --generic --transport | egrep 'disk|0x14|enclo'\"\n else:\n command = \"lsscsi --generic --transport | fgrep 'disk|0x14'\"\n pdata = self._run_command(command=command, message=message, logger=self._logger, shell=True)\n #\n # Format:\n # $ lsscsi --generic --transport\n # [0] [1] [2] [3] [4]\n # [0:0:0:0] disk sas:0x5000cca25103b471 /dev/sda /dev/sg0 \n # [0:0:1:0] disk sas:0x5000cca251029301 /dev/sdb /dev/sg1 \n # ...\n # [0:0:14:0] enclosu sas:0x5001636001caa0bd - /dev/sg14\n # [7:0:0:0] cd/dvd usb: 1-1.3:1.2 /dev/sr0 /dev/sg15\n #\n # Special Case:\n # Handle lines without a transport (spaces only). (screen scrapping danger)\n # [0:0:10:0] enclosu sas:0x50030480091d71fd - /dev/sg10\n # [1:0:0:0] disk <spaces> /dev/sdk /dev/sg11 <- INTEL disk!\n #\n # Another SNAFU! (and why I hate screen scrapping!!!)\n # [15:0:53597:0]disk sas:0x5000cca23b359649 /dev/sdg /dev/sg6 \n # [15:0:53598:0]disk sas:0x5000cca23b0c0a99 /dev/sdh /dev/sg7 \n # [15:0:53599:0]disk sas:0x5000cca23b0b7531 /dev/sdi /dev/sg8 \n # ...\n # [15:0:53686:0]enclosu sas:0x5000ccab040001bc - /dev/sg165\n # [15:0:53766:0]enclosu sas:0x5000ccab040001fc - /dev/sg144\n #\n # Evidently, the author of lsscsi did not think of consistent output! ;(\n #\n for line in pdata['stdout'].splitlines():\n dinfo = line.split()\n device = dict()\n if len(dinfo) < 5:\n m = re.search('(?P<device>disk|\\(0x14\\)|enclosu)', dinfo[0])\n if m:\n device['Device Type'] = m.group('device')\n sas_index = 1\n dev_index = 2\n sg_index = 3\n else:\n continue\n else:\n device['Device Type'] = dinfo[1]\n sas_index = 2\n dev_index = 3\n sg_index = 4\n\n # lsscsi does not understand 'Host Managed' device type.\n if '0x14' in device['Device Type']:\n device['Device Type'] = 'disk'\n\n # Parse remaining information.\n if 'sas:' in dinfo[sas_index]:\n device['SAS Address'] = dinfo[sas_index][4:]\n self._sas_addresses += 1\n else:\n device['SAS Address'] = \"\"\n\n # Note: Enclosure has no driver, so reports '-' for name.\n if '/dev/' in dinfo[dev_index]:\n if self._drives and not dinfo[dev_index] in self._drives:\n continue\n if self._exclude and dinfo[dev_index] in self._exclude:\n continue\n device['Linux Device Name'] = dinfo[dev_index]\n else:\n device['Linux Device Name'] = \"\"\n if '/dev/sg' in dinfo[sg_index]:\n device['SCSI Device Name'] = dinfo[sg_index]\n else:\n device['SCSI Device Name'] = \"\"\n\n self._devices.append(device)\n\n except RuntimeError as exc:\n self._logger.error(\"Failed to acquire SCSI devices: {0}\".format(exc))\n raise exc", "def get_devices_spt(self):\n\n #import pdb; pdb.set_trace()\n if self._drives or self.firmware_version or self.product_name or self.vendor_name or \\\n self.serial_number or self.target_port:\n user_options = True\n else:\n user_options = False\n try:\n # Note: Extra logic to optimize spt device directory scanning.\n if not user_options:\n if self._include_enclosures:\n message = \"Find SCSI Devices\"\n command = \"{tool} show devices dtype=direct,hostmanaged,enclosure\".format(tool=self.tool)\n else:\n message = \"Find SCSI Disk Drives\"\n command = \"{tool} show devices dtype=direct,hostmanaged\".format(tool=self.tool)\n # Use common execute below.\n else:\n # Request enclosures separately.\n if self._include_enclosures:\n message = \"Find SCSI Enclosures\"\n command = \"{tool} show devices dtype=enclosure ofmt=json\".format(tool=self.tool)\n pdata = self._run_command(command=command, message=message,\n logger=self._logger, shell=False, expected_failure=True)\n if pdata['exit_code'] == self.EXIT_STATUS_SUCCESS and pdata['stdout']:\n devices = json.loads(pdata['stdout'])\n self.parse_devices_spt(devices)\n\n message = \"Find SCSI Disk Drives\"\n # Selective drives or all direct access (disk drives).\n if self._drives:\n command = \"{tool} show edt dtype=direct,hostmanaged devices={drives}\"\\\n .format(tool=self.tool, drives=\",\".join(self._drives))\n else:\n command = \"{tool} show devices dtype=direct,hostmanaged\".format(tool=self.tool)\n # Apply optional parameters.\n if self.product_name:\n command += \" pid={product}\".format(product=self.product_name)\n if self.vendor_name:\n command += \" vid={vendor}\".format(vendor=self.vendor_name)\n if self.serial_number:\n command += \" serial={serial}\".format(serial=self.serial_number)\n if self.target_port:\n command += \" tport={target}\".format(target=self.target_port)\n if self.firmware_version:\n command += \" fw_version={firmware}\".format(firmware=self.firmware_version)\n\n # Add common spt options, we want JSON output!\n if self._exclude:\n command += \" exclude={drives}\".format(drives=\",\".join(self._exclude))\n command += \" ofmt=json\"\n # Finally, execute spt and parse its' JSON output (if any).\n pdata = self._run_command(command=command, message=message,\n logger=self._logger, shell=False, expected_failure=True)\n # spt emits warning status (1) and no JSON output if no devices found.\n if pdata['exit_code'] == self.EXIT_STATUS_SUCCESS and pdata['stdout']:\n devices = json.loads(pdata['stdout'])\n self.parse_devices_spt(devices)\n\n except RuntimeError as exc:\n self._logger.error(\"Failed to acquire SCSI devices: {0}\".format(exc))\n raise exc\n\n except ValueError as exc:\n self._logger.error(\"Failed to parse spts' JSON output: {0}\".format(exc))\n raise exc", "def get_partition_details(disk, partition):\n details = {}\n script = [\n 'select disk {}'.format(disk['Number']),\n 'select partition {}'.format(partition['Number']),\n 'detail partition']\n\n # Diskpart details\n try:\n # Run script\n result = run_diskpart(script)\n except subprocess.CalledProcessError:\n pass\n else:\n # Get volume letter or RAW status\n output = result.stdout.decode().strip()\n tmp = re.search(r'Volume\\s+\\d+\\s+(\\w|RAW)\\s+', output)\n if tmp:\n if tmp.group(1).upper() == 'RAW':\n details['FileSystem'] = RAW\n else:\n details['Letter'] = tmp.group(1)\n # Remove empty lines from output\n tmp = [s.strip() for s in output.splitlines() if s.strip() != '']\n # Split each line on ':' skipping those without ':'\n tmp = [s.split(':') for s in tmp if ':' in s]\n # Add key/value pairs to the details variable and return dict\n details.update({key.strip(): value.strip() for (key, value) in tmp})\n\n # Get MBR type / GPT GUID for extra details on \"Unknown\" partitions\n guid = PARTITION_UIDS.get(details.get('Type').upper(), {})\n if guid:\n details.update({\n 'Description': guid.get('Description', '')[:29],\n 'OS': guid.get('OS', 'Unknown')[:27]})\n\n if 'Letter' in details:\n # Disk usage\n try:\n tmp = psutil.disk_usage('{}:\\\\'.format(details['Letter']))\n except OSError as err:\n details['FileSystem'] = 'Unknown'\n details['Error'] = err.strerror\n else:\n details['Used Space'] = human_readable_size(tmp.used)\n\n # fsutil details\n cmd = [\n 'fsutil',\n 'fsinfo',\n 'volumeinfo',\n '{}:'.format(details['Letter'])\n ]\n try:\n result = run_program(cmd)\n except subprocess.CalledProcessError:\n pass\n else:\n output = result.stdout.decode().strip()\n # Remove empty lines from output\n tmp = [s.strip() for s in output.splitlines() if s.strip() != '']\n # Add \"Feature\" lines\n details['File System Features'] = [s.strip() for s in tmp\n if ':' not in s]\n # Split each line on ':' skipping those without ':'\n tmp = [s.split(':') for s in tmp if ':' in s]\n # Add key/value pairs to the details variable and return dict\n details.update({key.strip(): value.strip() for (key, value) in tmp})\n\n # Set Volume Name\n details['Name'] = details.get('Volume Name', '')\n\n # Set FileSystem Type\n if details.get('FileSystem', '') not in ['RAW', 'Unknown']:\n details['FileSystem'] = details.get('File System Name', 'Unknown')\n\n return details", "def get_vdisk_map_by_uid(self, vdisk_uid):\n\n if not vdisk_uid:\n msg = _(\"Invalid parameter: vdisk_uid is None \")\n LOG.exception(msg)\n raise SVCInvalidVdiskUID\n\n map_info = {}\n\n cmd = 'svcinfo lshostvdiskmap -delim :'\n\n output, err_output = self._svc_command(cmd)\n\n if err_output:\n msg = (_(\"get_vdisk_map_scsi_id() failure: svc cmd: %(cmd)s\"\n \" error: %(error)s\") % {'cmd': cmd, 'error': err_output})\n LOG.exception(msg)\n ex_args = {'cmd': cmd, 'e': err_output}\n raise SVCCommandException(**ex_args)\n\n if len(output) < 2:\n msg = _(\"_add_svc_hostmap() failed to create \"\n \"vdisk host mapping on storage\")\n LOG.warn(msg)\n return map_info\n\n header = output.pop(0).split(':')\n scsi_id_idx = header.index(SVC_KEY_MAP_SCSI_ID)\n vdisk_uid_idx = header.index(SVC_KEY_VDISK_UID)\n hostname_idx = header.index(SVC_KEY_HOST_NAME)\n\n for line in output:\n info = line.split(':')\n host = info[hostname_idx]\n vdisk_UID = info[vdisk_uid_idx]\n scsi_id = info[scsi_id_idx]\n\n # we may have one LUN mapped to multiple hosts during LPM\n if vdisk_UID == vdisk_uid:\n map_info[host] = scsi_id\n\n if len(map_info) == 0:\n map_info = None\n\n LOG.debug(\"vdisk hostmap: %(map_info)s\"\n % {'map_info': str(map_info)})\n return map_info", "def show_device_information_only(self):\n\n # TODO: Optimize formatting later!\n if self._sas_addresses:\n if self._report_header:\n print(\"\\n\")\n print(\" Linux SCSI Firmware Drive Block Curr\")\n print(\" Device Device Vendor Product Revision Capacity Length Temp Serial Number SAS Address\")\n print(\"---------- ---------- -------- ---------------- -------- ------------ ------ ---- -------------- ------------------\")\n\n for device in self._devices:\n if not device['Device Type'].startswith(\"disk\"):\n continue\n print('{dsf:<10} {sdsf:<10} {vid:<8} {pid:<16} {fw:<8} {capacity:>12} {blocklen:>4} {temp:<4} {serial:<14} {sas:<18}'\n .format(dsf=device['Linux Device Name'],\n sdsf=device['SCSI Device Name'],\n vid=device['Vendor Identification'],\n pid=device['Product Identification'],\n fw=device['Firmware Version'],\n capacity=device['Drive Capacity'],\n blocklen=device['Block Length'],\n temp=device['Current Temperature'],\n serial=device['Serial Number'],\n sas=device['SAS Address']))\n else:\n if self._report_header:\n print(\"\\n\")\n print(\" Linux SCSI Firmware Drive Block Curr\")\n print(\" Device Device Vendor Product Revision Capacity Length Temp Serial Number\")\n print(\"---------- ---------- -------- ---------------- -------- ------------ ------ ---- -------------\")\n\n for device in self._devices:\n if not device['Device Type'].startswith(\"disk\"):\n continue\n print('{dsf:<10} {sdsf:<10} {vid:<8} {pid:<16} {fw:<8} {capacity:>12} {blocklen:>4} {temp:<4} {serial:<14}'\n .format(dsf=device['Linux Device Name'],\n sdsf=device['SCSI Device Name'],\n vid=device['Vendor Identification'],\n pid=device['Product Identification'],\n fw=device['Firmware Version'],\n capacity=device['Drive Capacity'],\n blocklen=device['Block Length'],\n temp=device['Current Temperature'],\n serial=device['Serial Number']))\n\n if self._report_header:\n print(\"\\n\")", "def mpt():\n lbl_drives = ['device','mountpoint','fstype']\n disks = [d[0:3] for d in psutil.disk_partitions()]\n drives = [dict(zip(lbl_drives,ds)) for ds in disks]\n return [d['mountpoint']for d in drives]", "def get_volume_info(self, uid):\n LOG.debug(\"Entering\")\n cmd = \"svcinfo lsvdisk -bytes -filtervalue vdisk_UID=%s -delim :\" % uid\n output = self._svc_command(cmd)[0]\n\n if len(output) != 2:\n raise SVCVolumeNotFound(\n _(\"Couldn't find volume information for UID %s\") % uid)\n\n header = output[0].split(':')\n values = output[1].split(':')\n index = header.index(SVC_KEY_VDISK_ID)\n diskId = values[index]\n index = header.index(SVC_KEY_VDISK_NAME)\n name = values[index]\n index = header.index(SVC_KEY_VOLUME_GROUP)\n volumeGroup = values[index]\n index = header.index(SVC_KEY_VDISK_CAPACITY)\n capacity = values[index]\n\n info = {SVC_KEY_VDISK_ID: diskId,\n SVC_KEY_VDISK_NAME: name,\n SVC_KEY_VOLUME_GROUP: volumeGroup,\n SVC_KEY_VDISK_CAPACITY: capacity}\n\n LOG.debug(\"Exiting\")\n return info", "def vdisk_in_flashcopy(self, diskname):\n LOG.debug(\"Entering\")\n cmd = ''.join([\"svcinfo lsfcmap -filtervalue \",\n \"target_vdisk_name=%s -delim :\" % diskname])\n output = self._svc_command(cmd)[0]\n\n if len(output) != 2:\n return(100, None)\n\n header = output[0].split(':')\n values = output[1].split(':')\n index = header.index('progress')\n progress = values[index]\n index = header.index('id')\n map_id = values[index]\n\n LOG.debug(\"Exiting (progress = %s, map_id = %s)\" % (progress, map_id))\n return progress, map_id", "def get_block_device_list(vars = {}, log = sys.stderr):\n\n # make sure we can access to the files/directories in /proc\n if not os.access(PROC_PARTITIONS_PATH, os.F_OK):\n return None\n\n # table with valid scsi/sata/ide/raid block device names\n valid_blk_names = {}\n # add in valid sd and hd block device names\n for blk_prefix in ('sd','hd'):\n for blk_num in map (\\\n lambda x: chr(x), range(ord('a'),ord('z')+1)):\n devicename=\"%s%c\" % (blk_prefix, blk_num)\n valid_blk_names[devicename]=None\n\n # add in valid scsi raid block device names\n for M in range(0,1+1):\n for N in range(0,7+1):\n devicename = \"cciss/c%dd%d\" % (M,N)\n valid_blk_names[devicename]=None\n\n for devicename in valid_blk_names.keys():\n # devfs under 2.4 (old boot cds) used to list partitions\n # in a format such as scsi/host0/bus0/target0/lun0/disc\n # and /dev/sda, etc. were just symlinks\n try:\n devfsname= os.readlink( \"/dev/%s\" % devicename )\n valid_blk_names[devfsname]=None\n except OSError:\n pass\n\n # only do this once every system boot\n if not os.access(DEVICES_SCANNED_FLAG, os.R_OK):\n\n # this is ugly. under devfs, device\n # entries in /dev/scsi/.. and /dev/ide/...\n # don't show up until you attempt to read\n # from the associated device at /dev (/dev/sda).\n # so, lets run sfdisk -l (list partitions) against\n # most possible block devices, that way they show\n # up when it comes time to do the install.\n devicenames = valid_blk_names.keys()\n devicenames.sort()\n for devicename in devicenames:\n os.system( \"sfdisk -l /dev/%s > /dev/null 2>&1\" % devicename )\n\n # touch file\n fb = open(DEVICES_SCANNED_FLAG,\"w\")\n fb.close()\n\n devicelist= {}\n\n partitions_file= file(PROC_PARTITIONS_PATH,\"r\")\n line_count= 0\n for line in partitions_file:\n line_count= line_count + 1\n\n # skip the first two lines always\n if line_count < 2:\n continue\n\n parts= string.split(line)\n\n if len(parts) < 4:\n continue\n\n device= parts[3]\n\n # skip and ignore any partitions\n if not valid_blk_names.has_key(device):\n continue\n\n try:\n major= int(parts[0])\n minor= int(parts[1])\n blocks= int(parts[2])\n except ValueError, err:\n continue\n\n gb_size= blocks/BLOCKS_PER_GB\n\n # check to see if the blk device is readonly\n try:\n # can we write to it?\n dev_name= \"/dev/%s\" % device\n fb = open(dev_name,\"w\")\n fb.close()\n readonly=False\n except IOError, e:\n # check if EROFS errno\n if errno.errorcode.get(e.errno,None) == 'EROFS':\n readonly=True\n else:\n # got some other errno, pretend device is readonly\n readonly=True\n\n devicelist[dev_name]= {'major': major,'minor': minor,'blocks': blocks, 'size': gb_size, 'readonly': readonly}\n return devicelist", "def get_device_map():\n ret = []\n vlist = subprocess.check_output(['ceph-volume', 'lvm', 'list',\n '--format=json'])\n for osd_id, data in json.loads(vlist.decode('utf8')).items():\n osd_id = normalize_osd_id(osd_id)\n for elem in data:\n for device in elem['devices']:\n ret.append({'id': osd_id, 'path': device})\n return ret", "def get_lun_storage_info(lun_id):\n host = ll_hosts.get_spm_host(config.HOSTS)\n host_ip = ll_hosts.get_host_ip(host)\n executor = rhevm_helpers.get_host_executor(\n host_ip, config.VDC_ROOT_PASSWORD\n )\n # Execute 'pvscan' to display the latest volume info\n storage_resources.pvscan(host)\n logger.info(\"Executing command 'pvs | grep %s'\", lun_id)\n status, output, err = executor.run_cmd(\n shlex.split(PVS_SHOW_LUN_INFO % lun_id)\n )\n if status:\n logger.info(\n \"Status was False executing 'pvs | grep %s'. Err: %s\",\n lun_id, err\n )\n return 0, 0\n\n # Format the output into the 6 expected display parameters (PV, VG,\n # Format, LV Attributes, Physical size and Physical free size)\n formatted_output = shlex.split(output)\n logger.info(\n \"The output received when running pvs on LUN id %s is: %s\"\n % (lun_id, formatted_output)\n )\n # The 2nd last displayed data output is needed - Physical size\n lun_size = formatted_output[-2]\n lun_size = lun_size.replace(\"g\", \"\")\n lun_free_space = formatted_output[-1]\n lun_free_space = lun_free_space.replace(\"g\", \"\")\n lun_size_bytes = float(lun_size) * config.GB\n logger.info(\"The LUN size in bytes is '%s'\", str(lun_size_bytes))\n lun_free_bytes = float(lun_free_space) * config.GB\n logger.info(\"The LUN free space in bytes is '%s'\", str(lun_free_bytes))\n\n return int(lun_size_bytes), int(lun_free_bytes)", "def get_generic_scsi_dev(self, nIndex):\n\t\treturn handle_to_object(call_sdk_function('PrlVmCfg_GetGenericScsiDev', self.handle, nIndex))", "def get_disk_info(virt_type, instance, block_device_info=None,\n image_meta=None, rescue=False):\n\n disk_bus = get_disk_bus_for_device_type(virt_type, image_meta, \"disk\")\n cdrom_bus = get_disk_bus_for_device_type(virt_type, image_meta, \"cdrom\")\n mapping = get_disk_mapping(virt_type, instance,\n disk_bus, cdrom_bus,\n block_device_info,\n image_meta, rescue)\n\n return {'disk_bus': disk_bus,\n 'cdrom_bus': cdrom_bus,\n 'mapping': mapping}", "def get_generic_scsi_device(self, nIndex):\n\t\treturn handle_to_object(call_sdk_function('PrlSrvCfg_GetGenericScsiDevice', self.handle, nIndex))", "def usableparts(self):\n # First get the partition type-id for all hard disk partitions\n partid = {}\n for pline in self.fdiskl():\n partid[pline[0]] = pline[4]\n ups = {}\n for s in self.xlist(\"get-blkinfo\")[1]:\n mo = re.match(r'(/dev/[^:]*):(?: LABEL=\"([^\"]*)\")?(?:'\n ' UUID=\"([^\"]*)\")?(?: TYPE=\"([^\"]*)\")?', s)\n if mo:\n dev, label, uuid, fstype = mo.groups()\n if fstype in (None, \"linux_raid_member\", \"LVM2_member\"):\n continue\n if dev.startswith(\"/dev/loop\"):\n continue\n rem = None\n if dev.startswith(\"/dev/sd\"):\n if partid.get(dev) == \"fd\":\n # This test seems to be necessary because blkid\n # sometimes returns an fs-type, rather than\n # linux_raid_member\", for the the first device\n # in a formatted raid array\n continue\n rem = self.xlist(\"removable\", dev)[1][0].strip() == \"1\"\n ups[dev] = (fstype, label, uuid, rem)\n return ups", "def show_device_information_enclosures(self):\n\n if self._report_header:\n print(\"\\n\")\n print(\" Linux SCSI Firmware Drive Block Curr Enc SCSI Enc Enc Slot\")\n print(\" Device Device Vendor Product Revision Capacity Length Temp Serial Number SAS Address Device Slot Description\")\n print(\"---------- ---------- -------- ---------------- -------- ------------ ------ ---- -------------- ------------------ ---------- ---- -----------------\")\n\n for device in self._devices:\n if not device['Device Type'].startswith(\"disk\"):\n continue\n print('{dsf:<10} {sdsf:<10} {vid:<8} {pid:<16} {fw:<8} {capacity:>12} {blocklen:>4} {temp:<4} {serial:<14} {sas:<18} {edsf:<10} {slot:<3} {text:<16}'\n .format(dsf=device['Linux Device Name'],\n sdsf=device['SCSI Device Name'],\n vid=device['Vendor Identification'],\n pid=device['Product Identification'],\n fw=device['Firmware Version'],\n capacity=device['Drive Capacity'],\n blocklen=device['Block Length'],\n temp=device['Current Temperature'],\n serial=device['Serial Number'],\n sas=device['SAS Address'],\n edsf=device['Enclosure Device'],\n slot=device['Enclosure Slot'],\n text=device['Slot Description']))\n\n if self._report_header:\n print(\"\\n\")", "def connect_instance_disk_to_mgmt(self, instance):\n for stg_elem, vios in self._get_bootdisk_iter(instance):\n msg_args = {'disk_name': stg_elem.name, 'vios_name': vios.name}\n\n # Create a new mapping. NOTE: If there's an existing mapping on\n # the other VIOS but not this one, we'll create a second mapping\n # here. It would take an extreme sequence of events to get to that\n # point, and the second mapping would be harmless anyway. The\n # alternative would be always checking all VIOSes for existing\n # mappings, which increases the response time of the common case by\n # an entire GET of VIOS+VIO_SMAP.\n LOG.debug(\"Mapping boot disk %(disk_name)s to the management \"\n \"partition from Virtual I/O Server %(vios_name)s.\",\n msg_args, instance=instance)\n try:\n tsk_map.add_vscsi_mapping(self.host_uuid, vios, self.mp_uuid,\n stg_elem)\n # If that worked, we're done. add_vscsi_mapping logged.\n return stg_elem, vios\n except Exception:\n LOG.exception(\"Failed to map boot disk %(disk_name)s to the \"\n \"management partition from Virtual I/O Server \"\n \"%(vios_name)s.\", msg_args, instance=instance)\n # Try the next hit, if available.\n # We either didn't find the boot dev, or failed all attempts to map it.\n raise npvmex.InstanceDiskMappingFailed(instance_name=instance.name)", "def __init__(self,guid='',name=''):\n extent.Extent.__init__(self,guid,name,'Physical LUN/Disk')\n self.exttype='physical'\n self.alias=''\n self.alwaysadd=True\n self.direct=True\n self.serial='' # SCSI Serial ID\n self.model='' # SCSI Model\n self.vendor='' # SCSI Vendor\n self.revision='' # SCSI revision\n self.raid=RaidLevel.none # underline raid '0' '1' '5' '6' '10'\n self.sectorsize=512\n self.paths=VSACollection(san.SanPath,self,'paths',desc='Paths To Disk',icon='link_g.png') # list of paths leading to this physical disk (local or remote), key = dev/hbtl\n self.partitions=VSACollection(SanPartition,self,'partitions',icon='vm_g.png') # list of paths leading to this physical disk (local or remote), key = dev/hbtl\n self.usedin=RefDict(extent.Extent,self,'usedin',desc='Owned by Storage Extents',icon='hard_disk.png',call=lambda self:[v for v in self.usedby if v.exttype<>'partition'])\n self.assigned=[] # list of chunks assigned from this disk\n self.primordial=1 # primary extent, based on physical HW (e.g. not a logical volume)\n self.cachepvds=[] # list of provider names which have a matching cache logical volume (found in discovery, process extents)\n self.cachedrdev=None # points to cache DRBD device\n self._flush()", "def __get_info(self, disk):\n # Checking the disk exists\n\tFNULL = open(os.devnull, 'w') # Defining /dev/null\n\tcheck = subprocess.call(['df', disk], stdout=FNULL,\\\n stderr=subprocess.STDOUT)\n # Disk not found\n if check != 0:\n sys.stdout.write(\"^fg(%s)!E! DiskNotFound^fg()\" % Colors.CL_BASE08)\n # Disk found\n else:\n # Executing command, parsing output and removing empty elements\n cmd = subprocess.Popen(['df', '-H', disk], stdout=subprocess.PIPE)\n cmd_out, cmd_err = cmd.communicate()\n cmd_outparsed = cmd_out.split(' ')\n cmd_outparsed = filter(None, cmd_outparsed)\n\n # Getting information\n disk_name = disk\n disk_size = cmd_outparsed[7]\n disk_used = cmd_outparsed[8]\n disk_available = cmd_outparsed[9]\n disk_percentage = cmd_outparsed[10].translate(None, \"%\")\n disk_percentage = int(disk_percentage)\n\n # Disk Name: ~\n sys.stdout.write(\"^fg(%s)[^fg()\" % Colors.CL_BASE02)\n if disk_name == DISK_DATA:\n sys.stdout.write(\"^fg(%s)~^fg()\" % Colors.CL_BASE0D)\n sys.stdout.write(\"^fg(%s): ^fg()\" % Colors.CL_BASE03)\n # Disk Name: /\n elif disk_name == DISK_ROOT:\n sys.stdout.write(\"^fg(%s)/^fg()\" % Colors.CL_BASE0D)\n sys.stdout.write(\"^fg(%s): ^fg()\" % Colors.CL_BASE03)\n \n # Disk Percentage: Good\n if 0 <= disk_percentage <= 60:\n sys.stdout.write(\"^fg(%s)%s%%^fg()\" % (Colors.CL_BASE0B, disk_percentage))\n sys.stdout.write(\"^fg(%s) %s^fg()\" % (Colors.CL_BASE0B, disk_used))\n sys.stdout.write(\"^fg(%s)/^fg()\" % Colors.CL_BASE03)\n # Disk Percentage: Fair\n elif 61 <= disk_percentage <= 90:\n sys.stdout.write(\"^fg(%s)%s%%^fg()\" % (Colors.CL_BASE09, disk_percentage))\n sys.stdout.write(\"^fg(%s) %s^fg()\" % (Colors.CL_BASE09, disk_used))\n sys.stdout.write(\"^fg(%s)/^fg()\" % Colors.CL_BASE03)\n # Disk Percentage: Weak\n elif 91 <= disk_percentage <= 100:\n sys.stdout.write(\"^fg(%s)%s%%^fg()\" % (Colors.CL_BASE08, disk_percentage))\n sys.stdout.write(\"^fg(%s) %s^fg()\" % (Colors.CL_BASE08, disk_used))\n sys.stdout.write(\"^fg(%s)/^fg()\" % Colors.CL_BASE03)\n sys.stdout.write(\"^fg(%s)%s^fg()\" % (Colors.CL_BASE0D, disk_size))\n sys.stdout.write(\"^fg(%s)]^fg()\" % Colors.CL_BASE02)", "def show_mappings(project):\n # with BMI(_username, _password, project) as bmi:\n # ret = bmi.show_mounted()\n # if ret[constants.STATUS_CODE_KEY] == 200:\n # table = PrettyTable(field_names=['Target', 'Block Device'])\n # mappings = ret[constants.RETURN_VALUE_KEY]\n # for k, v in mappings.iteritems():\n # table.add_row([k, v])\n # click.echo(table.get_string())\n # else:\n # click.echo(ret[constants.MESSAGE_KEY])\n click.echo(\"Need to Re-Implement\")", "def _get_dmi(cls, data):\n\t\tdata['pdi'] = cls._get_pdi(data, 14)\n\t\tdata['mdi'] = cls._get_mdi(data, 14)\n\t\tdata['dx'] = cls._get_dx(data, 14)\n\t\tdata['adx'] = data['dx_6_ema']\n\t\tdata['adxr'] = data['adx_6_ema']\n\t\treturn data", "def step2(self):\n self.mr_vds = []\n # Checking UGood availablity here, because after reboot cycle,\n # init will be called.\n for mr in self.mrs:\n self.log.info(\"Check for UGood PD's availablity on controller:%d\"\n % (mr.ctrl_id))\n pds = mr.cli.list_all_drives(pd_type=\"SAS\",\n state=\"UGood\", media_type=\"HDD\",\n sector_size='512B')\n if len(pds) < 10:\n raise SALError(\"%d UGood HDD PD's are not available for RAID \"\n \"Creation on Controller:%d\"\n % (10 - len(pds), mr.ctrl_id))\n else:\n self.log.info(\"Number of UGood HDD PD's available is %d on \"\n \"controller:%d\" % (len(pds), mr.ctrl_id))\n for mr in self.mrs:\n self.vds = []\n self.log.info(\"Create R0, R1, R5 with diff properties on \"\n \"Controller: %d\" % (mr.ctrl_id))\n self.vds.append(mr.cli.add_vd(raid=0, vd_size=\"3000MB\",\n WB=\"WB\", RA='ra', cache=\"direct\"))\n self.vds.append(mr.cli.add_vd(raid=1, vd_size=\"3000MB\",\n WB=\"WT\", RA='nora', cache=\"cached\"))\n self.vds.append(mr.cli.add_vd(raid=5, vd_size=\"3000MB\",\n WB=\"WB\", RA='ra', cache=\"direct\"))\n for vd in self.vds:\n while True:\n time.sleep(5)\n if (mr.cli.init_progress(vd) == -1):\n self.log.info(\"FGI completed for VD:%s\"\n % (vd))\n break\n for vd in self.vds:\n if mr.cli.vd_get_info(vd)['state'] != 'OPTL':\n raise SALError(\"RAID-%s VD%s is not optimal\"\n % (mr.cli.vd_get_info(vd)['raid'], vd))\n else:\n self.log.info(\"RAID-%s VD%s is optimal\"\n % (mr.cli.vd_get_info(vd)['raid'], vd))\n\n self.mr_vds.extend(self.vds)\n self.log.info(\"Created R0, R1, R5 successfully on controller: %d\"\n % (mr.ctrl_id))", "def get_all_disk():\n\t\tdisks = []\n\t\tdisks_lines = linux.exe_shell(\"lsblk -o NAME,VENDOR|grep -P '^sd.*[A-Z]'\")\n\t\tfor line in disks_lines.splitlines():\n\t\t\tdisk_t = line.split()\n\t\t\tif len(disk_t) > 1 and \"LSI\" not in disk_t[1]:\n\t\t\t\tdisks.append(disk_t[0])\n\t\tds = []\n\t\tfor i in disks:\n\t\t\td_t = DiskFromLsiSas3(\"\", i)\n\t\t\td_t.fill_attrs()\n\t\t\tds.append(d_t)\n\t\treturn ds", "def get_partition_info(disk, host_disk):\n\n volume = \"/dev/\"+disk\n DISKINFO[volume] = {}\n DISKINFO[volume][\"Name\"] = volume\n DISKINFO[volume][\"Type\"] = \"Partition\"\n DISKINFO[volume][\"HostDevice\"] = host_disk\n DISKINFO[volume][\"Partitions\"] = []\n DISKINFO[host_disk][\"Partitions\"].append(volume)\n DISKINFO[volume][\"Vendor\"] = get_vendor(disk)\n DISKINFO[volume][\"Product\"] = \"Host Device: \"+DISKINFO[host_disk][\"Product\"]\n DISKINFO[volume][\"RawCapacity\"], DISKINFO[volume][\"Capacity\"] = get_capacity()\n DISKINFO[volume][\"Description\"] = get_description(disk)\n DISKINFO[volume][\"Flags\"] = get_capabilities(disk)\n DISKINFO[volume][\"FileSystem\"] = get_file_system(disk)\n DISKINFO[volume][\"Partitioning\"] = \"N/A\"\n DISKINFO[volume][\"UUID\"] = get_uuid(disk)\n DISKINFO[volume][\"ID\"] = get_id(disk)\n DISKINFO[volume][\"BootRecord\"], DISKINFO[volume][\"BootRecordStrings\"] = get_boot_record(disk)\n\n return volume", "def device_info(node):\n\n if \"cpu\" in node and \"total_mbufs\" in node[\"cpu\"]:\n total_mbufs = node[\"cpu\"][\"total_mbufs\"]\n if total_mbufs != 0:\n print(\"Total Number of Buffers: {}\".format(total_mbufs))\n\n vpp = VppPCIUtil(node)\n vpp.get_all_devices()\n linkup_devs = vpp.get_link_up_devices()\n if len(linkup_devs):\n print(\"\\nDevices with link up (can not be used with VPP):\")\n vpp.show_vpp_devices(linkup_devs, show_header=False)\n # for dev in linkup_devs:\n # print (\" \" + dev)\n kernel_devs = vpp.get_kernel_devices()\n if len(kernel_devs):\n print(\"\\nDevices bound to kernel drivers:\")\n vpp.show_vpp_devices(kernel_devs, show_header=False)\n else:\n print(\"\\nNo devices bound to kernel drivers\")\n\n dpdk_devs = vpp.get_dpdk_devices()\n if len(dpdk_devs):\n print(\"\\nDevices bound to DPDK drivers:\")\n vpp.show_vpp_devices(dpdk_devs, show_interfaces=True, show_header=False)\n else:\n print(\"\\nNo devices bound to DPDK drivers\")\n\n other_devs = vpp.get_other_devices()\n if len(other_devs):\n print(\"\\nDevices not bound to Kernel or DPDK drivers:\")\n vpp.show_vpp_devices(other_devs, show_interfaces=True, show_header=False)\n else:\n print(\"\\nNo devices not bound to Kernel or DPDK drivers\")\n\n vpputl = VPPUtil()\n interfaces = vpputl.get_hardware(node)\n if interfaces == {}:\n return\n\n print(\"\\nDevices in use by VPP:\")\n\n if len(interfaces.items()) < 2:\n print(\"None\")\n return\n\n print(\n \"{:30} {:4} {:4} {:7} {:4} {:7}\".format(\n \"Name\", \"Numa\", \"RXQs\", \"RXDescs\", \"TXQs\", \"TXDescs\"\n )\n )\n for intf in sorted(interfaces.items()):\n name = intf[0]\n value = intf[1]\n if name == \"local0\":\n continue\n numa = rx_qs = rx_ds = tx_qs = tx_ds = \"\"\n if \"numa\" in value:\n numa = int(value[\"numa\"])\n if \"rx queues\" in value:\n rx_qs = int(value[\"rx queues\"])\n if \"rx descs\" in value:\n rx_ds = int(value[\"rx descs\"])\n if \"tx queues\" in value:\n tx_qs = int(value[\"tx queues\"])\n if \"tx descs\" in value:\n tx_ds = int(value[\"tx descs\"])\n\n print(\n \"{:30} {:>4} {:>4} {:>7} {:>4} {:>7}\".format(\n name, numa, rx_qs, rx_ds, tx_qs, tx_ds\n )\n )", "def get_disk_details(disk):\n details = {}\n script = [\n 'select disk {}'.format(disk['Number']),\n 'detail disk']\n\n # Run\n try:\n result = run_diskpart(script)\n except subprocess.CalledProcessError:\n pass\n else:\n output = result.stdout.decode().strip()\n # Remove empty lines\n tmp = [s.strip() for s in output.splitlines() if s.strip() != '']\n # Set disk name\n details['Name'] = tmp[4]\n # Split each line on ':' skipping those without ':'\n tmp = [s.split(':') for s in tmp if ':' in s]\n # Add key/value pairs to the details variable and return dict\n details.update({key.strip(): value.strip() for (key, value) in tmp})\n\n return details", "def step6(self):\n if len(self.mrs) == 3:\n count = 3\n elif len(self.mrs) == 2:\n count = 2\n else:\n count = 1\n for mr in self.mrs[0:count]:\n self.log.info(\"Display boot drive on controller:%d\"\n % (mr.ctrl_id))\n vd_id = mr.cli.bootdrive_vd_get()\n if (int(vd_id) == -1): # -1 : No boot VD.\n self.log.info(\"No boot VD found on controller: %d\"\n % (mr.ctrl_id))\n else:\n self.log.info(\"VD ID of the boot VD: %d\"\n % int((vd_id)))", "def show_device_information_long(self):\n\n for device in self._devices:\n print(\"\")\n if device['Device Type'].startswith(\"enclosu\"):\n if device.get('Device Type'):\n print(\"{0:>32}: {1}\".format(\"Device Type\", device['Device Type']))\n if device['Device Type Description']:\n print(\"{0:>32}: {1}\".format(\"Device Description\", device['Device Type Description']))\n if device.get('SCSI Device Name'):\n print(\"{0:>32}: {1}\".format(\"SCSI Device Name\", device['SCSI Device Name']))\n if device.get('Product Identification'):\n print(\"{0:>32}: {1}\".format(\"Product Identification\", device['Product Identification']))\n if device.get('Vendor Identification'):\n print(\"{0:>32}: {1}\".format(\"Vendor Identification\", device['Vendor Identification']))\n if device.get('Firmware Version'):\n print(\"{0:>32}: {1}\".format(\"Firmware Version\", device['Firmware Version']))\n if device.get('Serial Number'):\n print(\"{0:>32}: {1}\".format(\"Serial Number\", device['Serial Number']))\n if device.get('SAS Address'):\n print(\"{0:>32}: {1}\".format(\"SAS Address\", device['SAS Address']))\n else:\n if device.get('Device Type'):\n print(\"{0:>32}: {1}\".format(\"Device Type\", device['Device Type']))\n if device['Device Type Description']:\n print(\"{0:>32}: {1}\".format(\"Device Description\", device['Device Type Description']))\n if device.get('Linux Device Name'):\n print(\"{0:>32}: {1}\".format(\"Linux Device Name\", device['Linux Device Name']))\n if device.get('SCSI Device Name'):\n print(\"{0:>32}: {1}\".format(\"SCSI Device Name\", device['SCSI Device Name']))\n if device.get('Product Identification'):\n print(\"{0:>32}: {1}\".format(\"Product Identification\", device['Product Identification']))\n if device.get('Vendor Identification'):\n print(\"{0:>32}: {1}\".format(\"Vendor Identification\", device['Vendor Identification']))\n if device.get('Firmware Version'):\n print(\"{0:>32}: {1}\".format(\"Firmware Version\", device['Firmware Version']))\n if device.get('Serial Number'):\n print(\"{0:>32}: {1}\".format(\"Serial Number\", device['Serial Number']))\n if device.get('Drive Capacity'):\n print(\"{0:>32}: {1}\".format(\"Drive Capacity\", device['Drive Capacity']))\n if device.get('Block Length'):\n print(\"{0:>32}: {1}\".format(\"Block Length\", device['Block Length']))\n if device.get('Power On Hours'):\n print(\"{0:>32}: {1}\".format(\"Power On Hours\", device['Power On Hours']))\n if device.get('Current Temperature'):\n print(\"{0:>32}: {1}\".format(\"Current Temperature\", device['Current Temperature']))\n if device.get('SAS Address'):\n print(\"{0:>32}: {1}\".format(\"SAS Address\", device['SAS Address']))\n if device.get('Enclosure Device'):\n print(\"{0:>32}: {1}\".format(\"Enclosure Device\", device['Enclosure Device']))\n if device.get('Enclosure Slot'):\n print(\"{0:>32}: {1}\".format(\"Enclosure Slot\", device['Enclosure Slot']))\n if device.get('Slot Description'):\n print(\"{0:>32}: {1}\".format(\"Slot Desciption\", device['Slot Description']))\n\n if len(self._devices):\n print(\"\")" ]
[ "0.6226752", "0.62065095", "0.613328", "0.5941464", "0.58446807", "0.5758732", "0.57254314", "0.5669364", "0.5603092", "0.5584567", "0.55693907", "0.5547329", "0.553814", "0.55317974", "0.55209994", "0.5405666", "0.5393464", "0.5392122", "0.5388479", "0.5382682", "0.53686225", "0.53394", "0.53099704", "0.5270263", "0.5264361", "0.52484417", "0.5245146", "0.5174927", "0.5137749", "0.5128073" ]
0.69272226
0
Ensures that a websafe key is valid and of the desired kind.
def _raiseIfWebsafeKeyNotValid(websafeKey, kind): # Check that websafeKey is not None if not websafeKey: raise endpoints.BadRequestException( "Websafe key not provided for '%s'" % kind) # Try to decode the websafe key into a real key try: key = ndb.Key(urlsafe=websafeKey) except: raise endpoints.BadRequestException( "Websafe key provided for '%s' could not be decoded: %s" % (kind, websafeKey)) # Ensure that the key is of the desired kind if key.kind() != kind: raise endpoints.BadRequestException( "Websafe key is not of the '%s' kind: %s" % (kind, websafeKey)) # If all is well, return the key return key
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def isValidKey(key):\n return True", "def validate_key(key):\r\n try:\r\n secret.Secret(key)\r\n except secret.Secret.InvalidSecret as e:\r\n raise KeyIsInvalid(e.message)", "def _check_key(self, key):\n raise NotImplementedError", "def test_valid_key(self):\n f = lws.valid_data_key\n assert f('string', int, r'string') is False\n assert f('string', str, r'test') is False\n assert f(123, int, '123') is False\n assert f(123.00, float, '123') is False\n assert f('123', str, r'[0-9]*') is True", "def validate_handshake_public_key(cls, public_key: bytes) -> None:\n ...", "def check_type( string_key ) : \r\n\r\n if type( string_key ) != type( '' ) : \r\n\r\n # raise self.__class__( \"'%s': EGI wants the key to be four _characters_ (not %s) !\" % (type(string_key), ) ) \r\n raise Eggog( \"'%s': EGI wants the key to be four _characters_ (not %s) !\" % (type(string_key), ) ) \r\n \r\n else :\r\n \r\n return True", "def check_key(self):\n\n if self.type == \"RSA\" and self.size < 1024:\n raise HostkeyError(\"RSA keys must at least be 1024 bits.\")\n elif self.type == \"DSA\" and self.size != 1024:\n raise HostkeyError(\"DSA keys can only be 1024 bits.\")\n elif self.type == \"ECDSA\" and self.size not in [256, 384, 521]: # yes, that is *really* 521 bits, not a typo!\n raise HostkeyError(\"ECDSA key must be either 256, 384 or 521 bits (yes, 521 not 512!)\")\n elif self.type ==\"ED25519\" and self.size != 128:\n raise HostkeyError(\"ED25519 keys have a fixed size, which cannot be altered.\") # can't really happen, size is ignored for ED25519\n\n # if privkey is already there check size\n self.key_exists = False\n self.key_current_size = 0\n if os.path.exists(self.fullpath):\n self.key_exists = True\n if self.type == \"ED25519\":\n self.curve = \"EC25519\"\n self.key_current_size = 128 # somewhat erbitrary, attack complexity on ED25519 is larger that brute forcing a 128bit key\n self.key_exists = True\n elif self.type == \"RSA1\":\n self.key_exists = True\n self.key_current_size = 1024\n else:\n try:\n with open(self.fullpath, \"rb\") as key_file:\n self.privkey = crypto_serialization.load_pem_private_key(key_file.read(), password=None, backend=crypto_default_backend())\n except IOError:\n raise HostkeyError(get_exception())\n\n if self.type == \"DSA\" or self.type == \"RSA\":\n self.key_current_size = self.privkey.key_size\n elif self.type == \"ED25519\":\n self.key_current_size = 128\n elif self.type == \"ECDSA\":\n self.pubkey = self.privkey.public_key()\n if self.pubkey.curve.name == \"secp256r1\":\n self.key_current_size = 256\n elif self.pubkey.curve.name == \"secp384r1\":\n self.key_current_size = 384\n elif self.pubkey.curve.name == \"secp521r1\":\n self.key_current_size = 521\n else:\n self.curve = self.pubkey.curve.name", "def self_check() -> None:\n assert len(ZBASE32_ALPHABET) == 32\n\n # Test vector from https://github.com/matusf/z-base-32/blob/0.1.2/src/lib.rs\n assert zbase32_encode(b\"asdasd\") == \"cf3seamuco\"\n assert zbase32_decode(\"cf3seamuco\") == b\"asdasd\"\n\n # Test vector from https://www.uriports.com/blog/setting-up-openpgp-web-key-directory/\n # assert zbase32_encode(hashlib.sha1(b\"yourmail\").digest()) == \"hacabazoakmnagxwmkjerb9yehuwehbm\"\n # -> this hash is wrong, and I don't know what username gives the SHA1\n # e61980e2f0c2962c19f45a928207e0472744702b\n\n # Test vector from https://metacode.biz/openpgp/web-key-directory\n assert zbase32_encode(hashlib.sha1(b\"test-wkd\").digest()) == \"4hg7tescnttreaouu4z1izeuuyibwww1\"\n\n # Test vector from https://datatracker.ietf.org/doc/draft-koch-openpgp-webkey-service/\n assert (\n get_wkd_advanced_url(\"[email protected]\")\n == \"https://openpgpkey.example.org/.well-known/openpgpkey/example.org/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe\" # noqa\n )\n assert (\n get_wkd_direct_url(\"[email protected]\")\n == \"https://example.org/.well-known/openpgpkey/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe\"\n )\n\n # Test vector from https://wiki.gnupg.org/WKD\n assert (\n get_wkd_direct_url(\"[email protected]\")\n == \"https://intevation.de/.well-known/openpgpkey/hu/it5sewh54rxz33fwmr8u6dy4bbz8itz4?l=bernhard.reiter\"\n )", "def verify_hack_key(self):\r\n\t\tself.percent_english = Dict_Control(self.my_code).check_key()\r\n\t\t#If more than half the words are english, the key will pass. \r\n\t\tif self.percent_english > 50:\r\n\t\t\tself.hack_plausible = True", "def check_auth_publickey(self, username, key):\n return AUTH_FAILED", "def _check_key(key): # type: (str) -> None\n if not key:\n raise ValueError('Key must not be empty.')\n if '.' in key:\n raise ValueError('Key must not contain dots.')", "def validate_key(self, key: keyType) -> bool:\n if isinstance(key, (dict,bool)):\n raise Exception\n if key is None:\n raise Exception\n # Numerical key object has no len(),\n # so explicitly specify which types are not allowed to use empty value as keys\n if isinstance(key, (str, tuple, set, list)) and (len(key) == 0):\n raise Exception\n return True", "def _validate_key(self, key):\n if isinstance(key, str):\n key = unicode(key, 'utf-8')\n elif not isinstance(key, unicode):\n raise TypeError(\n \"`key` must be `str` or `unicode`, not `{}`\".format(\n key.__class__.__name__)\n )\n return key", "def test_invalid_chars_ssck(self):\r\n valid_base = SlashSeparatedCourseKey(u'org.dept-1%2', u'course.sub-2%3', u'run.faster-4%5')\r\n for key in SlashSeparatedCourseKey.KEY_FIELDS:\r\n with self.assertRaises(InvalidKeyError):\r\n # this ends up calling the constructor where the legality check should occur\r\n valid_base.replace(**{key: u'funny thing'})", "def _is_valid_key(self, key):\r\n\r\n # Check the length\r\n if len(key) > 250:\r\n return False\r\n\r\n # Check that there are no spaces or control characters\r\n for char in key:\r\n if ord(char) < 33 or ord(char) == 127:\r\n return False\r\n\r\n return True", "def validate(self, key, val):\n return True", "def validate(self, key, val):\n return True", "def checkKey(self):\n # TO DO for checking API authentication\n if self.apikey is None:\n return False\n else:\n return True", "def test_validate_with_invalid_key_format_type(self):\n key_format_type = \"invalid\"\n kwargs = {'key_format_type': key_format_type}\n\n self.assertRaisesRegex(\n TypeError, \"invalid key format type\", Digest, **kwargs)", "def validate(self):\n if not self.key or not self.certificates:\n raise ValueError(\"Key or certificate missing in Keypair\")", "def test_key_none(self):\n try:\n AlphaVantage()\n self.fail(msg='A None api key must raise an error')\n except ValueError:\n self.assertTrue(True)", "def test_key_none(self):\n try:\n AlphaVantage()\n self.fail(msg='A None api key must raise an error')\n except ValueError:\n self.assertTrue(True)", "def check_len( string_key ) : \r\n\r\n if len( string_key ) != 4 :\r\n\r\n raise Eggog( \"'%s': EGI wants the key to be exactly four characters!\" % (string_key, ) ) \r\n \r\n else :\r\n \r\n return True", "def test_getKey_nokey(self):\n filename = os.path.join(os.getcwd(), 'sekrit')\n key = crypto.getKey(filename)\n self.failUnlessIsInstance(key, basestring,\n \"key isn't a string! type=%r\" % type(key))", "def is_key(v) -> bool:\n try:\n _validate(v, prefixes=[b\"edsk\", b\"edpk\", b\"spsk\", b\"p2sk\", b\"sppk\", b\"p2pk\"])\n except (ValueError, TypeError):\n return False\n return True", "def test_invalid_chars_location(self):\r\n course_key = SlashSeparatedCourseKey(u'org.dept-1%2', u'course.sub-2%3', u'run.faster-4%5')\r\n valid_base = course_key.make_usage_key('tomato-again%9', 'block-head:sub-4%9')\r\n for key in SlashSeparatedCourseKey.KEY_FIELDS:\r\n with self.assertRaises(InvalidKeyError):\r\n # this ends up calling the constructor where the legality check should occur\r\n valid_base.replace(**{key: u'funny thing'})", "def _validateKey(self, key, cls = None):\n\n key_class_types = [self._BaseKey__class, self._LocalKey__class,\n self._MsgKey__class, ErrorMsgManager]\n\n if cls:\n if inspect.isclass(cls) and cls in key_class_types:\n classes = [cls]\n else:\n return None\n else:\n classes = key_class_types\n return any([isinstance(key, cls) for cls in classes])", "def test_key_type(self):\n self.failureResultOf(self.producer.send_messages(\"topic\", key=\"key\", msgs=[b\"msg\"]), TypeError)", "def creation_validation(**_):\n\n for property_key in constants.KEYPAIR_REQUIRED_PROPERTIES:\n utils.validate_node_property(property_key, ctx.node.properties)\n\n key_file = _get_path_to_key_file()\n key_file_in_filesystem = _search_for_key_file(key_file)\n\n if ctx.node.properties['use_external_resource']:\n if not key_file_in_filesystem:\n raise NonRecoverableError(\n 'External resource, but the key file does not exist locally.')\n try:\n _get_key_pair_by_id(ctx.node.properties['resource_id'])\n except NonRecoverableError as e:\n raise NonRecoverableError(\n 'External resource, '\n 'but the key pair does not exist in the account: '\n '{0}'.format(str(e)))\n else:\n if key_file_in_filesystem:\n raise NonRecoverableError(\n 'Not external resource, '\n 'but the key file exists locally.')\n try:\n _get_key_pair_by_id(ctx.node.properties['resource_id'])\n except NonRecoverableError:\n pass\n else:\n raise NonRecoverableError(\n 'Not external resource, '\n 'but the key pair exists in the account.')", "def valid_key(self): \n self.so.ValidKey.restype = c_bool\n result = self.so.ValidKey()\n return result" ]
[ "0.66728324", "0.6390312", "0.6318158", "0.62363094", "0.6178258", "0.61090034", "0.59969664", "0.59606546", "0.59150714", "0.5909401", "0.5852684", "0.5826533", "0.5803395", "0.5795646", "0.5785924", "0.5733037", "0.5733037", "0.57124954", "0.57062685", "0.56519485", "0.563369", "0.563369", "0.5631203", "0.5601875", "0.55847937", "0.55493295", "0.5546983", "0.55395055", "0.55312544", "0.5512068" ]
0.8084664
0
Create Announcement & assign to memcache; used by memcache cron job & putAnnouncement().
def _cacheAnnouncement(): confs = Conference.query(ndb.AND( Conference.seatsAvailable <= 5, Conference.seatsAvailable > 0) ).fetch(projection=[Conference.name]) if confs: # If there are conferences close to being sold out, # format announcement and set it in memcache announcement = '%s %s' % ( 'Last chance to attend! The following conferences ' 'are nearly sold out:', ', '.join(conf.name for conf in confs)) memcache.set(MEMCACHE_ANNOUNCEMENTS_KEY, announcement) else: # If there are no sold out conferences, # delete the memcache announcements entry announcement = "" memcache.delete(MEMCACHE_ANNOUNCEMENTS_KEY) return announcement
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def getAnnouncement(self, request):\n announcement = memcache.get(MEMCACHE_ANNOUNCEMENTS_KEY) or \"\"\n return StringMessage(data=announcement)", "def cache_announcement():\n confs = Conference.query(ndb.AND(\n Conference.seatsAvailable <= 5, Conference.seatsAvailable > 0\n )).fetch(projection=[Conference.name])\n\n if confs:\n # If there are almost sold out conferences,\n # format announcement and set it in memcache\n announcement = ANNOUNCEMENT_TPL % (\n ', '.join(conf.name for conf in confs))\n memcache.set(MEMCACHE_ANNOUNCEMENTS_KEY, announcement)\n else:\n # If there are no sold out conferences,\n # delete the memcache announcements entry\n announcement = \"\"\n memcache.delete(MEMCACHE_ANNOUNCEMENTS_KEY)\n\n return announcement", "def _Announce(self):\n key = self._GetServerKey(self.peer_id)\n logging.debug('Encrypting announcement.')\n value = self._Encrypt('%s:%d' % (self.host, self.port))\n logging.debug('Posting announcement.')\n self._dht.Put(key, value)", "def get_announcement(self, request):\n return StringMessage(\n data=memcache.get(MEMCACHE_ANNOUNCEMENTS_KEY) or \"\")", "def add_announcement(self, cid, c_name, content, date):\n query = \"INSERT INTO announcement(cid, c_name, content, date) VALUES (%s, %s, %s, %s)\"\n inputs = (cid, c_name, content, date, )\n return self.database_manager.execute_query(query, inputs)", "async def announce(self, ctx, *, msg):\n if self._announce_msg is not None:\n await self.bot.say(\"Already announcing, wait until complete to\"\n \" issue a new announcement.\")\n else:\n self._announce_msg = msg", "def DORadvertise_announce(key, value, ttlval, timeout=None):\r\n\r\n post_params = {'command': 'announce', 'key': key, 'value': value,\r\n 'lifetime': str(int(ttlval))}\r\n\r\n _DORadvertise_command(post_params, timeout=timeout)\r\n\r\n return None", "def add_announcement():\n '''\n ann_id = db.Announcements.insert( name=request.vars.name )\n ann = db.post(ann_id)\n '''\n\n logger.info(\"api:add_announcement ===>>> category: %r\" % (request.vars.category))\n\n if request.vars.category not in ['urgent', 'event', \"shutdown\"]:\n return response.json({})\n\n else:\n\n vars = request.vars\n logger.info(\"vars.end_date: %r\" % (vars.end_date))\n ann_id = db.Announcements.insert(\n name = vars.name,\n latitude = vars.latitude,\n longitude = vars.longitude,\n description = vars.description,\n category = vars.category,\n end_date = vars.end_date\n )\n ann = db.Announcements(ann_id)\n\n logger.info(\"api:add_announcement ==> ann= %r\" % (ann))\n #logger.info(\"api:add_announcement_category ==> ann= %r\" % (ann.category))\n\n\n return response.json(ann)", "def announce(self, request_announce):\n return self.client.call('POST',\n self.name + 'announce',\n payload=request_announce)", "def announceGenerate(self):", "def add_comment_to_announcement():\n vars = request.vars\n logger.info(\"vars.comment_text: %r\" % (vars.comment_text))\n comment_id = db.Comments.insert(\n comment_text = vars.comment_text,\n score = 1,\n ann_id= vars.ann_id,\n )\n comment = db.Announcements(comment_id)\n\n logger.info(\"api:add_comment_to_announcement ==> comment= %r\" % (comment))\n\n return response.json(comment)", "async def announce(self, ctx):\n\n await ctx.message.delete()\n if len(ctx.message.content.split()) < 2:\n return await ctx.send('You must inform all parameters!')\n\n announce_channel = discord.utils.get(ctx.guild.channels, id=announcement_channel_id)\n msg = ctx.message.content.split('!announce', 1)\n await announce_channel.send(msg[1])", "async def announcement(ctx, bot: typing.Union[discord.Member, discord.User]):\n plonked = await is_plonked(ctx.author.id)\n if plonked:\n return\n \n data = await make_request(\"https://www.motiondevelopment.top/api/v1.2/bots/\", bot.id)\n print(data)\n announcement = data[\"annoucements\"]\n print(bool(announcement))\n e = discord.Embed(color=0xfecdea, title=f'Announcement ID: {announcement[\"post_id\"]}')\n if announcement != False:\n e.add_field(\n name=f'{announcement[\"post_title\"]}',\n value=announcement[\"post_body\"]\n )\n e.description = f\"Post created by {data['owner_name']} ({data['owner_id']})\"\n else: \n e.description = 'This bot doesn\\'t have an announcement. :cry:'\n\n await em(ctx, embed=e)", "def prepare_announce(self, request_announce):\n return self.client.call('POST',\n self.name + 'prepare-announce',\n payload=request_announce)", "def add(self, missing_dates, uncached_announcements):\n raise NotImplementedError('AbstractCache is an abstract base class')", "def announce(self):\n m = rtorrent9.rpc.Multicall(self)\n self.multicall_add(m, \"d.tracker_announce\")\n\n return m.call()[-1]", "def persist(self) -> None:\n\n values = {\n \"time\": self.event.time,\n \"priority\": self.event.priority,\n \"argument\": self.event.argument,\n \"kwargs\": self.event.kwargs\n }\n\n cherrypy.engine.publish(\n \"cache:set\",\n self.cache_key,\n values,\n self.time_remaining\n )", "def prep_advertiser(self):\n self.prep_consumer()\n self.advertiser = Advertiser.objects.get(id=113)\n self.consumer = Consumer.objects.get(id=113)\n self.consumer.consumer_create_datetime = self.three_days_ago\n self.consumer.email_subscription.add(1)\n self.consumer.save()", "def add_assignment(cls, mentor_id, title, start_date, end_date, file_name, group='0'):\n new = Assignment(mentor_id, title, start_date, end_date, file_name, group)\n db.session.add(new)\n db.session.commit()", "def v2centralizedadvertise_announce(key, value, ttlval):\r\n # do basic argument checking / munging\r\n key = str(key)\r\n value = str(value)\r\n\r\n if not type(ttlval) is int and not type(ttlval) is long:\r\n raise TypeError(\"Invalid type '\"+str(type(ttlval))+\"' for ttlval.\")\r\n\r\n if ttlval < 1:\r\n raise ValueError(\"The argument ttlval must be positive, not '\"+str(ttlval)+\"'\")\r\n\r\n \r\n # build the tuple to send, then convert to a string because only strings\r\n # (bytes) can be transmitted over the network...\r\n datatosend = ('PUT',key,value,ttlval)\r\n datastringtosend = serialize_serializedata(datatosend)\r\n\r\n \r\n # send the data over a timeout socket using the session library, then\r\n # get a response from the server.\r\n sockobj = timeout_openconn(v2servername,v2serverport, timeout=10)\r\n try:\r\n session_sendmessage(sockobj, datastringtosend)\r\n rawresponse = session_recvmessage(sockobj)\r\n finally:\r\n # BUG: This raises an error right now if the call times out ( #260 )\r\n # This isn't a big problem, but it is the \"wrong\" exception\r\n sockobj.close()\r\n \r\n # We should check that the response is 'OK'\r\n try:\r\n response = serialize_deserializedata(rawresponse)\r\n if response != 'OK':\r\n raise CentralAdvertiseError(\"Centralized announce failed with '\"+response+\"'\")\r\n except ValueError, e:\r\n raise CentralAdvertiseError(\"Received unknown response from server '\"+rawresponse+\"'\")", "def touch_member(self, data, ttl=None, permanent=False):", "def add(self, artifact, ttl):\n self.purge()\n \n target = ( isinstance(artifact, DNSArtifact)\n and str(artifact.remote_address) or artifact.name\n ).lower().rstrip('.')\n fqdns = [ name.lower().rstrip('.') for name in artifact.onames ]\n\n if target in self.index:\n self.index[target].update(fqdns)\n return\n \n association = Association(target, fqdns, ttl)\n self.index[target] = association\n (self.new_expiry or self.expiry).append(association)\n \n return", "def announce(self, discovery):\n try:\n url = self.__discovery_url()\n logger.debug(\"making announce request to %s\", url)\n response = None\n response = self.client.put(url,\n data=to_json(discovery),\n headers={\"Content-Type\": \"application/json\"},\n timeout=0.8)\n\n if response.status_code is 200:\n self.last_seen = datetime.now()\n except (requests.ConnectTimeout, requests.ConnectionError):\n logger.debug(\"announce\", exc_info=True)\n finally:\n return response", "def test_announcments_add_model(self):\n announcement = Announcement(title = \"Title\", description = \"Description\", date = datetime(2015, 6, 5, 8, 10, 10, 10))\n db.session.add(announcement)\n db.session.commit()\n self.assertEqual(Announcement.query.filter_by(title = \"Title\", description = \"Description\", date = datetime(2015, 6, 5, 8, 10, 10, 10)).count(), 1)", "def centralizedadvertise_announce(key, value, ttlval):\r\n # do basic argument checking / munging\r\n key = str(key)\r\n value = str(value)\r\n\r\n if not type(ttlval) is int and not type(ttlval) is long:\r\n raise TypeError(\"Invalid type '\"+str(type(ttlval))+\"' for ttlval.\")\r\n\r\n if ttlval < 1:\r\n raise ValueError(\"The argument ttlval must be positive, not '\"+str(ttlval)+\"'\")\r\n\r\n \r\n # build the tuple to send, then convert to a string because only strings\r\n # (bytes) can be transmitted over the network...\r\n datatosend = ('PUT',key,value,ttlval)\r\n datastringtosend = serialize_serializedata(datatosend)\r\n\r\n \r\n # send the data over a timeout socket using the session library, then\r\n # get a response from the server.\r\n sockobj = timeout_openconn(servername,serverport, timeout=10)\r\n try:\r\n session_sendmessage(sockobj, datastringtosend)\r\n rawresponse = session_recvmessage(sockobj)\r\n finally:\r\n # BUG: This raises an error right now if the call times out ( #260 )\r\n # This isn't a big problem, but it is the \"wrong\" exception\r\n sockobj.close()\r\n \r\n # We should check that the response is 'OK'\r\n try:\r\n response = serialize_deserializedata(rawresponse)\r\n if response != 'OK':\r\n raise CentralAdvertiseError(\"Centralized announce failed with '\"+response+\"'\")\r\n except ValueError, e:\r\n raise CentralAdvertiseError(\"Received unknown response from server '\"+rawresponse+\"'\")", "def arp_announce(self):\n pass", "def newTask(name, description, assigner, id=None, priority=None, submitter_email=None, whose=None):\n if whose:\n user_id = jutdaapi.find_user(whose)\n if not user_id:\n raise ValueError('bad whose assignment: '+str(whose))\n #title = name + ' for: '+assigner.title()\n # that was the old scheme\n title = '('+assigner.title()+') '+name\n\n if priority != None:\n #priority = (int(priority) + 2) / 2\n priority = int(priority)\n RA_queue = 3\n #if assigner != 'no one':\n # description += '<tasktrackermeta assigner=\"'+assigner+'\"/>'\n if isinstance(id, str):\n description += '<tasktrackermeta id=\"'+id+'\"/>'\n ticket_id = jutdaapi.create_ticket(RA_queue, title, description,\n priority=priority, submitter_email=submitter_email)\n # Is there a race condition here? In this kind of database\n # I would assume not.\n time.sleep(1)\n ticket = jutdaapi.get_detailed_ticket(ticket_id)\n t = ticketToTask(ticket)\n return t", "def mempool_assert_announcement(condition: ConditionWithArgs, announcements: Set[bytes32]) -> Optional[Err]:\n announcement_hash = bytes32(condition.vars[0])\n if announcement_hash not in announcements:\n return Err.ASSERT_ANNOUNCE_CONSUMED_FAILED\n\n return None", "def save(self, *args, **kwargs):\n super(News, self).save(*args, **kwargs)\n pigeonpost_queue.send(sender=self, defer_for=6*60*60)", "def obj_create(self, bundle, **kwargs):\n logger.info(\"Creating a new acknowledgement...\")\n #Create the object\n bundle.obj = Acknowledgement()\n #hydrate\n bundle = self.full_hydrate(bundle)\n \n #Set the customer\n try:\n logger.info(\"Setting customer...\")\n bundle.obj.customer = Customer.objects.get(pk=bundle.data[\"customer\"][\"id\"])\n bundle.obj.discount = bundle.obj.customer.discount\n except:\n logger.error(\"Customer with ID {0} could not be found.\".format(bundle.data['customer']['id']))\n raise\n \n #Set the employee\n try:\n logger.info(\"Setting employee...\")\n bundle.obj.employee = bundle.request.user\n except User.DoesNotExist:\n logger.error(\"User with ID {0} could not be found\".format(bundle.data['employee']['id']))\n raise\n except KeyError:\n logger.critical(\"Missing employee ID.\")\n raise\n \n #Set Status\n bundle.obj.status = \"ACKNOWLEDGED\"\n \n #Set the project or create a new one\n if \"project\" in bundle.data:\n try:\n project = Project.objects.get(pk=bundle.data['project']['id'])\n except KeyError, Project.DoesNotExist:\n try:\n project = Project()\n project.codename = bundle.data['project']['codename']\n project.save()\n except KeyError:\n project = None\n \n bundle.obj.project = project\n \n #Create items without saving them \n logger.info(\"Creating items...\")\n self.items = [Item.create(acknowledgement=bundle.obj,\n commit=False,\n **product) for product in bundle.data[\"items\"]]\n \n #Calculate the total price\n logger.info(\"Calculating balance of the order...\")\n bundle.obj.calculate_totals(self.items)\n bundle = self.save(bundle)\n \n #Save the items\n logger.info(\"Saving the items to the database...\")\n for item in self.items:\n item.acknowledgement = bundle.obj\n item.save()\n \n log_message = \"Ack {0} created on {1}. Schedule to be delivered on {1}\"\n log_message = log_message.format(bundle.obj.id,\n bundle.obj.time_created.strftime('%B %d, %Y'),\n bundle.obj.delivery_date.strftime('%B %d, %Y'))\n log = Log(message=log_message,\n delivery_date=bundle.obj.delivery_date,\n acknowledgement=bundle.obj)\n log.save()\n #Create and upload the pdfs to the \n #S3 system. The save the pdfs as\n #Attributes of the acknowledgement\n logger.info(\"Creating PDF documents...\")\n bundle.obj.create_and_upload_pdfs()\n \n \n #Add the url of the pdf to the outgoing data\n #only for when an acknowledgement is create\n try:\n ack = bundle.obj.acknowledgement_pdf\n production = bundle.obj.production_pdf\n bundle.data['pdf'] = {'acknowledgement': ack.generate_url(),\n 'production': production.generate_url()}\n except AttributeError: \n logger.warn('Missing acknowledgement or production pdf')\n \n #Conditionally email ack to Decoroom\n if \"decoroom\" in bundle.obj.customer.name.lower():\n try:\n logger.info(\"Emailing Decoroom Co., Ltd. the order details...\")\n bundle.obj.email_decoroom()\n except Exception as e:\n logger.error(\"Unable to mail decoroom.\")\n logger.error(e)\n \n \n \n logger.info(u\"Acknowledgement #{0} created for {1}\".format(bundle.obj.id, \n bundle.obj.customer.name)) \n return bundle" ]
[ "0.6307641", "0.624139", "0.61396885", "0.5956356", "0.5670804", "0.5596737", "0.5565057", "0.5479931", "0.5426172", "0.5361342", "0.5328143", "0.5242326", "0.51843905", "0.5149838", "0.50455034", "0.5040873", "0.50037575", "0.4943809", "0.49090248", "0.4906064", "0.4883745", "0.48719388", "0.48582053", "0.4849399", "0.48446873", "0.4843126", "0.4827799", "0.4795237", "0.4785876", "0.4783023" ]
0.6339458
0
Register or unregister user for selected conference.
def _conferenceRegistration(self, request, reg=True): retval = None # Get user profile prof = self._getProfileFromUser() # Check if conference given in the websafeConferenceKey exists wsck = request.websafeConferenceKey conf = _getEntityByWebsafeKey(wsck, 'Conference') # Register if reg: # Check if user already registered, otherwise add if wsck in prof.conferenceKeysToAttend: raise ConflictException( "You have already registered for this conference.") # Check if seats available if conf.seatsAvailable <= 0: raise ConflictException( "There are no seats available.") # Register user, deduct one seat prof.conferenceKeysToAttend.append(wsck) conf.seatsAvailable -= 1 retval = True # Unregister else: # Check if user already registered if wsck in prof.conferenceKeysToAttend: # Unregister user, add back one seat prof.conferenceKeysToAttend.remove(wsck) conf.seatsAvailable += 1 retval = True else: retval = False # Update the datastore and return prof.put() conf.put() return BooleanMessage(data=retval)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def conference_registration(self, request, reg=True):\n prof = self.profile_service.get_profile_from_user() # get user Profile\n\n # check if conf exists given websafeConfKey\n # get conference; check that it exists\n wsck = request.websafeConferenceKey\n conf = ndb.Key(urlsafe=wsck).get()\n if not conf:\n raise endpoints.NotFoundException(\n 'No conference found with key: %s' % wsck)\n\n # register\n if reg:\n # check if user already registered otherwise add\n if wsck in prof.conferenceKeysToAttend:\n raise ConflictException(\n \"You have already registered for this conference\")\n\n # check if seats avail\n if conf.seatsAvailable <= 0:\n raise ConflictException(\"There are no seats available.\")\n\n # register user, take away one seat\n prof.conferenceKeysToAttend.append(wsck)\n conf.seatsAvailable -= 1\n retval = True\n\n # unregister\n else:\n # check if user already registered\n if wsck in prof.conferenceKeysToAttend:\n\n # unregister user, add back one seat\n prof.conferenceKeysToAttend.remove(wsck)\n conf.seatsAvailable += 1\n retval = True\n else:\n retval = False\n\n # write things back to the datastore & return\n prof.put()\n conf.put()\n return BooleanMessage(data=retval)", "def register_for_conference(self, request):\n return self.conference_registration(request)", "def registerForConference(self, request):\n return self._conferenceRegistration(request)", "def unregister_from_conference(self, request):\n return self.conference_registration(request, reg=False)", "def unregisterFromConference(self, request):\n return self._conferenceRegistration(request, reg=False)", "def register_user():\n pass", "def set_user_register(self, *args, **kwargs):\n return _uhd_swig.usrp_sink_set_user_register(self, *args, **kwargs)", "def register(self, voice=False):\n payload = {\"type\": \"register\", \"username\": self.username, \"voice\": voice}\n self._send_command(payload)", "def set_user_register(self, *args, **kwargs):\n return _uhd_swig.usrp_sink_sptr_set_user_register(self, *args, **kwargs)", "def event_register_participant(request, event_id, participant_id):\n if request.method == 'PUT':\n event = get_object_or_404(Event, pk=event_id, is_registration_open=True)\n participant = get_object_or_404(Participant, pk=participant_id)\n try:\n attendance = Attendance(participant=participant, event=event, datetime_register=datetime.now())\n attendance.save()\n except Exception as e:\n print e\n content = {'detail': config.PARTICIPANT_ALREADY_REGISTERED_TO_EVENT}\n return Response(content, status=status.HTTP_400_BAD_REQUEST)\n serializer = AttendanceSerializer(attendance)\n return Response(serializer.data, status=status.HTTP_202_ACCEPTED)", "async def register(ctx, *args):\n user = ctx.message.author\n user_mention = ctx.author.mention\n chan_mention = \"<#876850365730021386>\"\n \n if user in self.data[\"users.json\"]:\n await ctx.message.channel.send(user_mention+\", you are already registered. :blue_heart:\")\n else:\n self.data[\"users_asked_to_be_registered.json\"].append(user)\n await ctx.message.channel.send(user_mention+\", do you accept the \"+chan_mention+\n \" (Indie Library Terms of Service). Command .accept if you do. :blue_heart:\")", "def register(self, form):\n new_user = self.create_inactive_user(form)\n signals.user_registered.send(\n sender=self.__class__, user=new_user, request=self.request\n )\n return new_user", "def toggle_subscription(self):\n user = self.context['request'].user\n # pylint: disable=no-member\n profile = UserProfile.objects.get(\n user=user)\n club = self.context['club']\n\n if club in profile.subscriptions.all():\n club.subscribed_users.remove(profile)\n else:\n club.subscribed_users.add(profile)", "def create_conference_session(self, request):\n user = endpoints.get_current_user()\n\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n\n self.session_service.create_conference_session(request, user)\n return request", "def set_user_register(self, *args, **kwargs):\n return _uhd_swig.usrp_source_set_user_register(self, *args, **kwargs)", "def _register_user(request_form):\n idnr = request_form['idnr']\n\n if user_exists(idnr):\n raise UserAlreadyExistsError(idnr)\n\n response = elster_client.send_unlock_code_request_with_elster(request_form, request.remote_addr)\n request_id = escape(response['elster_request_id'])\n\n create_user(idnr, request_form['dob'].strftime(\"%d.%m.%Y\"), request_id)", "def register(session: Session, email: str, password: str, given_name: str, last_name: str,\n phone: str) -> RegisterResult:\n # Validate the data integrity of the parameters\n if email is None:\n return RegisterResult.BAD_USERNAME\n if password is None or not passwordService.validate(password):\n return RegisterResult.BAD_PASSWORD\n\n # Check to see if the user already exists\n existing_user = session.query(User) \\\n .filter(User.email == email) \\\n .first()\n if existing_user is not None:\n return RegisterResult.USERNAME_ALREADY_REGISTERED\n\n # Everything seems fine, so we go ahead and create the user & the linked account.\n password_hash = passwordService.hash(password)\n new_user = User(role=UserType.VOLUNTEER, password=password_hash, first_name=given_name, last_name=last_name,\n mobile_number=phone, email=email, preferred_hours={}, experience_years=0, possibleRoles=[\"Basic\"],\n qualifications=[],\n availabilities={\"Friday\": [], \"Monday\": [], \"Sunday\": [], \"Tuesday\": [], \"Saturday\": [],\n \"Thursday\": [], \"Wednesday\": []})\n session.add(new_user)\n session.flush()\n return RegisterResult.SUCCESS", "async def register(websocket):\n app['websockets'].add(websocket)\n await notify_users()", "def on_signup(self, data):\n self.participant_id = data[\"participant\"][\"id\"]", "def add_user(params, batch_registration=False):\n # Make sure the username is unique\n db = api.db.get_conn()\n if is_blacklisted_username(params[\"username\"]) or db.users.find_one(\n {\"username\": params[\"username\"]},\n collation=Collation(locale=\"en\", strength=CollationStrength.PRIMARY),\n ):\n raise PicoException(\"There is already a user with this username.\", 409)\n if db.teams.find_one(\n {\"team_name\": params[\"username\"]},\n collation=Collation(locale=\"en\", strength=CollationStrength.PRIMARY),\n ):\n raise PicoException(\"There is already a team with this username.\", 409)\n\n # If gid is specified, force affiliation to that team's name\n email_whitelist = None\n if params.get(\"gid\", None):\n group = api.group.get_group(gid=params[\"gid\"])\n group_settings = api.group.get_group_settings(gid=group[\"gid\"])\n params[\"affiliation\"] = group[\"name\"]\n email_whitelist = group_settings[\"email_filter\"]\n\n # If rid is specified and gid and email match,\n # get teacher status from registration token.\n # Additionally, invited users are automatically validated.\n user_is_teacher = params[\"usertype\"] == \"teacher\"\n user_was_invited = False\n join_group_as_teacher = False\n if params.get(\"rid\", None):\n key = api.token.find_key_by_token(\"registration_token\", params[\"rid\"])\n if params.get(\"gid\") != key[\"gid\"]:\n raise PicoException(\n \"Registration token group and supplied gid do not match.\"\n )\n if params[\"email\"] != key[\"email\"]:\n raise PicoException(\n \"Registration token email does not match the supplied one.\"\n )\n join_group_as_teacher = key[\"teacher\"]\n user_was_invited = True\n api.token.delete_token(key, \"registration_token\")\n\n # If not invited, validate the user's email against the whitelist\n else:\n if not verify_email_in_whitelist(params[\"email\"], email_whitelist):\n raise PicoException(\n \"Your email does not belong to the whitelist. \"\n + \"Please see the registration form for details.\"\n )\n\n # If CAPTCHAs are enabled, validate the submission if not batch registration\n if (\n api.config.get_settings()[\"captcha\"][\"enable_captcha\"]\n and not batch_registration\n and not _validate_captcha(params)\n ):\n raise PicoException(\"Incorrect captcha!\")\n\n # Create a team for the new user and set its count to 1\n tid = api.team.create_team(\n {\n \"team_name\": params[\"username\"],\n \"password\": api.common.hash_password(\"-\"),\n \"affiliation\": params[\"affiliation\"],\n }\n )\n db.teams.update_one({\"tid\": tid}, {\"$set\": {\"size\": 1}})\n\n # The first registered user automatically becomes an admin\n user_is_admin = False\n if db.users.count() == 0:\n user_is_admin = True\n user_is_teacher = True\n\n # Insert the new user in the DB\n uid = api.common.token()\n settings = api.config.get_settings()\n user = {\n \"uid\": uid,\n \"firstname\": params[\"firstname\"],\n \"lastname\": params[\"lastname\"],\n \"username\": params[\"username\"],\n \"email\": params[\"email\"],\n \"password_hash\": api.common.hash_password(params[\"password\"]),\n \"tid\": tid,\n \"usertype\": params[\"usertype\"],\n \"country\": params[\"country\"],\n \"demo\": params[\"demo\"],\n \"teacher\": user_is_teacher,\n \"admin\": user_is_admin,\n \"disabled\": False,\n \"verified\": (not settings[\"email\"][\"email_verification\"] or user_was_invited),\n \"extdata\": {},\n \"completed_minigames\": [],\n \"unlocked_walkthroughs\": [],\n \"tokens\": 0,\n }\n db.users.insert_one(user)\n\n # Determine the user team's initial eligibilities\n initial_eligibilities = [\n scoreboard[\"sid\"]\n for scoreboard in api.scoreboards.get_all_scoreboards()\n if api.scoreboards.is_eligible(user, scoreboard)\n ]\n db.teams.find_one_and_update(\n {\"tid\": tid}, {\"$set\": {\"eligibilities\": initial_eligibilities}}\n )\n\n # If gid was specified, add the newly created team to the group\n if params.get(\"gid\", None):\n api.group.join_group(params[\"gid\"], tid, teacher=join_group_as_teacher)\n\n # If email verification is enabled and user wasn't invited, send\n # validation email\n if settings[\"email\"][\"email_verification\"] and not user_was_invited:\n api.email.send_user_verification_email(params[\"username\"])\n\n return uid", "async def accept(ctx, *args):\n user = ctx.message.author\n user_mention = \"<@\"+str(user.id)+\">\"\n\n if user in self.data[\"users_asked_to_be_registered.json\"]:\n self.data[\"users.json\"].append(user)\n self.data[\"users_asked_to_be_registered.json\"].remove(user)\n await ctx.message.channel.send(user_mention+\", you have been successfully registered. :blue_heart:\")\n else:\n await ctx.message.channel.send(user_mention+\", have not commanded .register yet. \"\n \"Please do so first. :blue_heart:\")", "def preregister_user(username, full_name):\n registration_hash = secrets.token_hex(32)\n with get_conn().cursor() as cur:\n cur.execute(\"\"\"\n INSERT INTO unregistered_users (\n username, full_name, registration_hash\n ) VALUES (\n '%s', '%s', '%s'\n )\"\"\" % (username, full_name, registration_hash,))\n return True", "def on_user_signed_up(sender, request: HttpRequest, user: User, **_):\n Event.new(EventAction.SIGN_UP).from_http(request)", "def update_conference(self, request):\n return self.conference_service.update_conference_object(request)", "def register():\r\n form = RegisterForm(request.form)\r\n\r\n if request.method == 'POST' and form.validate():\r\n new_user = User(form.email.data, form.password.data)\r\n g.session.add(new_user)\r\n g.session.commit()\r\n\r\n new_profile = Profile(form.first_name.data, form.last_name.data, new_user.id)\r\n g.session.add(new_profile)\r\n g.session.commit()\r\n # TODO: make it async\r\n if current_app.config[\"REQUIRE_EMAIL_CONFIRMATION\"]:\r\n send_confirmation(new_user)\r\n new_user.init_folders()\r\n logout_user()\r\n return redirect(url_for(\".login\"))\r\n return render_template(\"account/register_user.pug\", form=form)", "def join_conference(request, conference_name):\n resp = VoiceResponse()\n resp.dial(hangupOnStar=True).conference(conference_name)\n return HttpResponse(str(resp))", "def register(user_name: str, password: str, age=20):\n auth.register(user_name, password)\n user_handler.register(user_name, age)", "async def unregister(websocket):\n USERS.remove(websocket)\n GAME.new_game()\n await update_players()", "def tournament_registration(request, tournament_id):\n tournament = get_object_or_404(Tournament, id=tournament_id)\n \n if tournament.is_online():\n form = OnlineTournamentRegistrationForm(request.POST, initial={\n 'tournament': tournament\n })\n else:\n form = TournamentRegistrationForm(request.POST, initial={\n 'tournament': tournament\n })\n\n if form.is_valid():\n if tournament.is_online():\n tenhou_nickname = form.cleaned_data.get('tenhou_nickname')\n if OnlineTournamentRegistration.objects.filter(tournament=tournament, tenhou_nickname=tenhou_nickname).exists():\n messages.success(request, _('You already registered to the tournament!'))\n return redirect(tournament.get_url())\n\n instance = form.save(commit=False)\n instance.tournament = tournament\n \n # it supports auto load objects only for Russian tournaments right now\n \n try:\n instance.player = Player.objects.get(first_name_ru=instance.first_name.title(), \n last_name_ru=instance.last_name.title())\n except (Player.DoesNotExist, Player.MultipleObjectsReturned):\n # TODO if multiple players are here, let's try to filter by city\n pass\n \n try:\n instance.city_object = City.objects.get(name_ru=instance.city)\n except City.DoesNotExist:\n pass\n\n if tournament.registrations_pre_moderation:\n instance.is_approved = False\n message = _('Your registration was accepted! It will be visible on the page after administrator approvement.')\n else:\n instance.is_approved = True\n message = _('Your registration was accepted!')\n\n instance.save()\n \n messages.success(request, message)\n else:\n messages.success(request, _('Please, allow to store personal data'))\n\n return redirect(tournament.get_url())", "def set_user_register(self, *args, **kwargs):\n return _uhd_swig.usrp_source_sptr_set_user_register(self, *args, **kwargs)" ]
[ "0.6628083", "0.64813596", "0.6294456", "0.6073893", "0.58963567", "0.5738126", "0.55700624", "0.55217904", "0.54235077", "0.541907", "0.5362311", "0.53576905", "0.5338589", "0.52861226", "0.5157034", "0.51566017", "0.5153176", "0.5135656", "0.5130247", "0.5127976", "0.5102711", "0.5102175", "0.5093956", "0.5064531", "0.506041", "0.50293416", "0.50259125", "0.5017605", "0.5016001", "0.5014886" ]
0.67306554
0
Copy relevant fields from Conference to ConferenceForm.
def _copyConferenceToForm(self, conf, displayName): cf = ConferenceForm() for field in cf.all_fields(): if hasattr(conf, field.name): # Convert Date to date string; just copy others if field.name.endswith('Date'): setattr(cf, field.name, str(getattr(conf, field.name))) else: setattr(cf, field.name, getattr(conf, field.name)) elif field.name == "websafeKey": setattr(cf, field.name, conf.key.urlsafe()) if displayName: setattr(cf, 'organizerDisplayName', displayName) cf.check_initialized() return cf
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _copyConferenceToForm(self, conf, displayName):\n cf = ConferenceForm()\n for field in cf.all_fields():\n if hasattr(conf, field.name):\n # convert Date to date string; just copy others\n if field.name.endswith('Date'):\n setattr(cf, field.name, str(getattr(conf, field.name)))\n else:\n setattr(cf, field.name, getattr(conf, field.name))\n elif field.name == \"websafeKey\":\n setattr(cf, field.name, conf.key.urlsafe())\n if displayName:\n setattr(cf, 'organizerDisplayName', displayName)\n cf.check_initialized()\n return cf", "def _copySpeakerToForm(self, speaker):\n sf = SpeakerForm()\n for field in sf.all_fields():\n if hasattr(speaker, field.name):\n setattr(sf, field.name, getattr(speaker, field.name))\n elif field.name == \"websafeKey\":\n setattr(sf, field.name, speaker.key.urlsafe())\n sf.check_initialized()\n return sf", "def update_c_mandatory_fields(request, campaign_id):\n # print(request.POST)\n campaign = Campaign.objects.get(id=campaign_id)\n form = CampaignForm(request.POST, instance = campaign)\n # print(form)\n if form.is_valid():\n form.save()\n return redirect('add_campaign_spec', id=campaign_id)\n else:\n # return redirect('clonecampaign', id=campaign_id)\n print(form.errors)\n return redirect(reverse('edit_campaign', kwargs={'campaign_id':campaign_id}))", "def clone(self):\n\n # recrea la instancia del form\n form_clone = FormularioDinamico(\n lang=self.lang,\n name=\"clon de %s\" % self.name,\n is_active=self.is_active,\n form_title=self.form_title,\n form_template=self.form_template,\n send_email=self.send_email,\n from_email=self.from_email,\n recipient_list=self.recipient_list,\n subject_template=self.subject_template,\n body_template=self.body_template,\n error_class=self.error_class,\n required_css_class=self.required_css_class,\n autorespond=self.autorespond,\n autorespond_subject_template=self.autorespond_subject_template,\n autorespond_body_template=self.autorespond_body_template,\n autorespond_email_field=self.autorespond_email_field\n )\n\n form_clone.save()\n\n content_type = ContentType.objects.get_for_model(form_clone)\n\n # recrea todos los fields\n for field in self.get_fields():\n field_clone = DynaFormField(\n content_type=content_type,\n object_pk=form_clone.pk,\n field_name=field.field_name,\n field_label=field.field_label,\n field_type=field.field_type,\n field_widget=field.field_widget,\n field_help=field.field_help,\n is_required=field.is_required,\n is_hidden=field.is_hidden,\n default_value=field.default_value,\n choices=field.choices,\n choices_delimiter=field.choices_delimiter,\n choices_queryset=field.choices_queryset,\n choices_queryset_filter=field.choices_queryset_filter,\n choices_queryset_empty_label=field.choices_queryset_empty_label,\n choices_queryset_label=field.choices_queryset_label,\n choices_related_field=field.choices_related_field,\n field_order=field.field_order\n )\n\n field_clone.save()", "def prepare(self, form):\n \n return form", "def _copySessionToForm(self, session):\n sf = SessionForm()\n for field in sf.all_fields():\n if hasattr(session, field.name):\n # Convert date field to date string\n if field.name == 'date':\n setattr(sf, field.name, str(getattr(session, field.name)))\n # Convert time field to time string\n elif field.name == 'startTime':\n setattr(sf, field.name,\n getattr(session, field.name).strftime('%H:%M'))\n # Convert typeOfSession string field to enum\n elif field.name == 'typeOfSession':\n setattr(sf, field.name,\n getattr(SessionType, getattr(session, field.name)))\n # Copy other fields verbatim\n else:\n setattr(sf, field.name, getattr(session, field.name))\n # Ensure that the SessionForm contains websafeKey\n elif field.name == \"websafeKey\":\n setattr(sf, field.name, session.key.urlsafe())\n sf.check_initialized()\n return sf", "def getConference(self, request):\n # Get Conference object from request; bail if not found\n conf = _getEntityByWebsafeKey(request.websafeConferenceKey,\n 'Conference')\n prof = conf.key.parent().get()\n # Return ConferenceForm\n return self._copyConferenceToForm(conf, getattr(prof, 'displayName'))", "def get_conference(self, request):\n # get Conference object from request; bail if not found\n conf = ndb.Key(urlsafe=request.websafeConferenceKey).get()\n if not conf:\n raise endpoints.NotFoundException(\n 'No conference found with key: %s' %\n request.websafeConferenceKey)\n prof = conf.key.parent().get()\n # return ConferenceForm\n return self.conference_service.copy_conference_to_form(ConferenceForm(),\n conf,\n prof.displayName)", "def copyform(form, settings):\r\n for name, value in form.iteritems():\r\n setattr(settings, name, value)\r\n settings.commit()", "def get_conferences_to_attend(self, request):\n prof = self.profile_service.get_profile_from_user() # get user Profile\n conf_keys = [ndb.Key(urlsafe=wsck) for wsck in\n prof.conferenceKeysToAttend]\n conferences = ndb.get_multi(conf_keys)\n\n # get organizers\n organisers = [ndb.Key(Profile, conf.organizerUserId) for conf in\n conferences]\n profiles = ndb.get_multi(organisers)\n\n # put display names in a dict for easier fetching\n names = {}\n for profile in profiles:\n names[profile.key.id()] = profile.displayName\n\n # return set of ConferenceForm objects per Conference\n return ConferenceForms(items=[\n self.conference_service.copy_conference_to_form(\n ConferenceForm(), conf, names[conf.organizerUserId]\n ) for conf in conferences])", "def _replace_fields(self):\n for name, value in self._cleaned_data.items():\n setattr(self, name, value)", "def copyCommonFields(self):\n self.fetchDataToForm(self.selected_row, self.selected_column, fields = \"Recent\")", "def copy_form(self, from_module, form, to_module, rename=False):\n copy_source = deepcopy(form.to_json())\n # only one form can be a release notes form, so set them to False explicitly when copying\n copy_source['is_release_notes_form'] = False\n copy_source['enable_release_notes'] = False\n if 'unique_id' in copy_source:\n del copy_source['unique_id']\n\n if rename:\n for lang, name in copy_source['name'].items():\n with override(lang):\n copy_source['name'][lang] = _('Copy of {name}').format(name=name)\n\n copy_form = to_module.add_insert_form(from_module, FormBase.wrap(copy_source))\n to_app = to_module.get_app()\n save_xform(to_app, copy_form, form.source.encode('utf-8'))\n\n return copy_form", "def get_conferences_created(self, request):\n # make sure user is authed\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = self.auth.get_user_id(user)\n\n # create ancestor query for all key matches for this user\n confs = Conference.query(ancestor=ndb.Key(Profile, user_id))\n prof = ndb.Key(Profile, user_id).get()\n # return set of ConferenceForm objects per Conference\n return ConferenceForms(items=[\n self.conference_service.copy_conference_to_form(ConferenceForm(),\n conf,\n prof.displayName)\n for conf in confs])", "def __init__(self, corp_app, field_objs, *args, **kwargs):\n self.corp_app = corp_app\n self.field_objs = field_objs\n super(CorpMembForm, self).__init__(*args, **kwargs)\n\n for field in field_objs:\n if field.field_type not in ['section_break', 'page_break']:\n if field.field_name:\n field_key = field.field_name\n else:\n field_key = \"field_%s\" % field.id\n\n # if field is display only, remove it from the form\n # for example, on the edit page, we\n # set corporate_membership_type\n # and payment_method as the display only fields\n if hasattr(field, 'display_only') and field.display_only:\n del self.fields[field_key]\n else:\n # get field class and set field initial\n self.fields[field_key] = field.get_field_class()\n if ((not field.field_name) \\\n or field.field_name == 'authorized_domains') \\\n and self.instance:\n initial = field.get_value(self.instance)\n if field.field_type in [\n 'MultipleChoiceField/django.forms.CheckboxSelectMultiple',\n 'MultipleChoiceField']:\n if initial:\n self.fields[field_key].initial = [\n item.strip() for item in initial.split(',')]\n else:\n self.fields[field_key].initial = initial\n\n #self.fields['captcha'] = CaptchaField(label=_('Type the code below'))", "def _copyProfileToForm(self, prof):\n pf = ProfileForm()\n for field in pf.all_fields():\n if hasattr(prof, field.name):\n # Convert t-shirt string to Enum; just copy others\n if field.name == 'teeShirtSize':\n setattr(pf, field.name,\n getattr(TeeShirtSize, getattr(prof, field.name)))\n else:\n setattr(pf, field.name, getattr(prof, field.name))\n pf.check_initialized()\n return pf", "def set_up_fields(self, fields):\n self.fields = {\n 'name': self.name,\n 'email': self.email\n }\n for key in fields.keys():\n # special keys first, not to be used in the template\n if key.upper() == 'CC':\n self.is_cc = fields[key]\n elif key.upper() == 'BCC':\n self.is_bcc = fields[key]\n else:\n self.fields[key] = fields[key]", "def _createConferenceObject(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = user.email()\n if not request.name:\n raise endpoints.BadRequestException(\n \"Conference 'name' field required\")\n # Copy ConferenceForm/ProtoRPC Message into dict\n data = {\n field.name: getattr(request, field.name) for field in\n request.all_fields()\n }\n del data['websafeKey']\n del data['organizerDisplayName']\n # Add default values for those missing (both data model and\n # outbound Message)\n for df in CONF_DEFAULTS:\n if data[df] in (None, []):\n data[df] = CONF_DEFAULTS[df]\n setattr(request, df, CONF_DEFAULTS[df])\n # Convert dates from strings to Date objects; set month based\n # on start_date\n if data['startDate']:\n data['startDate'] = datetime.strptime(\n data['startDate'][:10], \"%Y-%m-%d\").date()\n data['month'] = data['startDate'].month\n else:\n data['month'] = 0\n if data['endDate']:\n data['endDate'] = datetime.strptime(\n data['endDate'][:10], \"%Y-%m-%d\").date()\n # Set seatsAvailable to be same as maxAttendees on creation\n if data[\"maxAttendees\"] > 0:\n data[\"seatsAvailable\"] = data[\"maxAttendees\"]\n # Get the user profile key, then set the conference's parent\n # to that value.\n # NOTE: The original code made a call to allocate_ids in order to\n # generate an ID for the conference. Since the profiles utilize\n # strings (email addresses) for their IDs, resulting in no risk\n # of colliding with NDB's auto-generated numeric IDs, I decided\n # to let NDB generate the conference ID automatically.\n # https://cloud.google.com/appengine/docs/python/ndb/entities?hl=en#numeric_keys\n p_key = ndb.Key(Profile, user_id)\n data['parent'] = p_key\n data['organizerUserId'] = request.organizerUserId = user_id\n # Create Conference, send email to organizer confirming\n # creation of Conference and return (modified) ConferenceForm\n Conference(**data).put()\n taskqueue.add(params={'email': user.email(),\n 'conferenceInfo': repr(request)},\n url='/tasks/send_confirmation_email'\n )\n return request", "def individual_formfields():\n # Instantiate Consent Tracker\n consent = s3db.auth_Consent(processing_types = VOL_CONSENT_OPTIONS)\n\n formfields = [utable.first_name,\n utable.last_name,\n Field(\"addr_L3\",\n label = T(\"Location\"),\n requires = IS_IN_SET(districts_and_uk),\n ),\n Field(\"addr_street\",\n label = T(\"Street Address\"),\n ),\n Field(\"addr_postcode\",\n label = T(\"Postcode\"),\n ),\n Field(\"mobile\",\n label = T(\"Contact Number (Preferred)\"),\n requires = IS_PHONE_NUMBER_MULTI(),\n comment = DIV(_class = \"tooltip\",\n _title = \"%s|%s\" % (T(\"Contact Number (Preferred)\"),\n T(\"Ideally a Mobile Number, so that we can send you Text Messages.\")),\n ),\n ),\n Field(\"home\",\n label = T(\"Contact Number (Secondary)\"),\n requires = IS_EMPTY_OR(IS_PHONE_NUMBER_MULTI()),\n ),\n utable.email,\n utable[passfield],\n # Password Verification Field\n Field(\"password_two\", \"password\",\n label = auth_messages.verify_password,\n requires = IS_EXPR(\"value==%s\" % \\\n repr(request.vars.get(passfield)),\n error_message = auth_messages.mismatched_password,\n ),\n ),\n\n # Skills\n s3db.hrm_multi_skill_id(empty = False,\n label = T(\"Volunteer Offer\"),\n ),\n Field(\"skills_details\",\n label = T(\"Please specify details\"),\n ),\n Field(\"certificates\", \"list:string\",\n label = T(\"Qualifications\"),\n requires = IS_IN_SET(certificates, multiple=True),\n widget = S3MultiSelectWidget(header=\"\",\n selectedList=3),\n ),\n Field(\"experience\",\n label = T(\"Skills and Experience\"),\n widget = lambda f, v: \\\n s3_comments_widget(f, v, _placeholder = \"e.g. Co-ordination, Event Management, PCV qualified.\")\n ),\n Field(\"resources\",\n label = T(\"Offers of Resources\"),\n widget = lambda f, v: \\\n s3_comments_widget(f, v, _placeholder = \"e.g. Minibus.\")\n ),\n Field(\"where_operate\", \"list:string\",\n label = T(\"Where would you be willing to volunteer?\"),\n requires = IS_IN_SET(districts, multiple=True),\n widget = S3MultiSelectWidget(header=\"\",\n selectedList=3),\n ),\n Field(\"travel\", \"integer\",\n label = T(\"Willing to Travel?\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"slots\", \"list:string\",\n label = T(\"Times\"),\n requires = IS_IN_SET(slots, multiple=True),\n widget = S3MultiSelectWidget(header=\"\",\n selectedList=3),\n ),\n Field(\"significant_physical\", \"integer\",\n label = T(\"That require significant physical activity (including lifting and carrying) and may involve being outdoors (e.g. clean up of affected properties)\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"some_physical\", \"integer\",\n label = T(\"That require some physical activity and may involve being outdoors (e.g. door knocking)\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"little_physical\", \"integer\",\n label = T(\"That require little physical activity and are based indoors (e.g. preparing refreshments)\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"health_details\",\n label = T(\"If you wish, you can give us some further information on any fitness, medical or mobility issues that might limit the kind of activities you are able to volunteer for; this will help us to suggest suitable opportunities for you\"),\n ),\n Field(\"faith_requirements\", \"integer\",\n label = T(\"Do you have any faith requirements that you would like help with if you are coming to Support Cumbria?\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"faith_requirements_details\",\n label = T(\"If Yes please outline\"),\n ),\n Field(\"emergency_contact_name\",\n label = T(\"Contact Name\"),\n requires = IS_NOT_EMPTY(),\n ),\n Field(\"emergency_contact_number\",\n label = T(\"Contact Number\"),\n requires = IS_PHONE_NUMBER_MULTI(),\n ),\n Field(\"emergency_contact_relationship\",\n label = T(\"Relationship\"),\n requires = IS_NOT_EMPTY(),\n ),\n Field(\"workplace\", \"integer\",\n label = T(\"Are you volunteering under your workplace volunteering scheme?\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"workplace_details\",\n label = T(\"If yes please name your employer\"),\n ),\n Field(\"dbs\", \"integer\",\n label = T(\"Are you DBS checked?\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n #Field(\"convictions\", \"integer\",\n # label = T(\"Do you have any unspent convictions?\"),\n # comment = T(\"Please tick 'Yes' if you have any convictions that are not yet spent under the Rehabilitation of Offenders Act 1974. The term 'convictions' is used to refer to any sentence or disposal issued by a court. If all your convictions are spent, you can tick 'No'. If you're not sure if your convictions are unspent or spent, you can use a tool available at www.disclosurecalculator.org.uk and read guidance at hub.unlock.org.uk/roa\"),\n # requires = IS_IN_SET({0: T(\"No\"),\n # 1: T(\"Yes\"),\n # }),\n # widget = lambda f, v: \\\n # SQLFORM.widgets.radio.widget(f, v,\n # style=\"divs\"),\n # ),\n # Consent (GDPR + FOC)\n Field(\"consent\",\n label = T(\"Consent\"),\n widget = consent.widget,\n ),\n ]\n\n required_fields = [\"first_name\",\n \"last_name\",\n \"addr_L3\",\n \"addr_street\",\n \"addr_postcode\",\n \"mobile\",\n \"emergency_contact\",\n \"where_operate\",\n ]\n\n return formfields, required_fields", "def _createConferenceObject(self, request):\n # preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = getUserId(user)\n\n if not request.name:\n raise endpoints.BadRequestException(\"Conference 'name' field required\")\n\n # copy ConferenceForm/ProtoRPC Message into dict\n data = {field.name: getattr(request, field.name) for field in request.all_fields()}\n del data['websafeKey']\n del data['organizerDisplayName']\n\n # add default values for those missing (both data model & outbound Message)\n for df in DEFAULTS:\n if data[df] in (None, []):\n data[df] = DEFAULTS[df]\n setattr(request, df, DEFAULTS[df])\n\n # convert dates from strings to Date objects; set month based on start_date\n if data['startDate']:\n data['startDate'] = datetime.strptime(data['startDate'][:10], \"%Y-%m-%d\").date()\n data['month'] = data['startDate'].month\n else:\n data['month'] = 0\n if data['endDate']:\n data['endDate'] = datetime.strptime(data['endDate'][:10], \"%Y-%m-%d\").date()\n\n # set seatsAvailable to be same as maxAttendees on creation\n if data[\"maxAttendees\"] > 0:\n data[\"seatsAvailable\"] = data[\"maxAttendees\"]\n # generate Profile Key based on user ID and Conference\n # ID based on Profile key get Conference key from ID\n p_key = ndb.Key(Profile, user_id)\n c_id = Conference.allocate_ids(size=1, parent=p_key)[0]\n c_key = ndb.Key(Conference, c_id, parent=p_key)\n data['key'] = c_key\n data['organizerUserId'] = request.organizerUserId = user_id\n\n # create Conference, send email to organizer confirming\n # creation of Conference & return (modified) ConferenceForm\n Conference(**data).put()\n taskqueue.add(params={'email': user.email(),\n 'conferenceInfo': repr(request)},\n url='/tasks/send_confirmation_email'\n )\n return request", "def getConference(self, request):\n # get Conference object from request; bail if not found\n conf = ndb.Key(urlsafe=request.websafeConferenceKey).get()\n if not conf:\n raise endpoints.NotFoundException(\n 'No conference found with key: %s' % request.websafeConferenceKey)\n prof = conf.key.parent().get()\n # return ConferenceForm\n return self._copyConferenceToForm(conf, getattr(prof, 'displayName'))", "def get_form(self, form_class):\n form = super(InvitationCreateView, self).get_form(form_class)\n form.fields[\"receiver\"].queryset = self.get_available_invitees()\n form.fields[\"team\"].queryset = self.get_available_teams()\n form.fields[\"team\"].empty_label = None\n return form", "def update_fields(self):\n if hasattr(self.day, \"body_composition\"):\n for f in self.get_fields():\n name = f.get_field().name\n value = getattr(self.day.body_composition, name, None)\n if value is not None:\n f.set_field(value)\n else:\n f.set_field(\"\")", "def make_fields(self):\n for name, prop in self.edit:\n instance_value = self.model.get(name)\n post_value = self.data[name] if (self.data and self.data.has_key(name)) else instance_value\n form_field_class = self.get_field_type(prop)\n form_field = form_field_class(model=self.model, property=prop, name=name, instance_value=instance_value, post_value=post_value)\n self.add(form_field)", "def test_form_has_fields(self):\r\n self.form = SubscriptionForm()\r\n expect = ['name', 'cpf', 'email', 'phone']\r\n self.assertSequenceEqual(expect, list(self.form.fields))", "def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n placeholders = {\n 'company_name': 'Company name',\n 'street_address1': 'Street Address 1',\n 'street_address2': 'Street Address 2',\n 'country': 'Country or State',\n 'postcode': 'Postcode',\n 'town_or_city': 'Town or City',\n 'payment': 'Paid for number of months',\n 'setting_daystart': 'Hour when your day starts',\n 'setting_dayend': 'hour when your day ends'\n }\n\n for field in self.fields:\n if field != 'country':\n if self.fields[field].required:\n placeholder = f'{placeholders[field]} *'\n else:\n placeholder = placeholders[field]\n self.fields[field].widget.attrs['placeholder'] = placeholder\n if field == 'setting_daystart' or field == 'setting_dayend' or field == 'payment':\n self.fields[field].widget.attrs['class'] = 'width-numbers'\n else:\n self.fields[field].widget.attrs['class'] = 'profile-form-input'\n self.fields[field].label = placeholder", "def __init__(self, *args, **kwargs):\n\n\t\tsuper(AssonaForm, self).__init__(*args, **kwargs)\n\n\t\tfor field in self.fields:\n\t\t\tif field == 'versicherungskarte':\n\t\t\t\tself.fields[field].widget.attrs.update({\n\t\t\t\t\t'class': 'form-control-file'\n\t\t\t\t})\n\t\t\telse:\n\t\t\t\tself.fields[field].widget.attrs.update({\n\t\t\t\t\t'class': 'form-control'\n\t\t\t\t})", "def get_form(self):\n # setup request layer\n self.request = TestRequest()\n # get add view\n form = getMultiAdapter((self.experiments, self.request),\n name=\"newProjection\")\n # update the form once to initialise all widgets\n form.update()\n # go through all widgets on the form and update the request with default values\n data = {}\n for widget in form.widgets.values():\n data[widget.name] = widget.value\n data.update({\n 'form.widgets.IDublinCore.title': u\"My CC Experiment\",\n 'form.widgets.IDublinCore.description': u'This is my experiment description',\n 'form.widgets.species_distribution_models': unicode(self.sdmexp.UID()),\n 'form.widgets.species_distribution_models.model': [unicode(self.sdmmodel.UID())],\n 'form.widgets.future_climate_datasets': [unicode(self.future.UID())]\n })\n self.request.form.update(data)\n form = getMultiAdapter((self.experiments, self.request),\n name=\"newProjection\")\n return form", "def update_models_from_create_event_form(form, venue, event):\n\n # Venue Info\n event.venue.name = form.venue_name.data\n event.venue.address = form.address.data\n event.venue.city = form.city.data\n event.venue.state = CreateEventForm.convert_choice_to_value(\n form.state.data, \"STATES\"\n )\n event.venue.zip_code = form.zip_code.data\n\n # Event info\n event.title = form.title.data\n event_type = EventType.query.get(form.event_type.data)\n event_category = EventCategory.query.get(form.category.data)\n start_time = CreateEventForm.convert_choice_to_value(form.start_time.data, \"TIMES\")\n end_time = CreateEventForm.convert_choice_to_value(form.end_time.data, \"TIMES\")\n event.start_datetime = datetime.combine(form.start_date.data, start_time)\n event.end_datetime = datetime.combine(form.end_date.data, end_time)", "def populate_form(self, **kwargs):\n for name, value in kwargs.items():\n self.populate_field(name, value)" ]
[ "0.7727503", "0.6019656", "0.5861441", "0.57970494", "0.57781494", "0.57305723", "0.5715702", "0.56819946", "0.5602228", "0.5507136", "0.5495653", "0.5426627", "0.5412202", "0.53416955", "0.5293119", "0.5292693", "0.5292534", "0.5275663", "0.52391994", "0.52208537", "0.52118427", "0.52099943", "0.5197609", "0.51803076", "0.5153143", "0.50826246", "0.5079311", "0.5069185", "0.5055291", "0.5053549" ]
0.77495456
0
Retrieve all conferences matching one or more given topics.
def _getConferencesByTopicSearch(self, request): # Generate list of filters from the topic arguments filters = [Conference.topics == topic for topic in request.topics] if not filters: raise endpoints.BadRequestException( 'At least one topic must be specified' ) # Retrieve all conferences matching one or more of the topic filters conferences = Conference.query( ndb.OR(*filters)).order(Conference.name).fetch() return conferences
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def getConferencesByTopicSearch(self, request):\n conferences = self._getConferencesByTopicSearch(request)\n # Need to fetch organiser displayName from profiles\n # Get all keys and use get_multi for speed\n organisers = [\n (ndb.Key(Profile, conf.organizerUserId)) for conf in conferences\n ]\n profiles = ndb.get_multi(organisers)\n # Put display names in a dict for easier fetching\n names = {}\n for profile in profiles:\n names[profile.key.id()] = profile.displayName\n # Return individual ConferenceForm object per Conference\n # Return individual ConferenceForm object per Conference\n return ConferenceForms(\n items=[\n self._copyConferenceToForm(conf, names[conf.organizerUserId])\n for conf in conferences\n ]\n )", "def get_all_subscriptions_by_topic(self, topic, next_token=None):\r\n params = {'ContentType' : 'JSON',\r\n 'TopicArn' : topic}\r\n if next_token:\r\n params['NextToken'] = next_token\r\n response = self.make_request('ListSubscriptionsByTopic', params,\r\n '/', 'GET')\r\n body = response.read()\r\n if response.status == 200:\r\n return json.loads(body)\r\n else:\r\n boto.log.error('%s %s' % (response.status, response.reason))\r\n boto.log.error('%s' % body)\r\n raise self.ResponseError(response.status, response.reason, body)", "def get_subscriptions(self, topic_name):\r\n resp = self._make_request('get',\r\n 'topics/%s/subscriptions' % topic_name)\r\n return resp.json()", "def resolve_topics(self, info, **kwargs):\n return Topic.objects.all()", "def _get_all_filtered_channels(self, topics_without_signature):\n mpe_address = self.get_mpe_address()\n event_signature = self.ident.w3.sha3(\n text=\"ChannelOpen(uint256,uint256,address,address,address,bytes32,uint256,uint256)\").hex()\n topics = [event_signature] + topics_without_signature\n logs = self.ident.w3.eth.getLogs(\n {\"fromBlock\": self.args.from_block, \"address\": mpe_address, \"topics\": topics})\n abi = get_contract_def(\"MultiPartyEscrow\")\n event_abi = abi_get_element_by_name(abi, \"ChannelOpen\")\n channels_ids = [get_event_data(event_abi, l)[\n \"args\"][\"channelId\"] for l in logs]\n return channels_ids", "def get_topics():\n topics, _ = base_query(db_session)\n return jsonify([p.serialize for p in topics])", "def schools_by_topic(mongo_collection, topic):\n res = []\n returned_values = mongo_collection.find({\"topics\": {\"$all\": [topic]}})\n for value in returned_values:\n res.append(value)\n return res", "def get_conferences():\n conferences = []\n for confId in range(1, 8):\n c = requests.get(CONFERENCES_URL.format(confId)).json()\n conf_list = c.get('conferences')\n if conf_list is None or len(conf_list) == 0:\n continue\n conferences.append(Conference.from_json(conf_list[0]))\n return conferences", "def list_topics(project_id):\n project_path = f\"projects/{project_id}\"\n for topic in PUBLISHER_CLIENT.list_topics(request={\"project\": project_path}):\n print(topic)", "def check_exists(cls, topics):\n\t\tresult = []\n\t\tfor known_feed in cls.get([cls.create_key(url) for url in set(topics)]):\n\t\t\tif known_feed is not None:\n\t\t\t\tresult.append(known_feed.topic)\n\t\treturn result", "def listen(self, topics):\n logging.debug(f'Listen to {list(map(lambda x: x.name, topics))}')\n\n for topic in map(lambda x: x.name, topics):\n try:\n self.subscribe(topic)\n logging.debug(f'Subscribed the {topic} topic')\n except Exception:\n logging.debug(f\"Can't subscribe the {topic} topic\")", "def get_topics(self, tags=None):\r\n params = {}\r\n if tags:\r\n params['tags'] = ','.join(tags)\r\n resp = self._make_request('get', 'topics', params=params)\r\n return resp.json()", "def get_subscribers(cls, topic, count, starting_at_callback=None):\n\t\tquery = cls.all()\n\t\tquery.filter('topic_hash =', utils.sha1_hash(topic))\n\t\tquery.filter('subscription_state = ', cls.STATE_VERIFIED)\n\t\tif starting_at_callback:\n\t\t\tquery.filter('callback_hash >=', utils.sha1_hash(starting_at_callback))\n\t\tquery.order('callback_hash')\n\n\t\treturn query.fetch(count)", "def get_topics(self):\n try:\n with self.__db_lock:\n sql = \"SELECT * FROM 'topics' ORDER BY 'name' ASC\"\n self.__cursor.execute(sql)\n topics = self.__cursor.fetchall()\n if topics is None or len(topics) == 0:\n return []\n return [topic[1] for topic in topics]\n except Exception as e:\n logging.error(\n \"Exception when trying to get topics list: {}\".format(e))\n return []", "def schools_by_topic(mongo_collection, topic):\n return mongo_collection.find({\"topics\": {\"$in\": [topic]}})", "def get_confs(req_topics=None, start_year=None, end_year=None):\n\n req_topics = req_topics or ','.join(topics.keys())\n start_year = start_year or min_year\n end_year = end_year or max_year\n\n # validate topics\n parsed_topics = req_topics.split(',')\n\n for topic in parsed_topics:\n if topic not in topics.keys():\n raise ValueError('Topic \\'{topic}\\' not valid'.format(topic=topic))\n\n # validate startYear\n try:\n start_year = int(start_year)\n except ValueError:\n raise ValueError('startYear must be valid integer.')\n\n if start_year > max_year or start_year < min_year:\n raise ValueError('startYear must be value between {min} and {max}.'.format(min=min_year, max=max_year))\n\n # validate endYear\n try:\n end_year = int(end_year)\n except ValueError:\n raise ValueError('endYear must be valid integer.')\n\n if end_year > max_year or end_year < min_year:\n raise ValueError('endYear must be value between {min} and {max}.'.format(min=min_year, max=max_year))\n\n if start_year > end_year:\n raise ValueError(\"startYear must not be greater than endYear.\")\n\n # aggregate data\n data = list()\n\n for year in range(start_year, end_year+1):\n for topic in parsed_topics:\n tmp = fetch(topic, year)\n tmp = normalize_confs(tmp, topic, year)\n data.extend(tmp)\n return data", "def _get_subscriptions(self, topic_arn):\n return self.conn.get_all_subscriptions_by_topic(topic_arn)['ListSubscriptionsByTopicResponse']['ListSubscriptionsByTopicResult']['Subscriptions']", "def topics(self, project: str) -> list:\n assert self.exists(project), f'Project {project} inesistente'\n\n cursor = self.projects(\n {\n 'url': project\n }\n )\n try:\n return cursor.next()['topics']\n except StopIteration:\n return []", "def get_teacher_topic_all():\n topic_data = query_db(\n \"SELECT topics.id, topics.name, classes.name FROM topics JOIN classes \"\n \"ON topics.class_id=classes.id WHERE teacher_id=?;\",\n [flask.session[\"id\"]],\n )\n topics = []\n for topic in topic_data:\n topic_dict_teacher = {}\n topic_dict_teacher[\"id\"] = topic[0]\n topic_dict_teacher[\"name\"] = flask.escape(str(topic[1]))\n topic_dict_teacher[\"class\"] = flask.escape(str(topic[2]))\n topics.append(topic_dict_teacher)\n return topics", "def describe_topics(self, topics=None):\n return self._client.describe_topics(topics)", "def get_topics(self):\r\n return [x[0] for x in get_published_topics()]", "def extract_queries(self, path_topics=\"../topics-rnd5.xml\"):\n \n topic_queries = []\n with open(path_topics, \"r\") as f:\n for line in f:\n match = re.match(\".*<query>([^<]*)<\\/query>.*\", line)\n if match:\n topic_queries.append(match.group(1))\n if len(topic_queries) != 50:\n sys.exit(\"There should be 50 topics, found {}\".format(\n len(topic_queries)))\n \n return topic_queries", "def get_all_subscriptions(cls, **kwargs):\n return Subscription.query.filter(**kwargs).all()", "async def subscribe(topics, on_close=None, on_error=None):\n loop = asyncio.get_event_loop()\n session = aiohttp.ClientSession()\n async with session.ws_connect('wss://api.huobi.pro/ws') as ws:\n keys = {\n topic: uuid.uuid4().hex\n for topic in topics\n }\n keyed_channels = {\n v: topics[k]\n for k, v in keys.items()\n }\n subscribed_chanels = {}\n for topic, config in topics.items():\n payload = {\n 'sub': topic,\n 'id': keys[topic]\n }\n await ws.send_str(encode_ws_payload(payload))\n async for msg in ws:\n if msg.type == aiohttp.WSMsgType.BINARY:\n data = decode_ws_payload(msg.data)\n\n ping = data.get('ping')\n if ping:\n reply = encode_ws_payload({'pong': ping})\n await ws.send_str(\n reply\n )\n\n subbed = data.get('subbed')\n if subbed:\n if data.get('status') == 'ok':\n subscribed_chanels[subbed] = keyed_channels[data['id']]\n\n ch = data.get('ch')\n if ch:\n cb = subscribed_chanels[ch].get('callback', lambda _: None)\n if asyncio.iscoroutinefunction(cb):\n await cb(data)\n else:\n loop.run_in_executor(None, partial(cb, data))\n elif msg.type == aiohttp.WSMsgType.CLOSED:\n if on_close:\n return on_close()\n return\n elif msg.type == aiohttp.WSMsgType.ERROR:\n if on_error:\n return on_error()\n return", "def conferences(self):\r\n return conferences.Conferences(self)", "def list(self, topic, **options):\n pass", "def test_get_full_topic_courses(self):\r\n course_id = None # Change me!!\r\n topic_id = None # Change me!!\r\n\r\n r = self.client.get_full_topic_courses(topic_id, course_id)", "def get_questions_of_topic(topic):\n\n dynamodb = boto3.resource(\"dynamodb\", region_name=\"eu-central-1\")\n question_table = dynamodb.Table(\"Questions\")\n\n fe = Attr(\"TopicId\").eq(topic.get(\"TopicId\"))\n response = question_table.scan(FilterExpression=fe)\n questions = response.get(\"Items\")\n return questions", "def topics(ctx):\n pass", "def topics(self):\n # use get_model to avoid circular dependency\n topic_model = apps.get_model('tags', 'Topic')\n return topic_model.objects.filter(tag__in=self.tags.all()).distinct()" ]
[ "0.6711613", "0.6230136", "0.61121726", "0.60039115", "0.5985098", "0.5940796", "0.58975244", "0.58861005", "0.5882982", "0.5880553", "0.5836876", "0.58033025", "0.57397395", "0.5726651", "0.5705699", "0.5686112", "0.56527764", "0.56140727", "0.54820526", "0.5472352", "0.54712605", "0.53320074", "0.5298793", "0.5282408", "0.5267355", "0.5241664", "0.5239345", "0.52040243", "0.51683694", "0.51324224" ]
0.790961
0
Return Announcement from memcache.
def getAnnouncement(self, request): announcement = memcache.get(MEMCACHE_ANNOUNCEMENTS_KEY) or "" return StringMessage(data=announcement)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_announcement(self, request):\n return StringMessage(\n data=memcache.get(MEMCACHE_ANNOUNCEMENTS_KEY) or \"\")", "def _cacheAnnouncement():\n confs = Conference.query(ndb.AND(\n Conference.seatsAvailable <= 5,\n Conference.seatsAvailable > 0)\n ).fetch(projection=[Conference.name])\n if confs:\n # If there are conferences close to being sold out,\n # format announcement and set it in memcache\n announcement = '%s %s' % (\n 'Last chance to attend! The following conferences '\n 'are nearly sold out:',\n ', '.join(conf.name for conf in confs))\n memcache.set(MEMCACHE_ANNOUNCEMENTS_KEY, announcement)\n else:\n # If there are no sold out conferences,\n # delete the memcache announcements entry\n announcement = \"\"\n memcache.delete(MEMCACHE_ANNOUNCEMENTS_KEY)\n return announcement", "def cache_announcement():\n confs = Conference.query(ndb.AND(\n Conference.seatsAvailable <= 5, Conference.seatsAvailable > 0\n )).fetch(projection=[Conference.name])\n\n if confs:\n # If there are almost sold out conferences,\n # format announcement and set it in memcache\n announcement = ANNOUNCEMENT_TPL % (\n ', '.join(conf.name for conf in confs))\n memcache.set(MEMCACHE_ANNOUNCEMENTS_KEY, announcement)\n else:\n # If there are no sold out conferences,\n # delete the memcache announcements entry\n announcement = \"\"\n memcache.delete(MEMCACHE_ANNOUNCEMENTS_KEY)\n\n return announcement", "def get_announcement(self, cid):\n cid = cid.upper()\n query = \"SELECT * from announcement where cid = %s\"\n inputs = (cid, )\n result = self.database_manager.execute_query(query, inputs)\n if result:\n announcement = \"Announcement for {} ({}): {}\".format(result[0][0], result[0][3], result[0][2])\n else:\n announcement = \"No announcement for this {}\".format(cid)\n return announcement", "def memcacheGetRecord(self, key):\n\n pickled = self.memcacheGet(key)\n return self.unpickleRecord(pickled) if pickled is not None else None", "def _cache_get(self, metric_name):\n pass", "def _memcache_get(*args, **kwargs):\n return ndb.get_context().memcache_get(*args, **kwargs)", "def get(self):\n return {'status': 'success', 'count': Announcement.query.count()}, 200", "def get_cache(self, key):\n return self.r.get(key)", "def get(self, key):\n return self.cache_data.get(key)", "def __getitem__(self, key):\n self._remove_expired()\n\n cache_entry = self._d.get(key, None)\n log.debug(\"__getitem__: {}\".format(cache_entry))\n\n return cache_entry", "def get(key):\n return Cache.cache_connector.get(key)", "def get_cache(self):\n return self.cache", "def _cache_get(self, metric_name):\n encoded_metric_name = self._encode(metric_name)\n with self.__env.begin(self.__metric_to_metadata_db, write=False) as txn:\n payload = txn.get(encoded_metric_name)\n\n if payload == self._EMPTY:\n return None, True\n\n if payload is not None:\n payload = self._decode(payload)\n\n if not payload:\n # cache miss\n return None, False\n\n # found something in the cache\n split = self.__split_payload(payload)\n\n if split is None:\n # invalid string => evict from cache\n with self.__env.begin(self.__metric_to_metadata_db, write=True) as txn:\n txn.delete(key=encoded_metric_name)\n return None, False\n\n # valid value => get id and metadata string\n # TODO: optimization: id is a UUID (known length)\n id_str, metadata_str, timestamp = split\n try:\n id = uuid.UUID(id_str)\n except Exception as e:\n logging.debug(str(e))\n with self.__env.begin(self.__metric_to_metadata_db, write=True) as txn:\n txn.delete(key=encoded_metric_name)\n return None, False\n\n # if the timestamp expired evict it in order to force\n # its recreation for the next time\n if self.__expired_timestamp(timestamp):\n with self.__env.begin(self.__metric_to_metadata_db, write=True) as txn:\n txn.delete(key=encoded_metric_name)\n\n metadata = self.metadata_from_str(metadata_str)\n return bg_metric.Metric(metric_name, id, metadata), True", "def get(self, key):\n # Initialize key variables\n result = self.cache.get(key)\n\n # Return\n return result", "def cache_get(self, key: str) -> Optional[bytes]:\n if self.cache is not None:\n return self.cache.get(key)\n return None", "def get(self):\n CACHE_KEY = 'topics'\n if not memcache.get(CACHE_KEY):\n logging.info('Populating cache.')\n topics = Topic.all().order('name')\n topic_list = []\n for topic in topics:\n topic_list.append(topic.ToDict())\n memcache.add(CACHE_KEY, simplejson.dumps(topic_list), 600)\n logging.info('Using cache.')\n logging.info(memcache.get(CACHE_KEY))\n self.response.headers['Content-Type'] = 'application/json'\n self.response.out.write(memcache.get(CACHE_KEY))", "def get(self, key):\n if key is None:\n return None\n return self.cache_data.get(key, None)", "def GetFromCache(self, filename):\n return memcache.get('%s%s' % (self.CACHE_PREFIX, filename))", "def GetFromCache(self, filename):\n return memcache.get('%s%s' % (self.CACHE_PREFIX, filename))", "def _retrieveCachedData(self):", "def memcacheGet(self, key):\n\n key = base64.b64encode(key)\n try:\n value = self._getMemcacheClient().get(key)\n except MemcacheError:\n log.error(\"Could not read from memcache, retrying\")\n try:\n value = self._getMemcacheClient(refresh=True).get(key)\n except MemcacheError:\n log.error(\"Could not read from memcache again, giving up\")\n del self.memcacheClient\n raise DirectoryMemcacheError(\"Failed to read from memcache\")\n return value", "def get(self, key):\n if key:\n return self.cache_data.get(key)\n else:\n return None", "def cache(self):\n return self.payload.setdefault(self._CACHE_ATTRIBUTE, {})", "def from_cache(cls, cached_token):\n return cls(\n access_token=cached_token[\"access_token\"],\n expires_in=int(cached_token[\"expires_in\"]),\n expires_at=int(cached_token[\"expires_at\"]),\n token_type=cached_token[\"token_type\"],\n audience=cached_token[\"audience\"],\n )", "def get(self, cache_id):\n return self.certificates.get(cache_id)", "def getData(self, local_cache):", "def fetch(self,key):\n try:\n return self.__content[key]\n except KeyError:\n return None", "def GetEntityViaMemcache(entity_key):\n entity = memcache.get(entity_key)\n if entity is not None:\n return entity\n key = ndb.Key(urlsafe=entity_key)\n entity = key.get()\n if entity is not None:\n memcache.set(entity_key, entity)\n return entity", "def get(self):\n CACHE_KEY = 'sources'\n if not memcache.get(CACHE_KEY):\n logging.info('Populating cache.')\n feeds = Feed.all().order('name')\n feed_list = []\n for feed in feeds:\n feed_list.append(feed.ToDict())\n memcache.add(CACHE_KEY, simplejson.dumps(feed_list), 600)\n logging.info('Using cache.')\n logging.info(memcache.get(CACHE_KEY))\n self.response.headers['Content-Type'] = 'application/json'\n self.response.out.write(memcache.get(CACHE_KEY))" ]
[ "0.74955446", "0.7363284", "0.705641", "0.58456796", "0.5752198", "0.56402856", "0.56222874", "0.5593154", "0.5495106", "0.5467809", "0.54591787", "0.54124725", "0.5330662", "0.53072345", "0.5287989", "0.527153", "0.52640253", "0.523533", "0.52221245", "0.52221245", "0.52177227", "0.51978356", "0.51977575", "0.5173499", "0.5169746", "0.5145714", "0.5130637", "0.51244736", "0.5121278", "0.51201475" ]
0.78246075
0
Return requested conference (by websafeConferenceKey).
def getConference(self, request): # Get Conference object from request; bail if not found conf = _getEntityByWebsafeKey(request.websafeConferenceKey, 'Conference') prof = conf.key.parent().get() # Return ConferenceForm return self._copyConferenceToForm(conf, getattr(prof, 'displayName'))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def getConference(self, request):\n # get Conference object from request; bail if not found\n conf = ndb.Key(urlsafe=request.websafeConferenceKey).get()\n if not conf:\n raise endpoints.NotFoundException(\n 'No conference found with key: %s' % request.websafeConferenceKey)\n prof = conf.key.parent().get()\n # return ConferenceForm\n return self._copyConferenceToForm(conf, getattr(prof, 'displayName'))", "def get_conference(self, request):\n # get Conference object from request; bail if not found\n conf = ndb.Key(urlsafe=request.websafeConferenceKey).get()\n if not conf:\n raise endpoints.NotFoundException(\n 'No conference found with key: %s' %\n request.websafeConferenceKey)\n prof = conf.key.parent().get()\n # return ConferenceForm\n return self.conference_service.copy_conference_to_form(ConferenceForm(),\n conf,\n prof.displayName)", "def conference(self, sid):\r\n return conferences.Conference(self, sid)", "def ConferenceRecord(self, default={}):\n return HEP.JSONReferenceObject(self.data.get('conference_record', default))", "async def get_conference(self, user_id):\n db = self['db_engine']\n async with db.acquire() as connection:\n query = select([Conference]).where(Conference.user_id == user_id).where(\n Conference.expired_at > datetime.now(timezone.utc))\n if conference := await(await connection.execute(query)).first():\n return conference", "async def get_conference_by_id(self, session_id):\n db = self['db_engine']\n async with db.acquire() as connection:\n query = select([Conference]).where(Conference.session_id == session_id).where(\n Conference.expired_at > datetime.now(timezone.utc))\n if conference := await(await connection.execute(query)).first():\n result = dict(conference)\n query = select([User.email, User.display_name]).where(User.id == conference.user_id)\n if user := await(await connection.execute(query)).first():\n result['user_data'] = UserData(conference.user_id, *user)\n return result", "def conference_country(self):\n return ConferenceReader(self.record).country", "def updateConference(self, request):\n return self._updateConferenceObject(request)", "def registerForConference(self, request):\n return self._conferenceRegistration(request)", "def _getConferenceSessions(self, request):\n # Ensure that websafeConferenceKey is a valid conference key\n confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey,\n 'Conference')\n # Retrieve all sessions that have a matching conference key\n sessions = Session.query(Session.conference == confKey).fetch()\n return sessions", "def join_conference(request, conference_name):\n resp = VoiceResponse()\n resp.dial(hangupOnStar=True).conference(conference_name)\n return HttpResponse(str(resp))", "def updateConference(self, request):\n return self._updateConferenceObject(request)", "def _createConferenceObject(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = user.email()\n if not request.name:\n raise endpoints.BadRequestException(\n \"Conference 'name' field required\")\n # Copy ConferenceForm/ProtoRPC Message into dict\n data = {\n field.name: getattr(request, field.name) for field in\n request.all_fields()\n }\n del data['websafeKey']\n del data['organizerDisplayName']\n # Add default values for those missing (both data model and\n # outbound Message)\n for df in CONF_DEFAULTS:\n if data[df] in (None, []):\n data[df] = CONF_DEFAULTS[df]\n setattr(request, df, CONF_DEFAULTS[df])\n # Convert dates from strings to Date objects; set month based\n # on start_date\n if data['startDate']:\n data['startDate'] = datetime.strptime(\n data['startDate'][:10], \"%Y-%m-%d\").date()\n data['month'] = data['startDate'].month\n else:\n data['month'] = 0\n if data['endDate']:\n data['endDate'] = datetime.strptime(\n data['endDate'][:10], \"%Y-%m-%d\").date()\n # Set seatsAvailable to be same as maxAttendees on creation\n if data[\"maxAttendees\"] > 0:\n data[\"seatsAvailable\"] = data[\"maxAttendees\"]\n # Get the user profile key, then set the conference's parent\n # to that value.\n # NOTE: The original code made a call to allocate_ids in order to\n # generate an ID for the conference. Since the profiles utilize\n # strings (email addresses) for their IDs, resulting in no risk\n # of colliding with NDB's auto-generated numeric IDs, I decided\n # to let NDB generate the conference ID automatically.\n # https://cloud.google.com/appengine/docs/python/ndb/entities?hl=en#numeric_keys\n p_key = ndb.Key(Profile, user_id)\n data['parent'] = p_key\n data['organizerUserId'] = request.organizerUserId = user_id\n # Create Conference, send email to organizer confirming\n # creation of Conference and return (modified) ConferenceForm\n Conference(**data).put()\n taskqueue.add(params={'email': user.email(),\n 'conferenceInfo': repr(request)},\n url='/tasks/send_confirmation_email'\n )\n return request", "def build_conference(self, data) -> dict:\n builder = ConferenceBuilder()\n builder.add_title(title=data.get(\"name\"), subtitle=data.get(\"subtitle\"))\n builder.set_short_description(value=data.get(\"description\", \"\"))\n builder.set_opening_date(get_value(data, \"dates[0]\"))\n builder.set_closing_date(get_value(data, \"dates[1]\"))\n builder.add_inspire_categories(data.get(\"field_of_interest\", []))\n builder.add_public_note(value=data.get(\"additional_info\", \"\"))\n builder.add_series(\n name=data.get(\"series_name\"), number=data.get(\"series_number\")\n )\n for address in data.get(\"addresses\"):\n builder.add_address(\n cities=[address.get(\"city\")],\n state=address.get(\"state\"),\n place_name=address.get(\"venue\"),\n country_code=country_name_to_code(address.get(\"country\")),\n )\n for contact in data.get(\"contacts\", []):\n builder.add_contact(**contact)\n for acr in data.get(\"acronyms\", []):\n builder.add_acronym(acr)\n for website in data.get(\"websites\", []):\n builder.add_url(website)\n for keyword in data.get(\"keywords\", []):\n builder.add_keyword(value=keyword)\n\n builder.record[\"$schema\"] = url_for(\n \"invenio_jsonschemas.get_schema\",\n schema_path=\"records/conferences.json\",\n _external=True,\n )\n\n return builder.record", "def conferences(self):\r\n return conferences.Conferences(self)", "def register_for_conference(self, request):\n return self.conference_registration(request)", "def get_conferences():\n conferences = []\n for confId in range(1, 8):\n c = requests.get(CONFERENCES_URL.format(confId)).json()\n conf_list = c.get('conferences')\n if conf_list is None or len(conf_list) == 0:\n continue\n conferences.append(Conference.from_json(conf_list[0]))\n return conferences", "def is_in_conference(self) -> bool:", "def get_conference_sessions(self, request):\n return self.session_service.get_conference_sessions(\n request.websafeConferenceKey)", "def get_current_conference_for_series(self, series):\n #print 'get_current_conference_for_series()'\n #print ' series: %s' % series\n current_year = datetime.today().year\n #print ' current_year: %s' % current_year\n resources = Resource.objects.filter(\n resource_type__name=ResourceType.CONFERENCE,\n conference_series=series, year__gte=current_year\n ).order_by('date', 'year', 'id')\n #print ' resources.count(): %s' % resources.count()\n if resources.count():\n # Use the next future resource\n #print ' using next future'\n return resources[0]\n else:\n # Use the most-recent past resource\n #print ' using most recent'\n resources = Resource.objects.filter(\n resource_type__name=ResourceType.CONFERENCE,\n conference_series=series, year__lt=current_year\n ).order_by('-date', '-year', '-id')\n #print ' resources.count(): %s' % resources.count()\n if resources.count():\n return resources[0]\n else:\n # Check if any resources have a NULL year\n resources = \\\n Resource.objects.filter(\n resource_type__name=ResourceType.CONFERENCE,\n conference_series=series, year=None).\\\n order_by('-date', '-year', '-id')\n if resources.count():\n return resources[0]\n else:\n # There are no conferences in this series\n return None", "def update_conference(self, request):\n return self.conference_service.update_conference_object(request)", "def _createConferenceObject(self, request):\n # preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = getUserId(user)\n\n if not request.name:\n raise endpoints.BadRequestException(\"Conference 'name' field required\")\n\n # copy ConferenceForm/ProtoRPC Message into dict\n data = {field.name: getattr(request, field.name) for field in request.all_fields()}\n del data['websafeKey']\n del data['organizerDisplayName']\n\n # add default values for those missing (both data model & outbound Message)\n for df in DEFAULTS:\n if data[df] in (None, []):\n data[df] = DEFAULTS[df]\n setattr(request, df, DEFAULTS[df])\n\n # convert dates from strings to Date objects; set month based on start_date\n if data['startDate']:\n data['startDate'] = datetime.strptime(data['startDate'][:10], \"%Y-%m-%d\").date()\n data['month'] = data['startDate'].month\n else:\n data['month'] = 0\n if data['endDate']:\n data['endDate'] = datetime.strptime(data['endDate'][:10], \"%Y-%m-%d\").date()\n\n # set seatsAvailable to be same as maxAttendees on creation\n if data[\"maxAttendees\"] > 0:\n data[\"seatsAvailable\"] = data[\"maxAttendees\"]\n # generate Profile Key based on user ID and Conference\n # ID based on Profile key get Conference key from ID\n p_key = ndb.Key(Profile, user_id)\n c_id = Conference.allocate_ids(size=1, parent=p_key)[0]\n c_key = ndb.Key(Conference, c_id, parent=p_key)\n data['key'] = c_key\n data['organizerUserId'] = request.organizerUserId = user_id\n\n # create Conference, send email to organizer confirming\n # creation of Conference & return (modified) ConferenceForm\n Conference(**data).put()\n taskqueue.add(params={'email': user.email(),\n 'conferenceInfo': repr(request)},\n url='/tasks/send_confirmation_email'\n )\n return request", "def _getConferenceSessionsByType(self, request):\n # Ensure that websafeConferenceKey is a valid conference key\n confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey,\n 'Conference')\n # Retrieve all sessions that have a matching conference key, by type\n sessions = Session.query(\n Session.conference == confKey,\n Session.typeOfSession == str(request.typeOfSession)\n ).fetch()\n return sessions", "def get(congregation_guid):\n try:\n return Congregation.objects.get(guid=congregation_guid)\n except ObjectDoesNotExist:\n url = f'https://apps.jw.org/api/public/meeting-search/weekly-meetings/{congregation_guid}'\n with urllib.request.urlopen(url) as response:\n print(response.read())", "def getConferencesByTopicSearch(self, request):\n conferences = self._getConferencesByTopicSearch(request)\n # Need to fetch organiser displayName from profiles\n # Get all keys and use get_multi for speed\n organisers = [\n (ndb.Key(Profile, conf.organizerUserId)) for conf in conferences\n ]\n profiles = ndb.get_multi(organisers)\n # Put display names in a dict for easier fetching\n names = {}\n for profile in profiles:\n names[profile.key.id()] = profile.displayName\n # Return individual ConferenceForm object per Conference\n # Return individual ConferenceForm object per Conference\n return ConferenceForms(\n items=[\n self._copyConferenceToForm(conf, names[conf.organizerUserId])\n for conf in conferences\n ]\n )", "def conference_title(self):\n try:\n conference_records = ConferencesRecord.get_linked_records_from_dict_field(\n self.record, \"publication_info.conference_record\"\n )\n conference_record = next(conference_records, None)\n return LiteratureReader(conference_record).title\n except TypeError:\n pass", "def getConferencesToAttend(self, request):\n prof = self._getProfileFromUser() # get user Profile\n conf_keys = [\n ndb.Key(urlsafe=wsck) for wsck in prof.conferenceKeysToAttend\n ]\n conferences = ndb.get_multi(conf_keys)\n # Get organizers\n organisers = [\n ndb.Key(Profile, conf.organizerUserId) for conf in conferences\n ]\n profiles = ndb.get_multi(organisers)\n # Put display names in a dict for easier fetching\n names = {}\n for profile in profiles:\n names[profile.key.id()] = profile.displayName\n # Return set of ConferenceForm objects per Conference\n return ConferenceForms(\n items=[\n self._copyConferenceToForm(conf, names[conf.organizerUserId])\n for conf in conferences\n ]\n )", "def is_conference_paper(self):\n if self.root.xpath(\"./conference-info\"):\n return True\n journal_issue = self.root.xpath(\n \"string(./RDF/Description/issueName[1])\"\n ).extract_first()\n if journal_issue:\n is_conference = re.findall(r\"proceedings|proc.\", journal_issue.lower())\n return bool(is_conference)\n return False", "def _conferenceRegistration(self, request, reg=True):\n retval = None\n # Get user profile\n prof = self._getProfileFromUser()\n # Check if conference given in the websafeConferenceKey exists\n wsck = request.websafeConferenceKey\n conf = _getEntityByWebsafeKey(wsck, 'Conference')\n # Register\n if reg:\n # Check if user already registered, otherwise add\n if wsck in prof.conferenceKeysToAttend:\n raise ConflictException(\n \"You have already registered for this conference.\")\n # Check if seats available\n if conf.seatsAvailable <= 0:\n raise ConflictException(\n \"There are no seats available.\")\n # Register user, deduct one seat\n prof.conferenceKeysToAttend.append(wsck)\n conf.seatsAvailable -= 1\n retval = True\n # Unregister\n else:\n # Check if user already registered\n if wsck in prof.conferenceKeysToAttend:\n # Unregister user, add back one seat\n prof.conferenceKeysToAttend.remove(wsck)\n conf.seatsAvailable += 1\n retval = True\n else:\n retval = False\n # Update the datastore and return\n prof.put()\n conf.put()\n return BooleanMessage(data=retval)", "def _get_subscription(self):\n response = requests.get(\n 'https://www.googleapis.com/youtube/v3/subscriptions',\n params={\n 'part': 'snippet',\n 'mine': 'true',\n 'forChannelId': self.channel_id\n },\n headers=self.headers\n )\n if response.status_code == 200:\n return response.json()\n return {}" ]
[ "0.78688127", "0.7338922", "0.6852561", "0.66271746", "0.6565396", "0.6170011", "0.59663105", "0.59223247", "0.58999187", "0.58343786", "0.57726926", "0.57710207", "0.56881297", "0.5676969", "0.56550586", "0.55980533", "0.5535486", "0.5489521", "0.5484447", "0.5409856", "0.5335382", "0.53034943", "0.52697337", "0.5171076", "0.5107901", "0.5100049", "0.5094293", "0.5013125", "0.49801326", "0.49754968" ]
0.73678344
1
Get list of conferences matching one or more of the given topics.
def getConferencesByTopicSearch(self, request): conferences = self._getConferencesByTopicSearch(request) # Need to fetch organiser displayName from profiles # Get all keys and use get_multi for speed organisers = [ (ndb.Key(Profile, conf.organizerUserId)) for conf in conferences ] profiles = ndb.get_multi(organisers) # Put display names in a dict for easier fetching names = {} for profile in profiles: names[profile.key.id()] = profile.displayName # Return individual ConferenceForm object per Conference # Return individual ConferenceForm object per Conference return ConferenceForms( items=[ self._copyConferenceToForm(conf, names[conf.organizerUserId]) for conf in conferences ] )
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _getConferencesByTopicSearch(self, request):\n # Generate list of filters from the topic arguments\n filters = [Conference.topics == topic for topic in request.topics]\n if not filters:\n raise endpoints.BadRequestException(\n 'At least one topic must be specified'\n )\n # Retrieve all conferences matching one or more of the topic filters\n conferences = Conference.query(\n ndb.OR(*filters)).order(Conference.name).fetch()\n return conferences", "def schools_by_topic(mongo_collection, topic):\n res = []\n returned_values = mongo_collection.find({\"topics\": {\"$all\": [topic]}})\n for value in returned_values:\n res.append(value)\n return res", "def get_all_subscriptions_by_topic(self, topic, next_token=None):\r\n params = {'ContentType' : 'JSON',\r\n 'TopicArn' : topic}\r\n if next_token:\r\n params['NextToken'] = next_token\r\n response = self.make_request('ListSubscriptionsByTopic', params,\r\n '/', 'GET')\r\n body = response.read()\r\n if response.status == 200:\r\n return json.loads(body)\r\n else:\r\n boto.log.error('%s %s' % (response.status, response.reason))\r\n boto.log.error('%s' % body)\r\n raise self.ResponseError(response.status, response.reason, body)", "def check_exists(cls, topics):\n\t\tresult = []\n\t\tfor known_feed in cls.get([cls.create_key(url) for url in set(topics)]):\n\t\t\tif known_feed is not None:\n\t\t\t\tresult.append(known_feed.topic)\n\t\treturn result", "def schools_by_topic(mongo_collection, topic):\n return mongo_collection.find({\"topics\": {\"$in\": [topic]}})", "def get_topics(self):\n try:\n with self.__db_lock:\n sql = \"SELECT * FROM 'topics' ORDER BY 'name' ASC\"\n self.__cursor.execute(sql)\n topics = self.__cursor.fetchall()\n if topics is None or len(topics) == 0:\n return []\n return [topic[1] for topic in topics]\n except Exception as e:\n logging.error(\n \"Exception when trying to get topics list: {}\".format(e))\n return []", "def _get_all_filtered_channels(self, topics_without_signature):\n mpe_address = self.get_mpe_address()\n event_signature = self.ident.w3.sha3(\n text=\"ChannelOpen(uint256,uint256,address,address,address,bytes32,uint256,uint256)\").hex()\n topics = [event_signature] + topics_without_signature\n logs = self.ident.w3.eth.getLogs(\n {\"fromBlock\": self.args.from_block, \"address\": mpe_address, \"topics\": topics})\n abi = get_contract_def(\"MultiPartyEscrow\")\n event_abi = abi_get_element_by_name(abi, \"ChannelOpen\")\n channels_ids = [get_event_data(event_abi, l)[\n \"args\"][\"channelId\"] for l in logs]\n return channels_ids", "def get_confs(req_topics=None, start_year=None, end_year=None):\n\n req_topics = req_topics or ','.join(topics.keys())\n start_year = start_year or min_year\n end_year = end_year or max_year\n\n # validate topics\n parsed_topics = req_topics.split(',')\n\n for topic in parsed_topics:\n if topic not in topics.keys():\n raise ValueError('Topic \\'{topic}\\' not valid'.format(topic=topic))\n\n # validate startYear\n try:\n start_year = int(start_year)\n except ValueError:\n raise ValueError('startYear must be valid integer.')\n\n if start_year > max_year or start_year < min_year:\n raise ValueError('startYear must be value between {min} and {max}.'.format(min=min_year, max=max_year))\n\n # validate endYear\n try:\n end_year = int(end_year)\n except ValueError:\n raise ValueError('endYear must be valid integer.')\n\n if end_year > max_year or end_year < min_year:\n raise ValueError('endYear must be value between {min} and {max}.'.format(min=min_year, max=max_year))\n\n if start_year > end_year:\n raise ValueError(\"startYear must not be greater than endYear.\")\n\n # aggregate data\n data = list()\n\n for year in range(start_year, end_year+1):\n for topic in parsed_topics:\n tmp = fetch(topic, year)\n tmp = normalize_confs(tmp, topic, year)\n data.extend(tmp)\n return data", "def get_subscriptions(self, topic_name):\r\n resp = self._make_request('get',\r\n 'topics/%s/subscriptions' % topic_name)\r\n return resp.json()", "def get_conferences():\n conferences = []\n for confId in range(1, 8):\n c = requests.get(CONFERENCES_URL.format(confId)).json()\n conf_list = c.get('conferences')\n if conf_list is None or len(conf_list) == 0:\n continue\n conferences.append(Conference.from_json(conf_list[0]))\n return conferences", "def topics(self, project: str) -> list:\n assert self.exists(project), f'Project {project} inesistente'\n\n cursor = self.projects(\n {\n 'url': project\n }\n )\n try:\n return cursor.next()['topics']\n except StopIteration:\n return []", "def get_subscribers(cls, topic, count, starting_at_callback=None):\n\t\tquery = cls.all()\n\t\tquery.filter('topic_hash =', utils.sha1_hash(topic))\n\t\tquery.filter('subscription_state = ', cls.STATE_VERIFIED)\n\t\tif starting_at_callback:\n\t\t\tquery.filter('callback_hash >=', utils.sha1_hash(starting_at_callback))\n\t\tquery.order('callback_hash')\n\n\t\treturn query.fetch(count)", "def list_topics(project_id):\n project_path = f\"projects/{project_id}\"\n for topic in PUBLISHER_CLIENT.list_topics(request={\"project\": project_path}):\n print(topic)", "def get_topics():\n topics, _ = base_query(db_session)\n return jsonify([p.serialize for p in topics])", "def _get_subscriptions(self, topic_arn):\n return self.conn.get_all_subscriptions_by_topic(topic_arn)['ListSubscriptionsByTopicResponse']['ListSubscriptionsByTopicResult']['Subscriptions']", "def get_topics(self, tags=None):\r\n params = {}\r\n if tags:\r\n params['tags'] = ','.join(tags)\r\n resp = self._make_request('get', 'topics', params=params)\r\n return resp.json()", "def resolve_topics(self, info, **kwargs):\n return Topic.objects.all()", "def listen(self, topics):\n logging.debug(f'Listen to {list(map(lambda x: x.name, topics))}')\n\n for topic in map(lambda x: x.name, topics):\n try:\n self.subscribe(topic)\n logging.debug(f'Subscribed the {topic} topic')\n except Exception:\n logging.debug(f\"Can't subscribe the {topic} topic\")", "def get_topics(self):\r\n return [x[0] for x in get_published_topics()]", "def describe_topics(self, topics=None):\n return self._client.describe_topics(topics)", "def extract_queries(self, path_topics=\"../topics-rnd5.xml\"):\n \n topic_queries = []\n with open(path_topics, \"r\") as f:\n for line in f:\n match = re.match(\".*<query>([^<]*)<\\/query>.*\", line)\n if match:\n topic_queries.append(match.group(1))\n if len(topic_queries) != 50:\n sys.exit(\"There should be 50 topics, found {}\".format(\n len(topic_queries)))\n \n return topic_queries", "def get_teacher_topic_all():\n topic_data = query_db(\n \"SELECT topics.id, topics.name, classes.name FROM topics JOIN classes \"\n \"ON topics.class_id=classes.id WHERE teacher_id=?;\",\n [flask.session[\"id\"]],\n )\n topics = []\n for topic in topic_data:\n topic_dict_teacher = {}\n topic_dict_teacher[\"id\"] = topic[0]\n topic_dict_teacher[\"name\"] = flask.escape(str(topic[1]))\n topic_dict_teacher[\"class\"] = flask.escape(str(topic[2]))\n topics.append(topic_dict_teacher)\n return topics", "def getSubscriptionTopics(self) -> List[str]:\n return self.cpp.getSubscriptionTopics()", "def topics(self):\n # use get_model to avoid circular dependency\n topic_model = apps.get_model('tags', 'Topic')\n return topic_model.objects.filter(tag__in=self.tags.all()).distinct()", "def conferences(self):\r\n return conferences.Conferences(self)", "def topics(ctx):\n pass", "def get_questions_of_topic(topic):\n\n dynamodb = boto3.resource(\"dynamodb\", region_name=\"eu-central-1\")\n question_table = dynamodb.Table(\"Questions\")\n\n fe = Attr(\"TopicId\").eq(topic.get(\"TopicId\"))\n response = question_table.scan(FilterExpression=fe)\n questions = response.get(\"Items\")\n return questions", "def topics(self):\n return topics.Topics(self)", "def list(self, topic, **options):\n pass", "def topics(self):\r\n return topics.Topics(self)" ]
[ "0.80247796", "0.6207768", "0.6082327", "0.60714567", "0.60658276", "0.5967323", "0.5963362", "0.59466356", "0.59359396", "0.59351933", "0.5891676", "0.5859516", "0.5841561", "0.5830634", "0.57371825", "0.5714454", "0.57058376", "0.56997335", "0.5605298", "0.54814893", "0.5423089", "0.5363475", "0.532482", "0.52794933", "0.5257338", "0.52056634", "0.5198464", "0.5186003", "0.51842296", "0.5157304" ]
0.67396516
1
Register user for selected conference.
def registerForConference(self, request): return self._conferenceRegistration(request)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def register_for_conference(self, request):\n return self.conference_registration(request)", "def register_user():\n pass", "def conference_registration(self, request, reg=True):\n prof = self.profile_service.get_profile_from_user() # get user Profile\n\n # check if conf exists given websafeConfKey\n # get conference; check that it exists\n wsck = request.websafeConferenceKey\n conf = ndb.Key(urlsafe=wsck).get()\n if not conf:\n raise endpoints.NotFoundException(\n 'No conference found with key: %s' % wsck)\n\n # register\n if reg:\n # check if user already registered otherwise add\n if wsck in prof.conferenceKeysToAttend:\n raise ConflictException(\n \"You have already registered for this conference\")\n\n # check if seats avail\n if conf.seatsAvailable <= 0:\n raise ConflictException(\"There are no seats available.\")\n\n # register user, take away one seat\n prof.conferenceKeysToAttend.append(wsck)\n conf.seatsAvailable -= 1\n retval = True\n\n # unregister\n else:\n # check if user already registered\n if wsck in prof.conferenceKeysToAttend:\n\n # unregister user, add back one seat\n prof.conferenceKeysToAttend.remove(wsck)\n conf.seatsAvailable += 1\n retval = True\n else:\n retval = False\n\n # write things back to the datastore & return\n prof.put()\n conf.put()\n return BooleanMessage(data=retval)", "def _conferenceRegistration(self, request, reg=True):\n retval = None\n # Get user profile\n prof = self._getProfileFromUser()\n # Check if conference given in the websafeConferenceKey exists\n wsck = request.websafeConferenceKey\n conf = _getEntityByWebsafeKey(wsck, 'Conference')\n # Register\n if reg:\n # Check if user already registered, otherwise add\n if wsck in prof.conferenceKeysToAttend:\n raise ConflictException(\n \"You have already registered for this conference.\")\n # Check if seats available\n if conf.seatsAvailable <= 0:\n raise ConflictException(\n \"There are no seats available.\")\n # Register user, deduct one seat\n prof.conferenceKeysToAttend.append(wsck)\n conf.seatsAvailable -= 1\n retval = True\n # Unregister\n else:\n # Check if user already registered\n if wsck in prof.conferenceKeysToAttend:\n # Unregister user, add back one seat\n prof.conferenceKeysToAttend.remove(wsck)\n conf.seatsAvailable += 1\n retval = True\n else:\n retval = False\n # Update the datastore and return\n prof.put()\n conf.put()\n return BooleanMessage(data=retval)", "def register(self, voice=False):\n payload = {\"type\": \"register\", \"username\": self.username, \"voice\": voice}\n self._send_command(payload)", "def create_conference_session(self, request):\n user = endpoints.get_current_user()\n\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n\n self.session_service.create_conference_session(request, user)\n return request", "def register_user(self):\n User.add_user(User(self.email.data, self.password.data))", "async def register(ctx, *args):\n user = ctx.message.author\n user_mention = ctx.author.mention\n chan_mention = \"<#876850365730021386>\"\n \n if user in self.data[\"users.json\"]:\n await ctx.message.channel.send(user_mention+\", you are already registered. :blue_heart:\")\n else:\n self.data[\"users_asked_to_be_registered.json\"].append(user)\n await ctx.message.channel.send(user_mention+\", do you accept the \"+chan_mention+\n \" (Indie Library Terms of Service). Command .accept if you do. :blue_heart:\")", "def register(self, form):\n new_user = self.create_inactive_user(form)\n signals.user_registered.send(\n sender=self.__class__, user=new_user, request=self.request\n )\n return new_user", "def register():\r\n form = RegisterForm(request.form)\r\n\r\n if request.method == 'POST' and form.validate():\r\n new_user = User(form.email.data, form.password.data)\r\n g.session.add(new_user)\r\n g.session.commit()\r\n\r\n new_profile = Profile(form.first_name.data, form.last_name.data, new_user.id)\r\n g.session.add(new_profile)\r\n g.session.commit()\r\n # TODO: make it async\r\n if current_app.config[\"REQUIRE_EMAIL_CONFIRMATION\"]:\r\n send_confirmation(new_user)\r\n new_user.init_folders()\r\n logout_user()\r\n return redirect(url_for(\".login\"))\r\n return render_template(\"account/register_user.pug\", form=form)", "def signup(self, request, user):\n pass", "def register(user_name: str, password: str, age=20):\n auth.register(user_name, password)\n user_handler.register(user_name, age)", "def on_signup(self, data):\n self.participant_id = data[\"participant\"][\"id\"]", "def _register(self,user,project):\n url = reverse(\"comicsite.views._register\", \n kwargs={\"site_short_name\":project.short_name})\n factory = RequestFactory()\n request = factory.get(url)\n request.user = user\n self.apply_standard_middleware(request)\n \n response = _register(request,project.short_name)\n \n \n self.assertEqual(response.status_code,\n 200,\n \"After registering as user %s at '%s', page did not\"\n \" load properly\" % (user.username,url))\n \n self.assertTrue(project.is_participant(user),\n \"After registering as user %s at '%s', user does not \"\n \" appear to be registered.\" % (user.username,url))", "def set_user_register(self, *args, **kwargs):\n return _uhd_swig.usrp_sink_set_user_register(self, *args, **kwargs)", "def user_register():\n \n data = user_obj.user_register(request.forms) \n return data", "def register(self):\r\n if self.fields_not_empty(request, [\"first_name\", \"last_name\", \"age\", \"CPR\", \"email\", \"phone_number\", \"password\", \"confirm_password\"]):\r\n return jsonify({\"error\": \"Some fields are empty\"}), 400\r\n user = self.create_user_object(request)\r\n if request.form.get(\"password\") != request.form.get(\"confirm_password\"):\r\n return jsonify({\"error\": \"Passwords did not match\"}), 400\r\n db.insert_one(user)\r\n return self.start_session(user)", "def register(self, form):\n new_user = form.save(commit=False)\n username_field = getattr(new_user, 'USERNAME_FIELD', 'username')\n # Save lowercased email as username.\n setattr(new_user, username_field, form.cleaned_data['email'].lower())\n new_user.first_name = form.cleaned_data['first_name']\n new_user.last_name = form.cleaned_data['last_name']\n new_user.save()\n new_user = authenticate(username=getattr(new_user, username_field), password=form.cleaned_data['password1'])\n login(self.request, new_user)\n user_registered.send(sender=self.__class__, user=new_user, request=self.request)\n profile, _ = Profile.objects.get_or_create(user=new_user)\n self.request.session['signed_up'] = True\n profile.payment_plan = int(form.cleaned_data['payment_plan'])\n profile.company_name = form.cleaned_data['company']\n profile.phone = form.cleaned_data['phone']\n profile.save(update_fields=['payment_plan', 'company_name', 'phone'])\n if profile.payment_plan != Profile.PAYMENT_PLAN_FREE:\n messages.add_message(self.request, messages.INFO,\n 'Congratulations! We won\\'t charge you for this plan for now.')\n return new_user", "def signup():", "def register(session: Session, email: str, password: str, given_name: str, last_name: str,\n phone: str) -> RegisterResult:\n # Validate the data integrity of the parameters\n if email is None:\n return RegisterResult.BAD_USERNAME\n if password is None or not passwordService.validate(password):\n return RegisterResult.BAD_PASSWORD\n\n # Check to see if the user already exists\n existing_user = session.query(User) \\\n .filter(User.email == email) \\\n .first()\n if existing_user is not None:\n return RegisterResult.USERNAME_ALREADY_REGISTERED\n\n # Everything seems fine, so we go ahead and create the user & the linked account.\n password_hash = passwordService.hash(password)\n new_user = User(role=UserType.VOLUNTEER, password=password_hash, first_name=given_name, last_name=last_name,\n mobile_number=phone, email=email, preferred_hours={}, experience_years=0, possibleRoles=[\"Basic\"],\n qualifications=[],\n availabilities={\"Friday\": [], \"Monday\": [], \"Sunday\": [], \"Tuesday\": [], \"Saturday\": [],\n \"Thursday\": [], \"Wednesday\": []})\n session.add(new_user)\n session.flush()\n return RegisterResult.SUCCESS", "def _register_user(request_form):\n idnr = request_form['idnr']\n\n if user_exists(idnr):\n raise UserAlreadyExistsError(idnr)\n\n response = elster_client.send_unlock_code_request_with_elster(request_form, request.remote_addr)\n request_id = escape(response['elster_request_id'])\n\n create_user(idnr, request_form['dob'].strftime(\"%d.%m.%Y\"), request_id)", "def register():\r\n form = RegistrationUserForm()\r\n if form.validate_on_submit():\r\n user = User(\r\n first_name=form.first_name.data,\r\n last_name=form.last_name.data,\r\n email=form.email.data,\r\n password=form.password.data)\r\n db.session.add(user)\r\n db.session.commit()\r\n professor = Professor(\r\n user_id=user.id,\r\n first_name=form.first_name.data,\r\n last_name=form.last_name.data,\r\n university=form.university.data\r\n )\r\n db.session.add(professor)\r\n db.session.commit()\r\n token = user.generate_confirmation_token()\r\n confirm_link = url_for('account.confirm', token=token, _external=True)\r\n get_queue().enqueue(\r\n send_email,\r\n recipient=user.email,\r\n subject='Confirm Your Account',\r\n template='account/email/confirm',\r\n user=user,\r\n confirm_link=confirm_link)\r\n flash('A confirmation link has been sent to {}.'.format(user.email),\r\n 'warning')\r\n return redirect(url_for('main.index'))\r\n return render_template('professor/signup.html', form=form)", "def register_user(request, extra_context=None):\r\n if request.user.is_authenticated():\r\n return redirect(reverse('dashboard'))\r\n if settings.FEATURES.get('AUTH_USE_CERTIFICATES_IMMEDIATE_SIGNUP'):\r\n # Redirect to branding to process their certificate if SSL is enabled\r\n # and registration is disabled.\r\n return external_auth.views.redirect_with_get('root', request.GET)\r\n\r\n context = {\r\n 'course_id': request.GET.get('course_id'),\r\n 'email': '',\r\n 'enrollment_action': request.GET.get('enrollment_action'),\r\n 'name': '',\r\n 'running_pipeline': None,\r\n 'platform_name': microsite.get_value(\r\n 'platform_name',\r\n settings.PLATFORM_NAME\r\n ),\r\n 'selected_provider': '',\r\n 'username': '',\r\n }\r\n\r\n if extra_context is not None:\r\n context.update(extra_context)\r\n\r\n if context.get(\"extauth_domain\", '').startswith(external_auth.views.SHIBBOLETH_DOMAIN_PREFIX):\r\n return render_to_response('register-shib.html', context)\r\n\r\n # If third-party auth is enabled, prepopulate the form with data from the\r\n # selected provider.\r\n if settings.FEATURES.get('ENABLE_THIRD_PARTY_AUTH') and pipeline.running(request):\r\n running_pipeline = pipeline.get(request)\r\n current_provider = provider.Registry.get_by_backend_name(running_pipeline.get('backend'))\r\n overrides = current_provider.get_register_form_data(running_pipeline.get('kwargs'))\r\n overrides['running_pipeline'] = running_pipeline\r\n overrides['selected_provider'] = current_provider.NAME\r\n context.update(overrides)\r\n\r\n return render_to_response('register.html', context)", "def register_user(self):\n response = self.client.post(self.register_url, self.register_data, format='json')\n return response", "def set_user_register(self, *args, **kwargs):\n return _uhd_swig.usrp_sink_sptr_set_user_register(self, *args, **kwargs)", "def event_register_participant(request, event_id, participant_id):\n if request.method == 'PUT':\n event = get_object_or_404(Event, pk=event_id, is_registration_open=True)\n participant = get_object_or_404(Participant, pk=participant_id)\n try:\n attendance = Attendance(participant=participant, event=event, datetime_register=datetime.now())\n attendance.save()\n except Exception as e:\n print e\n content = {'detail': config.PARTICIPANT_ALREADY_REGISTERED_TO_EVENT}\n return Response(content, status=status.HTTP_400_BAD_REQUEST)\n serializer = AttendanceSerializer(attendance)\n return Response(serializer.data, status=status.HTTP_202_ACCEPTED)", "def register():\n response = jsonify(msg='unauthorized'), 400\n if request.method == 'POST' and request.get_json():\n data = request.get_json()\n valid = validator.user_registration(data)\n\n if valid['status']:\n ecomap_user.register(data['first_name'],\n data['last_name'],\n data['email'],\n data['password'])\n msg = 'added %s %s' % (data['first_name'],\n data['last_name'])\n response = jsonify({'status_message': msg}), 201\n else:\n response = Response(json.dumps(valid),\n mimetype='application/json'), 400\n return response", "def create_candidate(self, data, header):\n return self.client.post(\n path='/api/v2/office/1/register/', data=json.dumps(data), content_type='application/json', headers=header)", "def register_ctf_user(self, username):\n self.browser.get(self.get_server_url())\n player_box = self.browser.find_element_by_id(\"ctf-player\")\n player_box.send_keys(username)\n player_box.send_keys(Keys.ENTER)", "def add_user(params, batch_registration=False):\n # Make sure the username is unique\n db = api.db.get_conn()\n if is_blacklisted_username(params[\"username\"]) or db.users.find_one(\n {\"username\": params[\"username\"]},\n collation=Collation(locale=\"en\", strength=CollationStrength.PRIMARY),\n ):\n raise PicoException(\"There is already a user with this username.\", 409)\n if db.teams.find_one(\n {\"team_name\": params[\"username\"]},\n collation=Collation(locale=\"en\", strength=CollationStrength.PRIMARY),\n ):\n raise PicoException(\"There is already a team with this username.\", 409)\n\n # If gid is specified, force affiliation to that team's name\n email_whitelist = None\n if params.get(\"gid\", None):\n group = api.group.get_group(gid=params[\"gid\"])\n group_settings = api.group.get_group_settings(gid=group[\"gid\"])\n params[\"affiliation\"] = group[\"name\"]\n email_whitelist = group_settings[\"email_filter\"]\n\n # If rid is specified and gid and email match,\n # get teacher status from registration token.\n # Additionally, invited users are automatically validated.\n user_is_teacher = params[\"usertype\"] == \"teacher\"\n user_was_invited = False\n join_group_as_teacher = False\n if params.get(\"rid\", None):\n key = api.token.find_key_by_token(\"registration_token\", params[\"rid\"])\n if params.get(\"gid\") != key[\"gid\"]:\n raise PicoException(\n \"Registration token group and supplied gid do not match.\"\n )\n if params[\"email\"] != key[\"email\"]:\n raise PicoException(\n \"Registration token email does not match the supplied one.\"\n )\n join_group_as_teacher = key[\"teacher\"]\n user_was_invited = True\n api.token.delete_token(key, \"registration_token\")\n\n # If not invited, validate the user's email against the whitelist\n else:\n if not verify_email_in_whitelist(params[\"email\"], email_whitelist):\n raise PicoException(\n \"Your email does not belong to the whitelist. \"\n + \"Please see the registration form for details.\"\n )\n\n # If CAPTCHAs are enabled, validate the submission if not batch registration\n if (\n api.config.get_settings()[\"captcha\"][\"enable_captcha\"]\n and not batch_registration\n and not _validate_captcha(params)\n ):\n raise PicoException(\"Incorrect captcha!\")\n\n # Create a team for the new user and set its count to 1\n tid = api.team.create_team(\n {\n \"team_name\": params[\"username\"],\n \"password\": api.common.hash_password(\"-\"),\n \"affiliation\": params[\"affiliation\"],\n }\n )\n db.teams.update_one({\"tid\": tid}, {\"$set\": {\"size\": 1}})\n\n # The first registered user automatically becomes an admin\n user_is_admin = False\n if db.users.count() == 0:\n user_is_admin = True\n user_is_teacher = True\n\n # Insert the new user in the DB\n uid = api.common.token()\n settings = api.config.get_settings()\n user = {\n \"uid\": uid,\n \"firstname\": params[\"firstname\"],\n \"lastname\": params[\"lastname\"],\n \"username\": params[\"username\"],\n \"email\": params[\"email\"],\n \"password_hash\": api.common.hash_password(params[\"password\"]),\n \"tid\": tid,\n \"usertype\": params[\"usertype\"],\n \"country\": params[\"country\"],\n \"demo\": params[\"demo\"],\n \"teacher\": user_is_teacher,\n \"admin\": user_is_admin,\n \"disabled\": False,\n \"verified\": (not settings[\"email\"][\"email_verification\"] or user_was_invited),\n \"extdata\": {},\n \"completed_minigames\": [],\n \"unlocked_walkthroughs\": [],\n \"tokens\": 0,\n }\n db.users.insert_one(user)\n\n # Determine the user team's initial eligibilities\n initial_eligibilities = [\n scoreboard[\"sid\"]\n for scoreboard in api.scoreboards.get_all_scoreboards()\n if api.scoreboards.is_eligible(user, scoreboard)\n ]\n db.teams.find_one_and_update(\n {\"tid\": tid}, {\"$set\": {\"eligibilities\": initial_eligibilities}}\n )\n\n # If gid was specified, add the newly created team to the group\n if params.get(\"gid\", None):\n api.group.join_group(params[\"gid\"], tid, teacher=join_group_as_teacher)\n\n # If email verification is enabled and user wasn't invited, send\n # validation email\n if settings[\"email\"][\"email_verification\"] and not user_was_invited:\n api.email.send_user_verification_email(params[\"username\"])\n\n return uid" ]
[ "0.7278912", "0.6624827", "0.6584575", "0.65791625", "0.6362521", "0.61771995", "0.61010563", "0.6078329", "0.6028847", "0.60168606", "0.6004366", "0.60029817", "0.59709775", "0.5970334", "0.5965763", "0.59528124", "0.5946669", "0.5870501", "0.5858669", "0.58469427", "0.58424914", "0.5842267", "0.58414584", "0.58252496", "0.58072853", "0.5782229", "0.5769551", "0.5767778", "0.57532537", "0.5740037" ]
0.70775634
1
Unregister user for selected conference.
def unregisterFromConference(self, request): return self._conferenceRegistration(request, reg=False)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def unregister_from_conference(self, request):\n return self.conference_registration(request, reg=False)", "async def unregister(self):\n\t\tif self.group != None:\n\t\t\tif self.group.in_game:\n\t\t\t\tfor team in self.group.game.teams:\n\t\t\t\t\tif self in team:\n\t\t\t\t\t\tself.group.game.teams.remove(team)\n\t\t\t\t\t\tbreak\n\n\t\t\tawait self.group.remove(self)\n\n\t\tshared.users.remove(self)", "def unregister_user(self, userID: str):\n requests.post('https://' + self.serverIp + '/unregister/' + userID, verify=False)", "def unregister(self):\n if self.hub.is_connected:\n self._is_registered = False\n self.hub.unregister(self._private_key)\n self._hub_id = None\n self._public_id = None\n self._private_key = None\n else:\n raise SAMPClientError(\n \"Unable to unregister from the SAMP Hub. Hub proxy not connected.\"\n )", "def unregister(self, user_id: int, client_name: str) -> None:\n with self.app.app_context():\n user: User = User.query.get(user_id)\n\n if not self.contains(user_id, client_name):\n raise ValueError(f'User {user.username!r} has no associated client '\n f'named {client_name!r}')\n\n remote_addr = self.clients[user_id][client_name].protocol.remote_address\n self.clients[user_id][client_name].protocol.close()\n del self.clients[user_id][client_name]\n logger.info(f'Unregistered client {client_name!r} of user '\n f'{user.username!r} ({util.format_addr(remote_addr)})')", "def unsubscribe(self):\r\n self._unregister()", "def sipserver_user_remove(self, user: str) -> None:\n self.remove_endpoint_from_sipserver(endpoint=user)", "def destroy(self):\r\n for user in self._users.copy():\r\n user.destroy()\r\n\r\n assert len(self._users) == 0\r\n\r\n self._interface.unregisterConnection(self)\r\n self._interface = None\r\n\r\n self._protocol.unregisterConnection(self)\r\n self._protocol = None", "async def unregister(websocket):\n app['websockets'].discard(websocket)\n await notify_users()", "def remove_user(self):\n self.currentuser = None\n self.carlocked = False", "async def unregister(websocket):\n USERS.remove(websocket)\n GAME.new_game()\n await update_players()", "def logout_user(session):\n del session['user']", "def event_unregister_participant(request, event_id, participant_id):\n if request.method == 'PUT':\n event = get_object_or_404(Event, pk=event_id, is_registration_open=True)\n participant = get_object_or_404(Participant, pk=participant_id)\n attendance = get_object_or_404(Attendance, participant=participant, event=event)\n attendance.delete()\n serializer = EventSimpleSerializer(event)\n return Response(serializer.data, status=status.HTTP_202_ACCEPTED)", "def unregister(self):\r\n self._unregister()", "def removeUserId(self, user_id):\n self.__register_user_ids.discard(user_id)", "def delete(self, event_id: int) -> Response:\n if current_user not in self.event.participants:\n return jsonify({\n \"status\": 404,\n \"message\": \"You are not registered as a participant\"\n })\n else:\n self.event.participants.remove(current_user)\n self.event.save_to_db()\n\n return jsonify({\n \"status\": 200,\n \"message\": \"Successfully unregister from participants\"\n })", "def disconnect_user(room: PublicChatRoom, user) -> bool:\n return room.disconnect_user(user)", "async def logout(self):\n try:\n user = self.request.session.get(\"user\")\n chat = self.request.session.get(\"chat\")\n active_sockets = self.request.app.active_sockets\n active_sockets.get_chat(chat).del_user(user)\n\n self.request.session.pop(\"user\")\n self.request.user = None\n self.request.chat = None\n\n return {\n \"Type\": \"account\",\n \"Command\": \"logout\",\n \"Status\": \"success\"\n }\n except KeyError:\n return {\"Type\": \"account\", \"Command\": \"logout\", \"Status\": \"error\"}", "def delete_conference(request, pk):\n conference = ConferenceLog.objects.get(pk=pk)\n conference.delete()\n return redirect('readapi:index')", "def unsubscribe(cls,sender,receiver):\n cls._unsubscribe(id(sender),receiver)", "def unsubscribe(self, update, context):\n # remove or update to the sqlite table.\n chat = update.message.chat\n self.db_manager.remove_user(chat.id)\n self.logger.info(\n 'Username: %s and chat_id: %s unsubscribed to the list.' % (chat.username, chat.id)\n )\n update.message.reply_text('You have successfully unsubscribed the notifications forever.')", "def unregister(self):\n idaapi.unregister_action(self.get_name())", "def unsubscribe(self, request):\n email = self.cleaned_data.get('email')\n subscriber = Subscriber.objects.get(email=email, mailing_list=self.mailing_list)\n subscriber.unsubscribe(request)", "def unsubscribe(id, userId):\n db = core.connect()\n theUser = db[userId]\n if id in theUser[\"streams\"]:\n theUser[\"streams\"].remove(id)\n db[userId] = theUser", "def _unregister(self):\n try:\n self._profilemgr_proxy.proxy.UnregisterProfile(\n HFP_DBUS_PROFILE_ENDPOINT)\n logger.debug(\"Unregistered HFP profile.\")\n except Exception:\n logger.exception(\"Error unregistering profile endpoint.\")\n\n self._profile = None", "async def unsubscribe(self):\n LOGGER.info('Subscription removed')\n await self._ros.send(self._unsubscribe_msg)", "def RemoveSubscription(self, observedUser):\n\n self.__Delete(\"/subscriptions/\"+observedUser)", "def unregister(self,listener):\n\t\tif listener in self.cartesianListeners:\n\t\t\tself.cartesianListeners.remove(listener)\n\t\tif listener in self.polarListeners:\n\t\t\tself.polarListeners.remove(listener)", "def unsubscribe(self, jid=None, username=None, domain=None):\n if jid is not None:\n self._pres_manager.flag_offline(jid)\n elif username is not None and domain is not None:\n self._pres_manager.flag_offline(xmpp.JID(node=username, domain=domain))\n self._roster.unsubscribe(jid=jid, username=username, domain=domain)", "def unsubscribe(self, inst):\r\n if inst in self._subscribers:\r\n self._subscribers.remove(inst)\r\n vprint(\"{} is unsubscribed from {}\".format(inst.name, self.name))" ]
[ "0.7598682", "0.6543255", "0.6435641", "0.63821507", "0.63548124", "0.63164395", "0.61379206", "0.6102971", "0.60985017", "0.6050816", "0.59991133", "0.597415", "0.59388", "0.5870086", "0.5869279", "0.58567876", "0.5786231", "0.57730377", "0.5772766", "0.5762881", "0.5759947", "0.57500327", "0.5675999", "0.5674756", "0.5643517", "0.5638951", "0.5626867", "0.5609843", "0.5607273", "0.5598384" ]
0.7392909
1
Copy relevant fields from Speaker to SpeakerForm.
def _copySpeakerToForm(self, speaker): sf = SpeakerForm() for field in sf.all_fields(): if hasattr(speaker, field.name): setattr(sf, field.name, getattr(speaker, field.name)) elif field.name == "websafeKey": setattr(sf, field.name, speaker.key.urlsafe()) sf.check_initialized() return sf
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def copyform(form, settings):\r\n for name, value in form.iteritems():\r\n setattr(settings, name, value)\r\n settings.commit()", "def _copySessionToForm(self, session):\n sf = SessionForm()\n for field in sf.all_fields():\n if hasattr(session, field.name):\n # Convert date field to date string\n if field.name == 'date':\n setattr(sf, field.name, str(getattr(session, field.name)))\n # Convert time field to time string\n elif field.name == 'startTime':\n setattr(sf, field.name,\n getattr(session, field.name).strftime('%H:%M'))\n # Convert typeOfSession string field to enum\n elif field.name == 'typeOfSession':\n setattr(sf, field.name,\n getattr(SessionType, getattr(session, field.name)))\n # Copy other fields verbatim\n else:\n setattr(sf, field.name, getattr(session, field.name))\n # Ensure that the SessionForm contains websafeKey\n elif field.name == \"websafeKey\":\n setattr(sf, field.name, session.key.urlsafe())\n sf.check_initialized()\n return sf", "def _copyProfileToForm(self, prof):\n pf = ProfileForm()\n for field in pf.all_fields():\n if hasattr(prof, field.name):\n # Convert t-shirt string to Enum; just copy others\n if field.name == 'teeShirtSize':\n setattr(pf, field.name,\n getattr(TeeShirtSize, getattr(prof, field.name)))\n else:\n setattr(pf, field.name, getattr(prof, field.name))\n pf.check_initialized()\n return pf", "def _copyConferenceToForm(self, conf, displayName):\n cf = ConferenceForm()\n for field in cf.all_fields():\n if hasattr(conf, field.name):\n # Convert Date to date string; just copy others\n if field.name.endswith('Date'):\n setattr(cf, field.name, str(getattr(conf, field.name)))\n else:\n setattr(cf, field.name, getattr(conf, field.name))\n elif field.name == \"websafeKey\":\n setattr(cf, field.name, conf.key.urlsafe())\n if displayName:\n setattr(cf, 'organizerDisplayName', displayName)\n cf.check_initialized()\n return cf", "def _copyConferenceToForm(self, conf, displayName):\n cf = ConferenceForm()\n for field in cf.all_fields():\n if hasattr(conf, field.name):\n # convert Date to date string; just copy others\n if field.name.endswith('Date'):\n setattr(cf, field.name, str(getattr(conf, field.name)))\n else:\n setattr(cf, field.name, getattr(conf, field.name))\n elif field.name == \"websafeKey\":\n setattr(cf, field.name, conf.key.urlsafe())\n if displayName:\n setattr(cf, 'organizerDisplayName', displayName)\n cf.check_initialized()\n return cf", "def prepare(self, form):\n \n return form", "def individual_formfields():\n # Instantiate Consent Tracker\n consent = s3db.auth_Consent(processing_types = VOL_CONSENT_OPTIONS)\n\n formfields = [utable.first_name,\n utable.last_name,\n Field(\"addr_L3\",\n label = T(\"Location\"),\n requires = IS_IN_SET(districts_and_uk),\n ),\n Field(\"addr_street\",\n label = T(\"Street Address\"),\n ),\n Field(\"addr_postcode\",\n label = T(\"Postcode\"),\n ),\n Field(\"mobile\",\n label = T(\"Contact Number (Preferred)\"),\n requires = IS_PHONE_NUMBER_MULTI(),\n comment = DIV(_class = \"tooltip\",\n _title = \"%s|%s\" % (T(\"Contact Number (Preferred)\"),\n T(\"Ideally a Mobile Number, so that we can send you Text Messages.\")),\n ),\n ),\n Field(\"home\",\n label = T(\"Contact Number (Secondary)\"),\n requires = IS_EMPTY_OR(IS_PHONE_NUMBER_MULTI()),\n ),\n utable.email,\n utable[passfield],\n # Password Verification Field\n Field(\"password_two\", \"password\",\n label = auth_messages.verify_password,\n requires = IS_EXPR(\"value==%s\" % \\\n repr(request.vars.get(passfield)),\n error_message = auth_messages.mismatched_password,\n ),\n ),\n\n # Skills\n s3db.hrm_multi_skill_id(empty = False,\n label = T(\"Volunteer Offer\"),\n ),\n Field(\"skills_details\",\n label = T(\"Please specify details\"),\n ),\n Field(\"certificates\", \"list:string\",\n label = T(\"Qualifications\"),\n requires = IS_IN_SET(certificates, multiple=True),\n widget = S3MultiSelectWidget(header=\"\",\n selectedList=3),\n ),\n Field(\"experience\",\n label = T(\"Skills and Experience\"),\n widget = lambda f, v: \\\n s3_comments_widget(f, v, _placeholder = \"e.g. Co-ordination, Event Management, PCV qualified.\")\n ),\n Field(\"resources\",\n label = T(\"Offers of Resources\"),\n widget = lambda f, v: \\\n s3_comments_widget(f, v, _placeholder = \"e.g. Minibus.\")\n ),\n Field(\"where_operate\", \"list:string\",\n label = T(\"Where would you be willing to volunteer?\"),\n requires = IS_IN_SET(districts, multiple=True),\n widget = S3MultiSelectWidget(header=\"\",\n selectedList=3),\n ),\n Field(\"travel\", \"integer\",\n label = T(\"Willing to Travel?\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"slots\", \"list:string\",\n label = T(\"Times\"),\n requires = IS_IN_SET(slots, multiple=True),\n widget = S3MultiSelectWidget(header=\"\",\n selectedList=3),\n ),\n Field(\"significant_physical\", \"integer\",\n label = T(\"That require significant physical activity (including lifting and carrying) and may involve being outdoors (e.g. clean up of affected properties)\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"some_physical\", \"integer\",\n label = T(\"That require some physical activity and may involve being outdoors (e.g. door knocking)\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"little_physical\", \"integer\",\n label = T(\"That require little physical activity and are based indoors (e.g. preparing refreshments)\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"health_details\",\n label = T(\"If you wish, you can give us some further information on any fitness, medical or mobility issues that might limit the kind of activities you are able to volunteer for; this will help us to suggest suitable opportunities for you\"),\n ),\n Field(\"faith_requirements\", \"integer\",\n label = T(\"Do you have any faith requirements that you would like help with if you are coming to Support Cumbria?\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"faith_requirements_details\",\n label = T(\"If Yes please outline\"),\n ),\n Field(\"emergency_contact_name\",\n label = T(\"Contact Name\"),\n requires = IS_NOT_EMPTY(),\n ),\n Field(\"emergency_contact_number\",\n label = T(\"Contact Number\"),\n requires = IS_PHONE_NUMBER_MULTI(),\n ),\n Field(\"emergency_contact_relationship\",\n label = T(\"Relationship\"),\n requires = IS_NOT_EMPTY(),\n ),\n Field(\"workplace\", \"integer\",\n label = T(\"Are you volunteering under your workplace volunteering scheme?\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n Field(\"workplace_details\",\n label = T(\"If yes please name your employer\"),\n ),\n Field(\"dbs\", \"integer\",\n label = T(\"Are you DBS checked?\"),\n requires = IS_IN_SET({0: T(\"No\"),\n 1: T(\"Yes\"),\n }),\n widget = lambda f, v: \\\n SQLFORM.widgets.radio.widget(f, v,\n style=\"divs\"),\n ),\n #Field(\"convictions\", \"integer\",\n # label = T(\"Do you have any unspent convictions?\"),\n # comment = T(\"Please tick 'Yes' if you have any convictions that are not yet spent under the Rehabilitation of Offenders Act 1974. The term 'convictions' is used to refer to any sentence or disposal issued by a court. If all your convictions are spent, you can tick 'No'. If you're not sure if your convictions are unspent or spent, you can use a tool available at www.disclosurecalculator.org.uk and read guidance at hub.unlock.org.uk/roa\"),\n # requires = IS_IN_SET({0: T(\"No\"),\n # 1: T(\"Yes\"),\n # }),\n # widget = lambda f, v: \\\n # SQLFORM.widgets.radio.widget(f, v,\n # style=\"divs\"),\n # ),\n # Consent (GDPR + FOC)\n Field(\"consent\",\n label = T(\"Consent\"),\n widget = consent.widget,\n ),\n ]\n\n required_fields = [\"first_name\",\n \"last_name\",\n \"addr_L3\",\n \"addr_street\",\n \"addr_postcode\",\n \"mobile\",\n \"emergency_contact\",\n \"where_operate\",\n ]\n\n return formfields, required_fields", "def _createSpeakerObject(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = user.email()\n if not request.name:\n raise endpoints.BadRequestException(\n \"Speaker 'name' field required\")\n # Copy SpeakerForm/ProtoRPC Message into dict\n data = {\n field.name: getattr(request, field.name) for field in\n request.all_fields()\n }\n del data['websafeKey']\n # Add default values for those missing (both data model and\n # outbound Message)\n for df in SPEAKER_DEFAULTS:\n if data[df] in (None, []):\n data[df] = SPEAKER_DEFAULTS[df]\n # Create Speaker and return SpeakerForm\n speaker = Speaker(**data)\n speaker.put()\n return self._copySpeakerToForm(speaker)", "def populate_form(self, **kwargs):\n for name, value in kwargs.items():\n self.populate_field(name, value)", "def getSpeakers(self, request):\n speakers = Speaker.query().order(Speaker.name).fetch()\n # Return individual SpeakerForm object per Speaker\n return SpeakerForms(\n items=[self._copySpeakerToForm(speaker) for speaker in speakers]\n )", "def _replace_fields(self):\n for name, value in self._cleaned_data.items():\n setattr(self, name, value)", "def make_form(self):", "def getSpeaker(self, request):\n # Get Speaker object from request; bail if not found\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey,\n 'Speaker')\n # Return SpeakerForm\n return self._copySpeakerToForm(speaker)", "def make_fields(self):\n for name, prop in self.edit:\n instance_value = self.model.get(name)\n post_value = self.data[name] if (self.data and self.data.has_key(name)) else instance_value\n form_field_class = self.get_field_type(prop)\n form_field = form_field_class(model=self.model, property=prop, name=name, instance_value=instance_value, post_value=post_value)\n self.add(form_field)", "def __init__(self, *args, **kwargs):\n super(AddEventForm, self).__init__(*args)\n\n if kwargs.get('current_user') is not None:\n self.fields['speakers'].initial = kwargs.get('current_user')\n\n self.fields['speakers'].label_from_instance = self.label_from_instance", "def __init__(self, *args, **kwargs):\n # defaults to \"text\" and \"identity\"\n \n self.text_name = kwargs.pop('text_name', 'text')\n self.identity_name = kwargs.pop('identity_name', 'identity')\n \n super(EnvayaSMSIncomingForm, self).__init__(*args, **kwargs)\n \n #The following two fields are non-mandatory because actions other than 'incoming' won't have them.\n self.fields[self.text_name] = forms.CharField(required=False)\n self.fields[self.identity_name] = forms.CharField(required=False)\n\n self.fields['phone_number'] = forms.CharField() #which envaya phone forwarded us the msg?\n self.fields['action'] = forms.CharField() #what is the action?", "def form_tweaks(self):\n pass", "def form_valid(self, form):\n gamer = getGamer(self.request.user, self.kwargs['slugmaster'], self.kwargs['pkgamer'])\n form.instance.master = gamer.master\n form.instance.gamer = gamer\n\n return super(WordCreateView, self).form_valid(form)", "def save(self):\n if self.pumping_test_form:\n self.pumping_test_form.save()\n if self.form:\n if self.pumping_test_form:\n self.form.instance.pumping_test = self.pumping_test_form.instance\n self.form.save()\n self.well.hydrogeology_parameter = self.form.instance", "def create(self, validated_data):\n return Speaker.objects.create(**validated_data)", "def __fill_applicant_form(self, profile):\r\n\r\n actions = ActionChains(self.driver)\r\n actions.send_keys(profile['name'] + Keys.TAB + \\\r\n profile['email'] + Keys.TAB + \\\r\n profile['phone_number'] + Keys.TAB)\r\n actions.perform()", "def set_form(self, form):\n self.parameters = form", "def fillData(self):\n self.textname.SetValue(c.getUserName())\n self.textphon.SetValue(str(c.getPhone()))\n self.textcarr.SetValue(c.getCarrier())\n self.smsfin.SetValue(c.getSmsFinished())\n self.smserr.SetValue(c.getSmsError())", "def make_fields(self):\n #Let's first get fields in material_information printer_information\n metadata = GUI.api.get_metadata()\n field_correct_form = filter(lambda field: field['form_name']=='material_information' or field['form_name'] == 'printer_information', metadata)\n rows_w_fields = []\n for field in field_correct_form:\n #make label\n row = []\n key = field['field_name']\n type = field['field_type']\n row.append(sg.Text(text = field['field_label'], key=key+\"_label\"))#keys for labels are key_label (ex. record_id_label)\n if(type == 'radio' or type == \"dropdown\"):\n options = utils.get_options(field)\n row.append(sg.Combo(options, key=key, disabled= True, metadata=True, enable_events=True))\n elif(type == \"yesno\"):\n options = [\"Yes\", \"No\"]\n row.append(sg.Combo(options, key=key, disabled= True, metadata=True, enable_events=True))\n elif(type == \"text\"):\n row.append(sg.Input(key=key, disabled=True, metadata=True))\n else:#descirptive\n row[0] = sg.Text(text = field['field_label'], key=key, metadata=True)#we only need text in this case\n rows_w_fields.append(row)\n return rows_w_fields", "def test_form_has_fields(self):\r\n self.form = SubscriptionForm()\r\n expect = ['name', 'cpf', 'email', 'phone']\r\n self.assertSequenceEqual(expect, list(self.form.fields))", "def copy_form(self, from_module, form, to_module, rename=False):\n copy_source = deepcopy(form.to_json())\n # only one form can be a release notes form, so set them to False explicitly when copying\n copy_source['is_release_notes_form'] = False\n copy_source['enable_release_notes'] = False\n if 'unique_id' in copy_source:\n del copy_source['unique_id']\n\n if rename:\n for lang, name in copy_source['name'].items():\n with override(lang):\n copy_source['name'][lang] = _('Copy of {name}').format(name=name)\n\n copy_form = to_module.add_insert_form(from_module, FormBase.wrap(copy_source))\n to_app = to_module.get_app()\n save_xform(to_app, copy_form, form.source.encode('utf-8'))\n\n return copy_form", "def get_form_kwargs(self):\n kwargs = super(ProfileUpdate, self).get_form_kwargs()\n self.profile = self.get_object()\n extra_kwargs = {'profile': self.profile}\n kwargs.update(extra_kwargs)\n return kwargs", "def make_from_instance(instance):\n data = model_to_dict(instance)\n data['id'] = instance.id\n if instance.location:\n data['latitude'] = round(instance.location.y, 7)\n data['longitude'] = round(instance.location.x, 7)\n else:\n data['latitude'] = None\n data['longitude'] = None\n return GeneralInformationForm(\n initial=data\n )", "def duplicateSettings(self, otherField):\n self.name = otherField.name\n self.enName = otherField.enName\n self.format = otherField.format\n self.prefix = otherField.prefix\n self.suffix = otherField.suffix\n self.html = otherField.html\n self.isRequired = otherField.isRequired\n self.hidden = otherField.hidden\n self.numLines = otherField.numLines\n self.initDefault = otherField.initDefault\n self.linkAltField = otherField.linkAltField\n self.parentLevel = otherField.parentLevel\n self.useFileInfo = otherField.useFileInfo\n self.showInDialog = otherField.showInDialog", "def update_fields(self):\n if hasattr(self.day, \"body_composition\"):\n for f in self.get_fields():\n name = f.get_field().name\n value = getattr(self.day.body_composition, name, None)\n if value is not None:\n f.set_field(value)\n else:\n f.set_field(\"\")" ]
[ "0.5991479", "0.5911988", "0.58958936", "0.5770801", "0.57492906", "0.5703371", "0.56609124", "0.5593124", "0.5583714", "0.5578275", "0.55456966", "0.5397008", "0.53902537", "0.5387579", "0.5369967", "0.53496546", "0.5334173", "0.5330454", "0.530139", "0.5282616", "0.51898223", "0.51817733", "0.51368594", "0.5116736", "0.5098547", "0.50849473", "0.50633055", "0.50310576", "0.5029701", "0.5009965" ]
0.8053849
0
Create a speaker, returning SpeakerForm/request.
def _createSpeakerObject(self, request): # Preload necessary data items user = endpoints.get_current_user() if not user: raise endpoints.UnauthorizedException('Authorization required') user_id = user.email() if not request.name: raise endpoints.BadRequestException( "Speaker 'name' field required") # Copy SpeakerForm/ProtoRPC Message into dict data = { field.name: getattr(request, field.name) for field in request.all_fields() } del data['websafeKey'] # Add default values for those missing (both data model and # outbound Message) for df in SPEAKER_DEFAULTS: if data[df] in (None, []): data[df] = SPEAKER_DEFAULTS[df] # Create Speaker and return SpeakerForm speaker = Speaker(**data) speaker.put() return self._copySpeakerToForm(speaker)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create(self, validated_data):\n return Speaker.objects.create(**validated_data)", "def _copySpeakerToForm(self, speaker):\n sf = SpeakerForm()\n for field in sf.all_fields():\n if hasattr(speaker, field.name):\n setattr(sf, field.name, getattr(speaker, field.name))\n elif field.name == \"websafeKey\":\n setattr(sf, field.name, speaker.key.urlsafe())\n sf.check_initialized()\n return sf", "def getSpeaker(self, request):\n # Get Speaker object from request; bail if not found\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey,\n 'Speaker')\n # Return SpeakerForm\n return self._copySpeakerToForm(speaker)", "def create_speaker(conn, speaker):\n\n sql = ''' INSERT INTO speaker(name,gender,native)\n VALUES(?,?,?) '''\n cur = conn.cursor()\n cur.execute(sql, speaker)\n return cur.lastrowid", "def create_appointment():\n\n form = AppointmentForm()\n\n if form.validate_on_submit():\n\n appointment = Appointment(\n title = form.title.data,\n description = form.description.data,\n location = form.location.data,\n start = form.start.data,\n client = form.client.data,\n user = current_user\n )\n\n try:\n db.session.add(appointment)\n db.session.commit()\n\n flash('Successfully created the appointment.')\n\n return redirect(url_for('appointment.read_appointments'))\n except:\n flash('Error creating the appointment')\n\n return render_template('appointments/form.html.j2', form=form, title='Create appointment')", "def createRoom():\n\n global room_creation_params\n\n # get POST body\n profid = request.args.get(\"pid\")\n questions = request.args.get('q')\n\n # keep params for callback\n room_creation_params[\"profid\"] = profid\n room_creation_params[\"questions\"] = questions\n\n return redirect(stripe_test_URL) # redirect to stripe payment confirmation page", "def perform_create(self, serializer):\r\n serializer.save(user_type=\"SPEAKER\")", "def _createSessionObject(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = user.email()\n # Get the conference entity\n conf = _getEntityByWebsafeKey(request.websafeConferenceKey,\n 'Conference')\n # Ensure that the current user is the conference organizer\n if user_id != conf.organizerUserId:\n raise endpoints.UnauthorizedException(\n 'Only the conference organizer can create a new session')\n # Verify that the speaker exists\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey, 'Speaker')\n # Ensure that the user submitted the required name property\n if not request.name:\n raise endpoints.BadRequestException(\n \"Session 'name' field required\")\n # Copy SessionForm/ProtoRPC Message into dict\n data = {\n field.name: getattr(request, field.name) for field in\n request.all_fields()\n }\n # Remove data that isn't destined for the Session entity\n del data['websafeConferenceKey']\n del data['websafeSpeakerKey']\n del data['websafeKey']\n # Add default values for those missing in the data model\n for df in SESSION_DEFAULTS:\n if data[df] in (None, []):\n data[df] = SESSION_DEFAULTS[df]\n # Ensure the string version of typeOfSession is what is stored\n # in the NDB model\n data['typeOfSession'] = str(data['typeOfSession'])\n # Convert date from string to Date object\n if data['date'] is not None:\n try:\n data['date'] = datetime.strptime(\n data['date'][:10], '%Y-%m-%d').date()\n except:\n raise endpoints.BadRequestException(\n \"Invalid 'date' value\")\n # Convert startTime from string to Time object\n if data['startTime'] is not None:\n try:\n data['startTime'] = datetime.strptime(\n data['startTime'], '%H:%M').time()\n except:\n raise endpoints.BadRequestException(\n \"Invalid 'startTime' value\")\n # Create Session\n session = Session(**data)\n session.conference = conf.key\n session.speaker = speaker.key\n session.put()\n # Add the session key to the speaker's sessions list\n speaker.sessions.append(session.key)\n speaker.put()\n # Add a task to task queue which checks if the speaker of this session\n # should be the new featured speaker\n taskqueue.add(params={'websafeSpeakerKey': request.websafeSpeakerKey,\n 'websafeConferenceKey': request.websafeConferenceKey},\n url='/tasks/update_featured_speaker'\n )\n # Return SessionForm object\n return self._copySessionToForm(session)", "def post(self):\n s = ScuttlebuttService()\n try:\n topic_dict = simplejson.loads(self.request.body)\n topic = s.CreateTopic(topic_dict)\n self.response.headers['Content-Type'] = 'application/json'\n self.response.out.write(simplejson.dumps(topic.ToDict()))\n except simplejson.JSONDecodeError:\n # HTTP 400 for bad syntax.\n self.response.set_status(\n 400, 'Failed to create topic. Invalid JSON: %s' % self.request.body)\n except Exception, e:\n # HTTP 422 for syntactically correct but semantically wrong.\n self.response.set_status(422, 'Error creating topic: %s' % e)", "def post(self):\r\n data = request.form\r\n return create(data=data)", "def create(self, body):\n\t\tif self.has_permission('RightTPI') is False:\n\t\t\tself.no_access()\n\n\t\tid_survey = uuid.uuid4()\n\t\tid_language_content = MultiLang.set(body['name'], True)\n\n\t\twith Database() as db:\n\t\t\tdb.insert(Table(id_survey, id_language_content, body['survey_type']))\n\t\t\tdb.commit()\n\n\t\treturn {\n\t\t\t'id_survey': id_survey,\n\t\t\t'message': 'survey successfully created'\n\t\t}", "def create_participant(name='Not Brian', email='[email protected]') ->\\\n Participant:\n participant = Participant(name=name, email=email)\n return participant", "def create_subject(name=\"Basket Weaving\"):\n subj = Subject(name=name)\n subj.save()\n return subj", "def insert_talk(request):\n try:\n is_loggedin, username = get_session_variables(request)\n # User is not logged in\n if not logged_in(request):\n return HttpResponseRedirect('/register/login')\n\n # User is logged in\n else:\n if request.method == 'POST':\n form = AddSpeakerForm(request.POST)\n\n # Invalid form imput\n if not form.is_valid():\n error = \"Invalid inputs\"\n return render_to_response('achievement/new_speaker.html', \\\n {'form':form, \\\n 'error':error, \\\n 'is_loggedin':is_loggedin, \\\n 'username':username}, \\\n RequestContext(request))\n\n # Form is valid\n else:\n # Get the new achievement_id\n achievement_id = get_achievement_id(request)\t\n achievement_type = \"Speaker\"\n\n # Saving inputs\n achievement_obj = Achievement(achievement_id, \\\n achievement_type, \\\n username)\n achievement_obj.save()\n contribution_obj = form.save(commit = False)\n contribution_obj.achievement_id = achievement_obj\n contribution_obj.achieve_typ = achievement_type\n user_obj = get_object_or_404(User_info, username = username)\n contribution_obj.username = user_obj\n contribution_obj.save()\n return render_to_response('achievement/success.html', \\\n {'achievement_type':achievement_type, \\\n 'is_loggedin':is_loggedin, \\\n 'username':username}, \\\n RequestContext(request))\n # Method is not POST\n else:\n return render_to_response('achievement/new_speaker.html', \\\n {'form': AddSpeakerForm, \\\n 'is_loggedin':is_loggedin, \\\n 'username':username}, \\\n RequestContext(request))\n except KeyError:\n return error_key(request)", "def create_appointment():\n\n msg = render_template('date')\n return question(msg)", "def getSpeakers(self, request):\n speakers = Speaker.query().order(Speaker.name).fetch()\n # Return individual SpeakerForm object per Speaker\n return SpeakerForms(\n items=[self._copySpeakerToForm(speaker) for speaker in speakers]\n )", "def make_petition(request):\n form_class = SongPetitionForm\n template_name = 'petition/make_petition'\n\n if request.method == 'POST':\n petition = SongPetition(song_title=request.user)\n form = SongPetitionForm(request.POST, instance=petition)\n # If all the data has been filled in correctly it saves the petition into the DB\n if form.is_valid():\n form.save()\n # Feedback message telling that the petition has been created\n messages.success(request, \"Petition succesfully created\")\n return redirect ('petition:list')\n\n else:\n form = SongPetitionForm()\n return render(request, 'petition/make_petition.html', {'form': form})", "def _createConferenceObject(self, request):\n # preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = getUserId(user)\n\n if not request.name:\n raise endpoints.BadRequestException(\"Conference 'name' field required\")\n\n # copy ConferenceForm/ProtoRPC Message into dict\n data = {field.name: getattr(request, field.name) for field in request.all_fields()}\n del data['websafeKey']\n del data['organizerDisplayName']\n\n # add default values for those missing (both data model & outbound Message)\n for df in DEFAULTS:\n if data[df] in (None, []):\n data[df] = DEFAULTS[df]\n setattr(request, df, DEFAULTS[df])\n\n # convert dates from strings to Date objects; set month based on start_date\n if data['startDate']:\n data['startDate'] = datetime.strptime(data['startDate'][:10], \"%Y-%m-%d\").date()\n data['month'] = data['startDate'].month\n else:\n data['month'] = 0\n if data['endDate']:\n data['endDate'] = datetime.strptime(data['endDate'][:10], \"%Y-%m-%d\").date()\n\n # set seatsAvailable to be same as maxAttendees on creation\n if data[\"maxAttendees\"] > 0:\n data[\"seatsAvailable\"] = data[\"maxAttendees\"]\n # generate Profile Key based on user ID and Conference\n # ID based on Profile key get Conference key from ID\n p_key = ndb.Key(Profile, user_id)\n c_id = Conference.allocate_ids(size=1, parent=p_key)[0]\n c_key = ndb.Key(Conference, c_id, parent=p_key)\n data['key'] = c_key\n data['organizerUserId'] = request.organizerUserId = user_id\n\n # create Conference, send email to organizer confirming\n # creation of Conference & return (modified) ConferenceForm\n Conference(**data).put()\n taskqueue.add(params={'email': user.email(),\n 'conferenceInfo': repr(request)},\n url='/tasks/send_confirmation_email'\n )\n return request", "def create(self, validated_data):\n if (AdoptionRequest.objects\n .filter(adoption_proposal=validated_data['adoption_proposal'])\n .exists()):\n return (AdoptionRequest.objects\n .get(adoption_proposal=validated_data['adoption_proposal']))\n else:\n adoption_request = AdoptionRequest.objects.create(**validated_data)\n person_devices = find_devices(adoption_request.adoption_proposal\n .owner.id)\n title = 'SOSFIDO'\n message = ('Se ha enviado una petición a tu propuesta de ' +\n 'adopción de tu mascota ' +\n adoption_request.adoption_proposal.pet_name)\n data_notification = {'proposal_id': adoption_request\n .adoption_proposal.id}\n image_notification = 'https://s3.amazonaws.com/uploads.hipchat.' + \\\n 'com/529035/4489842/wDKOTjEIqyr1pPX/' + \\\n 'logo_sosfido1.png'\n send_notification_mobile(person_devices, title, message,\n data_notification, image_notification)\n return adoption_request", "def post():\n\n errors = check_petitions_keys2(request)\n if errors:\n return raise_error(400, \"Invalid {} key\".format(', '.join(errors)))\n details = request.get_json()\n createdBy = details['createdBy']\n office = details['office']\n body = details['body']\n\n if details['office'].isalpha() is False \\\n or details['createdBy'].isalpha() is False:\n return raise_error(400, \"input is in wrong format\")\n petition = PetitionsModel().save(createdBy, office, body)\n return make_response(jsonify({\n \"status\": \"201\",\n \"message\": \"petition filed successfully\",\n \"petition\": petition\n }), 201)", "def post(self):\n teacher = self.request.get(\"teacher\")\n temail = self.request.get(\"temail\")\n tphone = self.request.get(\"tphone\")\n specialty = self.request.get(\"specialty\")\n\n if teacher and temail and tphone and specialty:\n\n #create a new teacher object and store it in the database\n teacher = Teacher(\n teacher=teacher,\n temail=temail,\n tphone=tphone, \n specialty=specialty)\n teacher.put()\n\n id = teacher.key().id()\n self.redirect(\"/teacher/%s\" % id)\n else:\n error = \"Please include a teacher, an email, a phone number, and a specialty.\"\n self.render_form(teacher, temail, tphone, specialty, error)", "def create(self, validated_data):\n return Meet.objects.create(**validated_data)", "def create(self, validated_data):\n return Meet.objects.create(**validated_data)", "def participant_create(request):\n if request.method == 'POST':\n serializer = ParticipantSerializer(data=request.data)\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n else:\n participant = get_object_or_404(Participant, email=serializer.data['email'])\n serializer = ParticipantSerializer(participant, data=request.data, partial=True)\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data, status=status.HTTP_202_ACCEPTED)\n else:\n content = {'detail': config.PARTICIPANT_ALREADY_REGISTERED_OR_BAD_REQUEST}\n return Response(content, status=status.HTTP_400_BAD_REQUEST)", "def cmd_conversation_create(client, args):\n create_message = client.create_message(args.recipient, args.body)\n generate_output({'create_message': create_message})", "def add_speaker(self, name, audio_sample):\n if name in self.speakers:\n print(\"Error: %s is already a speaker\" % name)\n return\n \n self.speakers.append(name)\n\n features_left, features_right = self.extract_features(audio_sample)\n\n if self.both_channels:\n self.save_data(name, (features_left, features_right))\n else:\n self.save_data(name, (features_left))\n\n self.train_model()", "def post(self):\n try:\n body = json.loads(request.data.decode(\"utf-8\"))\n trainer = Trainer(body[\"firstName\"], body[\"lastName\"],\n body[\"email\"])\n trainer = TrainerService.create_trainer(trainer)\n return jsonify(trainer.json()), 201\n except ValueError:\n return INVALID_ID_ERROR, 400 # Bad Request\n except ResourceNotFound as r:\n return r.message, 404", "def create_session():\n app = Application.query.filter_by(id=request.json['app']).first()\n questionnaire = Questionnaire.query.filter_by(id=request.json['questionnaire']).first()\n expected_emotions = request.json['expected_emotions']\n\n # validate application type\n if not app:\n return {'status': 'error', 'message': 'Invalid application.'}, 400\n\n new_session = Session(app=app, expected_emotions=expected_emotions, questionnaire=questionnaire)\n\n db.session.add(new_session)\n db.session.commit()\n\n result = session_schema.dump(new_session).data\n\n return jsonify({'status': 'success', 'message': 'Created new session for application with id of {}.'.format(request.json['app']), 'data': result}), 201", "def create(self, *args, **kwargs):\n\n talk = kwargs.get('talk', None)\n comment = kwargs.get('comment', None)\n\n if talk is None and comment is None:\n raise ValidationError('Either talk or comment should not be null.')\n elif talk is not None and comment is not None:\n raise ValidationError('Either talk or comment should be null.')\n\n return super().create(*args, **kwargs)", "def create_teacher(username, password, email, preferred_language,skype_id,name, phone_number, country,availability):\n person.create_person(username,password,email,preferred_language,skype_id,name,phone_number,country)\n teacher_account_id = person.get_last()\n query = 'INSERT INTO teacher VALUES( %s,%s );'\n args = (teacher_account_id, availability)\n database.connection.save_data(query, args)" ]
[ "0.7516256", "0.6617961", "0.6517369", "0.60076547", "0.58922905", "0.58491045", "0.5713825", "0.57095647", "0.5580512", "0.5575898", "0.5537071", "0.54411596", "0.54135704", "0.54083437", "0.53903764", "0.5356573", "0.53268087", "0.5320126", "0.5311113", "0.53081363", "0.52962637", "0.52894694", "0.52894694", "0.52887094", "0.5267905", "0.52614486", "0.5255995", "0.5218763", "0.5172046", "0.51521814" ]
0.7869294
0
Check if the specified speaker is speaking at multiple sessions in the specified conference, and create memcache entry if so.
def _updateFeaturedSpeaker(websafeSpeakerKey, websafeConferenceKey): # Validate the websafe key arguments. Exception is raised if either # call fails. speaker = _getEntityByWebsafeKey(websafeSpeakerKey, 'Speaker') confKey = _raiseIfWebsafeKeyNotValid(websafeConferenceKey, 'Conference') # Get all sessions by the specified speaker at the specified # conference. Use a projection query, since the only information we're # interested in from the session entities is their name. sessionsBySpeaker = Session.query( Session.speaker == speaker.key, Session.conference == confKey ).fetch(projection=[Session.name]) # If there are fewer than two sessions, return immediately since # there is nothing left to do if len(sessionsBySpeaker) < 2: return # Put the session names into a list, alphabetically sessionNames = sorted([s.name for s in sessionsBySpeaker]) # Generate the featured speaker message featuredSpeakerMsg = ( 'Our featured speaker is {}, who will be speaking at the following ' 'sessions: {}'.format(speaker.name, ', '.join(sessionNames)) ) # Set the memcache entry to the new featured speaker message memcache.set(MEMCACHE_FEATURED_SPEAKER_KEY, featuredSpeakerMsg)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _getSessionsBySpeaker(self, request):\n # Ensure that the speaker key is valid and that the speaker exists\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey, 'Speaker')\n # Return all of the speaker's sessions\n return ndb.get_multi(speaker.sessions)", "def _getConferenceSessions(self, request):\n # Ensure that websafeConferenceKey is a valid conference key\n confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey,\n 'Conference')\n # Retrieve all sessions that have a matching conference key\n sessions = Session.query(Session.conference == confKey).fetch()\n return sessions", "def get_speaker_sessions(self, request):\n return self.session_service.get_speaker_sessions(\n request.websafeSpeakerKey)", "def speaker_is_occupied(self, session):\n self.occupied.append((session.starts_at, session.ends_at))", "def increment_participants(self):\n participants = memcache.incr(\"%s:participants\" % self.key_for_self(), initial_value=self.participants)\n\n if participants is None:\n # Memcache may be down and returning None for incr. Don't update the model in this case.\n return False\n\n self.participants = participants\n return True", "def makeSessions(self):\n \n sessionDict = {}\n self.sessions = []\n for sample in self.samples:\n session = sessionDict.get(sample.avId, None)\n if session:\n # Is this still the same session?\n elapsedA = sample.date - session.endDate\n assert elapsedA >= 0\n elapsedB = sample.timeInGame - session.timeInGame\n if elapsedB < 0:\n # Reported time-in-game smaller than last sample.\n session = None\n\n elif elapsedA > 1800 or elapsedB > 1800:\n # Too much time elapsed between consecutive\n # samples.\n session = None\n \n elif not session.sameHardware(sample):\n # New hardware.\n session = None\n \n if not session:\n session = Session(sample)\n sessionDict[sample.avId] = session\n self.sessions.append(session)\n else:\n session.addSample(sample)\n\n for session in self.sessions:\n session.calcFrameRate()", "def is_in_conference(self) -> bool:", "def get_conference_sessions(self, request):\n return self.session_service.get_conference_sessions(\n request.websafeConferenceKey)", "def add_talk(talk):\n # Check if this user is already registered\n exists = check_attendee_exists(talk.userId, talk.profile)\n if not exists[0]:\n return False\n\n talk.put()\n return True", "def cache_announcement():\n confs = Conference.query(ndb.AND(\n Conference.seatsAvailable <= 5, Conference.seatsAvailable > 0\n )).fetch(projection=[Conference.name])\n\n if confs:\n # If there are almost sold out conferences,\n # format announcement and set it in memcache\n announcement = ANNOUNCEMENT_TPL % (\n ', '.join(conf.name for conf in confs))\n memcache.set(MEMCACHE_ANNOUNCEMENTS_KEY, announcement)\n else:\n # If there are no sold out conferences,\n # delete the memcache announcements entry\n announcement = \"\"\n memcache.delete(MEMCACHE_ANNOUNCEMENTS_KEY)\n\n return announcement", "def _cacheAnnouncement():\n confs = Conference.query(ndb.AND(\n Conference.seatsAvailable <= 5,\n Conference.seatsAvailable > 0)\n ).fetch(projection=[Conference.name])\n if confs:\n # If there are conferences close to being sold out,\n # format announcement and set it in memcache\n announcement = '%s %s' % (\n 'Last chance to attend! The following conferences '\n 'are nearly sold out:',\n ', '.join(conf.name for conf in confs))\n memcache.set(MEMCACHE_ANNOUNCEMENTS_KEY, announcement)\n else:\n # If there are no sold out conferences,\n # delete the memcache announcements entry\n announcement = \"\"\n memcache.delete(MEMCACHE_ANNOUNCEMENTS_KEY)\n return announcement", "def _getConferenceSessionsByType(self, request):\n # Ensure that websafeConferenceKey is a valid conference key\n confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey,\n 'Conference')\n # Retrieve all sessions that have a matching conference key, by type\n sessions = Session.query(\n Session.conference == confKey,\n Session.typeOfSession == str(request.typeOfSession)\n ).fetch()\n return sessions", "async def get_conference_by_id(self, session_id):\n db = self['db_engine']\n async with db.acquire() as connection:\n query = select([Conference]).where(Conference.session_id == session_id).where(\n Conference.expired_at > datetime.now(timezone.utc))\n if conference := await(await connection.execute(query)).first():\n result = dict(conference)\n query = select([User.email, User.display_name]).where(User.id == conference.user_id)\n if user := await(await connection.execute(query)).first():\n result['user_data'] = UserData(conference.user_id, *user)\n return result", "def get_sessions_by_speaker_in_wishlist(self, user):\n sessions = self.wishlist_sessions(user)\n\n speaker_keys = []\n for s in sessions:\n sk = getattr(s, 'speakerKeys', [])\n speaker_keys += sk\n\n if not speaker_keys:\n return ConferenceSessionForms()\n\n sessions = ConferenceSession.query(\n ConferenceSession.speakerKeys.IN(speaker_keys)\n ).fetch()\n\n return ConferenceSessionForms(\n items=[self.copy_entity_to_form(ConferenceSessionForm(), s)\n for s in sessions])", "def write_speaker_to_coma(self, spktuple):\n for speaker in self.cparse.iter('Speaker'):\n try:\n if speaker.find('Sigle').text == spktuple[0]:\n #write all metadata to description tag\n desc = speaker.find('Description')\n\n for i in spktuple[1].iteritems():\n s = etree.SubElement(desc, 'Key', {'Name': i[0]})\n s.text = str(i[1])\n except:\n continue\n return", "def getSessionsBySpeaker(self, request):\n sessions = self._getSessionsBySpeaker(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def test_phonebook_with_duplicate_entries_is_inconsostent(self):\n self.phonebook.add(\"Bob\", \"12345\")\n self.phonebook.add(\"Mary\", \"12345\")\n self.assertFalse(self.phonebook.is_consistent())", "def _createSessionObject(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n user_id = user.email()\n # Get the conference entity\n conf = _getEntityByWebsafeKey(request.websafeConferenceKey,\n 'Conference')\n # Ensure that the current user is the conference organizer\n if user_id != conf.organizerUserId:\n raise endpoints.UnauthorizedException(\n 'Only the conference organizer can create a new session')\n # Verify that the speaker exists\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey, 'Speaker')\n # Ensure that the user submitted the required name property\n if not request.name:\n raise endpoints.BadRequestException(\n \"Session 'name' field required\")\n # Copy SessionForm/ProtoRPC Message into dict\n data = {\n field.name: getattr(request, field.name) for field in\n request.all_fields()\n }\n # Remove data that isn't destined for the Session entity\n del data['websafeConferenceKey']\n del data['websafeSpeakerKey']\n del data['websafeKey']\n # Add default values for those missing in the data model\n for df in SESSION_DEFAULTS:\n if data[df] in (None, []):\n data[df] = SESSION_DEFAULTS[df]\n # Ensure the string version of typeOfSession is what is stored\n # in the NDB model\n data['typeOfSession'] = str(data['typeOfSession'])\n # Convert date from string to Date object\n if data['date'] is not None:\n try:\n data['date'] = datetime.strptime(\n data['date'][:10], '%Y-%m-%d').date()\n except:\n raise endpoints.BadRequestException(\n \"Invalid 'date' value\")\n # Convert startTime from string to Time object\n if data['startTime'] is not None:\n try:\n data['startTime'] = datetime.strptime(\n data['startTime'], '%H:%M').time()\n except:\n raise endpoints.BadRequestException(\n \"Invalid 'startTime' value\")\n # Create Session\n session = Session(**data)\n session.conference = conf.key\n session.speaker = speaker.key\n session.put()\n # Add the session key to the speaker's sessions list\n speaker.sessions.append(session.key)\n speaker.put()\n # Add a task to task queue which checks if the speaker of this session\n # should be the new featured speaker\n taskqueue.add(params={'websafeSpeakerKey': request.websafeSpeakerKey,\n 'websafeConferenceKey': request.websafeConferenceKey},\n url='/tasks/update_featured_speaker'\n )\n # Return SessionForm object\n return self._copySessionToForm(session)", "def _check_conversation(self):\n _conversation_info = redis_hash_to_dict(self.application.redis, ConversationInfo, self._conversation_uuid)\n if _conversation_info == None:\n return (None, None, [])\n\n _key = ConversationUserData.__tablename__ + \".app_uuid.\" + self._app_uuid + \".user_uuid.\" + \\\n self._user_uuid + \".conversation_uuid.\" + self._conversation_uuid\n _conversation_user_data_uuid = self.application.redis.get(_key)\n if _conversation_user_data_uuid == None:\n return (_conversation_info, None, [])\n\n _conversation_user_data = redis_hash_to_dict(self.application.redis, ConversationUserData, _conversation_user_data_uuid)\n if _conversation_user_data == None:\n return (_conversation_info, None, [])\n\n _key = ConversationUserData.__tablename__ + \".conversation_uuid.\" + self._conversation_uuid\n _members = self.application.redis.smembers(_key)\n return (_conversation_info, _conversation_user_data, _members)", "def get_sessions_by_speakers_in_wishlist(self, request):\n user = endpoints.get_current_user()\n return self.wishlist_service.get_sessions_by_speaker_in_wishlist(\n user)", "def __init__(self):\n self._conferences = set()", "def cacheChats(self):\n logger.debug(\"Async cacheChats() -- this may take a while\")\n self.chats = OrderedDict()\n\n # First get all fresh chats\n chats = []\n for chat in self.skype.Chats:\n\n # filter chats older than 6 months\n if time.time() - chat.ActivityTimestamp > 3600 * 24 * 180:\n continue\n\n chats.append(chat)\n\n chats = sorted(chats, key=lambda c: c.ActivityTimestamp, reverse=True)\n\n for chat in chats:\n # Encode ids in b64 so they are easier to pass in URLs\n m = hashlib.md5()\n m.update(chat.Name)\n self.chats[m.hexdigest()] = chat", "def filter_by_session (self, site_id, participant_id, session_id):\n\n query = \"\"\"\n select distinct * where {\n\n BIND (\"%s\" AS ?pid)\n BIND (\"%s\" as ?sessionid)\n \n ?participant austalk:id ?pid .\n ?rc rdf:type austalk:RecordedComponent .\n ?rc olac:speaker ?participant .\n ?rc austalk:session ?sessionid .\n\n ?rc austalk:prototype ?prototype .\n ?prototype austalk:name ?name .\n ?prototype austalk:shortname ?shortname .\n \n optional { ?rc austalk:audiorating ?audiorating .}\n optional { ?rc austalk:videorating ?videorating .}\n optional { ?rc austalk:comment ?comment .}\n \n }\n \"\"\" % (participant_id, session_id)\n\n\n sparql_results = self.query (query)\n results = []\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n for field in ['audiorating', 'videorating', 'comment']:\n if field not in result:\n result[field] = {'value': ''}\n\n comp = Component (\n client = self.client,\n identifier = result[\"rc\"][\"value\"],\n participantId = result[\"pid\"][\"value\"],\n sessionId = result[\"sessionid\"][\"value\"],\n prototype = result[\"prototype\"][\"value\"],\n audiorating = result[\"audiorating\"][\"value\"],\n videorating = result[\"videorating\"][\"value\"],\n comment = result[\"comment\"][\"value\"],\n name = result[\"name\"][\"value\"],\n componentId = result[\"shortname\"][\"value\"],\n site = site_id,\n )\n comp.details()\n results.append(comp)\n return results", "def __init__(self, session_key):\n super(Talk, self).__init__()\n self.session_key = session_key\n self.tcp_clients = []\n self.participants = []\n if self.session_key in Talk.talk_sessions:\n raise NameError(\"There already exists a session with that number\")\n Talk.talk_sessions[self.session_key] = self\n self.action_queue = Queue.Queue()\n self.talk_queue = Queue.Queue()\n self.udp_server = None", "def filter_by_participant (self, participant):\n \n sparql_results = self.query (\"\"\"\n select distinct ?rs ?session ?name ?number ?pid ?sitename\n where {\n BIND (<%s> AS ?participant)\n \n ?rs rdf:type austalk:RecordedSession .\n ?rs olac:speaker ?participant .\n \n ?participant austalk:id ?pid .\n ?participant austalk:recording_site ?site .\n ?site rdfs:label ?sitename .\n \n ?rs austalk:prototype ?session .\n ?session austalk:name ?name .\n ?session austalk:id ?number .\n }\n ORDER BY ?name\"\"\" % participant.identifier)\n \n results = []\n\n for result in sparql_results[\"results\"][\"bindings\"]:\n results.append (Session (\n client = self.client,\n identifier = result[\"rs\"][\"value\"],\n prototype = result[\"session\"][\"value\"],\n name = result[\"name\"][\"value\"],\n number = result[\"number\"][\"value\"],\n site = result[\"sitename\"][\"value\"],\n participantId = result[\"pid\"][\"value\"]))\n\n return results", "def add_information_about_person(self, session_info):\n\n session_info = dict(session_info)\n name_id = session_info[\"name_id\"]\n issuer = session_info.pop(\"issuer\")\n self.cache.set(name_id, issuer, session_info, session_info[\"not_on_or_after\"])\n return name_id", "def _match_voice_wrapper(self, cluster, wav_name, db_entry, gender):\n results = self.get_db().match_voice(wav_name, db_entry, gender)\n for res in results:\n self[cluster].add_speaker(res, results[res])", "def _conferenceRegistration(self, request, reg=True):\n retval = None\n # Get user profile\n prof = self._getProfileFromUser()\n # Check if conference given in the websafeConferenceKey exists\n wsck = request.websafeConferenceKey\n conf = _getEntityByWebsafeKey(wsck, 'Conference')\n # Register\n if reg:\n # Check if user already registered, otherwise add\n if wsck in prof.conferenceKeysToAttend:\n raise ConflictException(\n \"You have already registered for this conference.\")\n # Check if seats available\n if conf.seatsAvailable <= 0:\n raise ConflictException(\n \"There are no seats available.\")\n # Register user, deduct one seat\n prof.conferenceKeysToAttend.append(wsck)\n conf.seatsAvailable -= 1\n retval = True\n # Unregister\n else:\n # Check if user already registered\n if wsck in prof.conferenceKeysToAttend:\n # Unregister user, add back one seat\n prof.conferenceKeysToAttend.remove(wsck)\n conf.seatsAvailable += 1\n retval = True\n else:\n retval = False\n # Update the datastore and return\n prof.put()\n conf.put()\n return BooleanMessage(data=retval)", "def conference_registration(self, request, reg=True):\n prof = self.profile_service.get_profile_from_user() # get user Profile\n\n # check if conf exists given websafeConfKey\n # get conference; check that it exists\n wsck = request.websafeConferenceKey\n conf = ndb.Key(urlsafe=wsck).get()\n if not conf:\n raise endpoints.NotFoundException(\n 'No conference found with key: %s' % wsck)\n\n # register\n if reg:\n # check if user already registered otherwise add\n if wsck in prof.conferenceKeysToAttend:\n raise ConflictException(\n \"You have already registered for this conference\")\n\n # check if seats avail\n if conf.seatsAvailable <= 0:\n raise ConflictException(\"There are no seats available.\")\n\n # register user, take away one seat\n prof.conferenceKeysToAttend.append(wsck)\n conf.seatsAvailable -= 1\n retval = True\n\n # unregister\n else:\n # check if user already registered\n if wsck in prof.conferenceKeysToAttend:\n\n # unregister user, add back one seat\n prof.conferenceKeysToAttend.remove(wsck)\n conf.seatsAvailable += 1\n retval = True\n else:\n retval = False\n\n # write things back to the datastore & return\n prof.put()\n conf.put()\n return BooleanMessage(data=retval)", "def _update_subscribers(self):\n try:\n campaign = self.campaigns.latest('when')\n except StudyGuideCampaign.DoesNotExist:\n pass\n else:\n for student in utils.students_for_event(self.event):\n subscriber, created = StudyGuideCampaignSubscriber.objects.get_or_create(\n campaign=campaign,\n user=student.user)\n if created: # only add if it's not there already\n campaign.subscribers.add(subscriber)" ]
[ "0.5446472", "0.51810056", "0.5084203", "0.5033835", "0.49242762", "0.48869962", "0.4859192", "0.48236132", "0.4786593", "0.47050533", "0.46819773", "0.46688312", "0.46350497", "0.45954737", "0.45568752", "0.4550118", "0.45501018", "0.4542919", "0.45304805", "0.452583", "0.45191735", "0.45109996", "0.44467556", "0.4433786", "0.44242808", "0.44226032", "0.44117713", "0.44101897", "0.4408384", "0.4404256" ]
0.56610495
0
Return the current featured speaker message from memcache.
def getFeaturedSpeaker(self, request): message = memcache.get(MEMCACHE_FEATURED_SPEAKER_KEY) or "" return StringMessage(data=message)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_featured_speaker(self, request):\n return StringMessage(\n data=memcache.get(MEMCACHE_FEATURED_SPEAKER_KEY) or \"\")", "def _updateFeaturedSpeaker(websafeSpeakerKey, websafeConferenceKey):\n # Validate the websafe key arguments. Exception is raised if either\n # call fails.\n speaker = _getEntityByWebsafeKey(websafeSpeakerKey, 'Speaker')\n confKey = _raiseIfWebsafeKeyNotValid(websafeConferenceKey,\n 'Conference')\n # Get all sessions by the specified speaker at the specified\n # conference. Use a projection query, since the only information we're\n # interested in from the session entities is their name.\n sessionsBySpeaker = Session.query(\n Session.speaker == speaker.key,\n Session.conference == confKey\n ).fetch(projection=[Session.name])\n # If there are fewer than two sessions, return immediately since\n # there is nothing left to do\n if len(sessionsBySpeaker) < 2:\n return\n # Put the session names into a list, alphabetically\n sessionNames = sorted([s.name for s in sessionsBySpeaker])\n # Generate the featured speaker message\n featuredSpeakerMsg = (\n 'Our featured speaker is {}, who will be speaking at the following '\n 'sessions: {}'.format(speaker.name, ', '.join(sessionNames))\n )\n # Set the memcache entry to the new featured speaker message\n memcache.set(MEMCACHE_FEATURED_SPEAKER_KEY, featuredSpeakerMsg)", "def get_speaker(self):\n return self._speaker", "def get_speaker(self):\n if self._speaker == None:\n self._speaker = self.get_best_speaker()\n return self._speaker", "def _cacheAnnouncement():\n confs = Conference.query(ndb.AND(\n Conference.seatsAvailable <= 5,\n Conference.seatsAvailable > 0)\n ).fetch(projection=[Conference.name])\n if confs:\n # If there are conferences close to being sold out,\n # format announcement and set it in memcache\n announcement = '%s %s' % (\n 'Last chance to attend! The following conferences '\n 'are nearly sold out:',\n ', '.join(conf.name for conf in confs))\n memcache.set(MEMCACHE_ANNOUNCEMENTS_KEY, announcement)\n else:\n # If there are no sold out conferences,\n # delete the memcache announcements entry\n announcement = \"\"\n memcache.delete(MEMCACHE_ANNOUNCEMENTS_KEY)\n return announcement", "def getAnnouncement(self, request):\n announcement = memcache.get(MEMCACHE_ANNOUNCEMENTS_KEY) or \"\"\n return StringMessage(data=announcement)", "def get_greetings(self):\n greetings = memcache.get(\"greetings\")\n if greetings is not None:\n return greetings\n else:\n greetings = self.render_greetings()\n if not memcache.add(\"greetings\", greetings, 10):\n logging.error(\"Memcache set failed.\")\n return greetings", "def get_incoming_message(self):\n\n if self.incoming_message['is_used']:\n return {}\n\n self.incoming_message['is_used'] = True\n return self.incoming_message['message']", "def cache_announcement():\n confs = Conference.query(ndb.AND(\n Conference.seatsAvailable <= 5, Conference.seatsAvailable > 0\n )).fetch(projection=[Conference.name])\n\n if confs:\n # If there are almost sold out conferences,\n # format announcement and set it in memcache\n announcement = ANNOUNCEMENT_TPL % (\n ', '.join(conf.name for conf in confs))\n memcache.set(MEMCACHE_ANNOUNCEMENTS_KEY, announcement)\n else:\n # If there are no sold out conferences,\n # delete the memcache announcements entry\n announcement = \"\"\n memcache.delete(MEMCACHE_ANNOUNCEMENTS_KEY)\n\n return announcement", "def get_announcement(self, request):\n return StringMessage(\n data=memcache.get(MEMCACHE_ANNOUNCEMENTS_KEY) or \"\")", "def get(self, id):\n\n\t\treturn MessageStore.get(id)", "def get(self):\n return self.sent.get()", "def memcacheGetRecord(self, key):\n\n pickled = self.memcacheGet(key)\n return self.unpickleRecord(pickled) if pickled is not None else None", "def get(self):\n CACHE_KEY = 'topics'\n if not memcache.get(CACHE_KEY):\n logging.info('Populating cache.')\n topics = Topic.all().order('name')\n topic_list = []\n for topic in topics:\n topic_list.append(topic.ToDict())\n memcache.add(CACHE_KEY, simplejson.dumps(topic_list), 600)\n logging.info('Using cache.')\n logging.info(memcache.get(CACHE_KEY))\n self.response.headers['Content-Type'] = 'application/json'\n self.response.out.write(memcache.get(CACHE_KEY))", "def get_discovery_message(self):\n return self.messages[\"discovery\"].get()", "def get_message(self, sender=None):\n if sender == None:\n if self.public_key == None:\n return None\n participant = self.public_key\n else:\n participant = sender\n following = [tx.follow for block in self.__chain for tx in block.chipsactions if tx.sender == participant] \n tx_recipient2 = [tx.message for block in self.__chain for tx in block.messsactions if tx.follower in following]\n print(\"tx_recipient2\")\n print(tx_recipient2)\n return tx_recipient2", "def get_message(self):\n return self.__mess", "def get(self) -> Optional[Message]:\n self._recv()\n if not self.inq:\n return None\n return self.inq.popleft()", "async def poll_message(self):\n message_cache = self.message_cache\n if (message_cache is not None) and message_cache:\n return message_cache.pop()\n \n if not self.message_request_more:\n return\n \n message_cache = await self.client.message_get_chunk(self.source_channel, after = self.last_message_id)\n self.message_cache = message_cache\n \n if len(message_cache) < 100:\n self.message_request_more = False\n \n if message_cache:\n return message_cache.pop()", "def pull(self):\n \n data = self.s.recv(1024)\n if data:\n info = json.loads(data.decode()) \n print(\"DATA FROM BROKER : \", info)\n \n return info.get(\"topic\"), info.get(\"value\")\n pass", "def get(self, msgid):\r\n return self.trans.get(msgid, str(msgid))", "def get(self):\n return {\"oldest_msg\": oldest_msg()}", "def _get_cached_response(self, message_id, from_node):\n return self.operator.get_message_item(message_id, from_node)", "def _get_message(self, sender_message):\n # type: (str) -> Message or None\n st_re = self.SENDER_TEXT.search(sender_message)\n if st_re is None:\n return None\n else:\n return Message(speaker=st_re.group(1), text=st_re.group(2).strip())", "def get(self):\n return {\"newest_msg\": newest_msg()}", "def _get_message(self):\n return self.__message", "def get_message(self):\n return self.message", "def get_message(self):\n return self.message", "def read_message(self, topic_name, offset):\n try:\n return self.topics[topic_name][offset]\n except Exception:\n return None", "def pop_message(self, queue_name):\r\n messages = self.pop_messages(queue_name, count=1)\r\n if messages['item_count'] > 0:\r\n return messages['items'][0]\r\n else:\r\n return None" ]
[ "0.7784476", "0.58692867", "0.58229965", "0.5695719", "0.56409705", "0.5568219", "0.54920876", "0.54669094", "0.5427401", "0.5424601", "0.5384481", "0.5369317", "0.5333994", "0.5280963", "0.5277826", "0.5231841", "0.5128428", "0.50919986", "0.50889385", "0.50869083", "0.5072661", "0.5071199", "0.5068012", "0.5047879", "0.50168663", "0.50050944", "0.4993946", "0.4993946", "0.49869344", "0.4978353" ]
0.7579103
1
Add a session to the user's wishlist, returning a boolean.
def _addSessionToWishlist(self, request): # Preload necessary data items user = endpoints.get_current_user() if not user: raise endpoints.UnauthorizedException('Authorization required') # Verify that the session actually exists session = _getEntityByWebsafeKey(request.websafeSessionKey, 'Session') profile = self._getProfileFromUser() if session.key not in profile.sessionWishlist: profile.sessionWishlist.append(session.key) profile.put() return BooleanMessage(data=True)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def add_session_to_wishlist(self, request):\n return self.wishlist_service.add_session_to_wishlist(\n request.websafeSessionKey, endpoints.get_current_user())", "def add_session_to_wishlist(self, websafe_session_key, user):\n wl_key = self.get_wishlist_key(user)\n\n wishlist = wl_key.get()\n\n if websafe_session_key in wishlist.sessionKeys:\n raise ConflictException(\n \"You already have this session in your wishlist.\")\n\n wishlist.sessionKeys.append(websafe_session_key)\n wishlist.put()\n\n return self.to_message(wishlist)", "def addSessionToWishlist(self, request):\n return self._addSessionToWishlist(request)", "def add_user_to_session(self,session_id,client_id,display_name):\n self.sessions[session_id][\"USERS\"][client_id] = {\n \"display_name\" :display_name,\n \"permissions\" : {\n \"add_to_queue\" : True,\n \"playback\" : True,\n \"skip\" : True,\n \"edit_queue\" : True\n }\n }", "def add_to_wish_list(self, user_id, caption):\n with self.connection:\n return self.cursor.execute(\"INSERT INTO 'wish_list' (`user_id`, `wish_list`) VALUES(?,?)\", (user_id,caption))", "def add_to_fav(show_id, name):\n db = get_db()\n db.execute(\n 'INSERT INTO shows_users (show_id, user_id)'\n ' VALUES (?, ?)',\n (show_id, session['user_id'])\n )\n\n flash('\\\"%s\\\" has been successfully added to your favourite TV Shows!' % name)\n db.commit()\n return redirect(request.referrer)", "def _removeSessionFromWishlist(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n profile = self._getProfileFromUser()\n # Get actual session key from websafe key\n sessionKey = _raiseIfWebsafeKeyNotValid(request.websafeSessionKey,\n 'Session')\n # If the key is in the profile's session wishlist, remove it\n if sessionKey in profile.sessionWishlist:\n profile.sessionWishlist.remove(sessionKey)\n profile.put()\n retval = True\n else:\n retval = False\n return BooleanMessage(data=retval)", "def add_to_wishlist(request, product_id):\n redirect_url = request.POST.get('redirect_url')\n\n user = get_object_or_404(UserProfile, user=request.user)\n wishlist = Wishlist.objects.get_or_create(user=user)\n wishlist_user = wishlist[0]\n\n product = Product.objects.get(pk=product_id)\n if request.POST:\n existingWishlistItem = WishlistItem.objects.filter(\n wishlist=wishlist_user, product=product).exists()\n if existingWishlistItem:\n messages.error(request, \"Item already in your wishlist\")\n return redirect(redirect_url)\n\n else:\n added_item = WishlistItem(\n wishlist=wishlist_user, product=product, date_added=timezone.now())\n added_item.save()\n messages.success(request, \"Product added to your wishlist\")\n return redirect(redirect_url)\n else:\n messages.error(request, \"Click 'Add to wishlist' to add a item \")\n return render(request, 'home/index.html')", "def add_favourite(recipe_id):\r\n if \"user\" in session:\r\n user = coll_users.find_one(\r\n {\"username_lower\": session[\"user\"]})[\"_id\"]\r\n coll_users.update_one(\r\n {\"_id\": ObjectId(user)},\r\n {\"$push\": {\"user_favs\": ObjectId(recipe_id)}})\r\n coll_recipes.update(\r\n {\"_id\": ObjectId(recipe_id)}, {\"$inc\": {\"favourites\": 1}})\r\n return redirect(url_for(\r\n \"recipes.recipe_detail\",\r\n recipe_id=recipe_id))\r\n else:\r\n flash(\"You must be logged in to perform that action!\")\r\n return redirect(url_for(\"users.login\"))", "def add_unsaved_likes_to_user(sender, user, request, **kwargs):\n session_likes = request.session.get('likes')\n if session_likes:\n user.userprofile.liked_products.add(*session_likes)", "def add_to_wishlist(request, product_id):\n product = get_object_or_404(Product, pk=product_id)\n wishlist = get_object_or_404(Wishlist, user=request.user)\n\n if product not in wishlist.products.all():\n wishlist.products.add(product)\n messages.info(request,\n f\"{product.name} has been added to your wishlist.\")\n else:\n messages.error(request,\n \"Error, you already have this item in your wishlist!\")\n return redirect(reverse(\"product_detail\", args=[product_id]))", "def connect_user(self, user):\n\t\tis_user_added = False\n\t\tif not user in self.users.all():\n\t\t\tself.users.add(user)\n\t\t\tself.save()\n\t\t\tis_user_added = True\n\t\telif user in self.users.all():\n\t\t\tis_user_added = True\n\t\treturn is_user_added", "def add_session(self, session):\n with self._sessions_lock:\n if session.session_id in self.sessions:\n raise KeyError(\"non-unique session id %s for %s\" % (session.session_id, session))\n self.sessions[session.session_id] = session\n\n return session", "def add_talk(talk):\n # Check if this user is already registered\n exists = check_attendee_exists(talk.userId, talk.profile)\n if not exists[0]:\n return False\n\n talk.put()\n return True", "def remove_session_from_wishlist(self, websafe_session_key, user):\n wishlist = self.get_wishlist_key(user).get()\n if wishlist is None or wishlist.sessionKeys is []:\n raise ConflictException(\"This session is not in your wishlist.\")\n\n if websafe_session_key not in wishlist.sessionKeys:\n raise ConflictException(\n \"This session is not in your wishlist.\")\n\n wishlist.sessionKeys.remove(websafe_session_key)\n wishlist.put()\n\n return self.to_message(wishlist)", "def add_session(self, session_id):\n sessions = self.get_sessions()\n if session_id not in sessions:\n sessions.append(session_id)\n self.ref_cache.set(self.sid, sessions)", "def get_sessions_in_wishlist(self, request):\n user = endpoints.get_current_user()\n return self.wishlist_service.get_sessions_in_wishlist(user)", "def remove_session_from_wishlist(self, request):\n return self.wishlist_service.remove_session_from_wishlist(\n request.websafeSessionKey, endpoints.get_current_user())", "def add_to_fav(request, q_id):\n if request.method == 'POST':\n Quotes.objects.add_to_user_fav(request.session['id'], q_id)\n return redirect('/quotes')", "def addsession_url_link(update, context):\n\t\n\turl_link = update.message.text\n\n\turl_link_valid = check_new_url(url_link)\n\tif (not url_link_valid):\n\t\tupdate.message.reply_text('Oops, your link is not a valid URL-link.\\n'\n\t\t'Please, retype the URL-link in a good format.')\n\t\treturn URL_LINK\n\n\tcontext.user_data['url-str'] = url_link\n\tupdate.message.reply_text('Thank you.\\n')\n\n\t# query to finalize adding new session\n\tkeyboard = [\n\t\t[ \n\t\t\tInlineKeyboardButton(\"Yes\", callback_data='Yes'), \n\t\t\tInlineKeyboardButton(\"No\", callback_data='No'),\n\t\t]\n\t,]\n\treply_markup = InlineKeyboardMarkup(keyboard)\n\n\tupdate.message.reply_text('So, you want to add a session with the following parameters, right?\\n\\n'\n\t\t'Date and time: {}\\nURL-link: {}\\n'.format(context.user_data['datetime-str'], context.user_data['url-str']),\n\t\treply_markup = reply_markup)\n\t\n\treturn PROCESS_NEW_SESSION", "def addsession(update, context):\n\tupdate.message.reply_text('Ok, for this I will need two items:\\n\\n'\n\t\t'1. Date and time of the registration opening (your local time)\\n' \n\t\t'2. URL-link to the session page\\n\\n' \n\t\t'You can always cancel the input of a new session by typing /cancel.')\n\n\tupdate.message.reply_text('Let\\'s start with the first one.\\n\\n' \n\t\t'When does the registration open for your swimming session?\\n' \n\t\t'Please, send me the date and time in the following format:\\n\\n'\n\t\t'dd/mm/yyyy hh:mm')\n\n\treturn DATETIME", "def testing_favourites(self, email, password, song):\n add_to_favourites = Favourites(self.driver, email, password, song)\n add_to_favourites.open_login_page()\n add_to_favourites.set_user_inputs()\n add_to_favourites.open_home_page()\n add_to_favourites.selecting_a_random_song()\n assert add_to_favourites.check_the_song_is_displayed() is True, 'Song is not displaying'\n add_to_favourites.add_to_wishlist()\n add_to_favourites.go_to_favourites_page()\n assert add_to_favourites.check_the_song_is_added_to_wishlist() == song, 'Not the song'", "def add_to_wishlist(self, beer_id: str) -> Dict:\n method = \"user/wishlist/add\"\n auth = self._get_access_token()\n params = {\"bid\": beer_id}\n return self._do_get(method, auth, params)", "def addsession(cls, session, username, passwd):\n sessionkey = cls.sessionkey(session)\n tmpdict = dict({'username': username, 'password': passwd})\n sessionmgr.update(dict({sessionkey: tmpdict}))", "def test_add_remove_from_wishlist(self):\n url = reverse('add-to-wishlist')\n data = {\n 'igdb': self.game.igdb,\n 'name': self.game.name,\n 'slug': self.game.slug,\n 'cover_id': self.game.cover_id,\n 'backdrop_id': self.game.backdrop_id\n }\n\n add = self.client.post(url, data, format='json')\n self.assertEqual(True, add.data['value'])\n\n remove = self.client.post(url, data, format='json')\n self.assertEqual(False, remove.data['value'])", "def _add_to_cart(obj, sessionid):\n product_id = obj.get('prod_id')\n product = get_product(product_id)\n\n data = {}\n data['product'] = product.product\n data['sessionid'] = sessionid\n data['quantity'] = obj.get('quantity',1)\n data['guests'] = obj.get('guests', 1)\n data['tables'] = obj.get('tables', 1)\n data['wedding'] = obj.get('wedding', 1)\n\n add_to_cart(data)\n\n return True", "def user_in_session():\n return 'user_id' in login_session", "def add_to_bag(request, item_id):\n\n product = get_object_or_404(Product, pk=item_id)\n quantity = int(request.POST.get('quantity'))\n redirect_url = request.POST.get('redirect_url')\n bag = request.session.get('bag', {})\n\n if item_id in list(bag.keys()):\n bag[item_id] += quantity\n messages.success(request,\n f'Updated {product.name} quantity to {bag[item_id]}!')\n else:\n bag[item_id] = quantity\n messages.success(request, f'Added {product.name} to your bag!')\n\n # remove product from wishlist if added from it\n if redirect_url == '/wishlist/':\n profile = get_object_or_404(UserProfile, user=request.user)\n wishlist = Wishlist.objects.get(user_profile=profile)\n wishitems = WishLineItem.objects.get(\n product=product, wishlist=wishlist.id)\n wishitems.delete()\n messages.success(\n request, f'Removed {product.name} from your wishlist!')\n\n request.session['bag'] = bag\n return redirect(redirect_url)", "def new_watched_movie(username: str, movie_id: int) -> bool:\n with connection:\n all_movies = connection.execute(MOVIES_IDS, (movie_id,)).fetchone()\n all_usernames = connection.execute(USERS_IDS, (username,)).fetchone()\n if all_usernames is not None and all_movies is not None:\n connection.execute(ADD_WATCHED_MOVIE, (all_usernames[0], movie_id))\n return True\n return False", "def wishlist(request):\n items = []\n user = get_object_or_404(UserProfile, user=request.user)\n wishlist = Wishlist.objects.get_or_create(user=user)\n wishlist_user = wishlist[0]\n existingWishlist = WishlistItem.objects.filter(\n wishlist=wishlist_user).exists()\n\n if existingWishlist:\n user_wishlist = get_list_or_404(WishlistItem, wishlist=wishlist_user)\n for obj in user_wishlist:\n product = get_object_or_404(Product, name=obj)\n items.append(product)\n context = {\n 'wishlist': True,\n 'products': items\n }\n return render(request, 'wishlist/wishlist.html', context)\n\n else:\n context = {\n 'wishlist': False,\n }\n return render(request, 'wishlist/wishlist.html', context)" ]
[ "0.8124176", "0.7818929", "0.7583546", "0.6178471", "0.6127408", "0.5941033", "0.5855412", "0.581283", "0.57922465", "0.57513714", "0.57402354", "0.56839013", "0.5678189", "0.56736964", "0.5617717", "0.5528249", "0.5518363", "0.5493492", "0.54671645", "0.54640555", "0.5460476", "0.5430389", "0.5406839", "0.53967935", "0.5387633", "0.5370143", "0.5312268", "0.53054667", "0.52975076", "0.52640426" ]
0.83459866
0
Copy relevant fields from Session to SessionForm.
def _copySessionToForm(self, session): sf = SessionForm() for field in sf.all_fields(): if hasattr(session, field.name): # Convert date field to date string if field.name == 'date': setattr(sf, field.name, str(getattr(session, field.name))) # Convert time field to time string elif field.name == 'startTime': setattr(sf, field.name, getattr(session, field.name).strftime('%H:%M')) # Convert typeOfSession string field to enum elif field.name == 'typeOfSession': setattr(sf, field.name, getattr(SessionType, getattr(session, field.name))) # Copy other fields verbatim else: setattr(sf, field.name, getattr(session, field.name)) # Ensure that the SessionForm contains websafeKey elif field.name == "websafeKey": setattr(sf, field.name, session.key.urlsafe()) sf.check_initialized() return sf
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def form_valid(self, form):\n label = form.cleaned_data[\"label\"]\n\n if \"objects\" not in self.request.session:\n self.request.session[\"objects\"] = OrderedDict()\n if \"forms\" not in self.request.session:\n self.request.session[\"forms\"] = OrderedDict()\n\n self.request.session[\"objects\"].update({label: form.halomod_obj})\n self.request.session[\"forms\"].update({label: form.data})\n\n return super().form_valid(form)", "def _forms_to_session_data(self, forms):\n session_data = []\n\n for key in forms:\n data = self._get_normalized_form_data(forms[key], key)\n\n # Store a key/data pair in the session data\n session_data.append([str(key), data])\n\n return session_data", "def copyform(form, settings):\r\n for name, value in form.iteritems():\r\n setattr(settings, name, value)\r\n settings.commit()", "def _session_data_to_forms(self, user, department, session_data):\n forms = OrderedDict()\n\n if not session_data:\n key, form = self._create_certificate_form(user, department)\n forms.update({key: form})\n else:\n for index, (key, form_data) in enumerate(session_data):\n for data_key in form_data:\n if data_key == 'quantity':\n try:\n form_data[data_key] = decimal.Decimal(form_data[data_key])\n except decimal.InvalidOperation:\n form_data[data_key] = ''\n\n form = CertificateForm(user, department, initial=form_data,\n prefix='gc{}'.format(key), index=index + 1)\n forms.update({key: form})\n\n return forms", "def _copySpeakerToForm(self, speaker):\n sf = SpeakerForm()\n for field in sf.all_fields():\n if hasattr(speaker, field.name):\n setattr(sf, field.name, getattr(speaker, field.name))\n elif field.name == \"websafeKey\":\n setattr(sf, field.name, speaker.key.urlsafe())\n sf.check_initialized()\n return sf", "def add_sessionid_in_form(self, form):\n if self.session:\n form(self.div(self.session.sessionid_in_form(self, self.request, self.response), class_='nagare-generated nagare-session-data'))", "def prepare(self, form):\n \n return form", "def _replace_fields(self):\n for name, value in self._cleaned_data.items():\n setattr(self, name, value)", "def current_session_view(request):\n if request.method == 'POST':\n form = CurrentSessionForm(request.POST)\n if form.is_valid():\n session = form.cleaned_data['current_session']\n term = form.cleaned_data['current_term']\n AcademicSession.objects.filter(name=session).update(current=True)\n AcademicSession.objects.exclude(name=session).update(current=False)\n AcademicTerm.objects.filter(name=term).update(current=True)\n AcademicTerm.objects.exclude(name=term).update(current=False)\n\n else:\n form = CurrentSessionForm(initial={\n \"current_session\": AcademicSession.objects.get(current=True),\n \"current_term\": AcademicTerm.objects.get(current=True)\n })\n\n\n return render(request, 'corecode/current_session.html', {\"form\":form})", "def to_form(self):\n form = UserGameForm()\n form.user_name = self.user.get().name\n form.game_key = self.game_key.urlsafe()\n form.game_over = self.game_over\n form.win_status = self.win_status\n moves_count = self.moves_count\n return form", "def get_form_data(self, request):\n return request.session[self.id]['form_data']", "def _copyProfileToForm(self, prof):\n pf = ProfileForm()\n for field in pf.all_fields():\n if hasattr(prof, field.name):\n # Convert t-shirt string to Enum; just copy others\n if field.name == 'teeShirtSize':\n setattr(pf, field.name,\n getattr(TeeShirtSize, getattr(prof, field.name)))\n else:\n setattr(pf, field.name, getattr(prof, field.name))\n pf.check_initialized()\n return pf", "def copy_form(self, from_module, form, to_module, rename=False):\n copy_source = deepcopy(form.to_json())\n # only one form can be a release notes form, so set them to False explicitly when copying\n copy_source['is_release_notes_form'] = False\n copy_source['enable_release_notes'] = False\n if 'unique_id' in copy_source:\n del copy_source['unique_id']\n\n if rename:\n for lang, name in copy_source['name'].items():\n with override(lang):\n copy_source['name'][lang] = _('Copy of {name}').format(name=name)\n\n copy_form = to_module.add_insert_form(from_module, FormBase.wrap(copy_source))\n to_app = to_module.get_app()\n save_xform(to_app, copy_form, form.source.encode('utf-8'))\n\n return copy_form", "def parse_login_form_fields (self, form_soup):\n login_input_fields = {}\n login_inputs = form_soup.find_all('input')\n # gather all form fields, set an empty string as the default value\n for item in login_inputs:\n keys = dict(item.attrs).keys()\n if 'name' in keys and 'value' not in keys:\n login_input_fields[item['name']] = ''\n elif 'name' in keys and 'value' in keys:\n login_input_fields[item['name']] = item['value']\n return login_input_fields", "def get_current_fields(self):\n return self.form.fields.copy()", "def copyCommonFields(self):\n self.fetchDataToForm(self.selected_row, self.selected_column, fields = \"Recent\")", "def form_data(self, password1, password2=None):\n form_data = {'old_password': self.pwd}\n if password2 is None:\n password2 = password1\n form_data[self.password_field + '1'] = password1\n form_data[self.password_field + '2'] = password2\n return form_data", "def to_dict(self, session=False):\n data = {}\n for name, field in self._fields.items():\n if not session and name not in self._permanent_fields:\n continue\n data[name] = field.serialise(self, name)\n return data", "def __extract_form_fields(self, soup):\n fields = OrderedDict()\n for input in soup.find_all('input', attrs={'name': True}):\n if 'type' not in input.attrs:\n input.attrs['type'] = 'text'\n # Single element name/value fields\n if input.attrs['type'].lower() in ('text', 'email', 'hidden', 'password', 'submit', 'image'):\n value = ''\n if 'value' in input.attrs:\n value = input.attrs['value']\n fields[input.attrs['name']] = value\n continue\n\n # Checkboxes and radios\n if input.attrs['type'].lower() in ('checkbox', 'radio'):\n value = ''\n if input.has_attr(\"checked\"):\n if input.has_attr('value'):\n value = input.attrs['value']\n else:\n value = 'on'\n if value:\n fields[input.attrs['name']] = value\n continue\n\n # Textareas\n for textarea in soup.find_all('textarea', attrs={'name': True}):\n fields[textarea.attrs['name']] = textarea.string or ''\n\n # Select fields\n for select in soup.find_all('select', attrs={'name': True}):\n value = ''\n options = select.find_all('option')\n is_multiple = select.has_attr('multiple')\n selected_options = [\n option for option in options\n if option.has_attr('selected') and option.has_attr('value')\n ]\n\n # If no select options, go with the first one\n if not selected_options and options:\n selected_options = [options[0]]\n\n if not is_multiple:\n if len(selected_options) == 1:\n if selected_options[0].has_attr('value'):\n value = selected_options[0]['value']\n else:\n value = ''\n else:\n value = [\n option['value'] for option in selected_options \n if option.has_attr('value')\n ]\n\n fields[select['name']] = value\n\n return fields", "def _copyConferenceToForm(self, conf, displayName):\n cf = ConferenceForm()\n for field in cf.all_fields():\n if hasattr(conf, field.name):\n # Convert Date to date string; just copy others\n if field.name.endswith('Date'):\n setattr(cf, field.name, str(getattr(conf, field.name)))\n else:\n setattr(cf, field.name, getattr(conf, field.name))\n elif field.name == \"websafeKey\":\n setattr(cf, field.name, conf.key.urlsafe())\n if displayName:\n setattr(cf, 'organizerDisplayName', displayName)\n cf.check_initialized()\n return cf", "def test_checkout_form_prefilled(self):\n self.login_test_user()\n self.fill_session_cart()\n response = self.client.get(self.CHECKOUT_URL)\n form_fields_w_values = response.context['form'].initial\n self.assertEqual(len(form_fields_w_values.items()),\n len(self.CHECKOUT_FIELDS))\n for k, v in form_fields_w_values.items():\n self.assertEqual(v, self.user_for_tests[k])", "def get_form_kwargs(self):\n\n press_release_obj = PressRelease.objects.get(\n id=self.kwargs['press_release_pk'])\n\n kwargs = super(PressReleaseUpdateView, self).get_form_kwargs()\n kwargs.update(\n {'field': self.kwargs['field'],\n 'press_release_obj': press_release_obj}\n )\n\n return kwargs", "def _copyConferenceToForm(self, conf, displayName):\n cf = ConferenceForm()\n for field in cf.all_fields():\n if hasattr(conf, field.name):\n # convert Date to date string; just copy others\n if field.name.endswith('Date'):\n setattr(cf, field.name, str(getattr(conf, field.name)))\n else:\n setattr(cf, field.name, getattr(conf, field.name))\n elif field.name == \"websafeKey\":\n setattr(cf, field.name, conf.key.urlsafe())\n if displayName:\n setattr(cf, 'organizerDisplayName', displayName)\n cf.check_initialized()\n return cf", "def build_session_attributes(session):\n if 'attributes' in session.keys():\n if session['attributes']:\n session_attributes = session['attributes']\n else:\n # called from test\n session_attributes = {}\n session_attributes['state'] = 'started'\n session_attributes['accepted_questions'] = []\n session_attributes['rejected_questions'] = []\n else:\n # called from tap\n session_attributes = {}\n session_attributes['state'] = 'started'\n session_attributes['accepted_questions'] = []\n session_attributes['rejected_questions'] = []\n\n return session_attributes", "def _form_data(self, response):\n SQFI_audit_type = response.xpath(self.filters[6]).extract_first()\n SQFI_audit_type_val = response.xpath(self.filters[7]).extract_first()\n food_sector_categories = response.xpath(self.filters[8]).extract_first()\n food_sector_categories_val = response.xpath(self.filters[9]).extract()\n audit_rating = response.xpath(self.filters[10]).extract_first()\n audit_rating_val = response.xpath(self.filters[11]).extract()\n country = response.xpath(self.filters[12]).extract_first()\n country_val = response.xpath(self.filters[13]).extract()\n form_data = {\n SQFI_audit_type: SQFI_audit_type_val,\n food_sector_categories: food_sector_categories_val,\n audit_rating: audit_rating_val,\n country: country_val,\n }\n return form_data", "def delivery_details(request):\n user = request.user\n if user.is_authenticated:\n delivery_details = {\n 'first_name': user.first_name, \n 'last_name': user.last_name,\n 'address_line_1': user.profile.address_line_1,\n 'address_line_2': user.profile.address_line_2,\n 'address_line_3 ': user.profile.address_line_3,\n 'town_or_city': user.profile.town_or_city,\n 'county': user.profile.county,\n 'postcode': user.profile.postcode\n }\n delivery_form = DeliveryForm(request.POST or delivery_details)\n if request.method == \"POST\":\n if delivery_form.is_valid():\n request.session['first_name'] = delivery_form.cleaned_data['first_name']\n request.session['last_name'] = delivery_form.cleaned_data['last_name']\n request.session['address_line_1'] = delivery_form.cleaned_data['address_line_1']\n request.session['address_line_2'] = delivery_form.cleaned_data['address_line_2']\n request.session['address_line_3'] = delivery_form.cleaned_data['address_line_3']\n request.session['town_or_city'] = delivery_form.cleaned_data['town_or_city']\n request.session['county'] = delivery_form.cleaned_data['county']\n request.session['postcode'] = delivery_form.cleaned_data['postcode']\n return redirect(reverse('card_details'))\n else:\n delivery_details = {\n 'first_name' : request.session.get('first_name', None), \n 'last_name' : request.session.get('last_name', None),\n 'address_line_1' : request.session.get('address_line_1', None),\n 'address_line_2' : request.session.get('address_line_2', None), \n 'address_line_3' : request.session.get('address_line_3', None), \n 'town_or_city' : request.session.get('town_or_city', None), \n 'county' : request.session.get('county', None), \n 'postcode' : request.session.get('postcode', None) \n }\n delivery_form = DeliveryForm(request.POST or delivery_details)\n if request.method == \"POST\":\n if delivery_form.is_valid():\n request.session['first_name'] = delivery_form.cleaned_data['first_name']\n request.session['last_name'] = delivery_form.cleaned_data['last_name']\n request.session['address_line_1'] = delivery_form.cleaned_data['address_line_1']\n request.session['address_line_2'] = delivery_form.cleaned_data['address_line_2']\n request.session['address_line_3'] = delivery_form.cleaned_data['address_line_3']\n request.session['town_or_city'] = delivery_form.cleaned_data['town_or_city']\n request.session['county'] = delivery_form.cleaned_data['county']\n request.session['postcode'] = delivery_form.cleaned_data['postcode']\n return redirect(reverse('card_details'))\n return render(request, 'delivery_details.html', {'delivery_form': delivery_form})", "def formfields():\n\n T = current.T\n request = current.request\n\n auth = current.auth\n auth_settings = auth.settings\n auth_messages = auth.messages\n\n utable = auth_settings.table_user\n passfield = auth_settings.password_field\n\n # Last name is required\n utable.last_name.requires = IS_NOT_EMPTY(error_message=T(\"input required\"))\n\n # Don't check for duplicate email (will be done in onvalidation)\n # => user might choose to use the current email address of the account\n # => if registration key or code are invalid, we don't want to give away\n # any existing email addresses\n utable.email.requires = [IS_EMAIL(error_message = auth_messages.invalid_email),\n IS_LOWER(),\n ]\n\n # Instantiate Consent Tracker\n consent = ConsentTracking(processing_types=[\"STORE\", \"RULES_ISS\"])\n\n # Form fields\n formfields = [utable.first_name,\n utable.last_name,\n utable.email,\n utable[passfield],\n Field(\"password_two\", \"password\",\n label = auth_messages.verify_password,\n requires = IS_EXPR(\"value==%s\" % \\\n repr(request.vars.get(passfield)),\n error_message = auth_messages.mismatched_password,\n ),\n comment = DIV(_class = \"tooltip\",\n _title = \"%s|%s\" % (auth_messages.verify_password,\n T(\"Enter the same password again\"),\n ),\n ),\n ),\n Field(\"code\",\n label = T(\"Registration Code\"),\n requires = IS_NOT_EMPTY(),\n ),\n Field(\"consent\",\n label = T(\"Consent\"),\n widget = consent.widget,\n ),\n ]\n\n\n # Required fields\n required_fields = [\"first_name\",\n \"last_name\",\n ]\n\n return formfields, required_fields", "def populate_form(self, **kwargs):\n for name, value in kwargs.items():\n self.populate_field(name, value)", "def to_form(self, message):\n form = GameForm()\n form.urlsafe_key = self.key.urlsafe()\n form.user_name = self.user.get().name\n form.game_state = self.game_state\n form.game_over = self.game_over\n form.history = self.history\n form.message = message\n return form", "def from_dict(self, data, session=False):\n for name, field in self._fields.items():\n if not session and name not in self._permanent_fields:\n continue\n # Check if Field has been added since it was saved - leave default\n if name not in data:\n continue\n field.deserialise(self, name, data[name])\n return self" ]
[ "0.6098217", "0.6088412", "0.6072761", "0.60642654", "0.6026982", "0.59829164", "0.5759055", "0.5618044", "0.561432", "0.55417347", "0.5487526", "0.54755", "0.544686", "0.5294741", "0.52286774", "0.52258146", "0.5166349", "0.5164965", "0.51523036", "0.51498306", "0.5126769", "0.51251554", "0.5104614", "0.50888383", "0.5062417", "0.50392693", "0.50326735", "0.5018545", "0.50105274", "0.5001941" ]
0.85230947
0
Retrieve all sessions associated with a conference.
def _getConferenceSessions(self, request): # Ensure that websafeConferenceKey is a valid conference key confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey, 'Conference') # Retrieve all sessions that have a matching conference key sessions = Session.query(Session.conference == confKey).fetch() return sessions
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_conference_sessions(self, request):\n return self.session_service.get_conference_sessions(\n request.websafeConferenceKey)", "def get_sessions_by_type(self, request):\n return self.session_service.get_conference_sessions_by_type(\n request.websafeConferenceKey, request.sessionType)", "def getConferenceSessions(self, request):\n sessions = self._getConferenceSessions(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def _getSessionsBySpeaker(self, request):\n # Ensure that the speaker key is valid and that the speaker exists\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey, 'Speaker')\n # Return all of the speaker's sessions\n return ndb.get_multi(speaker.sessions)", "def _getConferenceSessionsByType(self, request):\n # Ensure that websafeConferenceKey is a valid conference key\n confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey,\n 'Conference')\n # Retrieve all sessions that have a matching conference key, by type\n sessions = Session.query(\n Session.conference == confKey,\n Session.typeOfSession == str(request.typeOfSession)\n ).fetch()\n return sessions", "def sessions(self):\n logger.debug(\"Get sessions\")\n return self._raw_api.sessions.get()", "def conferences(self):\r\n return conferences.Conferences(self)", "def get_sessions_by_type_and_filters(self, request):\n return self.session_service.get_sessions_by_type_and_filters(\n request.websafeConferenceKey,\n request.typeOfSession, request.filters)", "def get_speaker_sessions(self, request):\n return self.session_service.get_speaker_sessions(\n request.websafeSpeakerKey)", "def get_sessions(self):\n\n return self.all_sessions", "def getSessionsBySpeaker(self, request):\n sessions = self._getSessionsBySpeaker(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "async def get_conference_by_id(self, session_id):\n db = self['db_engine']\n async with db.acquire() as connection:\n query = select([Conference]).where(Conference.session_id == session_id).where(\n Conference.expired_at > datetime.now(timezone.utc))\n if conference := await(await connection.execute(query)).first():\n result = dict(conference)\n query = select([User.email, User.display_name]).where(User.id == conference.user_id)\n if user := await(await connection.execute(query)).first():\n result['user_data'] = UserData(conference.user_id, *user)\n return result", "def get_conferences():\n conferences = []\n for confId in range(1, 8):\n c = requests.get(CONFERENCES_URL.format(confId)).json()\n conf_list = c.get('conferences')\n if conf_list is None or len(conf_list) == 0:\n continue\n conferences.append(Conference.from_json(conf_list[0]))\n return conferences", "def all (self):\n sparql_results = self.query (\"\"\"\n select distinct ?rs ?session ?name ?number ?pid ?sitename\n where {\n \n ?rs rdf:type austalk:RecordedSession .\n ?rs olac:speaker ?participant .\n \n ?participant austalk:id ?pid .\n ?participant austalk:recording_site ?site .\n ?site rdfs:label ?sitename .\n \n ?rs austalk:prototype ?session .\n ?session austalk:name ?name .\n ?session austalk:id ?number .\n }\n ORDER BY ?name\"\"\")\n\n results = []\n\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n results.append (Session (\n client = self.client,\n identifier = result[\"rs\"][\"value\"],\n prototype = result[\"session\"][\"value\"],\n name = result[\"name\"][\"value\"],\n number = result[\"number\"][\"value\"],\n site = result[\"sitename\"][\"value\"],\n participantId = result[\"pid\"][\"value\"]))\n\n return results", "def session_list(self, endpoint_name=None):\n if endpoint_name is None:\n _, body = self.request('/v1.1/endpoint/sessions', 'GET')\n else:\n _, body = self.request('/v1.1/endpoints/%s/sessions' % endpoint_name, 'GET')\n return body", "def GetSessions(firebase: firebase) -> None:\n\n global sessions\n obj_key_list = []\n \n result = firebase.get('/session', None)\n \n if result is None:\n print(\"no sessions found\")\n return\n \n for i in result.keys():\n obj_key_list.append(i)\n \n for i in obj_key_list:\n session = Session()\n session.setId(i)\n session.setCourseId(result[i]['courseid'])\n session.setDOW(result[i]['DOW'])\n session.setSessionNumber(result[i]['session_number'])\n session.setSessionDate(result[i]['session_date'])\n session.setSessionTimeStart(result[i]['session_time_start'])\n session.setSessionTimeEnd(result[i]['session_time_end'])\n sessions.append(session)", "def get_sessions_list():\n sessions = Session.query.all()\n result = sessions_schema.dump(sessions).data\n return jsonify({'status': 'success', 'message': None, 'data': result}), 200", "def getConferenceSessionsByType(self, request):\n sessions = self._getConferenceSessionsByType(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def sessions(self):\n return list(Session.get_sessions(self))", "def sessions(self):\n return utils.listItems(self, '/status/sessions')", "def get_sessions(self):\n return self.current_sessions", "def get_all_sessions(self) -> list:\n sessions = list()\n for stream_id in self.streams.keys():\n tcpsession, session_position, network_tuple = self.streams[stream_id]\n sessions.append(tcpsession.get_session(session_position - 1))\n return sessions", "def conference(self, sid):\r\n return conferences.Conference(self, sid)", "def sessions(self):\n return self._sessions", "def sessions(self):\n for session_id in self.get_sessions(): \n session = Session(self.session_cache, self.sid, session_id)\n yield session", "def sessions(self, *args, **kwargs):\r\n return self._get('Sessions', *args, **kwargs)", "def get_sessions(sessions, time_feat_dict):\n filt = Session.filter_time_func(time_feat_dict)\n return [s for s in sessions if filt(shortstr2time(s['start']))]", "def filter_by_session (self, site_id, participant_id, session_id):\n\n query = \"\"\"\n select distinct * where {\n\n BIND (\"%s\" AS ?pid)\n BIND (\"%s\" as ?sessionid)\n \n ?participant austalk:id ?pid .\n ?rc rdf:type austalk:RecordedComponent .\n ?rc olac:speaker ?participant .\n ?rc austalk:session ?sessionid .\n\n ?rc austalk:prototype ?prototype .\n ?prototype austalk:name ?name .\n ?prototype austalk:shortname ?shortname .\n \n optional { ?rc austalk:audiorating ?audiorating .}\n optional { ?rc austalk:videorating ?videorating .}\n optional { ?rc austalk:comment ?comment .}\n \n }\n \"\"\" % (participant_id, session_id)\n\n\n sparql_results = self.query (query)\n results = []\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n for field in ['audiorating', 'videorating', 'comment']:\n if field not in result:\n result[field] = {'value': ''}\n\n comp = Component (\n client = self.client,\n identifier = result[\"rc\"][\"value\"],\n participantId = result[\"pid\"][\"value\"],\n sessionId = result[\"sessionid\"][\"value\"],\n prototype = result[\"prototype\"][\"value\"],\n audiorating = result[\"audiorating\"][\"value\"],\n videorating = result[\"videorating\"][\"value\"],\n comment = result[\"comment\"][\"value\"],\n name = result[\"name\"][\"value\"],\n componentId = result[\"shortname\"][\"value\"],\n site = site_id,\n )\n comp.details()\n results.append(comp)\n return results", "def find_sessions(sfe):\n print(\"-\" * 20 + \" find_sessions started\")\n isessions = sfe.list_iscsisessions()\n json_isessions = isessions.to_json()\n return json_isessions", "def get_sessions(url: str, token: str) -> List[Session]:\n sessions_url = f'{url}api/sessions'\n response = requests.get(sessions_url, params={'token': token})\n assert(response.status_code == 200)\n sessions_raw = json.loads(response.text)\n sessions = []\n for session_raw in sessions_raw:\n session = Session(\n path = session_raw['path'],\n last_activity = dateutil.parser.isoparse(session_raw['kernel']['last_activity']),\n execution_state = session_raw['kernel']['execution_state']\n )\n assert(session['execution_state'] in valid_execution_states)\n sessions.append(session)\n\n sessions.sort(key=lambda session: session['last_activity'], reverse=True)\n return sessions" ]
[ "0.8155603", "0.72048026", "0.7024871", "0.7000029", "0.69593084", "0.6693818", "0.66356134", "0.66315794", "0.65811247", "0.6544365", "0.64583755", "0.643887", "0.64130956", "0.63868886", "0.63120866", "0.62812674", "0.6271837", "0.6205866", "0.6112134", "0.6100491", "0.6069305", "0.60190016", "0.6000727", "0.599727", "0.5972812", "0.5942718", "0.5865109", "0.58010185", "0.5751455", "0.56871325" ]
0.800364
1
Retrieve all sessions associated with a conference, by type.
def _getConferenceSessionsByType(self, request): # Ensure that websafeConferenceKey is a valid conference key confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey, 'Conference') # Retrieve all sessions that have a matching conference key, by type sessions = Session.query( Session.conference == confKey, Session.typeOfSession == str(request.typeOfSession) ).fetch() return sessions
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_sessions_by_type(self, request):\n return self.session_service.get_conference_sessions_by_type(\n request.websafeConferenceKey, request.sessionType)", "def get_sessions_by_type_and_filters(self, request):\n return self.session_service.get_sessions_by_type_and_filters(\n request.websafeConferenceKey,\n request.typeOfSession, request.filters)", "def getConferenceSessionsByType(self, request):\n sessions = self._getConferenceSessionsByType(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def _getConferenceSessions(self, request):\n # Ensure that websafeConferenceKey is a valid conference key\n confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey,\n 'Conference')\n # Retrieve all sessions that have a matching conference key\n sessions = Session.query(Session.conference == confKey).fetch()\n return sessions", "def get_conference_sessions(self, request):\n return self.session_service.get_conference_sessions(\n request.websafeConferenceKey)", "def _getSessionsBySpeaker(self, request):\n # Ensure that the speaker key is valid and that the speaker exists\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey, 'Speaker')\n # Return all of the speaker's sessions\n return ndb.get_multi(speaker.sessions)", "def getConferenceSessions(self, request):\n sessions = self._getConferenceSessions(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def getSessionsBySpeaker(self, request):\n sessions = self._getSessionsBySpeaker(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def conferences(self):\r\n return conferences.Conferences(self)", "def get_speaker_sessions(self, request):\n return self.session_service.get_speaker_sessions(\n request.websafeSpeakerKey)", "def get_sessions_by_types_in_wishlist(self, user):\n sessions = self.wishlist_sessions(user)\n\n types = [getattr(s, 'typeOfSession') for s in sessions]\n\n sessions = ConferenceSession.query(\n ConferenceSession.typeOfSession.IN(types)\n ).fetch()\n\n return ConferenceSessionForms(\n items=[self.copy_entity_to_form(ConferenceSessionForm(), s)\n for s in sessions])", "def get_conferences():\n conferences = []\n for confId in range(1, 8):\n c = requests.get(CONFERENCES_URL.format(confId)).json()\n conf_list = c.get('conferences')\n if conf_list is None or len(conf_list) == 0:\n continue\n conferences.append(Conference.from_json(conf_list[0]))\n return conferences", "def sessions(self):\n logger.debug(\"Get sessions\")\n return self._raw_api.sessions.get()", "async def get_conference_by_id(self, session_id):\n db = self['db_engine']\n async with db.acquire() as connection:\n query = select([Conference]).where(Conference.session_id == session_id).where(\n Conference.expired_at > datetime.now(timezone.utc))\n if conference := await(await connection.execute(query)).first():\n result = dict(conference)\n query = select([User.email, User.display_name]).where(User.id == conference.user_id)\n if user := await(await connection.execute(query)).first():\n result['user_data'] = UserData(conference.user_id, *user)\n return result", "def all (self):\n sparql_results = self.query (\"\"\"\n select distinct ?rs ?session ?name ?number ?pid ?sitename\n where {\n \n ?rs rdf:type austalk:RecordedSession .\n ?rs olac:speaker ?participant .\n \n ?participant austalk:id ?pid .\n ?participant austalk:recording_site ?site .\n ?site rdfs:label ?sitename .\n \n ?rs austalk:prototype ?session .\n ?session austalk:name ?name .\n ?session austalk:id ?number .\n }\n ORDER BY ?name\"\"\")\n\n results = []\n\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n results.append (Session (\n client = self.client,\n identifier = result[\"rs\"][\"value\"],\n prototype = result[\"session\"][\"value\"],\n name = result[\"name\"][\"value\"],\n number = result[\"number\"][\"value\"],\n site = result[\"sitename\"][\"value\"],\n participantId = result[\"pid\"][\"value\"]))\n\n return results", "def get_sessions(sessions, time_feat_dict):\n filt = Session.filter_time_func(time_feat_dict)\n return [s for s in sessions if filt(shortstr2time(s['start']))]", "def get_sessions(self):\n\n return self.all_sessions", "def get_sessions_list():\n sessions = Session.query.all()\n result = sessions_schema.dump(sessions).data\n return jsonify({'status': 'success', 'message': None, 'data': result}), 200", "def session_list(self, endpoint_name=None):\n if endpoint_name is None:\n _, body = self.request('/v1.1/endpoint/sessions', 'GET')\n else:\n _, body = self.request('/v1.1/endpoints/%s/sessions' % endpoint_name, 'GET')\n return body", "def sessions(self):\n return utils.listItems(self, '/status/sessions')", "def filter_by_participant (self, participant):\n \n sparql_results = self.query (\"\"\"\n select distinct ?rs ?session ?name ?number ?pid ?sitename\n where {\n BIND (<%s> AS ?participant)\n \n ?rs rdf:type austalk:RecordedSession .\n ?rs olac:speaker ?participant .\n \n ?participant austalk:id ?pid .\n ?participant austalk:recording_site ?site .\n ?site rdfs:label ?sitename .\n \n ?rs austalk:prototype ?session .\n ?session austalk:name ?name .\n ?session austalk:id ?number .\n }\n ORDER BY ?name\"\"\" % participant.identifier)\n \n results = []\n\n for result in sparql_results[\"results\"][\"bindings\"]:\n results.append (Session (\n client = self.client,\n identifier = result[\"rs\"][\"value\"],\n prototype = result[\"session\"][\"value\"],\n name = result[\"name\"][\"value\"],\n number = result[\"number\"][\"value\"],\n site = result[\"sitename\"][\"value\"],\n participantId = result[\"pid\"][\"value\"]))\n\n return results", "def filter_by_session (self, site_id, participant_id, session_id):\n\n query = \"\"\"\n select distinct * where {\n\n BIND (\"%s\" AS ?pid)\n BIND (\"%s\" as ?sessionid)\n \n ?participant austalk:id ?pid .\n ?rc rdf:type austalk:RecordedComponent .\n ?rc olac:speaker ?participant .\n ?rc austalk:session ?sessionid .\n\n ?rc austalk:prototype ?prototype .\n ?prototype austalk:name ?name .\n ?prototype austalk:shortname ?shortname .\n \n optional { ?rc austalk:audiorating ?audiorating .}\n optional { ?rc austalk:videorating ?videorating .}\n optional { ?rc austalk:comment ?comment .}\n \n }\n \"\"\" % (participant_id, session_id)\n\n\n sparql_results = self.query (query)\n results = []\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n for field in ['audiorating', 'videorating', 'comment']:\n if field not in result:\n result[field] = {'value': ''}\n\n comp = Component (\n client = self.client,\n identifier = result[\"rc\"][\"value\"],\n participantId = result[\"pid\"][\"value\"],\n sessionId = result[\"sessionid\"][\"value\"],\n prototype = result[\"prototype\"][\"value\"],\n audiorating = result[\"audiorating\"][\"value\"],\n videorating = result[\"videorating\"][\"value\"],\n comment = result[\"comment\"][\"value\"],\n name = result[\"name\"][\"value\"],\n componentId = result[\"shortname\"][\"value\"],\n site = site_id,\n )\n comp.details()\n results.append(comp)\n return results", "def conference(self, sid):\r\n return conferences.Conference(self, sid)", "def get_subscribed_clients(self, mtype):\n return self.hub.get_subscribed_clients(self.get_private_key(), mtype)", "def get_all_sessions(self) -> list:\n sessions = list()\n for stream_id in self.streams.keys():\n tcpsession, session_position, network_tuple = self.streams[stream_id]\n sessions.append(tcpsession.get_session(session_position - 1))\n return sessions", "def sessions(self):\n return list(Session.get_sessions(self))", "def sessions(self, *args, **kwargs):\r\n return self._get('Sessions', *args, **kwargs)", "def get_conference_series(self):\n conference_type = \\\n ResourceType.objects.getFromName(ResourceType.CONFERENCE)\n cursor = connection.cursor()\n cursor.execute(\"\"\"\n SELECT conference_series, COUNT(id) AS num_in_series\n FROM ieeetags_resource\n WHERE conference_series <> ''\n AND resource_type_id = %s\n GROUP BY conference_series\n \"\"\", [conference_type.id])\n return cursor.fetchall()", "def list(self, datasource_name=None, event_name=None, requested_after=None, requested_before=None, session_type=None):\n query = {\n 'dataSourceName': datasource_name,\n 'eventName': event_name,\n 'requestedBefore': requested_before,\n 'requestedAfter': requested_after,\n 'sessionType': session_type\n }\n response, _, headers = self._client.request_with_headers('GET', 'sessions', params=query)\n\n return [SessionResponse(item, headers) for item in response.get('items', [])]", "def get_subscribed_clients(self, private_key, mtype):\n return self._samp_hub.getSubscribedClients(private_key, mtype)" ]
[ "0.839267", "0.76810974", "0.74179673", "0.72810936", "0.7257645", "0.6629272", "0.6575053", "0.6291302", "0.61481535", "0.60031515", "0.59337205", "0.5886013", "0.5719723", "0.57138413", "0.57089686", "0.5639327", "0.5602851", "0.5507092", "0.5500614", "0.54822844", "0.547502", "0.54704434", "0.54438287", "0.54178697", "0.53411376", "0.53352463", "0.53340137", "0.53143334", "0.5313005", "0.529542" ]
0.83530176
1
Retrieve all sessions matching one or more given highlights.
def _getSessionsByHighlightSearch(self, request): # Generate list of filters from the highlight arguments filters = [Session.highlights == hl for hl in request.highlights] if not filters: raise endpoints.BadRequestException( 'At least one highlight must be specified' ) # Retrieve all sessions that match one or more of the highlight filters sessions = Session.query(ndb.OR(*filters)).order(Session.name).fetch() return sessions
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def getSessionsByHighlightSearch(self, request):\n sessions = self._getSessionsByHighlightSearch(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def get(self, request):\n # Retrieve the user from the request if they have been authenticated\n current_user = request.user\n # Get all highlights for the current user\n highlights = Highlights.objects.filter(profile=current_user.profile)\n paginator = self.pagination_class()\n highlights_page = paginator.paginate_queryset(highlights, request)\n highlights = self.serializer_class(highlights_page, many=True, context={'request': request})\n paginated_highlights = paginator.get_paginated_response(highlights.data)\n return Response({\n 'message': HIGHLIGHT_MSGS['HIGHLIGHTS_FOUND'],\n 'highlights': paginated_highlights.data['results'],\n 'count': paginated_highlights.data['count'],\n 'next': paginated_highlights.data['next'],\n 'previous': paginated_highlights.data['previous']\n },\n status=status.HTTP_200_OK)", "def _define_highlights(self):\n for ansi_code in dict.fromkeys([*self._colors.values(),\n *self._colors_special.values()]):\n code_safe = ansi_code.replace(';', '_')\n fg, bg, special = ansi_to_vim_color(ansi_code)\n args = ''\n if fg is not None:\n args += 'ctermfg=' + fg\n if bg is not None:\n args += ' ctermbg=' + bg\n if special: # special is never None\n args += ' cterm=' + special\n if args:\n cmd = f'hi color{code_safe} {args}'\n logger.debug(cmd)\n self._vim.command(cmd)", "def highlightSearch(self, wordList=None, regExpList=None):\n backColor = self.palette().brush(QPalette.Active,\n QPalette.Highlight)\n foreColor = self.palette().brush(QPalette.Active,\n QPalette.HighlightedText)\n if wordList is None:\n wordList = []\n if regExpList is None:\n regExpList = []\n for regExp in regExpList:\n for match in regExp.finditer(self.toPlainText()):\n matchText = match.group()\n if matchText not in wordList:\n wordList.append(matchText)\n selections = []\n for word in wordList:\n while self.find(word):\n extraSel = QTextEdit.ExtraSelection()\n extraSel.cursor = self.textCursor()\n extraSel.format.setBackground(backColor)\n extraSel.format.setForeground(foreColor)\n selections.append(extraSel)\n cursor = QTextCursor(self.document())\n self.setTextCursor(cursor) # reset main cursor/selection\n self.setExtraSelections(selections)", "def get_sessions(sessions, time_feat_dict):\n filt = Session.filter_time_func(time_feat_dict)\n return [s for s in sessions if filt(shortstr2time(s['start']))]", "def all (self):\n sparql_results = self.query (\"\"\"\n select distinct ?rs ?session ?name ?number ?pid ?sitename\n where {\n \n ?rs rdf:type austalk:RecordedSession .\n ?rs olac:speaker ?participant .\n \n ?participant austalk:id ?pid .\n ?participant austalk:recording_site ?site .\n ?site rdfs:label ?sitename .\n \n ?rs austalk:prototype ?session .\n ?session austalk:name ?name .\n ?session austalk:id ?number .\n }\n ORDER BY ?name\"\"\")\n\n results = []\n\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n results.append (Session (\n client = self.client,\n identifier = result[\"rs\"][\"value\"],\n prototype = result[\"session\"][\"value\"],\n name = result[\"name\"][\"value\"],\n number = result[\"number\"][\"value\"],\n site = result[\"sitename\"][\"value\"],\n participantId = result[\"pid\"][\"value\"]))\n\n return results", "def whatsNewHighlight(*args, highlightColor: Union[List[float, float, float], bool]=None,\n highlightOn: bool=True, showStartupDialog: bool=True, q=True, query=True,\n **kwargs)->Union[None, Any]:\n pass", "def get_highlights(self, ann_db, asset_id):\n if not asset_id:\n return None\n\n conn = sqlite3.connect(ann_db)\n cur = conn.cursor()\n a_id = (asset_id,)\n highlights = []\n for row, heading, created, location in cur.execute(\n \"\"\"SELECT ZANNOTATIONSELECTEDTEXT, \n ZFUTUREPROOFING5, \n ZANNOTATIONCREATIONDATE, \n ZANNOTATIONLOCATION \n FROM ZAEANNOTATION \n WHERE ZANNOTATIONASSETID=? \n AND ZANNOTATIONSELECTEDTEXT <> '' \n AND ZANNOTATIONDELETED=0\n \"\"\", a_id):\n chapter = int(location.split('[')[0].split('/')[2].replace(',', ''))\n try:\n ref_in_chapter = int(location.split('!')[1].split('/')[2].replace(',', ''))\n except ValueError:\n ref_in_chapter = 0\n highligt = Highlight(row, heading, float(created), chapter, ref_in_chapter)\n highlights.append(highligt)\n conn.close()\n return highlights", "def clients_search_by_term_view(request):\n # Check connected\n if not check_connected(request):\n raise exc.HTTPForbidden()\n\n searchTerm = request.params[\"searchTerm\"] if \"searchTerm\" in request.params else None\n old_clients = request.params['old_clients'] == 'true' if 'old_clients' in request.params else False\n\n clients = _multipleAttributesClientSearch(request, searchTerm, old_clients=old_clients)\n \n return Utils.serialize_many(clients)", "def get_highlights(lines, split_on='@highlight'):\n text = \" \".join(lines)\n splits = text.split(split_on)\n return splits[:1], splits[1:]", "def highlight_syntax(self) -> List[SyntaxHighlight]:\n raise NotImplementedError", "def get_highlight_text(start, start_index, end, end_index):\n #TODO: this has nothing to do with highlights, whatever those are\n #TODO: this method is ridiculous, how is it usable? the arguments could\n # mean two different things.\n text = \"\"\n words = WordInSentence.query.filter(\n (WordInSentence.sentence_id > start) &\n (WordInSentence.sentence_id < end) |\n (WordInSentence.sentence_id == start) &\n (WordInSentence.position >= start_index) |\n (WordInSentence.sentence_id == end) &\n (WordInSentence.position <= end_index)).\\\n order_by(WordInSentence.sentence_id).\\\n order_by(WordInSentence.position).all()\n\n for word in words:\n if not app.config[\"PUNCTUATION_ALL\"] in word.surface:\n text += \" \" # Don't put spaces in front of punctuation.\n\n text += word.surface\n\n return text", "def getSessionByUsername(self, username):\n match = []\n for session in self.sessions:\n if (session.identifier[1] == username):\n match.append(session)\n return match", "def search_sessions(name: str, provider: Optional[str] = None) -> List[str]:\n sessions = session_list(provider=provider).sessions\n name = name.lower()\n return [s.id for s in sessions if s.id.lower().startswith(name)]", "def find_queries(self, senses):\n queries = set()\n for token in senses:\n queries.add(token)\n\n if (len(queries) == 0):\n focus = self.q_classifer.matched_groups['focus'].split()\n focus = [f for f in focus if f not in set(stopwords.words('english'))]\n queries = focus\n\n return queries", "def _getSessionsInWishlist(self):\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n profile = self._getProfileFromUser()\n # Fetch the entities and return them\n return ndb.get_multi(profile.sessionWishlist)", "def _group_sessions(self, sessions):\n session_dict = collections.defaultdict(list)\n for session in sessions:\n session_dict[session.query].append(session)\n return session_dict", "def highlight(self, message):\n user = self.ts.get_user(message)\n msg_list = self.ts.get_human_readable_message(message).split(' ')\n if len(msg_list) > 1:\n user_note = ' '.join(msg_list[1:])\n else:\n user_note = ''\n time_dict = self._get_live_time()\n if time_dict is not None:\n est_tz = pytz.timezone('US/Eastern')\n start_time_utc = time_dict['stream_start']\n start_time_est = est_tz.normalize(start_time_utc.replace(tzinfo=pytz.utc).astimezone(est_tz))\n time_str = 'Approximately {hours}, {minutes} and {seconds} into the stream.'.format(\n hours=time_dict['hour'], minutes=time_dict['minute'], seconds=time_dict['second'])\n\n spreadsheet_name, _ = self.spreadsheets['highlights']\n gc = gspread.authorize(self.credentials)\n sheet = gc.open(spreadsheet_name)\n ws = sheet.worksheet('Highlight List')\n records = ws.get_all_records() # Doesn't include the first row\n next_row = len(records) + 2\n ws.update_cell(next_row, 1, user)\n ws.update_cell(next_row, 2, str(start_time_est)[:-6])\n ws.update_cell(next_row, 3, time_str)\n ws.update_cell(next_row, 4, user_note)\n self._add_to_whisper_queue(user, 'The highlight has been added to the spreadsheet for review.')", "def highlight_series(s):\n return ['background-color: #eee' for v in s]", "def find_all(v):\n screen = G.DEVICE.snapshot(quality=ST.SNAPSHOT_QUALITY)\n return v.match_all_in(screen)", "def sessions(self, *args, **kwargs):\r\n return self._get('Sessions', *args, **kwargs)", "def _getSessionsBySpeaker(self, request):\n # Ensure that the speaker key is valid and that the speaker exists\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey, 'Speaker')\n # Return all of the speaker's sessions\n return ndb.get_multi(speaker.sessions)", "def highlight(search_me, word, userid):\n if db_is_loggedin.find_one({\"user\": userid}).get(\"is_loggedin\") == True:\n word = word.strip()\n results = wisdomaiengine.highlighter(word)\n search_term = db_search_terms.find_one({\"value\": search_me.lower()})\n search_id = search_term.get(\"_id\")\n data = {\"user\": userid,\n \"search_id\": search_id,\n \"highlighted_word\": word,\n \"results\": results,\n \"date_saved\": datetime.utcnow()}\n x = db_highlights.insert(data, check_keys=False)\n jsonob = jsonify(results=results)\n return jsonob\n else:\n msg = {\"status\" : { \"type\" : \"fail\" , \"message\" : \"Please log in\"}}\n return jsonify(msg)", "def test_query_highlight_list_of_fields(self):\n fields_to_highlight = [\"user_id\", \"data\"]\n id = get_rand_string()\n\n # Same data and user_id\n user_id = data = get_rand_string()\n\n self.conn.add(id=id, user_id=user_id, data=data)\n self.conn.commit()\n\n # Specify the fields to highlight as a list of fields\n response = self.conn.query(\"user_id:\" + user_id,\n highlight=fields_to_highlight)\n\n self.assertTrue(hasattr(response, \"highlighting\"),\n (\"No fields have been highlighted \"\n \"(exptected a 'highlighting' attribute), id:%s\") % (id))\n\n highlighting = response.highlighting\n\n self.assertTrue(id in highlighting,\n \"Document (id:%s) should've been highlighted\")\n\n self.assertTrue(len(highlighting[id]) == 2,\n (\"There should be two documents highlighted, \"\n \"id:%s, highlighting:%s\" % (id, highlighting)))\n\n for field in fields_to_highlight:\n self.assertTrue(field in highlighting[id],\n \"%s should be highlighted, highlighting:%s\" % (\n field,highlighting))\n\n # user_id and data are equal\n content = parseString(highlighting[id][field][0])\n highlighting_value = content.firstChild.firstChild.nodeValue\n self.assertEquals(highlighting_value, data,\n \"Highlighting didn't return %s value, expected:%s, got:%s\" % (\n field, data, highlighting_value))", "def search_synapses(cls, **kwargs):\n found = filter_register('all_syn', **kwargs)\n print(\"Indexes of found synapses:{}\".format(found))\n register = load_register('all_syn').iloc[found]\n return found, register", "def highlight_active(series):\n style = []\n for i in range(len(series)):\n if series[i] > 0:\n style.append(\"background-color: lightblue\")\n else:\n style.append(\"background-color: dimgrey\")\n\n return style", "def active():\n session = session_maker(\n app.config['MYSQL_USER'], app.config['MYSQL_PASS'], app.config['MYSQL_SERVER_PORT_3306_TCP_ADDR'],\n app.config['MYSQL_SERVER_PORT_3306_TCP_PORT'], app.config['DB'])\n\n\n print(\n tabulate(\n selection_list_active(session),\n headers=['number', 'sqlid', 'name', 'city', 'state']))", "def get_colors(lines):\n\n patt = re.compile('\\#\\w+')\n\n return [\n patt.search(line).group(0)\n for line in lines\n if patt.search(line)\n ]", "def findSessions(self, channel):\n found = []\n for ss in self.sessions:\n try:\n _channel = channel.decode(ss.encoding)\n if _channel == ss.name:\n found.append(ss)\n if ss.matchNick(_channel):\n found.append(ss)\n except UnicodeDecodeError:\n continue\n if found == []:\n found = [self.defaultSession]\n return found", "def find_all(session=None):\n return find(session=session)" ]
[ "0.64075655", "0.51973224", "0.5143006", "0.5126674", "0.49887824", "0.49578515", "0.49377143", "0.48807254", "0.48346448", "0.47855932", "0.47821313", "0.47626352", "0.47079334", "0.4697971", "0.46542692", "0.46254164", "0.46242422", "0.4623829", "0.46137628", "0.45952874", "0.45932218", "0.4581849", "0.45621845", "0.45601192", "0.45444447", "0.45422438", "0.45260713", "0.4521903", "0.4508509", "0.4503359" ]
0.80607086
0
Retrieve all sessions given by a particular speaker.
def _getSessionsBySpeaker(self, request): # Ensure that the speaker key is valid and that the speaker exists speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey, 'Speaker') # Return all of the speaker's sessions return ndb.get_multi(speaker.sessions)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_speaker_sessions(self, request):\n return self.session_service.get_speaker_sessions(\n request.websafeSpeakerKey)", "def getSessionsBySpeaker(self, request):\n sessions = self._getSessionsBySpeaker(request)\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def all (self):\n sparql_results = self.query (\"\"\"\n select distinct ?rs ?session ?name ?number ?pid ?sitename\n where {\n \n ?rs rdf:type austalk:RecordedSession .\n ?rs olac:speaker ?participant .\n \n ?participant austalk:id ?pid .\n ?participant austalk:recording_site ?site .\n ?site rdfs:label ?sitename .\n \n ?rs austalk:prototype ?session .\n ?session austalk:name ?name .\n ?session austalk:id ?number .\n }\n ORDER BY ?name\"\"\")\n\n results = []\n\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n results.append (Session (\n client = self.client,\n identifier = result[\"rs\"][\"value\"],\n prototype = result[\"session\"][\"value\"],\n name = result[\"name\"][\"value\"],\n number = result[\"number\"][\"value\"],\n site = result[\"sitename\"][\"value\"],\n participantId = result[\"pid\"][\"value\"]))\n\n return results", "def session_list(self, endpoint_name=None):\n if endpoint_name is None:\n _, body = self.request('/v1.1/endpoint/sessions', 'GET')\n else:\n _, body = self.request('/v1.1/endpoints/%s/sessions' % endpoint_name, 'GET')\n return body", "def get_sessions_by_speakers_in_wishlist(self, request):\n user = endpoints.get_current_user()\n return self.wishlist_service.get_sessions_by_speaker_in_wishlist(\n user)", "def filter_by_participant (self, participant):\n \n sparql_results = self.query (\"\"\"\n select distinct ?rs ?session ?name ?number ?pid ?sitename\n where {\n BIND (<%s> AS ?participant)\n \n ?rs rdf:type austalk:RecordedSession .\n ?rs olac:speaker ?participant .\n \n ?participant austalk:id ?pid .\n ?participant austalk:recording_site ?site .\n ?site rdfs:label ?sitename .\n \n ?rs austalk:prototype ?session .\n ?session austalk:name ?name .\n ?session austalk:id ?number .\n }\n ORDER BY ?name\"\"\" % participant.identifier)\n \n results = []\n\n for result in sparql_results[\"results\"][\"bindings\"]:\n results.append (Session (\n client = self.client,\n identifier = result[\"rs\"][\"value\"],\n prototype = result[\"session\"][\"value\"],\n name = result[\"name\"][\"value\"],\n number = result[\"number\"][\"value\"],\n site = result[\"sitename\"][\"value\"],\n participantId = result[\"pid\"][\"value\"]))\n\n return results", "def get_sessions_by_speaker_in_wishlist(self, user):\n sessions = self.wishlist_sessions(user)\n\n speaker_keys = []\n for s in sessions:\n sk = getattr(s, 'speakerKeys', [])\n speaker_keys += sk\n\n if not speaker_keys:\n return ConferenceSessionForms()\n\n sessions = ConferenceSession.query(\n ConferenceSession.speakerKeys.IN(speaker_keys)\n ).fetch()\n\n return ConferenceSessionForms(\n items=[self.copy_entity_to_form(ConferenceSessionForm(), s)\n for s in sessions])", "def get_speakers(self, request):\n return self.speaker_service.get_speakers()", "def _getConferenceSessions(self, request):\n # Ensure that websafeConferenceKey is a valid conference key\n confKey = _raiseIfWebsafeKeyNotValid(request.websafeConferenceKey,\n 'Conference')\n # Retrieve all sessions that have a matching conference key\n sessions = Session.query(Session.conference == confKey).fetch()\n return sessions", "def sessions(self):\n logger.debug(\"Get sessions\")\n return self._raw_api.sessions.get()", "def list_speaker(request, template=\"core/list_speaker.html\"):\n speakers = Speaker.objects.all()\n response = { 'speakers': speakers, 'show_all_info': False }\n return direct_to_template(request, template, response)", "def get_conference_sessions(self, request):\n return self.session_service.get_conference_sessions(\n request.websafeConferenceKey)", "def get_sessions(self):\n\n return self.all_sessions", "def get_sessions_by_type(self, request):\n return self.session_service.get_conference_sessions_by_type(\n request.websafeConferenceKey, request.sessionType)", "def get_sessions_by_type_and_filters(self, request):\n return self.session_service.get_sessions_by_type_and_filters(\n request.websafeConferenceKey,\n request.typeOfSession, request.filters)", "def get_user_messages(self, speaker):\n # type: (str) -> [Message]\n return list(filter(lambda m: m.speaker == speaker, self.messages))", "def get_sessions_list():\n sessions = Session.query.all()\n result = sessions_schema.dump(sessions).data\n return jsonify({'status': 'success', 'message': None, 'data': result}), 200", "def search_sessions(name: str, provider: Optional[str] = None) -> List[str]:\n sessions = session_list(provider=provider).sessions\n name = name.lower()\n return [s.id for s in sessions if s.id.lower().startswith(name)]", "def GetSessions(firebase: firebase) -> None:\n\n global sessions\n obj_key_list = []\n \n result = firebase.get('/session', None)\n \n if result is None:\n print(\"no sessions found\")\n return\n \n for i in result.keys():\n obj_key_list.append(i)\n \n for i in obj_key_list:\n session = Session()\n session.setId(i)\n session.setCourseId(result[i]['courseid'])\n session.setDOW(result[i]['DOW'])\n session.setSessionNumber(result[i]['session_number'])\n session.setSessionDate(result[i]['session_date'])\n session.setSessionTimeStart(result[i]['session_time_start'])\n session.setSessionTimeEnd(result[i]['session_time_end'])\n sessions.append(session)", "def sessions(self):\n return utils.listItems(self, '/status/sessions')", "def sessions(self, *args, **kwargs):\r\n return self._get('Sessions', *args, **kwargs)", "def get_all_sessions(self) -> list:\n sessions = list()\n for stream_id in self.streams.keys():\n tcpsession, session_position, network_tuple = self.streams[stream_id]\n sessions.append(tcpsession.get_session(session_position - 1))\n return sessions", "def sessions(self):\n return list(Session.get_sessions(self))", "def getDetailedSpeakers(self, speakers):\n\n new_speakers = []\n try:\n for speaker in speakers:\n for user in self.users_data:\n if speaker[\"username\"] == user[\"username\"]:\n new_speakers.append(user)\n return new_speakers\n except KeyError as e:\n print(e)\n return \"Invalid\"", "def list(self, datasource_name=None, event_name=None, requested_after=None, requested_before=None, session_type=None):\n query = {\n 'dataSourceName': datasource_name,\n 'eventName': event_name,\n 'requestedBefore': requested_before,\n 'requestedAfter': requested_after,\n 'sessionType': session_type\n }\n response, _, headers = self._client.request_with_headers('GET', 'sessions', params=query)\n\n return [SessionResponse(item, headers) for item in response.get('items', [])]", "def getSpeakers(self, request):\n speakers = Speaker.query().order(Speaker.name).fetch()\n # Return individual SpeakerForm object per Speaker\n return SpeakerForms(\n items=[self._copySpeakerToForm(speaker) for speaker in speakers]\n )", "def speaker_viewall(request):\n is_loggedin, username = get_session_variables(request)\n speaker_list = Speaker.objects.all()\n\n if speaker_list:\n return render_to_response('achievement/speaker_viewall.html', \\\n {'is_loggedin':is_loggedin, \\\n 'username':username, \\\n 'speaker_list':speaker_list}, \\\n RequestContext(request))\n else:\n return render_to_response('achievement/noview.html', \\\n {'is_loggedin':is_loggedin, \\\n 'username':username, \\\n 'type': 'Speaker'}, \\\n RequestContext(request))", "def getSessionByUsername(self, username):\n match = []\n for session in self.sessions:\n if (session.identifier[1] == username):\n match.append(session)\n return match", "def get_sessions(sessions, time_feat_dict):\n filt = Session.filter_time_func(time_feat_dict)\n return [s for s in sessions if filt(shortstr2time(s['start']))]", "def filter_by_session (self, site_id, participant_id, session_id):\n\n query = \"\"\"\n select distinct * where {\n\n BIND (\"%s\" AS ?pid)\n BIND (\"%s\" as ?sessionid)\n \n ?participant austalk:id ?pid .\n ?rc rdf:type austalk:RecordedComponent .\n ?rc olac:speaker ?participant .\n ?rc austalk:session ?sessionid .\n\n ?rc austalk:prototype ?prototype .\n ?prototype austalk:name ?name .\n ?prototype austalk:shortname ?shortname .\n \n optional { ?rc austalk:audiorating ?audiorating .}\n optional { ?rc austalk:videorating ?videorating .}\n optional { ?rc austalk:comment ?comment .}\n \n }\n \"\"\" % (participant_id, session_id)\n\n\n sparql_results = self.query (query)\n results = []\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n for field in ['audiorating', 'videorating', 'comment']:\n if field not in result:\n result[field] = {'value': ''}\n\n comp = Component (\n client = self.client,\n identifier = result[\"rc\"][\"value\"],\n participantId = result[\"pid\"][\"value\"],\n sessionId = result[\"sessionid\"][\"value\"],\n prototype = result[\"prototype\"][\"value\"],\n audiorating = result[\"audiorating\"][\"value\"],\n videorating = result[\"videorating\"][\"value\"],\n comment = result[\"comment\"][\"value\"],\n name = result[\"name\"][\"value\"],\n componentId = result[\"shortname\"][\"value\"],\n site = site_id,\n )\n comp.details()\n results.append(comp)\n return results" ]
[ "0.78710306", "0.7094641", "0.67780095", "0.67494416", "0.67155963", "0.6615327", "0.64614207", "0.6425885", "0.6333383", "0.6315707", "0.6194976", "0.61806315", "0.6179548", "0.61346084", "0.61164176", "0.60324794", "0.60129184", "0.5877937", "0.58242875", "0.5822374", "0.5812562", "0.5807763", "0.5756186", "0.5755378", "0.57523334", "0.5748542", "0.5736504", "0.5694652", "0.56806624", "0.5662731" ]
0.8179627
0
Retrieve all sessions in the user's wishlist.
def _getSessionsInWishlist(self): user = endpoints.get_current_user() if not user: raise endpoints.UnauthorizedException('Authorization required') profile = self._getProfileFromUser() # Fetch the entities and return them return ndb.get_multi(profile.sessionWishlist)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def wishlist_sessions(self, user):\n wishlist_key = self.get_wishlist_key(user)\n session_keys = [ndb.Key(urlsafe=wsck) for wsck in\n wishlist_key.get().sessionKeys]\n sessions = ndb.get_multi(session_keys)\n return sessions", "def get_sessions_in_wishlist(self, request):\n user = endpoints.get_current_user()\n return self.wishlist_service.get_sessions_in_wishlist(user)", "def get_sessions_by_speakers_in_wishlist(self, request):\n user = endpoints.get_current_user()\n return self.wishlist_service.get_sessions_by_speaker_in_wishlist(\n user)", "def get_sessions_by_types_in_wishlist(self, request):\n user = endpoints.get_current_user()\n return self.wishlist_service.get_sessions_by_types_in_wishlist(\n user)", "def get_sessions_in_wishlist(self, user):\n sessions = self.wishlist_sessions(user)\n\n session_service = SessionService()\n return ConferenceSessionForms(\n items=[\n session_service.copy_entity_to_form(\n ConferenceSessionForm(), session)\n for session in sessions])", "def get_sessions_by_speaker_in_wishlist(self, user):\n sessions = self.wishlist_sessions(user)\n\n speaker_keys = []\n for s in sessions:\n sk = getattr(s, 'speakerKeys', [])\n speaker_keys += sk\n\n if not speaker_keys:\n return ConferenceSessionForms()\n\n sessions = ConferenceSession.query(\n ConferenceSession.speakerKeys.IN(speaker_keys)\n ).fetch()\n\n return ConferenceSessionForms(\n items=[self.copy_entity_to_form(ConferenceSessionForm(), s)\n for s in sessions])", "def getSessionsInWishlist(self, request):\n sessions = self._getSessionsInWishlist()\n # Return individual SessionForm object per Session\n return SessionForms(\n items=[self._copySessionToForm(session) for session in sessions]\n )", "def get_sessions_by_types_in_wishlist(self, user):\n sessions = self.wishlist_sessions(user)\n\n types = [getattr(s, 'typeOfSession') for s in sessions]\n\n sessions = ConferenceSession.query(\n ConferenceSession.typeOfSession.IN(types)\n ).fetch()\n\n return ConferenceSessionForms(\n items=[self.copy_entity_to_form(ConferenceSessionForm(), s)\n for s in sessions])", "def addSessionToWishlist(self, request):\n return self._addSessionToWishlist(request)", "def get_sessions(self):\n\n return self.all_sessions", "def display_all_wishlists():\n # \"\"\" Display wishlists of all customers if created\"\"\"\n if Customer.display_all():\n message = [Customer.find_by_custid(k) for k in Customer.redis.keys()]\n return make_response(jsonify(message),status.HTTP_200_OK)\n else:\n message = {'Error' : 'No wishlist created for any customer'}\n return make_response(jsonify(message),status.HTTP_404_NOT_FOUND)", "def sessions(self):\n return list(Session.get_sessions(self))", "def get_sessions_list():\n sessions = Session.query.all()\n result = sessions_schema.dump(sessions).data\n return jsonify({'status': 'success', 'message': None, 'data': result}), 200", "def _addSessionToWishlist(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n # Verify that the session actually exists\n session = _getEntityByWebsafeKey(request.websafeSessionKey, 'Session')\n profile = self._getProfileFromUser()\n if session.key not in profile.sessionWishlist:\n profile.sessionWishlist.append(session.key)\n profile.put()\n return BooleanMessage(data=True)", "def sessions(self):\n return utils.listItems(self, '/status/sessions')", "def sessions(self):\n logger.debug(\"Get sessions\")\n return self._raw_api.sessions.get()", "def add_session_to_wishlist(self, request):\n return self.wishlist_service.add_session_to_wishlist(\n request.websafeSessionKey, endpoints.get_current_user())", "def get_sessions(self):\n return self.current_sessions", "def get_users_list(self, session):\n\n users = session.query(User.chat_id).filter(User.is_admin==False).all()\n return users", "def removeSessionFromWishlist(self, request):\n return self._removeSessionFromWishlist(request)", "def get_booked_sessions(self):\n return [session for session in self.sessions if session.is_booked()]", "def get_current_users(self):\n active_sessions = Session.objects.filter(expire_date__gte=timezone.now())\n user_id_list = []\n for session in active_sessions:\n data = session.get_decoded()\n user_id_list.append(data.get('_auth_user_id', None))\n # Query all logged in users based on id list\n return self.filter(id__in=user_id_list)", "def my_wishlist_view(request):\n data = {'success': False, 'msg': '', 'wishlist': []}\n if request.method == 'GET':\n # check if the user has already logged in.\n # if user has not logged in, return an error msg to frontend.\n # if user has logged in, let user view his/her wishlist\n if not request.session.get('login_flag', None):\n data['msg'] = 'user does not log in'\n return JsonResponse(data)\n # else use is logged in\n user_name = request.session.get('name', None)\n # return user_obj by user_name from login.models.User database\n try:\n user_obj = login.models.User.objects.get(name=user_name)\n except ObjectDoesNotExist:\n data['msg'] = 'does not have user: ' + str(user_name)\n return JsonResponse(data)\n\n data['success'] = True\n data['msg'] = 'successfully get wishlist of the current user'\n\n movie_id_list = list(\n models.Wish_list.objects.filter(user__exact=user_obj).order_by('movie').values_list('movie_id', flat=True))\n useful_keys = {'mid', 'name', 'region', 'released_date', 'average_rating', 'poster'}\n for mid in movie_id_list:\n movie_obj = models.Movie.objects.get(mid=mid)\n movie_dict = movie_to_dict(movie_obj, request)\n data['wishlist'].append({key: value for key, value in movie_dict.items() if key in useful_keys})\n\n return JsonResponse(data)\n\n else:\n data['msg'] = 'please use GET'\n return JsonResponse(data)", "def sessions(self):\n return self._sessions", "def user_wishlist(request):\n if request.user.is_authenticated:\n wishlist = get_object_or_404(Wishlist, user=request.user)\n\n # Pagination show 12 products per page\n paginator = Paginator(wishlist.products.all().order_by('id'), 12)\n\n page = request.GET.get('page')\n try:\n all_wishlist = paginator.page(page)\n except PageNotAnInteger:\n # If page is not an integer, deliver first page.\n all_wishlist = paginator.page(1)\n except EmptyPage:\n # If page is out of range (e.g. 9999),\n # deliver last page of results.\n all_wishlist = paginator.page(paginator.num_pages)\n\n # Pagination was inspired, modified and\n # adapted to this project from from this\n # # Credit code\n # https://www.youtube.com/watch?v=MAIFJ3_bcCY\n index = all_wishlist.number - 1\n max_index = len(paginator.page_range)\n start_index = index - 2 if index >= 2 else 0\n end_index = index + 3 if index <= max_index - 3 else max_index\n page_range = paginator.page_range[start_index:end_index]\n\n context = {'wishlist': all_wishlist,\n 'page_range': page_range, }\n\n else:\n context = {\n 'wishlist': [],\n }\n\n return context", "def select_wish_list(self, user_id):\n with self.connection:\n data = self.cursor.execute(\"SELECT wish_list FROM Wish_list WHERE user_id=?\", (user_id,)).fetchall()\n for row in data:\n new_row = ''.join(row)\n return new_row", "def _removeSessionFromWishlist(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n profile = self._getProfileFromUser()\n # Get actual session key from websafe key\n sessionKey = _raiseIfWebsafeKeyNotValid(request.websafeSessionKey,\n 'Session')\n # If the key is in the profile's session wishlist, remove it\n if sessionKey in profile.sessionWishlist:\n profile.sessionWishlist.remove(sessionKey)\n profile.put()\n retval = True\n else:\n retval = False\n return BooleanMessage(data=retval)", "def get_all_users(session):\n\n return session.query(User).all()", "def _sessions(self):\n return self.__sessions", "def remove_session_from_wishlist(self, request):\n return self.wishlist_service.remove_session_from_wishlist(\n request.websafeSessionKey, endpoints.get_current_user())" ]
[ "0.851661", "0.79517555", "0.7718727", "0.7539904", "0.74082303", "0.69785184", "0.6908265", "0.6861531", "0.6479205", "0.6430566", "0.61952454", "0.6161288", "0.61303025", "0.6128384", "0.6097037", "0.6083551", "0.60752475", "0.5979322", "0.59384423", "0.5929593", "0.5898026", "0.585059", "0.5824485", "0.58242124", "0.5777362", "0.5772263", "0.5714416", "0.5704339", "0.57015413", "0.56902325" ]
0.8228693
1
Removes a session from the user's wishlist, returning a boolean.
def _removeSessionFromWishlist(self, request): # Preload necessary data items user = endpoints.get_current_user() if not user: raise endpoints.UnauthorizedException('Authorization required') profile = self._getProfileFromUser() # Get actual session key from websafe key sessionKey = _raiseIfWebsafeKeyNotValid(request.websafeSessionKey, 'Session') # If the key is in the profile's session wishlist, remove it if sessionKey in profile.sessionWishlist: profile.sessionWishlist.remove(sessionKey) profile.put() retval = True else: retval = False return BooleanMessage(data=retval)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def remove_session_from_wishlist(self, request):\n return self.wishlist_service.remove_session_from_wishlist(\n request.websafeSessionKey, endpoints.get_current_user())", "def removeSessionFromWishlist(self, request):\n return self._removeSessionFromWishlist(request)", "def remove_session_from_wishlist(self, websafe_session_key, user):\n wishlist = self.get_wishlist_key(user).get()\n if wishlist is None or wishlist.sessionKeys is []:\n raise ConflictException(\"This session is not in your wishlist.\")\n\n if websafe_session_key not in wishlist.sessionKeys:\n raise ConflictException(\n \"This session is not in your wishlist.\")\n\n wishlist.sessionKeys.remove(websafe_session_key)\n wishlist.put()\n\n return self.to_message(wishlist)", "def remove_session(self, session):\n if session in self.sessions:\n self.sessions.remove(session)\n else:\n print(\"Sorry, you can't remove that session.\")", "def unbook_session(self, session):\n confirm = input(\"Are you sure you want to unbook this session? (y/n) \")\n if confirm.lower().startswith(\"y\") and session in self.sessions:\n session.remove_student()\n session.remove_subject()\n self.sessions.remove(session)", "def remove_session(self) -> None:\n pass", "def remove_wish_list(self, user_id, caption):\n with self.connection:\n data = self.cursor.execute(\"SELECT wish_list FROM Wish_list WHERE user_id=?\", (user_id,)).fetchall()\n for row in data:\n new_row = ''.join(row)\n if new_row == caption:\n return self.cursor.execute(f\"DELETE FROM wish_list WHERE wish_list=? AND user_id={user_id}\", (row))", "def delete_session(self, _id: int) -> bool:\n logger.info(\"deleting session: %s\", _id)\n session = self.sessions.pop(_id, None)\n result = False\n if session:\n logger.info(\"shutting session down: %s\", _id)\n session.data_collect()\n session.shutdown()\n result = True\n else:\n logger.error(\"session to delete did not exist: %s\", _id)\n return result", "def logout_user(session):\n del session['user']", "def remove(self, session: \"pwncat.manager.Session\"):", "def destroy_wishlist(request):\n wishlist = Wishlist.objects.get(pk=request.POST.get('id'))\n title = wishlist.name\n if wishlist.customer != request.user:\n request.user.message_set.create(message=\"You can't delete wishlists that aren't yours!\")\n return HttpResponseRedirect(reverse(\"get_wishlist\"))\n\n if request.method == \"POST\" and request.POST[\"action\"] == \"delete\":\n wishlist.delete()\n request.user.message_set.create(message=_(\"Successfully deleted wishlist '%s'\") % title)\n return HttpResponseRedirect(reverse(\"get_wishlist\"))", "def remove_favourite(recipe_id):\r\n if \"user\" in session:\r\n user = coll_users.find_one({\"username_lower\": session[\"user\"]})[\"_id\"]\r\n coll_users.update_one(\r\n {\"_id\": ObjectId(user)},\r\n {\"$pull\": {\"user_favs\": ObjectId(recipe_id)}})\r\n coll_recipes.update(\r\n {\"_id\": ObjectId(recipe_id)}, {\"$inc\": {\"favourites\": -1}})\r\n return redirect(url_for(\r\n \"recipes.recipe_detail\",\r\n recipe_id=recipe_id))\r\n else:\r\n flash(\"You must be logged in to perform that action!\")\r\n return redirect(url_for(\"users.login\"))", "def do_logout():\n del session[CURRENT_USER_KEY]", "def remove_user(self, user: discord.User) -> bool:\n\t\tif not self.user_has_entry(user):\n\t\t\treturn False\n\t\t\n\t\tdef data_interaction(cur: Cursor):\n\t\t\tsql = f\"DELETE FROM {StrikeConsts.STRIKE_TABLE} WHERE id=%s;\"\n\t\t\tcur.execute(sql, (user.id,))\n\t\t\t\n\t\t\treturn [True]\n\t\t\t\n\t\treturn self.connect_and_execute(data_interaction)[1][0]", "def _addSessionToWishlist(self, request):\n # Preload necessary data items\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n # Verify that the session actually exists\n session = _getEntityByWebsafeKey(request.websafeSessionKey, 'Session')\n profile = self._getProfileFromUser()\n if session.key not in profile.sessionWishlist:\n profile.sessionWishlist.append(session.key)\n profile.put()\n return BooleanMessage(data=True)", "def remove_from_wishlist(request, product_id):\n\n redirect_url = request.POST.get('redirect_url')\n\n user = get_object_or_404(UserProfile, user=request.user)\n wishlist = Wishlist.objects.get_or_create(user=user)\n wishlist_user = wishlist[0]\n if request.POST:\n product = Product.objects.get(pk=product_id)\n\n # look for item in the user's wishlistItem - returns true if it exists\n existingWishlistItem = WishlistItem.objects.filter(\n product=product).exists()\n\n if existingWishlistItem:\n product = WishlistItem.objects.get(product=product)\n product.delete()\n messages.success(request, \"Item removed from wishlist\")\n return redirect(redirect_url)\n\n if existingWishlistItem is None:\n messages.error(\n request, \"You can not delete a item thats not in the wishlist\")\n return redirect(redirect_url)\n else:\n messages.error(request, 'Item can not be deleted from your wishlist')\n return render(request, 'home/index.html')", "def rm_from_fav(show_id, name):\n\n db = get_db()\n db.execute(\n 'DELETE FROM shows_users WHERE show_id = ? and user_id = ?',\n (show_id, session['user_id'])\n )\n\n flash('\\\"%s\\\" has been successfully removed from your favourite TV Shows!' % name)\n db.commit()\n return redirect(request.referrer)", "def remove_from_wishlist(self, beer_id: str):\n method = \"user/wishlist/delete\"\n auth = self._get_access_token()\n params = {\"bid\": beer_id}\n return self._do_get(method, auth, params)", "def delete_wishlist(cust_id, wishlist_id):\n # \"\"\" Delete the wishlist with the provided id\"\"\"\n success = Customer.delete_by_id(cust_id, wishlist_id)\n return make_response('', status.HTTP_204_NO_CONTENT)", "def disconnect_user(self, user):\n\t\tis_user_removed = False\n\t\tif user in self.users.all():\n\t\t\tself.users.remove(user)\n\t\t\tself.save()\n\t\t\tis_user_removed = True\n\t\treturn is_user_removed", "def clear_wishlist(request):\n wishlist = get_object_or_404(Wishlist, user=request.user)\n try:\n wishlist.products.clear()\n messages.info(request,\n \"Success! Your wishlist has been cleared!\")\n return redirect(reverse(\"view_wishlist\"))\n except Exception as e:\n messages.error(request, f\"Error clearing wishlist {e}\")\n return redirect(reverse(\"products\"))", "def delete_session(self, session_token):\n\n return self.user_manager.delete_user_session(session_token)", "def delete_session(self, session):\n db = self.open()\n del db[session.id]", "def remove_session(self, session):\n with self._sessions_lock:\n if session.session_id not in self.sessions:\n logger.info(\"session id %s not found (sessions=%s)\", session.session_id, self.sessions.keys())\n else:\n del self.sessions[session.session_id]\n\n return session", "def destroy_session(self, session):\n self._obex_client_proxy.proxy.RemoveSession(session.path)", "def logout():\n if session.get('authed', False):\n for i in ['phone', 'authed', 'confirmation_code']:\n if session.has_key(i):\n del session[i]\n return redirect(my_url('index'))", "def delete_session(self,session_id):\n del self.sessions[session_id]", "def add_session_to_wishlist(self, request):\n return self.wishlist_service.add_session_to_wishlist(\n request.websafeSessionKey, endpoints.get_current_user())", "def delete_session(db, useremail):\n db.cursor().execute('DELETE FROM sessions WHERE useremail IS ?', [useremail])\n db.commit()", "def dropsession():\n session.pop('user', None)\n return redirect(url_for('login'))" ]
[ "0.81915665", "0.76984537", "0.76317555", "0.6835183", "0.6607075", "0.6478658", "0.6209266", "0.61053085", "0.6085781", "0.6084388", "0.6060252", "0.5823839", "0.57571155", "0.57422817", "0.57323414", "0.57050776", "0.5699716", "0.56986994", "0.5697193", "0.5694112", "0.5676643", "0.5668976", "0.5656363", "0.5610834", "0.557966", "0.5575929", "0.55643874", "0.5549331", "0.55481994", "0.554149" ]
0.8256191
0
Get list of sessions matching one or more of the given highlights.
def getSessionsByHighlightSearch(self, request): sessions = self._getSessionsByHighlightSearch(request) # Return individual SessionForm object per Session return SessionForms( items=[self._copySessionToForm(session) for session in sessions] )
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _getSessionsByHighlightSearch(self, request):\n # Generate list of filters from the highlight arguments\n filters = [Session.highlights == hl for hl in request.highlights]\n if not filters:\n raise endpoints.BadRequestException(\n 'At least one highlight must be specified'\n )\n # Retrieve all sessions that match one or more of the highlight filters\n sessions = Session.query(ndb.OR(*filters)).order(Session.name).fetch()\n return sessions", "def get_sessions(sessions, time_feat_dict):\n filt = Session.filter_time_func(time_feat_dict)\n return [s for s in sessions if filt(shortstr2time(s['start']))]", "def search_sessions(name: str, provider: Optional[str] = None) -> List[str]:\n sessions = session_list(provider=provider).sessions\n name = name.lower()\n return [s.id for s in sessions if s.id.lower().startswith(name)]", "def getSessionByUsername(self, username):\n match = []\n for session in self.sessions:\n if (session.identifier[1] == username):\n match.append(session)\n return match", "def findSessions(self, channel):\n found = []\n for ss in self.sessions:\n try:\n _channel = channel.decode(ss.encoding)\n if _channel == ss.name:\n found.append(ss)\n if ss.matchNick(_channel):\n found.append(ss)\n except UnicodeDecodeError:\n continue\n if found == []:\n found = [self.defaultSession]\n return found", "def get_highlights(lines, split_on='@highlight'):\n text = \" \".join(lines)\n splits = text.split(split_on)\n return splits[:1], splits[1:]", "def all (self):\n sparql_results = self.query (\"\"\"\n select distinct ?rs ?session ?name ?number ?pid ?sitename\n where {\n \n ?rs rdf:type austalk:RecordedSession .\n ?rs olac:speaker ?participant .\n \n ?participant austalk:id ?pid .\n ?participant austalk:recording_site ?site .\n ?site rdfs:label ?sitename .\n \n ?rs austalk:prototype ?session .\n ?session austalk:name ?name .\n ?session austalk:id ?number .\n }\n ORDER BY ?name\"\"\")\n\n results = []\n\n for result in sparql_results[\"results\"][\"bindings\"]:\n\n results.append (Session (\n client = self.client,\n identifier = result[\"rs\"][\"value\"],\n prototype = result[\"session\"][\"value\"],\n name = result[\"name\"][\"value\"],\n number = result[\"number\"][\"value\"],\n site = result[\"sitename\"][\"value\"],\n participantId = result[\"pid\"][\"value\"]))\n\n return results", "def _getSessionsInWishlist(self):\n user = endpoints.get_current_user()\n if not user:\n raise endpoints.UnauthorizedException('Authorization required')\n profile = self._getProfileFromUser()\n # Fetch the entities and return them\n return ndb.get_multi(profile.sessionWishlist)", "def find_queries(self, senses):\n queries = set()\n for token in senses:\n queries.add(token)\n\n if (len(queries) == 0):\n focus = self.q_classifer.matched_groups['focus'].split()\n focus = [f for f in focus if f not in set(stopwords.words('english'))]\n queries = focus\n\n return queries", "def highlightSearch(self, wordList=None, regExpList=None):\n backColor = self.palette().brush(QPalette.Active,\n QPalette.Highlight)\n foreColor = self.palette().brush(QPalette.Active,\n QPalette.HighlightedText)\n if wordList is None:\n wordList = []\n if regExpList is None:\n regExpList = []\n for regExp in regExpList:\n for match in regExp.finditer(self.toPlainText()):\n matchText = match.group()\n if matchText not in wordList:\n wordList.append(matchText)\n selections = []\n for word in wordList:\n while self.find(word):\n extraSel = QTextEdit.ExtraSelection()\n extraSel.cursor = self.textCursor()\n extraSel.format.setBackground(backColor)\n extraSel.format.setForeground(foreColor)\n selections.append(extraSel)\n cursor = QTextCursor(self.document())\n self.setTextCursor(cursor) # reset main cursor/selection\n self.setExtraSelections(selections)", "def get_sessions_by_speakers_in_wishlist(self, request):\n user = endpoints.get_current_user()\n return self.wishlist_service.get_sessions_by_speaker_in_wishlist(\n user)", "def _getSessionsBySpeaker(self, request):\n # Ensure that the speaker key is valid and that the speaker exists\n speaker = _getEntityByWebsafeKey(request.websafeSpeakerKey, 'Speaker')\n # Return all of the speaker's sessions\n return ndb.get_multi(speaker.sessions)", "def get_intersect_run() -> List[str]:\n subj_runs = get_runs_all_subjects()\n return list(set.intersection(*map(set, subj_runs.values())))", "def global_matches(self, text):\n matches = []\n n = len(text)\n for list in [keyword.kwlist,\n __builtin__.__dict__,\n self.namespace]:\n for word in list:\n if word[:n] == text:\n matches.append(word)\n return matches", "def sessions(self):\n return utils.listItems(self, '/status/sessions')", "def sessions(self, *args, **kwargs):\r\n return self._get('Sessions', *args, **kwargs)", "def get_colors(lines):\n\n patt = re.compile('\\#\\w+')\n\n return [\n patt.search(line).group(0)\n for line in lines\n if patt.search(line)\n ]", "def wishlist_sessions(self, user):\n wishlist_key = self.get_wishlist_key(user)\n session_keys = [ndb.Key(urlsafe=wsck) for wsck in\n wishlist_key.get().sessionKeys]\n sessions = ndb.get_multi(session_keys)\n return sessions", "def highlight_syntax(self) -> List[SyntaxHighlight]:\n raise NotImplementedError", "def get_active_sessions():\n\n # The output changes based on locales, force it to be YY-MM-DD\n # for the benefit of split()\n os.environ['LANG'] = 'en_GB.utf8'\n try:\n output = subprocess.check_output(['who']).rstrip()\n except subprocess.CalledProcessError:\n print 'UNKNOWN: unable to invoke who'\n sys.exit(NAGIOS_UNKNOWN)\n\n # Nothing to process\n if not output:\n return {}\n\n sessions = {}\n for line in output.split(\"\\n\"):\n fields = line.split()\n sessions[fields[1]] = {\n 'user': fields[0],\n 'date': fields[2],\n 'time': fields[3],\n 'source': fields[4][1:-1] if len(fields) >= 5 else None,\n }\n\n return sessions", "def get_sessions_by_types_in_wishlist(self, request):\n user = endpoints.get_current_user()\n return self.wishlist_service.get_sessions_by_types_in_wishlist(\n user)", "def highlight_series(s):\n return ['background-color: #eee' for v in s]", "def _group_sessions(self, sessions):\n session_dict = collections.defaultdict(list)\n for session in sessions:\n session_dict[session.query].append(session)\n return session_dict", "def clients_search_by_term_view(request):\n # Check connected\n if not check_connected(request):\n raise exc.HTTPForbidden()\n\n searchTerm = request.params[\"searchTerm\"] if \"searchTerm\" in request.params else None\n old_clients = request.params['old_clients'] == 'true' if 'old_clients' in request.params else False\n\n clients = _multipleAttributesClientSearch(request, searchTerm, old_clients=old_clients)\n \n return Utils.serialize_many(clients)", "def get_sessions(self, network_tuple: NetworkTuple) -> list:\n tcpsession = self.sessions[network_tuple]\n session_list = tcpsession.get_sessions_list()\n return session_list", "def get_sessions_list():\n sessions = Session.query.all()\n result = sessions_schema.dump(sessions).data\n return jsonify({'status': 'success', 'message': None, 'data': result}), 200", "def get_sessions(url: str, token: str) -> List[Session]:\n sessions_url = f'{url}api/sessions'\n response = requests.get(sessions_url, params={'token': token})\n assert(response.status_code == 200)\n sessions_raw = json.loads(response.text)\n sessions = []\n for session_raw in sessions_raw:\n session = Session(\n path = session_raw['path'],\n last_activity = dateutil.parser.isoparse(session_raw['kernel']['last_activity']),\n execution_state = session_raw['kernel']['execution_state']\n )\n assert(session['execution_state'] in valid_execution_states)\n sessions.append(session)\n\n sessions.sort(key=lambda session: session['last_activity'], reverse=True)\n return sessions", "def sessions(self):\n return list(Session.get_sessions(self))", "def in_(self, to, *edge_classes):\n records = self.client.command('SELECT in({0}) FROM {1}'\n .format(','.join(self.coerce_class_names(edge_classes))\n , self.coerce_class_names(to)))\n return [self.get_vertex(v) for v in records[0].oRecordData['in']] \\\n if records else []", "def find_all(v):\n screen = G.DEVICE.snapshot(quality=ST.SNAPSHOT_QUALITY)\n return v.match_all_in(screen)" ]
[ "0.79675597", "0.5494635", "0.5191284", "0.5032431", "0.5007258", "0.49804562", "0.48610216", "0.4822324", "0.48123288", "0.47999722", "0.47981682", "0.47678235", "0.47661048", "0.47624898", "0.47622135", "0.47583574", "0.4731185", "0.4729603", "0.46977532", "0.4695372", "0.4686059", "0.46783894", "0.46611592", "0.466013", "0.46524796", "0.46305284", "0.46297705", "0.4627549", "0.4618356", "0.46109447" ]
0.6377008
1
Copy relevant fields from Profile to ProfileForm.
def _copyProfileToForm(self, prof): pf = ProfileForm() for field in pf.all_fields(): if hasattr(prof, field.name): # Convert t-shirt string to Enum; just copy others if field.name == 'teeShirtSize': setattr(pf, field.name, getattr(TeeShirtSize, getattr(prof, field.name))) else: setattr(pf, field.name, getattr(prof, field.name)) pf.check_initialized() return pf
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_form_kwargs(self):\n kwargs = super(ProfileUpdate, self).get_form_kwargs()\n self.profile = self.get_object()\n extra_kwargs = {'profile': self.profile}\n kwargs.update(extra_kwargs)\n return kwargs", "def _doProfile(self, save_request=None):\n prof = self._getProfileFromUser()\n # If saveProfile(), process user-modifyable fields\n if save_request:\n for field in ('displayName', 'teeShirtSize'):\n if hasattr(save_request, field):\n val = getattr(save_request, field)\n if val:\n print(val)\n setattr(prof, field, str(val))\n prof.put()\n # Return ProfileForm\n return self._copyProfileToForm(prof)", "def __fill_applicant_form(self, profile):\r\n\r\n actions = ActionChains(self.driver)\r\n actions.send_keys(profile['name'] + Keys.TAB + \\\r\n profile['email'] + Keys.TAB + \\\r\n profile['phone_number'] + Keys.TAB)\r\n actions.perform()", "def save(self, *args, **kwargs):\n\n kwargs[\"commit\"] = False\n user = super(JOSProfileForm, self).save(*args, **kwargs)\n\n try:\n profile = get_profile_for_user(user)\n profile_form = self.get_profile_fields_form()\n profile_form(self.data, self.files, instance=profile).save()\n except ProfileNotConfigured:\n pass\n\n return user", "def user_profile(request):\n instance = Profile.objects.get(pk=request.user.pk)\n if request.method == \"POST\":\n\n form = ProfileForm(request.POST, request.FILES, instance=instance)\n form.save()\n return redirect(reverse('index'))\n messages.error(request, \"Profile Updated\")\n\n profile = ProfileForm(instance=instance)\n return render(request, 'profile.html', {'profile': profile, 'instance': instance})", "def add_profile_details(request):\n\n if request.method == 'POST':\n try:\n profile = Profile.objects.get(user=request.user)\n edit_profile = EditProfileForm(request.POST, instance=profile)\n if edit_profile.is_valid():\n profile.save()\n messages.success(request, 'Your profile has been updated')\n if profile.user_type == 'dismantler':\n return redirect(reverse('addlisting'))\n else:\n return redirect(reverse('listings'))\n except ObjectDoesNotExist:\n profile_form = EditProfileForm(request.POST)\n if profile_form.is_valid():\n profile = Profile.objects.create(\n user=request.user,\n user_type=profile_form.cleaned_data['user_type'],\n business_name=profile_form.cleaned_data['business_name'],\n phone=profile_form.cleaned_data['phone'],\n postcode=profile_form.cleaned_data['postcode'],\n city=profile_form.cleaned_data['city'],\n street_address=profile_form.cleaned_data['street_address'],\n street_address2=profile_form.cleaned_data[\n 'street_address2'],\n county=profile_form.cleaned_data['county'],\n country=profile_form.cleaned_data['country'],\n )\n profile.save()\n messages.success(request, 'Your profile has been saved')\n if profile.user_type == 'dismantler':\n return redirect(reverse('addlisting'))\n else:\n return redirect(reverse('listings'))\n else:\n try:\n profile = Profile.objects.get(user=request.user)\n profile_form = EditProfileForm(instance=profile)\n context = {\n 'profile': profile,\n 'profile_form': profile_form,\n }\n except ObjectDoesNotExist:\n profile_form = EditProfileForm()\n context = {\n 'profile_form': profile_form,\n }\n\n return render(request, 'profile.html', context)", "def save(self):\n # First save the parent form and get the user.\n new_user = super(SignupFormExtra, self).save()\n\n # Get the profile, the `save` method above creates a profile for each\n # user because it calls the manager method `create_user`.\n # See: https://github.com/django-userena-ce/django-userena-ce/blob/master/userena/managers.py#L65\n profile = new_user.my_profile\n profile.gender = self.cleaned_data['gender']\n profile.education = self.cleaned_data['education']\n profile.birthday = self.cleaned_data['birthday']\n profile.annual_income = self.cleaned_data['annual_income']\n profile.save()\n\n # Userena expects to get the new user from this form, so return the new\n # user.\n return new_user", "def save(self, profile_callback=None):\n\n # First, save the parent form\n new_user = super(BodbRegistrationForm, self).save(profile_callback=profile_callback)\n\n # Update user with first, last names\n new_user.first_name = self.cleaned_data['first_name']\n new_user.last_name = self.cleaned_data['last_name']\n new_user.save()\n\n # Update profile with affiliation\n profile = new_user.get_profile()\n profile.affiliation = self.cleaned_data['affiliation']\n profile.save()\n\n cache.set('%d.profile' % new_user.id, profile)\n\n return new_user", "def edit_profile(request):\n form = ProfileForm(instance=request.user.profile)\n if request.method == \"POST\":\n form = ProfileForm(data=request.POST, files=request.FILES,\n instance=request.user.profile)\n if form.is_valid():\n form.save()\n return redirect('profile')\n return render(request, 'accounts/forms.html', {'form': form})", "def getEditUser(self, request):\n\n userForm = UserEditForm(instance=request.user)\n profileForm = ProfileEditForm(instance=request.user.profile)\n return (userForm, profileForm)", "def profile(request):\n profile = get_object_or_404(Profile, user=request.user)\n albums = profile.albums.all()\n plc_albums = albums.exclude(is_public=False)\n pvt_albums = albums.exclude(is_public=True)\n sent_f_requests = FriendRequest.objects.filter(\n from_user=profile.user\n )\n rec_f_requests = FriendRequest.objects.filter(\n to_user=profile.user\n )\n\n if request.method == 'POST':\n form = ProfileModelForm(\n request.POST or None,\n request.FILES or None,\n instance=profile\n )\n\n if form.is_valid():\n form.save()\n messages.success(request, \"Profile updated successfully\")\n\n form = ProfileModelForm(instance=profile)\n\n template = 'profiles/profile.html'\n context = {\n 'profile': profile,\n 'form': form,\n 'albums': albums,\n 'plc_albums': plc_albums,\n 'pvt_albums': pvt_albums,\n 'sent_req': sent_f_requests,\n 'rec_req': rec_f_requests,\n }\n return render(request, template, context)", "def copyform(form, settings):\r\n for name, value in form.iteritems():\r\n setattr(settings, name, value)\r\n settings.commit()", "def edit_user_profile(request):\n user = request.user\n user_profile = UserProfile.objects.filter(user=user)[0]\n if request.method == 'POST':\n form = MemberProfileForm(request.POST)\n additional_form = MemberAdditionalProfileForm(request.POST)\n if form.is_valid() and additional_form.is_valid():\n cd = form.cleaned_data\n user.first_name = cd['first_name']\n user.last_name = cd['last_name']\n user.email = cd['email']\n user.save()\n if 'picture' in request.FILES:\n file = request.FILES['picture']\n user_profile.picture.save(file.name, file, save=True)\n user_profile.gravatar = additional_form.cleaned_data['gravatar']\n user_profile.save()\n return HttpResponseRedirect('/')\n else:\n form = MemberProfileForm(instance=request.user)\n additional_form = MemberAdditionalProfileForm(instance=user_profile)\n return render_to_response('edit_profile.html', locals())", "def manage_myprofile(request):\n profile = request.user.get_profile()\n users_image = profile.users_image\n if not profile:\n raise Http404\n if request.method == 'POST':\n profile_form = MyProfileForm(request.POST, instance = profile)\n address_contact_form = AddressForm(request.POST,\n instance = profile.address_contact, prefix = 'contact')\n address_permanent_form = AddressForm(request.POST,\n instance = profile.address_permanent, prefix = 'permanent')\n\n if profile_form.is_valid() and address_contact_form.is_valid() \\\n and address_permanent_form.is_valid():\n address_contact = address_contact_form.save()\n address_permanent = address_permanent_form.save()\n\n profile_form.save(address_contact = address_contact,\n address_permanent = address_permanent)\n messages.success(request,\n _('your profile details saved sucessfully'))\n else:\n profile_form = MyProfileForm(instance = profile)\n address_contact_form = AddressForm(instance = profile.address_contact,\n prefix = 'contact')\n address_permanent_form = AddressForm(instance\n = profile.address_permanent, prefix = 'permanent')\n\n return render(request, 'myprofile.html', {\n 'profile_form': profile_form,\n 'address_contact_form': address_contact_form,\n 'address_permanent_form': address_permanent_form,\n 'users_image': users_image\n },\n )# Create your views here.", "def edit_profile():\n form = EditProfileForm()\n if request.method == 'GET':\n form.first_name.data = current_user.first_name\n form.first_name.data = current_user.first_name\n form.last_name.data = current_user.last_name\n form.email.data = current_user.email\n form.address_1.data = current_user.address_1\n form.address_2.data = current_user.address_2\n form.city.data = current_user.city\n form.state.data = current_user.state\n form.zipcode.data = current_user.zipcode\n form.telephone.data = current_user.telephone\n if form.validate_on_submit():\n form.last_name.data = form.last_name.data\n current_user.first_name = form.first_name.data\n current_user.last_name = form.last_name.data\n current_user.email = form.email.data\n current_user.address_1 = form.address_1.data\n current_user.address_2 = form.address_2.data\n current_user.city = form.city.data\n current_user.state = form.state.data\n current_user.zipcode = form.zipcode.data\n current_user.telephone = form.telephone.data\n db.session.commit()\n flash(('Your changes have been saved.'))\n\n return redirect(url_for('edit_profile'))\n\n return render_template('edit_profile.html', title=('Edit Profile'),\n form=form)", "def get_form(self):\n kwargs = {\n \"instance\": self.profile if self.form_object == \"profile\" else self.user,\n \"prefix\": self.name,\n }\n\n if self.request.method == \"POST\":\n return self.form_class(self.request.POST, self.request.FILES, **kwargs)\n else:\n return self.form_class(**kwargs)", "def get_form(self, request, obj=None, **kwargs):\n defaults = {}\n if obj is None:\n defaults['form'] = self.add_form\n defaults.update(kwargs)\n return super(ProfileAdmin, self).get_form(request, obj, **defaults)", "def edit_profile(request):\r\n\r\n user = request.user\r\n profile = Profile.objects.for_user(user)\r\n\r\n if request.method != 'POST':\r\n profile_form = ProfileForm(instance=profile)\r\n user_form = UserForm(instance=user)\r\n else:\r\n profile_form = ProfileForm(request.POST, instance=profile)\r\n user_form = UserForm(request.POST, instance=user)\r\n\r\n if profile_form.is_valid() and user_form.is_valid():\r\n profile_form.save()\r\n user_form.save()\r\n\r\n return HttpResponseRedirect(reverse('epic.core.views.view_profile', kwargs={}))\r\n\r\n return render_to_response(\r\n 'core/edit_profile.html',\r\n {'profile_form': profile_form, 'user_form': user_form,},\r\n context_instance=RequestContext(request))", "def profile(request):\n profile = get_object_or_404(UserProfile, user=request.user)\n\n if request.method == 'POST':\n form = UserProfileForm(request.POST, instance=profile)\n user_email = UserForm(instance=request.user)\n if form.is_valid():\n form.save()\n messages.success(request, 'Profile updated successfully')\n else:\n messages.error(request, 'Update failed. Please ensure the form is valid.')\n # doesnt work yet: Email is not saved:\n\n if user_email.is_valid():\n user_email.save()\n messages.success(request, 'Profile updated successfully')\n else:\n messages.error(request, 'Update failed. Please ensure the form is valid.')\n else:\n form = UserProfileForm(instance=profile)\n user_email = UserForm(instance=request.user)\n\n template = 'profiles/profile.html'\n context = {\n 'form': form,\n 'user_email': user_email,\n 'on_profile_page': True,\n 'profile': profile\n }\n\n return render(request, template, context)", "def show_profile(request, profile_uuid):\n context = context_assign_user(request.user)\n context['profile_form'] = ProfileForm(instance=context['current_user'])\n if request.method == 'POST':\n form = ProfileForm(request.POST)\n if form.is_valid():\n Profile.objects.filter(pk=context['current_user'].id).update(bio=request.POST['bio'], palette=request.POST['palette'], iex_api_key=request.POST['iex_api_key'])\n messages.success(request, 'Your settings have been saved.')\n return redirect('dash:dashboard')\n errors = form.errors\n form = ProfileForm(request, request.POST)\n messages.warning(request, f\"There's a problem with the form: {errors}\")\n return render(request, 'dash/users/show_profile.html', context)", "def make_fields(self):\n for name, prop in self.edit:\n instance_value = self.model.get(name)\n post_value = self.data[name] if (self.data and self.data.has_key(name)) else instance_value\n form_field_class = self.get_field_type(prop)\n form_field = form_field_class(model=self.model, property=prop, name=name, instance_value=instance_value, post_value=post_value)\n self.add(form_field)", "def profile_edit_view(request):\n if request.method == 'POST':\n user_model_form = UserModelForm(\n request.POST,\n prefix=\"user\",\n instance=request.user\n )\n profile_model_form = ProfileSignupForm(\n request.POST,\n prefix=\"profile\",\n instance=request.user.profile\n )\n forms = [user_model_form, profile_model_form]\n\n if request.user.profile.is_mentor():\n mentor_form = MentorForm(\n request.POST,\n prefix=\"mentor\",\n instance=request.user.profile.mentor\n )\n forms.append(mentor_form)\n\n if request.user.profile.is_mentee():\n mentee_form = MenteeForm(\n request.POST,\n prefix=\"mentee\",\n instance=request.user.profile.mentee\n )\n forms.append(mentee_form)\n\n forms_is_valid_list = []\n for form in forms:\n forms_is_valid_list.append(form.is_valid())\n\n if all(forms_is_valid_list):\n for form in forms:\n form.save()\n return redirect(\"private_profile\")\n\n elif request.method == 'GET':\n forms = [\n UserModelForm(\n instance=request.user,\n prefix=\"user\"\n ),\n ProfileSignupForm(\n instance=request.user.profile,\n prefix=\"profile\"\n )\n ]\n\n if request.user.profile.is_mentor():\n forms.append(\n MentorForm(\n instance=request.user.profile.mentor,\n prefix=\"mentor\"\n )\n )\n\n if request.user.profile.is_mentee():\n forms.append(\n MenteeForm(\n instance=request.user.profile.mentee,\n prefix=\"mentee\"\n )\n )\n\n return render(\n request,\n 'mentorship_profile/profile_edit.html',\n {\n \"forms\": forms\n }\n )", "def edit_profile(profile_id):\n # This check is in place to avoid users trying to edit a profile via the dashboard\n # when they have not created one. If not the option is not displayed\n user = mongo.db.user.find_one({'username': session['username']})\n chck = mongo.db.profiles.find_one_or_404({'user_id': user['_id']})\n if chck: \n profile = mongo.db.profiles.find_one(\n {'_id': ObjectId(profile_id)})\n \n form=ProfileForm()\n form.headline.data = profile['headline']\n form.bio.data = profile['bio']\n form.xp.data = profile['xp']\n form.interests.data = profile['interests']\n form.stack.data = profile['stack']\n form.languages.data = profile['languages']\n form.frameworks.data = profile['frameworks']\n form.github.data = profile['github']\n form.linkedin.data = profile['linkedin']\n \n return render_template('pages/editprofile.html', form=form, profile=profile, legend='Edit your Profile')", "def createOrgProfileFromForm(self):\n\n if self.data.org:\n form = OrgProfileForm(self.data.POST, instance=self.data.org)\n else:\n form = OrgCreateProfileForm(self.data.POST)\n\n if not form.is_valid():\n return None\n\n if not self.data.org:\n form.cleaned_data['founder'] = self.data.user\n form.cleaned_data['scope'] = self.data.program\n form.cleaned_data['scope_path'] = self.data.program.key().name() \n key_name = '%s/%s' % (\n self.data.program.key().name(),\n form.cleaned_data['link_id']\n )\n entity = form.create(commit=True, key_name=key_name)\n self.data.profile.org_admin_for.append(entity.key())\n self.data.profile.put()\n else:\n entity = form.save(commit=True)\n\n return entity", "def form_valid(self, form, profile_form):\n self.object = form.save()\n # print(self.object.shopuserprofile)\n # print(self.object)\n # print(self.request.user)\n\n return HttpResponseRedirect(self.get_success_url())", "def get_context_with_form(self):\n self.context['form'] = {\n 'profile': ProfileEditForm(),\n 'avatar': AvatarForm(),\n 'cover': CoverForm(),\n 'introduction': IntroductionForm(),\n 'statement': StatementForm(),\n 'experience': ExperienceForm(),\n 'work': WorkForm(),\n 'image': ImageForm(),\n 'url': UrlForm(),\n 'education': EducationForm(),\n }\n self.context['form']['introduction'] = IntroductionForm(\n initial={'introduction': self.context['introduction'].introduction}\n )\n self.context['form']['statement'] = StatementForm(\n initial={'statement': self.context['statement'].statement}\n )\n return self.context", "def _replace_fields(self):\n for name, value in self._cleaned_data.items():\n setattr(self, name, value)", "def save(self):\n # First save the parent form and get the user.\n new_user = super(SignupFormExtra, self).save()\n\n new_user.first_name = self.cleaned_data['first_name']\n new_user.last_name = self.cleaned_data['last_name']\n new_user.save()\n\n # Userena expects to get the new user from this form, so return the new\n # user.\n return new_user", "def __init__(self, *args, **kwargs):\n super(ProfileForm, self).__init__(*args, **kwargs) \n for field in self.fields:\n self.fields[field].widget.attrs.update(\n {\n 'class': 'form-control',\n }\n )", "def save(self):\n data = self.cleaned_data\n data.pop('password_confirmation')\n user = User.objects.create_user(**data)\n profile = Profile(user=user)\n profile.save()" ]
[ "0.6958025", "0.68784887", "0.64856434", "0.63592464", "0.6284996", "0.6272869", "0.6269607", "0.6214233", "0.62111384", "0.6183875", "0.6135155", "0.612614", "0.61004287", "0.6085827", "0.60856766", "0.6062952", "0.59629256", "0.58970684", "0.5893775", "0.5886109", "0.58743477", "0.58734435", "0.58305913", "0.5796513", "0.5763562", "0.5753989", "0.5700853", "0.5689489", "0.56884724", "0.5687185" ]
0.79479015
0
Return Profile from datastore, creating new one if nonexistent.
def _getProfileFromUser(self): # Make sure user is authenticated user = endpoints.get_current_user() if not user: raise endpoints.UnauthorizedException('Authorization required') # Get Profile from datastore user_id = user.email() p_key = ndb.Key(Profile, user_id) profile = p_key.get() # Create new Profile if not there if not profile: profile = Profile( key = p_key, displayName = user.nickname(), mainEmail= user.email(), teeShirtSize = str(TeeShirtSize.NOT_SPECIFIED), ) profile.put() return profile
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get(self, name='default'):\r\n if name not in self._cache:\r\n full_path = os.path.join(self._base_folder, name)\r\n self._create_if_needed(full_path)\r\n self._cache[name] = Profile(full_path, name)\r\n return self._cache[name]", "def create_or_update_profile(sender, instance, created, **kwargs):\n if created:\n Profile.objects.get_or_create(user=instance)\n instance.profile.save()", "def get_full_profile(self) -> Profile:\n return Profile(**{**self.profile, **self.contact})", "async def create_profile_for_user(self, *, profile_create: ProfileCreate) -> ProfileInDB:\n created_profile = await self.db.fetch_one(query=CREATE_PROFILE_FOR_USER_QUERY, values=profile_create.dict())\n return ProfileInDB(**created_profile)", "def get_profile(profile_id):\n profile = Profile.objects.get(id=profile_id)\n return profile", "def create_pootle_profile(sender, instance, **kwargs):\n try:\n profile = instance.get_profile()\n except PootleProfile.DoesNotExist:\n profile = PootleProfile(user=instance)\n profile.save()", "def get_profile():\n if environ['DB_INSTANCE'] in request.url_root:\n profile_id = request.form['id']\n profile = ndb.Key(Profile, profile_id).get()\n if profile is not None:\n activity_data = json.loads(profile.activity_data)\n items = activity_data.get('items', [])\n item = items[0]\n return json.dumps(item)\n \n # else (not DB_INSTANCE)\n return ''", "def createProfile(self):\n if self.profile:\n return\n from soc.modules.gsoc.models.profile import GSoCProfile\n user = self.createUser()\n properties = {'link_id': user.link_id, 'student_info': None, 'user': user,\n 'parent': user, 'scope': self.program, 'status': 'active'}\n self.profile = seeder_logic.seed(GSoCProfile, properties)", "def createUserProfile(user):\n MyProfile.objects.get_or_create(user=user)", "def create_profile(sender, instance, created, **kwargs):\n if created:\n profile, created = UserProfile.objects.get_or_create(user=instance)", "def create_profile(sender, instance, created, **kwargs):\n if created: \n profile, new = UserProfile.objects.get_or_create(user=instance)", "def create(self, validated_data):\n request = self.context.get('request')\n profile = Profile(**validated_data)\n profile.user = request.user\n profile.save()\n return profile", "def load_profile(path, profile):\n profiles = load(path)\n try:\n return profiles[profile]\n except KeyError:\n return Profile(None, None, None, None)", "def create(self, validated_data):\r\n user_data = validated_data.pop('user')\r\n user = UserSerializer.create(UserSerializer(), validated_data = user_data)\r\n profile, created = Profile.objects.update_or_create(user = user,\r\n bio = validated_data.pop('bio'),\r\n location = validated_data.pop('location'),\r\n birth_date = validated_data.pop('birth_date'))\r\n return profile", "def create(profile, name):\n # Make sure it doesn't exist already.\n if exists(profile, name):\n msg = \"Instance profile '\" + str(name) + \"' already exists.\"\n raise ResourceAlreadyExists(msg)\n\n # Now we can create it.\n params = {}\n params[\"profile\"] = profile\n params[\"name\"] = name\n response = utils.do_request(instanceprofile, \"create\", params)\n\n # Check that it exists.\n instance_profile_data = polling_fetch(profile, name)\n if not instance_profile_data:\n msg = \"Instance profile '\" + str(name) + \"' not created.\"\n raise ResourceNotCreated(msg)\n\n # Send back the instance profile's info.\n return instance_profile_data", "def create_profile(username):\n user = User.objects.create(username=username)\n return Profile.objects.create(user=user)", "def create_profile(sender, **kwargs):\n\n # I import profile here cause i can't import it right in the top.\n from .profiles import Profile\n\n user = kwargs['instance']\n\n Profile.objects.get_or_create(user=user)", "def create_user_profile(instance, created, **_):\n if created:\n Profile.objects.create(user=instance)", "def create(self, validated_data):\n return ShopProfile.objects.create(**validated_data)", "def get(self, username):\n\t\tdb = getattr(g, 'db', None)\n\n\t\tqry = \"SELECT username,email,active,steamid FROM\\\n\t\t\tprofiles WHERE username = %s;\"\n\t\twith db as cursor:\n\t\t\tcursor.execute(qry, (username,))\n\n\t\treturn {'profile':cursor.fetchone()}", "def create(self, request, *args, **kwargs):\n response = super().create(request, *args, **kwargs)\n profile = response.data\n user_name = profile.get(\"username\")\n cache.set(f\"{USER_PROFILE_PREFIX}{user_name}\", profile)\n return response", "def profile_create(faker_obj=fake_init()):\n profile = faker_obj.simple_profile()\n user = User.objects.create(\n username=profile[\"username\"],\n email=profile[\"mail\"],\n password=profile[\"username\"][::-1],\n )\n return user.id", "def create_user_profile_callback(sender, instance, created, **kwargs):\n try:\n instance.get_profile()\n except UserProfile.DoesNotExist:\n UserProfile.objects.create(user=instance)", "def create_profile_for_new_users(sender, instance, created, **kwargs):\n if not created:\n return\n\n profile = Profile.objects.filter(user=instance).first()\n if profile is None:\n profile = Profile(user=instance)\n profile.save()", "def get_profile(user_id):\n\tudb = UserPageDB()\n\ttry:\n\t\t(name,) = udb.user_profile(user_id)\n\t\treturn Profile(name)\n\tfinally:\n\t\tudb.close()", "async def get_profile_by_username(self, *, username: str) -> ProfileInDB:\n profile = await self.db.fetch_one(query=GET_PROFILE_BY_USERNAME_QUERY, values={\"username\": username})\n if profile:\n return ProfileInDB(**profile)", "def create_profile(self, user):\n salt = sha.new(str(random.random())).hexdigest()[:5]\n activation_key = sha.new(salt+user.username).hexdigest()\n# prepend \"key_\" to the key_name, because key_names can't start with numbers\n registrationprofile = RegistrationProfile(user=user, activation_key=activation_key)\n db = DB_Session()\n db.add(registrationprofile)\n db.flush()\n db.refresh(registrationprofile)\n db.commit()\n db.close()\n return registrationprofile", "def create(self, validated_data):\n user_data = validated_data.pop('user')\n user = UserSerializer.create(UserSerializer(), validated_data=user_data)\n profile, created = Profile.objects.update_or_create(\n user=user,\n avatar=validated_data.pop('avatar'),\n biography=validated_data.pop('biography'),\n link=validated_data.pop('link') \n )\n return profile", "def get_profile_data(auth, db):\n\n id_team, user, team, money, color_prim, color_sec = analyze_init(auth, db)\n id_user, seats, fans, ranking, streak = analyze_team_page(auth, db, id_team)\n\n \n v_profile = profile.Profile(\n id_user, user, id_team, team, money, color_prim, \n color_sec, seats, fans, ranking, streak\n )\n\n if (db.profile.find_one({\"id\": int(id_team)}) is not None):\n db.profile.replace_one(\n {\"id\": int(id_team)}, v_profile.to_db_collection())\n else:\n db.profile.insert_one(v_profile.to_db_collection())\n\n print(show(\"profile\") + \" > Perfil actualizado\")\n\n return id_team", "def get_or_create(self, user):\n if user is None: return None\n prefs = UserPrefs.all().filter(\"user =\", user).get()\n if prefs is None:\n prefs = UserPrefs(user=user)\n prefs.put()\n return prefs" ]
[ "0.67713064", "0.6432612", "0.638044", "0.6346034", "0.6343368", "0.62816966", "0.6173148", "0.61137223", "0.61025006", "0.61001056", "0.6060391", "0.6048729", "0.60413766", "0.601181", "0.5955292", "0.5921494", "0.5829891", "0.58183765", "0.5805473", "0.5801193", "0.5772561", "0.5767348", "0.57504815", "0.57485306", "0.57350755", "0.57280165", "0.5727591", "0.57250595", "0.5724283", "0.5695092" ]
0.70515484
0
Create our CRITsDBAPI object. You may specify a full mongodb uri or the arguments individually.
def __init__(self, mongo_uri='', mongo_host='localhost', mongo_port=27017, mongo_user='', mongo_pass='', db_name='crits'): # If the user provided a URI, we will use that. Otherwise we will build # a URI from the other arguments. if mongo_uri != '': self.mongo_uri = mongo_uri else: # Build the authentication portion. Simple authentication only for # now. auth_str = '' if mongo_user != '': auth_str = mongo_user if mongo_pass != '' and mongo_user != '': auth_str = auth_str + ':' + mongo_pass if auth_str != '': auth_str = auth_str + '@' # Build the URI self.mongo_uri = 'mongodb://{}{}:{}'.format(auth_str, mongo_host, mongo_port) self.db_name = db_name self.client = None self.db = None
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_mongodb(config):\n\n \n mongo_url = \"mongodb://\"\n mongo_url += \",\".join(map(lambda srv: srv['host'] + \":\" + str(srv['port']), config['data']['mongoServers']))\n \n if 'replica' in config['data']:\n mongo_url += \"/?replicaSet={0}\".format(config['data']['replica'])\n\n client = MongoClient(mongo_url)\n\n return client", "def __init__(self, db_name):\n uri = re.sub('<username>', mongoUser, connection_uri)\n uri = re.sub('<password>', mongoPassword, uri)\n self.client = pymongo.MongoClient(uri)\n # Create your database\n self.db = self.client[db_name]", "def __init__(self, dst_mongodb_uri, dst_database, dst_collection, dry_run):\n self.client = pymongo.MongoClient(dst_mongodb_uri)\n self.dst_mongodb_uri = dst_mongodb_uri\n self.lookup_col = self.client[dst_database][dst_collection]\n self.dry_run = dry_run", "def __init__(self,public_key=Env.get(Env.PUBLIC_KEY)\n ,private_key=Env.get(Env.PRIVATE_KEY)\n ,org_id=Env.get(Env.ORG_ID)\n ,project_id=Env.get(Env.PROJECT_ID)\n ,base_url=\"https://cloud.mongodb.com\"):\n self.public_key = public_key \n self.private_key = private_key\n self.org_id = org_id \n self.project_id = project_id \n\n if isinstance(base_url,AtlasEnvironment):\n self.base_url = base_url.value\n else:\n self.base_url = base_url\n self.api_root = '{}{}'.format(base_url,ApiVersion.A1.value)", "def __init__(\r\n self, db, collection, host, port=27017, tz_aware=True, user=None, password=None, **kwargs\r\n ):\r\n self.database = pymongo.database.Database(\r\n pymongo.MongoClient(\r\n host=host,\r\n port=port,\r\n tz_aware=tz_aware,\r\n document_class=son.SON,\r\n **kwargs\r\n ),\r\n db\r\n )\r\n\r\n if user is not None and password is not None:\r\n self.database.authenticate(user, password)\r\n\r\n self.course_index = self.database[collection + '.active_versions']\r\n self.structures = self.database[collection + '.structures']\r\n self.definitions = self.database[collection + '.definitions']\r\n\r\n # every app has write access to the db (v having a flag to indicate r/o v write)\r\n # Force mongo to report errors, at the expense of performance\r\n # pymongo docs suck but explanation:\r\n # http://api.mongodb.org/java/2.10.1/com/mongodb/WriteConcern.html\r\n self.course_index.write_concern = {'w': 1}\r\n self.structures.write_concern = {'w': 1}\r\n self.definitions.write_concern = {'w': 1}", "def __init__(self, uri, dbname, collection=\"casbin_rule\"):\r\n client = MongoClient(uri)\r\n db = client[dbname]\r\n self._collection = db[collection]", "def __init__(self, **kwargs):\r\n\r\n super(MongoBackend, self).__init__(**kwargs)\r\n\r\n # Extract connection parameters from kwargs\r\n\r\n host = kwargs.get('host', 'localhost')\r\n port = kwargs.get('port', 27017)\r\n\r\n user = kwargs.get('user', '')\r\n password = kwargs.get('password', '')\r\n\r\n db_name = kwargs.get('database', 'track')\r\n collection_name = kwargs.get('collection', 'events')\r\n\r\n # Other mongo connection arguments\r\n extra = kwargs.get('extra', {})\r\n\r\n # By default disable write acknowledgments, reducing the time\r\n # blocking during an insert\r\n extra['w'] = extra.get('w', 0)\r\n\r\n # Make timezone aware by default\r\n extra['tz_aware'] = extra.get('tz_aware', True)\r\n\r\n # Connect to database and get collection\r\n\r\n self.connection = MongoClient(\r\n host=host,\r\n port=port,\r\n **extra\r\n )\r\n\r\n database = self.connection[db_name]\r\n\r\n if user or password:\r\n database.authenticate(user, password)\r\n\r\n self.collection = database[collection_name]\r\n\r\n self._create_indexes()", "def __init__(self,\n uri: str,\n database: str,\n col_name: str,\n *args,\n **kwargs):\n\n super().__init__(*args, **kwargs)\n self.client = MongoClient(uri)\n self.db = self.client[database]\n self.backends = [MongoBackend(uri, database, col_name=col_name)]\n self.col_name = col_name", "def __init__(self):\n client = pymongo.MongoClient(os.getenv(\"MONGO_URI\"))\n if DEBUG or os.getenv(\"CI\"):\n self.db = client[os.getenv(\"TESTDB\")]\n else:\n self.db = client[os.getenv(\"DB\")]", "def mongodb_init(cls, host=\"127.0.0.1\", port=27017, username=\"\", password=\"\", dbname=\"admin\"):\n if username and password:\n uri = \"mongodb://{username}:{password}@{host}:{port}/{dbname}\".format(username=quote_plus(username),\n password=quote_plus(password),\n host=quote_plus(host),\n port=port,\n dbname=dbname)\n else:\n uri = \"mongodb://{host}:{port}/{dbname}\".format(host=host, port=port, dbname=dbname)\n cls._mongo_client = motor.motor_asyncio.AsyncIOMotorClient(uri, connectTimeoutMS=5000, serverSelectionTimeoutMS=5000)\n #LoopRunTask.register(cls._check_connection, 2)\n SingleTask.call_later(cls._check_connection, 2) #模拟串行定时器,避免并发\n logger.info(\"create mongodb connection pool.\")", "def __init__(self, url = None, context = \"corbaserver\"):\n self._initOrb (url)\n self._makeClients (\"manipulation\", self.defaultClients, context)", "def __init__(\n self,\n client: pymongo.MongoClient,\n database_name: str,\n session_id: uuid.UUID,\n ):\n self.mongo_client = client\n self.session_id = session_id\n self.mongo_database = client.get_database(database_name)", "def __init__(self, host, port, dbname, bucketname=\"fs\", rsname=None):\n self.client = None\n self.host = host\n self.dbname = dbname\n self.bucketname = bucketname\n self.rsname = rsname\n try:\n if rsname:\n self.client = MongoReplicaSetClient(host, replicaSet=rsname)\n self.client.read_preference = ReadPreference.PRIMARY_PREFERRED\n #self.client.read_preference = ReadPreference.SECONDARY_PREFERRED\n #print self.client.seeds\n #print self.client.hosts\n #print self.client.read_preference\n #print self.client.primary\n #print self.client.secondaries\n #print self.client.arbiters\n else:\n self.client = MongoClient(host)\n self.db = Database(self.client, self.dbname)\n self.gridfs = GridFS(self.db, self.bucketname)\n except Exception, e:\n print e\n raise e", "def __init__(self, db_name='leaderboard'):\n key = os.getenv('ATLAS_KEY')\n self.valid = key is not None\n self.client = None\n self.database = None\n if self.valid:\n try:\n self.client = pymongo.MongoClient(key % db_name)\n self.database = self.client[db_name]\n except pymongo.errors.ConfigurationError:\n self.valid = False", "def __init__(self):\n client = MongoClient(\n settings.MONGODB_SERVER,\n settings.MONGODB_PORT\n )\n db = client[settings.MONGODB_DB]\n self.collection = db[settings.MONGODB_COLLECTION]", "def __init_db(self, db_name):\n\t\tclient = pymongo.MongoClient(self.__db_url)\n\t\treturn client[db_name]", "def _generate_client(self):\n mongoConf = self._config.get('Connectivity', 'MongoDB') # type: dict\n if mongoConf.get('username') and mongoConf.get('password'):\n return pymongo.MongoClient(\n \"mongodb://{0}:{1}@{2}:{3}/{4}\".format(\n mongoConf.get('username', ''),\n mongoConf.get('password', ''),\n mongoConf.get('host', 'localhost'),\n mongoConf.get('port', 27017),\n mongoConf.get('db', 'grease')\n ),\n w=1\n )\n else:\n return pymongo.MongoClient(\n host=mongoConf.get('host', 'localhost'),\n port=mongoConf.get('port', 27017),\n w=1\n )", "def get_db(db_config):\n hosts=[]\n db_uri=''\n\n for host in db_config['hosts']:\n hosts.append( host['host'] + \":\" + str(host['port'] ))\n\n db_uri = \"mongodb://\" + \\\n ','.join(hosts) + \\\n \"/?authSource=\" + db_config['auth_source'] + \\\n \"&replicaSet=\" + db_config['replica_set']\n\n\n db = MongoClient(\n db_uri,\n username = db_config['username'],\n password = db_config['password'],\n authMechanism = db_config['auth_mechanism'],\n ssl = (True if db_config['use_ssl'] else False),\n ssl_certfile = (db_config['ssl_certificate_file'] if db_config['ssl_certificate_file'] else None),\n ssl_ca_certs = (db_config['ssl_ca_file'] if db_config['ssl_ca_file'] else None),\n ssl_cert_reqs = (ssl.CERT_OPTIONAL if db_config['use_ssl'] else None),\n maxPoolSize = 5,\n wtimeout = 2500\n )[db_config['db_name']]\n \n return db", "def __init__(self, dbname='cards'):\n\n self._client = MongoClient()\n self._db = self._client[dbname]\n self._collection = self._db[dbname]\n self._keys = ['set', 'color', 'text', 'creator']", "def mongo_client():\n return MongoClient(STRING_CONNECTION)", "def __init__(self, api: MongoAPI):\n self.__api = api", "def __init__(self, client=None, timeout=300, database_name=None, host=None, port=None):\n self.client = MongoClient() if client is None else client\n self.db = self.client.cache\n self.timeout = timeout", "def mongodb_init():\n # Load environment variables\n dotenv_path = find_dotenv()\n load_dotenv(dotenv_path)\n\n # Connect to the db\n # DB will be created if it doesn't already exist\n client = pymongo.MongoClient(os.environ.get(\"DATABASE_URL\"), 27017)\n # 'tweetbase' is the database name\n db = client.tweetbase\n\n #Sets both the tweet ID and user ID strings as unique indexes\n db.tweets.create_index([(\"id_str\", 1),\n (\"user.id_str\", 1)],\n unique=True)\n\n # The db is only actually created when something is inserted\n # So this inserts a test document and immediately deletes it...\n # AND EVERYTHING ELSE\n # DO NOT RUN THIS ON A DB YOU WANT TO KEEP. SERIOUSLY.\n db.tweets.insert_one({\"id_str\": 1, \"user\": {\"id_str\": 5}})\n db.tweets.remove()", "def __init__(self, host=\"localhost\", port=27017):\n self._client = MongoClient(host, port)", "def __init__(self, collection, database=None, host=None, port=None,\r\n expand=False, **mongo_args):\r\n\r\n self.collection_name = collection\r\n self.database_name = database\r\n self.host = host\r\n self.port = port\r\n self.mongo_args = mongo_args\r\n self.expand = expand\r\n\r\n self.collection = None\r\n self.fields = None", "def create_api(self):\n return DJinnAPI(djenkins=self.dj, pipeline_results=self.db)", "def __init__(self, source='10.0.2.32', is_local=False):\n super().__init__(source, is_local)\n self.client = MongoClient(source)", "def __init__(self, collection, database=None, host=None, port=None,\r\n truncate=False, expand=False, **mongo_args):\r\n\r\n self.collection_name = collection\r\n self.database_name = database\r\n self.host = host\r\n self.port = port\r\n self.mongo_args = mongo_args\r\n self.expand = expand\r\n self.truncate = truncate\r\n\r\n self.collection = None\r\n self.fields = None", "def __init__(self, collection, database=None, host=None, port=None,\n expand=False, **mongo_args):\n\n self.collection_name = collection\n self.database_name = database\n self.host = host\n self.port = port\n self.mongo_args = mongo_args\n self.expand = expand\n\n self.collection = None\n self.fields = None", "def __init__(self, collection, database=None, host=None, port=None,\n truncate=False, expand=False, **mongo_args):\n\n self.collection_name = collection\n self.database_name = database\n self.host = host\n self.port = port\n self.mongo_args = mongo_args\n self.expand = expand\n self.truncate = truncate\n\n self.collection = None\n self.fields = None" ]
[ "0.66846025", "0.6621965", "0.65577734", "0.6492812", "0.64823735", "0.6253829", "0.61448485", "0.6078013", "0.5876883", "0.5867382", "0.58658755", "0.5850278", "0.58107173", "0.5787932", "0.57878685", "0.57876056", "0.5785224", "0.5769064", "0.5743249", "0.57387024", "0.57375634", "0.57332635", "0.57286066", "0.5725728", "0.5693413", "0.5676918", "0.56472945", "0.5640778", "0.56351656", "0.561532" ]
0.6649437
1
Search a collection for the query provided. Just a raw interface to mongo to do any query you want.
def find(self, collection, query): obj = getattr(self.db, collection) result = obj.find(query) return result
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def find(self, collection, query=None, filter=None):\n return self.__db[collection].find(query, filter)", "def find_document(collection: str, query: dict = None, regex: list = None) -> dict:\n if query is not None:\n return DB[collection].find_one(query)\n if regex is not None:\n return DB[collection].find_one({regex[0]: {'$regex': regex[1]}})\n raise Exception('Didnt specify a query or a regex')", "def query_mongo(query=None, collection=None, db=None, sort=settings.DEFAULT_SORT, number_of_results=settings.DEFAULT_NUMBER_OF_RESULTS, order=settings.DEFAULT_ORDER):\n if sort == 'rel':\n sort = 'pagerank'\n elif sort == 'dat':\n sort = 'date'\n elif sort == 'len':\n pass\n elif sort == 'cha':\n pass\n else:\n sort = None\n if query != None and collection != None and db != None:\n selected_collection = db.dbconnection[collection]\n cursor = selected_collection.find(query).sort(sort, order).limit(number_of_results)\n results_count = cursor.count()\n if collection == settings.COLLECTION_INDEX:\n results = {}\n for item in cursor:\n results[item['word']] = item['index']\n elif collection == settings.COLLECTION_DOCUMENTS:\n results = []\n for item in cursor:\n results.append(item)\n return results, results_count\n else:\n return [], 0", "def query_mongo(database_name, collection_name, query={},\n skip=0, sort=None, limit=settings.MONGO_LIMIT, return_keys=()): \n \n l=[]\n response_dict={}\n \n try:\n mc = MongoClient(host=settings.MONGO_HOST,\n port=settings.MONGO_PORT)\n \n db = mc[str(database_name)]\n collection = db[str(collection_name)]\n \n \n #print query\n if return_keys:\n return_dict={}\n for k in return_keys:\n return_dict[k]=1\n #print \"returndict=\",return_dict\n mysearchresult=collection.find(query, return_dict).skip(skip).limit(limit)\n else: \n mysearchresult=collection.find(query).skip(skip).limit(limit)\n \n if sort:\n mysearchresult.sort(sort)\n\n response_dict['code']=200\n response_dict['type']=\"search-results\"\n for d in mysearchresult:\n d['id'] = d['_id'].__str__()\n del d['_id']\n l.append(d)\n response_dict['results']=l\n \n except:\n print \"Error reading from Mongo\"\n print str(sys.exc_info())\n response_dict['code']=500\n response_dict['type']=\"Error\"\n response_dict['results']=[]\n response_dict['message']=str(sys.exc_info())\n \n return response_dict", "def find(self, query, collection_name='talent5__staff_collection'):\r\n collection = self.get_collection(collection_name)\r\n\r\n if collection is None:\r\n return None\r\n else:\r\n return collection.find(query, {'_id': False})", "def search(self, query, **kwargs):\n query = dict(query)\n if self._query:\n query = {'$and': [self._query, query]}\n cat = type(self)(\n metadatastore_db=self._metadatastore_db,\n asset_registry_db=self._asset_registry_db,\n query=query,\n find_kwargs=kwargs,\n handler_registry=self._handler_registry,\n transforms=self._transforms,\n root_map=self._root_map,\n filler_class=self._filler_class,\n name='search results',\n getenv=self.getenv,\n getshell=self.getshell,\n auth=self.auth,\n metadata=(self.metadata or {}).copy(),\n storage_options=self.storage_options)\n return cat", "async def query_documents(self, collection, query, filter_=None, offset=None,\n limit=None, fields=None, sort=None, facet_dict=None,\n cursor=None, def_type=None, query_fields=None,\n stats_fields=None):\n await self.ensure_collection(collection)\n\n # Query params which are not supported by JSON Request API yet\n # should go inside \"params\" attribute.\n # See https://lucene.apache.org/solr/guide/7_6/json-request-api.html\n # for more details.\n params = {\n key: value for key, value in [\n ('cursorMark', cursor),\n ('defType', def_type),\n ('qf', ' '.join(query_fields) if query_fields else ''),\n ('stats', 'true' if stats_fields else None),\n ('stats.field', stats_fields)\n ]\n if value is not None\n }\n json_data = {\n key: value for key, value in [\n ('query', query),\n ('filter', filter_),\n ('offset', offset),\n ('limit', limit),\n ('fields', fields),\n ('facet', facet_dict),\n ('sort', ','.join(sort) if sort else ''),\n ('params', params)\n ]\n if value is not None\n }\n\n try:\n response = await self.post(\n '/v2/collections/{}/query'.format(collection),\n json_data=json_data\n )\n json_response = json.loads(response.body.decode('utf-8'))\n query_response = json_response['response']\n stats = json_response.get('stats')\n solr_search_result = SolrSearchResult(\n num_found=query_response['numFound'],\n documents=query_response['docs'],\n cursor=json_response.get('nextCursorMark'),\n facet_results=json_response.get('facets', {}),\n stats_results=stats.get('stats_fields', {}) if stats else {}\n )\n logger.debug('Found {} and fetched {} documents from collection {}'\n .format(solr_search_result.num_found,\n len(solr_search_result.documents), collection))\n return solr_search_result\n except SolrError:\n logger.warning('Failed to execute query {} against collection {}'\n .format(json_data, collection))\n raise", "def search_collections(search_args):\n if search_args.topic:\n if search_args.guid:\n topic_records = db[search_args.topic].find({'headers.guid': search_args.guid})\n print_records(topic_records, search_args.topic)\n elif search_args.email:\n topic_records = db[search_args.topic].find({'value.email_address': search_args.email})\n print_records(topic_records, search_args.topic)\n else:\n print(\" Use --email or --guid to search for specific records. Use 'dump' to get entire collection\")\n print()\n else:\n list_of_collections = db.list_collection_names()\n if search_args.guid:\n for collection in list_of_collections:\n topic_records = db[collection].find({'headers.guid': search_args.guid})\n print_records(topic_records, collection)\n elif search_args.email:\n for collection in list_of_collections:\n topic_records = db[collection].find({'value.email_address': search_args.email})\n print_records(topic_records, collection)\n else:\n print(\" Use --email or --guid to search for specific records. Use 'dump' to get entire collection\")\n print()\n return", "def search(self, query):", "def search():\n if request.method == \"GET\":\n mongo_collection = mongo_database[\"questions\"]\n query = request.args.get(\"keyword\")\n result = mongo_collection.find({\"$text\": {\"$search\": query}})\n objects = []\n for object in result:\n objects.append(object)\n return render_template(\"search.html\", cards=objects)\n else:\n return start()", "def find(self, query_data=None):\n if query_data:\n return self.collection.find(query_data)\n return self.collection.find()", "def _search(self, query):\n return self._request(query)", "def run_search(col, search_name,routing, query_obj):\n\n\n # Sending Search request to MongoDB and printing the perf\n sys.stderr.write( \"Running search %s\\n\" % search_name)\n time1=datetime.datetime.now()\n\n # max number of results to fetch\n max_result=99999\n\n # definition of the cursor\n #################################\n cursor=col.aggregate(query_obj) #\n #################################\n\n ### Following commented code prints the execution plan! ###\n #import pprint\n #pp = pprint.PrettyPrinter()\n #pp.pprint(db.command('aggregate','small', pipeline=query_obj, explain=True))\n \n # fetch the data - We don't do anything with the result: we just want to record the time\n nbrows=0\n docs=[]\n for doc in cursor:\n docs.append(doc)\n nbrows+=1\n #print doc\n if nbrows>max_result: break\n \n time2=datetime.datetime.now()\n \n # The following print records KPIs to be gathered to build the benchmark results\n ##########################################################################################################\n print \"SEARCH: %s,%s,%d,%d\" % (search_name, routing, len(docs),int((time2-time1).total_seconds()*1000) ) #\n ##########################################################################################################", "def search():\n args = request.args.to_dict()\n query = QueryModel(args)\n result = repository.search_text(COLLECTION_NAME, query.value)\n return {\"texts\": result}", "def find_one(self, collection, query):\n obj = getattr(self.db, collection)\n result = obj.find_one(query)\n return result", "def search(self, query):\n return self._search_provider.search(self._normalize_query(query))", "def find_all(self, collection):\n obj = getattr(self.db, collection)\n result = obj.find()\n return result", "def find(self, **kwargs):\n url = self.build_url(dict_args_in_out=kwargs)\n\n rl = self._list(\n '%(url)s%(query)s' % {\n 'url': url,\n 'query': '?%s' % urllib.urlencode(kwargs) if kwargs else '',\n },\n self.collection_key)\n num = len(rl)\n\n if num == 0:\n return None\n elif num > 1:\n raise NoUniqueMatch\n else:\n return rl[0]", "def __search(self):\n query = self.__query.query()\n self.__engine.search(query)", "def search():\n kwargs = {k: parse(v) for k, v in request.args.to_dict().items()}\n query = kwargs.pop('q', None)\n cid = kwargs.pop('cid', None)\n\n if query:\n kwargs.setdefault('keywords', query)\n\n if cid:\n kwargs.setdefault('categoryId', cid)\n\n kwargs.setdefault('sortOrder', kwargs.pop('sort_order', 'EndTimeSoonest'))\n kwargs.setdefault('verb', 'findItemsAdvanced')\n limit = kwargs.pop('limit', 10)\n page = kwargs.pop('page', 1)\n finding = Finding(**kwargs)\n\n options = {'paginationInput': {'entriesPerPage': limit, 'pageNumber': page}}\n options.update(kwargs)\n\n try:\n response = finding.search(options)\n except ConnectionError as err:\n result = str(err)\n status = 500\n else:\n result = finding.parse(response)\n status = 200\n\n return jsonify(status, objects=result)", "def find(self, *args, **kwds):\n return self.collection.find(*args, **kwds)", "def search_documents(db, col, raw_result=False, scrub_result=False, **kwargs):\n aql = 'FOR d in ' + col.name\n bind_vars = {}\n if kwargs:\n aql += ' FILTER '\n count = 0\n for key in kwargs:\n if count > 0:\n aql += ' AND '\n\n if isinstance(kwargs[key]['value'], str):\n bind_vars[key] = '%' + str(kwargs[key]['value']) + '%'\n if kwargs[key]['op'] == 'like':\n aql += ('TRIM(UPPER(d.' + key +\n ')) LIKE TRIM(UPPER(@' + key + '))')\n else:\n raise Exception('unknown operation: ' + kwargs[key]['op'])\n else:\n bind_vars[key] = kwargs[key]['value']\n if kwargs[key]['op'] == 'lt-eq':\n aql += 'd.' + key + ' <= @' + key\n elif kwargs[key]['op'] == 'gt-eq':\n aql += 'd.' + key + ' >= @' + key\n else:\n raise Exception('unknown operation: ' + kwargs[key]['op'])\n\n count = count + 1\n\n aql += ' RETURN d'\n\n result = db.AQLQuery(aql, bindVars=bind_vars, rawResults=raw_result)\n\n if scrub_result:\n new_result = []\n for r in result:\n new_result.append(scrub_db_specific_data(r))\n result = new_result\n\n return result", "def find(self, queries):\n return self.__return_query('find', queries)", "def search(self, query=None, options=None):\n return self.create_search(query, options).search()", "def get_doc_by_keyword(self,collection,field_name,search_key,like=True):\n if like:\n # This finds the records in which the field just \"contains\" the search_key\n res = self.client['rephie'][collection].find(({field_name : {'$regex' : \".*\"+search_key+\".*\"}}))\n else:\n # This finds the records in which the field is equal to the search_key\n res = self.client['rephie'][collection].find({field_name : search_key})\n\n return self._make_result_list(res)", "def search():\n query = request.args['query']\n # find instances of the entered word in title, tags or ingredients\n results = mongo.db.places.find({\n '$or': [\n {'name': {'$regex': query, '$options': 'i'}},\n {'tags': {'$regex': query, '$options': 'i'}},\n {'city': {'$regex': query, '$options': 'i'}},\n ]\n })\n return render_template('search.html', query=query, results=results)", "def find(self, filter_: dict, projection: dict = None, db_name: str = None, collection_name: str = None, **kwargs) -> list[dict]:\n db: Database = self.db if db_name is None else self.client[db_name]\n coll: Collection = self.coll if collection_name is None else db[collection_name]\n\n return list(coll.find(filter_, projection=projection))", "async def find_one(self, collection_name: str, filter_dict: dict, *args, catch_error: bool = True, **kwargs):\n collection = self.database.get_collection(collection_name)\n result = None\n try:\n result = await collection.find_one(filter_dict, *args, **kwargs)\n self.debug_log(collection_name, f\"Found a document {result} in '{collection_name}' \"\n f\"with options '{args, kwargs}'\")\n except BaseException as e:\n self.error_log(collection_name, f\"Could not execute 'find_one'-command in '{collection_name}' \"\n f\"with options '{args, kwargs}'\",\n err=e)\n if not catch_error:\n raise e\n return result", "def find(cls, query_args, as_dict=False, as_cursor=False):\n cursor = cls.collection.find(query_args.query,\n query_args.select,\n sort=query_args.order_by,\n limit=query_args.limit)\n\n if as_cursor:\n return cursor\n else:\n return [record for record in cursor] if as_dict else [\n cls(record) for record in cursor\n ]", "def search():\n query = request.form.get(\"query\", None)\n recipes = mongo.db.recipes.find({\"$text\": {\"$search\": query}})\n return render_template(\"recipes/list.html\", recipes=recipes)" ]
[ "0.7252133", "0.7012578", "0.6853299", "0.6789239", "0.66799027", "0.6665288", "0.6664408", "0.65937394", "0.6434959", "0.6420416", "0.6398617", "0.62957525", "0.628199", "0.6266347", "0.625643", "0.62044865", "0.6122247", "0.6091136", "0.6090586", "0.60842377", "0.6064155", "0.60011506", "0.5984188", "0.59725124", "0.5947967", "0.5931973", "0.58816046", "0.5869922", "0.5839762", "0.58147985" ]
0.76442003
0
Search a collection for all available items.
def find_all(self, collection): obj = getattr(self.db, collection) result = obj.find() return result
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_all(collection):\n data = []\n for item in collection.find({}):\n data.append(item)\n return data", "def searchItems(name, allPages = False):\n return Gw2Spidy._paginatedRequest(allPages, 'item-search', name)", "def searchCollection(self, limit=100, **kwargs):\n results = self._CMR.get_search_results(url=self._SEARCH_COLLECTION_URL, limit=limit, **kwargs)\n return [Collection(result, self._MAAP_HOST) for result in results][:limit]", "def find_all(self):\n pass", "def get_many(collection, query: dict, selection=None) -> List[dict]:\n data = []\n if selection is None:\n for item in collection.find(query):\n data.append(item)\n else:\n for item in collection.find(query, selection):\n data.append(item)\n return data", "def get_all_documents(self, collection):\n\n return self.client[self.db][collection].find()", "def query_all(self):\n return multisearch.queries.QueryAll().connect(self)", "def get_collection_items(client, collection_id):\n\n try:\n items = client.collection(collection_id=collection_id).get_items()\n for item in items:\n print('{item.type.capitalize()} \"{item.name}\" is in the collection')\n\n except Exception as e:\n print(f\"An error has occurred: {e}\")\n return None", "def find_all(self):", "def _fetch_items(self):\n url = self._api.router.publication['search'].format(\n project_id=self.project_id\n )\n res_data = self._api.post(url, data=self.search_param)\n self.total = res_data['total']\n self._items = (\n Publication(item, self.project_id)\n for item in res_data['hits']\n )\n div = self.total // self.search_param['limit']\n reste = self.total % self.search_param['limit']\n self.total_page = div\n if reste != 0: self.total_page += 1\n self.search_param = self.search_param.next_page()", "def test_search_collection():\n col = Collection(search='forest', app=['gfw'])\n assert len(col) > 1", "def search(cls, collection, filters=None):\n apartments = collection.find(filters or {})\n return apartments", "def _collection_samples(collection_query, limit, config):\n just_cid = lambda obj : obj.get('meta', {}).get('concept-id')\n found_collections = scom.search_by_page(\"collections\",\n query=collection_query,\n filters=just_cid,\n page_state=scom.create_page_state(limit=limit),\n config=config)\n return found_collections[:limit]", "def all(self):\n self.scan()\n return self.entries", "def all(self) -> MongoDBQuerySet:\n\n self.cursor = self.collection.find()\n return self", "def fetch_all(): \n client, index_name = connection_es()\n res = client.search(index = index_name+\"*\")\n return res", "def list_all(mongo_collection):\n return [doc for doc in mongo_collection.find()]", "def find_all(session=None):\n return find(session=session)", "def search_collections(search_args):\n if search_args.topic:\n if search_args.guid:\n topic_records = db[search_args.topic].find({'headers.guid': search_args.guid})\n print_records(topic_records, search_args.topic)\n elif search_args.email:\n topic_records = db[search_args.topic].find({'value.email_address': search_args.email})\n print_records(topic_records, search_args.topic)\n else:\n print(\" Use --email or --guid to search for specific records. Use 'dump' to get entire collection\")\n print()\n else:\n list_of_collections = db.list_collection_names()\n if search_args.guid:\n for collection in list_of_collections:\n topic_records = db[collection].find({'headers.guid': search_args.guid})\n print_records(topic_records, collection)\n elif search_args.email:\n for collection in list_of_collections:\n topic_records = db[collection].find({'value.email_address': search_args.email})\n print_records(topic_records, collection)\n else:\n print(\" Use --email or --guid to search for specific records. Use 'dump' to get entire collection\")\n print()\n return", "def collections(self, query, page=1, per_page=10):\n url = \"/search/collections\"\n data = self._search(url, query, page=page, per_page=per_page)\n data[\"results\"] = CollectionModel.parse_list(data.get(\"results\"))\n return data", "def get_all_items(self):\n return self.api.state['items']", "async def get_all_collections():\n try:\n collections_query_result = get_db().AQLQuery(\n query=menu_queries.QUERY_ALL_COLLECTIONS\n )\n return {\"result\": collections_query_result.result}\n\n except DocumentNotFoundError as error:\n print(error)\n raise HTTPException(status_code=404, detail=\"Item not found\") from error\n except AQLQueryError as error:\n print(\"AQLQueryError: \", error)\n raise HTTPException(status_code=400, detail=error.errors) from error\n except KeyError as error:\n print(\"KeyError: \", error)\n raise HTTPException(status_code=400) from error", "def search_collections(self, keyword: str) -> List[str]:\n return [x for x in self.list_collections() if keyword.lower() in x]", "def test_find_collection(self):\n md = self.hiarc_util.get_test_metadata()\n c1 = self.hiarc_collections.create_collection(\n self.hiarc_util.create_collection(metadata=md))\n md[\"quotaCarrying\"] = False\n self.hiarc_collections.create_collection(\n self.hiarc_util.create_collection(metadata=md))\n self.hiarc_collections.create_collection(\n self.hiarc_util.create_collection())\n\n q = [{\n \"prop\": \"department\",\n \"op\": \"starts with\",\n \"value\": \"sal\"\n },\n {\n \"bool\": \"and\"\n },\n {\n \"parens\": \"(\"\n },\n {\n \"prop\": \"targetRate\",\n \"op\": \">=\",\n \"value\": 4.22\n },\n {\n \"bool\": \"and\"\n },\n {\n \"prop\": \"quotaCarrying\",\n \"op\": \"=\",\n \"value\": True\n },\n {\n \"parens\": \")\"\n }]\n\n qr = hiarc.FindCollectionsRequest(query=q)\n fc = self.hiarc_collections.find_collection(qr)\n assert len(fc) == 1\n assert self.hiarc_util.compare_dict_to_entity(fc[0], c1)", "def find_items(type_name, query, page_index):\n global es\n\n return es.search(index=index_name(type_name), body=query, from_=page_index * PAGE_SIZE, size=PAGE_SIZE)", "def get_items_to_find(self):\n self.items_to_find = ['sole', 'farina', 'innaffiatoio']", "def all(self):\n return self.filter()", "def all(self):\n return self.filter()", "def all(self):\n return self.filter()", "def get_all(self):\n return self.__items" ]
[ "0.66130906", "0.6577022", "0.6503677", "0.64557165", "0.62310874", "0.61904126", "0.6143242", "0.6072285", "0.60007083", "0.59808266", "0.5948115", "0.59253854", "0.58733493", "0.5871716", "0.5857312", "0.58287", "0.582313", "0.5817589", "0.57799107", "0.57534194", "0.57507545", "0.574257", "0.57425565", "0.5726938", "0.57211524", "0.5714542", "0.57138747", "0.57138747", "0.57138747", "0.5679512" ]
0.6775198
0
Search a collection for the query provided and return one result. Just a raw interface to mongo to do any query you want.
def find_one(self, collection, query): obj = getattr(self.db, collection) result = obj.find_one(query) return result
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def find(self, collection, query):\n obj = getattr(self.db, collection)\n result = obj.find(query)\n return result", "def find(self, collection, query=None, filter=None):\n return self.__db[collection].find(query, filter)", "def find_document(collection: str, query: dict = None, regex: list = None) -> dict:\n if query is not None:\n return DB[collection].find_one(query)\n if regex is not None:\n return DB[collection].find_one({regex[0]: {'$regex': regex[1]}})\n raise Exception('Didnt specify a query or a regex')", "def find(self, query, collection_name='talent5__staff_collection'):\r\n collection = self.get_collection(collection_name)\r\n\r\n if collection is None:\r\n return None\r\n else:\r\n return collection.find(query, {'_id': False})", "async def find_one(self, collection_name: str, filter_dict: dict, *args, catch_error: bool = True, **kwargs):\n collection = self.database.get_collection(collection_name)\n result = None\n try:\n result = await collection.find_one(filter_dict, *args, **kwargs)\n self.debug_log(collection_name, f\"Found a document {result} in '{collection_name}' \"\n f\"with options '{args, kwargs}'\")\n except BaseException as e:\n self.error_log(collection_name, f\"Could not execute 'find_one'-command in '{collection_name}' \"\n f\"with options '{args, kwargs}'\",\n err=e)\n if not catch_error:\n raise e\n return result", "def get_one(collection: Collection, query: Dict[str, Any]):\n data = collection.find_one(query)\n if data is None:\n raise CannotFindItemInDatabase(query, data, collection.name)\n return data", "def find_one(self, query, collection_name='talent5__staff_collection'):\r\n collection = self.get_collection(collection_name)\r\n\r\n if collection is None:\r\n return None\r\n else:\r\n return collection.find_one(query)", "def query_mongo(database_name, collection_name, query={},\n skip=0, sort=None, limit=settings.MONGO_LIMIT, return_keys=()): \n \n l=[]\n response_dict={}\n \n try:\n mc = MongoClient(host=settings.MONGO_HOST,\n port=settings.MONGO_PORT)\n \n db = mc[str(database_name)]\n collection = db[str(collection_name)]\n \n \n #print query\n if return_keys:\n return_dict={}\n for k in return_keys:\n return_dict[k]=1\n #print \"returndict=\",return_dict\n mysearchresult=collection.find(query, return_dict).skip(skip).limit(limit)\n else: \n mysearchresult=collection.find(query).skip(skip).limit(limit)\n \n if sort:\n mysearchresult.sort(sort)\n\n response_dict['code']=200\n response_dict['type']=\"search-results\"\n for d in mysearchresult:\n d['id'] = d['_id'].__str__()\n del d['_id']\n l.append(d)\n response_dict['results']=l\n \n except:\n print \"Error reading from Mongo\"\n print str(sys.exc_info())\n response_dict['code']=500\n response_dict['type']=\"Error\"\n response_dict['results']=[]\n response_dict['message']=str(sys.exc_info())\n \n return response_dict", "def query_mongo(query=None, collection=None, db=None, sort=settings.DEFAULT_SORT, number_of_results=settings.DEFAULT_NUMBER_OF_RESULTS, order=settings.DEFAULT_ORDER):\n if sort == 'rel':\n sort = 'pagerank'\n elif sort == 'dat':\n sort = 'date'\n elif sort == 'len':\n pass\n elif sort == 'cha':\n pass\n else:\n sort = None\n if query != None and collection != None and db != None:\n selected_collection = db.dbconnection[collection]\n cursor = selected_collection.find(query).sort(sort, order).limit(number_of_results)\n results_count = cursor.count()\n if collection == settings.COLLECTION_INDEX:\n results = {}\n for item in cursor:\n results[item['word']] = item['index']\n elif collection == settings.COLLECTION_DOCUMENTS:\n results = []\n for item in cursor:\n results.append(item)\n return results, results_count\n else:\n return [], 0", "def find(self, query_data=None):\n if query_data:\n return self.collection.find(query_data)\n return self.collection.find()", "def find(self, *args, **kwds):\n return self.collection.find(*args, **kwds)", "def find(self, **kwargs):\n url = self.build_url(dict_args_in_out=kwargs)\n\n rl = self._list(\n '%(url)s%(query)s' % {\n 'url': url,\n 'query': '?%s' % urllib.urlencode(kwargs) if kwargs else '',\n },\n self.collection_key)\n num = len(rl)\n\n if num == 0:\n return None\n elif num > 1:\n raise NoUniqueMatch\n else:\n return rl[0]", "async def query_documents(self, collection, query, filter_=None, offset=None,\n limit=None, fields=None, sort=None, facet_dict=None,\n cursor=None, def_type=None, query_fields=None,\n stats_fields=None):\n await self.ensure_collection(collection)\n\n # Query params which are not supported by JSON Request API yet\n # should go inside \"params\" attribute.\n # See https://lucene.apache.org/solr/guide/7_6/json-request-api.html\n # for more details.\n params = {\n key: value for key, value in [\n ('cursorMark', cursor),\n ('defType', def_type),\n ('qf', ' '.join(query_fields) if query_fields else ''),\n ('stats', 'true' if stats_fields else None),\n ('stats.field', stats_fields)\n ]\n if value is not None\n }\n json_data = {\n key: value for key, value in [\n ('query', query),\n ('filter', filter_),\n ('offset', offset),\n ('limit', limit),\n ('fields', fields),\n ('facet', facet_dict),\n ('sort', ','.join(sort) if sort else ''),\n ('params', params)\n ]\n if value is not None\n }\n\n try:\n response = await self.post(\n '/v2/collections/{}/query'.format(collection),\n json_data=json_data\n )\n json_response = json.loads(response.body.decode('utf-8'))\n query_response = json_response['response']\n stats = json_response.get('stats')\n solr_search_result = SolrSearchResult(\n num_found=query_response['numFound'],\n documents=query_response['docs'],\n cursor=json_response.get('nextCursorMark'),\n facet_results=json_response.get('facets', {}),\n stats_results=stats.get('stats_fields', {}) if stats else {}\n )\n logger.debug('Found {} and fetched {} documents from collection {}'\n .format(solr_search_result.num_found,\n len(solr_search_result.documents), collection))\n return solr_search_result\n except SolrError:\n logger.warning('Failed to execute query {} against collection {}'\n .format(json_data, collection))\n raise", "def find_one(cls, query, select=None, as_dict=False):\n record = cls.collection.find_one(query, select)\n\n return record if as_dict else cls(record)", "async def find_one(\n self, query: Union[dict, MotycQuery] = None, *,\n _id=None,\n inject_default_id: bool = None\n ) -> T:\n\n mongo_query = self.build_mongo_query(query, _id=_id)\n\n document = await self.collection.find_one(mongo_query)\n\n if document is None: raise NotFound(mongo_query)\n return self.parse_document(document, inject_default_id=inject_default_id)", "def get_collection():\n\n args = request.args.to_dict()\n\n # Convert size to int, for Python.\n if 'size' in args:\n args['size'] = int(args['size'])\n\n results = collection.search(request.args.get('q', '*'), **args)\n\n return jsonify(records=[r.dict for r in results])", "def profile_find(func, args, kwargs, func_result):\n collection = args[0]\n if (len(args) > 1):\n query = args[1]\n else:\n query = 'all'\n\n return _profile_query(collection, query, op='find')", "def mmo_execute_query_on_mongos(self, mmo_connection, query, execution_database, collection, find_one=False):\n mongos_server = self.mmo_mongos_servers(mmo_connection)[0]\n hostname, port = mongos_server[\"hostname\"], mongos_server[\"port\"]\n auth_dic = self.mmo_get_auth_details_from_connection(mmo_connection)\n c = self.mmo_connect_mongos(hostname, port, auth_dic[\"username\"], auth_dic[\"password\"], auth_dic[\"authentication_database\"])\n if find_one:\n query_output = c[execution_database][collection].find_one(query)\n else:\n query_output = c[execution_database][collection].find(query)\n return query_output", "def findRecord(db, col, query):\n\n # result = result.to_dict(\"records\")\n conn = MongoClient(\"localhost\", 27017)\n connObj = conn[db][col]\n exist_count = connObj.find(query, {'_id': 0}).count()\n conn.close()\n if exist_count >= 1:\n return True\n else:\n return False", "def find(self, queries):\n return self.__return_query('find', queries)", "def search_collections(search_args):\n if search_args.topic:\n if search_args.guid:\n topic_records = db[search_args.topic].find({'headers.guid': search_args.guid})\n print_records(topic_records, search_args.topic)\n elif search_args.email:\n topic_records = db[search_args.topic].find({'value.email_address': search_args.email})\n print_records(topic_records, search_args.topic)\n else:\n print(\" Use --email or --guid to search for specific records. Use 'dump' to get entire collection\")\n print()\n else:\n list_of_collections = db.list_collection_names()\n if search_args.guid:\n for collection in list_of_collections:\n topic_records = db[collection].find({'headers.guid': search_args.guid})\n print_records(topic_records, collection)\n elif search_args.email:\n for collection in list_of_collections:\n topic_records = db[collection].find({'value.email_address': search_args.email})\n print_records(topic_records, collection)\n else:\n print(\" Use --email or --guid to search for specific records. Use 'dump' to get entire collection\")\n print()\n return", "def find_one():\n fmter.tpl._straightline(\"one document\", 100)\n result = users.find_one({})\n print(type(result))\n ppt(result)\n \n fmter.tpl._straightline(\"none result\", 100)\n result = users.find_one({\"_id\": 100})\n print(type(result))\n ppt(result)", "def document_exists(collection, field, value):\n #print(\"Collection: \", collection, \" Field: \", field, \" Value: \", value)\n try:\n cursor = database[collection].find_one({field: value})\n if cursor is None:\n return {'status': 1, 'data': 0}\n return {'status': 1, 'data': cursor}\n except Exception as e:\n return {'status': -1, 'data': 'document_exists ' + str(e)}", "def find(cls, query_args, as_dict=False, as_cursor=False):\n cursor = cls.collection.find(query_args.query,\n query_args.select,\n sort=query_args.order_by,\n limit=query_args.limit)\n\n if as_cursor:\n return cursor\n else:\n return [record for record in cursor] if as_dict else [\n cls(record) for record in cursor\n ]", "def find_all(self, collection):\n obj = getattr(self.db, collection)\n result = obj.find()\n return result", "def search():\n if request.method == \"GET\":\n mongo_collection = mongo_database[\"questions\"]\n query = request.args.get(\"keyword\")\n result = mongo_collection.find({\"$text\": {\"$search\": query}})\n objects = []\n for object in result:\n objects.append(object)\n return render_template(\"search.html\", cards=objects)\n else:\n return start()", "def find_matching_collection(self, a_coll, recursive=False):\n conditions = [\n (\"name =\", a_coll.get_name()),\n ]\n the_id = a_coll.get_id()\n\n if not (the_id is None):\n conditions.append((\"id =\", the_id))\n\n stored_collection = self.load_collections(\n recursive=recursive, collection_conditions=conditions\n )\n\n # name unique, so length is always 1\n if len(stored_collection) > 0:\n return stored_collection[0]\n else:\n return None", "def get_doc_by_keyword(self,collection,field_name,search_key,like=True):\n if like:\n # This finds the records in which the field just \"contains\" the search_key\n res = self.client['rephie'][collection].find(({field_name : {'$regex' : \".*\"+search_key+\".*\"}}))\n else:\n # This finds the records in which the field is equal to the search_key\n res = self.client['rephie'][collection].find({field_name : search_key})\n\n return self._make_result_list(res)", "def search(self, query, **kwargs):\n query = dict(query)\n if self._query:\n query = {'$and': [self._query, query]}\n cat = type(self)(\n metadatastore_db=self._metadatastore_db,\n asset_registry_db=self._asset_registry_db,\n query=query,\n find_kwargs=kwargs,\n handler_registry=self._handler_registry,\n transforms=self._transforms,\n root_map=self._root_map,\n filler_class=self._filler_class,\n name='search results',\n getenv=self.getenv,\n getshell=self.getshell,\n auth=self.auth,\n metadata=(self.metadata or {}).copy(),\n storage_options=self.storage_options)\n return cat", "def search():\n args = request.args.to_dict()\n query = QueryModel(args)\n result = repository.search_text(COLLECTION_NAME, query.value)\n return {\"texts\": result}" ]
[ "0.79049706", "0.75223446", "0.7400582", "0.7219015", "0.7148035", "0.7044073", "0.70365006", "0.6900449", "0.68693525", "0.6687954", "0.66815037", "0.6643801", "0.6433155", "0.6354233", "0.6272778", "0.62644094", "0.61989784", "0.61437243", "0.6137244", "0.612182", "0.6096255", "0.60867447", "0.60716057", "0.60625374", "0.6055258", "0.60453093", "0.60431623", "0.6030652", "0.6024506", "0.6010289" ]
0.7865692
1
Search a collection for the distinct key values provided.
def find_distinct(self, collection, key): obj = getattr(self.db, collection) result = obj.distinct(key) return result
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _filter_search_values(key: str, values: list, collection: list):\n return_data = []\n for item in collection:\n if any(val in values for val in item[key]):\n return_data.append(item)\n return return_data", "def distinct(self, key):\n return self.database.command({'distinct': self.name,\n 'key': key})['values']", "def _findUniqueMappingKeys(mapping):\n\n uniqueMappingKeys = set()\n for key, entries in viewitems(mapping):\n if len(entries) == 1:\n uniqueMappingKeys.add(key)\n return uniqueMappingKeys", "def unique(self, key, lst=None):\n d = self.find(key, lst)\n vals = set(d.values())\n return sorted(list(vals))", "def searchNoWildcard(key, cursor):\n IDs = set()\n iterator = cursor.first()\n while iterator and iterator[0].decode(\"utf-8\") != key:\n iterator = cursor.next()\n\n if iterator != None:\n while iterator and iterator[0].decode(\"utf-8\") == key:\n IDs.add(iterator[1].decode(\"utf-8\"))\n iterator = cursor.next()\n return IDs", "def search_collections(self, keyword: str) -> List[str]:\n return [x for x in self.list_collections() if keyword.lower() in x]", "def _findUniqueMappingValues(mapping):\n uniqueMappingValues = set()\n for entries in viewvalues(mapping):\n if len(entries) == 1:\n uniqueMappingValues.update(entries)\n return uniqueMappingValues", "def extract_key_query_params(\n query_params: Dict[str, List[str]], param_key: str\n) -> Set[str]:\n return set(\n [\n item.lower()\n for sublist in [\n [value.lower() for value in query_params[key]]\n for key in query_params.keys()\n if key.lower() == param_key and query_params.get(key)\n ]\n for item in sublist\n ]\n )", "def searchWithWildcard(key, cursor):\n IDs = set()\n iterator = cursor.first()\n while iterator and not iterator[0].decode(\"utf-8\").startswith(key):\n iterator = cursor.next()\n\n if iterator != None:\n while iterator and iterator[0].decode(\"utf-8\").startswith(key):\n IDs.add(iterator[1].decode(\"utf-8\"))\n iterator = cursor.next()\n return IDs", "def profile_distinct(func, args, kwargs, func_result):\n collection = args[0]\n\n report_kvs = _profile_query(collection)\n report_kvs['Key'] = args[1]\n\n return report_kvs", "def filter_keys_in_set(ds, keys):\n logger.info(\"For each element in the dataset, keeping only values with keys: %s.\", ', '.join(keys))\n\n def filter_keys(x):\n return {k: v for k, v in x.items() if k in keys}\n\n return ds.map(filter_keys, num_parallel_calls=TF_AUTOTUNE)", "def filter_by_keys(self, keys):\n return list(filter(lambda item: item.keyword in set(keys), self._metadata))", "def findUniqueResults(ids, results):\n ordered = OrderedDict(sorted(data.items(), key=lambda t: t[0]))\n return list(ordered.values())", "def search_collections(search_args):\n if search_args.topic:\n if search_args.guid:\n topic_records = db[search_args.topic].find({'headers.guid': search_args.guid})\n print_records(topic_records, search_args.topic)\n elif search_args.email:\n topic_records = db[search_args.topic].find({'value.email_address': search_args.email})\n print_records(topic_records, search_args.topic)\n else:\n print(\" Use --email or --guid to search for specific records. Use 'dump' to get entire collection\")\n print()\n else:\n list_of_collections = db.list_collection_names()\n if search_args.guid:\n for collection in list_of_collections:\n topic_records = db[collection].find({'headers.guid': search_args.guid})\n print_records(topic_records, collection)\n elif search_args.email:\n for collection in list_of_collections:\n topic_records = db[collection].find({'value.email_address': search_args.email})\n print_records(topic_records, collection)\n else:\n print(\" Use --email or --guid to search for specific records. Use 'dump' to get entire collection\")\n print()\n return", "async def get_keys(self, collection):\n raise NotImplementedError", "def search(self, **kwargs):\n ret = self.addresses\n for key, val in kwargs.items():\n # Slightly odd syntax setting default values for key and val so that\n # v and k are not leaky cell variables.\n ret = list(\n filter(lambda aitem, v=val, k=key: v in getattr(aitem, k, \"\"), ret)\n )\n if not ret:\n raise KeyError(\"No addresses found matching criteria.\")\n return ret", "def collect(node: dict, exact: bool) -> Set[int]:\n return set(_collect(node, exact))", "def filter_collection(collection, filter_tuples):\n\n for filter_tuple in filter_tuples:\n collection = collection[collection[filter_tuple[0]] == filter_tuple[1]]\n\n return collection", "def search(collection_of_books: tuple, search_tag: str, search_keyword: str) -> list:\r\n found_books = []\r\n\r\n if search_tag == \"Shelf\" and search_keyword.isnumeric():\r\n found_books = [book for book in collection_of_books if search_keyword == book[\"Shelf\"]]\r\n\r\n else:\r\n for book in collection_of_books:\r\n if search_keyword.lower() in book[search_tag].lower():\r\n found_books.append(book)\r\n\r\n return found_books", "def distinct(self, cls, *args, **kwargs):\n m = mapper(cls)\n return self.impl.distinct(m.collection, *args, **kwargs)", "def union(self, key: str, skip_duplicates=False) -> list:\n result = []\n for items in self.get(key):\n for item in items:\n if skip_duplicates and item in result:\n continue\n result.append(item)\n return result", "def search(self, *args, **kwargs):\n # comparison = f\"__{kwargs.get('comparison')}__\" if kwargs.get('comparison') else '__eq__'\n comparison = '__{comparison}__'.format(comparison=kwargs.get('comparison')) if kwargs.get('comparison') else '__eq__'\n try:\n key, value = args[0], args[1]\n except IndexError:\n for key in kwargs.keys():\n if '__' in key:\n # comparison = f'__{key.split(\"__\")[1]}__'\n comparison = '__{comparison}__'.format(comparison=key.split(\"__\")[1])\n key, value = key.split(\"__\")[0], kwargs[key]\n return SearchableList(list(filter(lambda x: try_compare(x, key, comparison, value), self)))", "def number_in_set(c,s):\n return sum(v for k,v in c.items() if k in s)", "def _mappingGetValueSet(mapping, keys):\n setUnion = set()\n for k in keys:\n setUnion = setUnion.union(mapping[k])\n return setUnion", "def search(self, key, headers=Headers()):", "def search_preproc(list_pre, dic_pre):\n result = []\n for x in list_pre:\n shared_items = [x[k] == dic_pre[k] for k in x if k in dic_pre]\n result.append(sum(shared_items)==len(dic_pre))\n return [i for i, x in enumerate(result) if x]\n # maybe also searches if the files exist?", "def substring_search(word, collection):\n return [item for item in sorted(collection) if item.startswith(word)]", "def getResults(self, queryVec: dict):\n result = set()\n for word in queryVec.keys():\n if word in self.invertedIndex:\n result.update(self.invertedIndex[word].keys())\n #print (result)\n #print (self.corpusCount)\n return result", "def findall(self, **kwargs):\n found = []\n searches = kwargs.items()\n\n for obj in self.list():\n try:\n if all(getattr(obj, attr) == value\n for (attr, value) in searches):\n found.append(obj)\n except AttributeError:\n continue\n\n return found", "def findall(self, **kwargs):\n found = []\n searches = kwargs.items()\n\n for obj in self.list():\n try:\n if all(getattr(obj, attr) == value\n for (attr, value) in searches):\n found.append(obj)\n except AttributeError:\n continue\n\n return found" ]
[ "0.69668233", "0.58777803", "0.5819709", "0.5781459", "0.5655128", "0.5622824", "0.56214416", "0.56086063", "0.5562973", "0.5557258", "0.5502194", "0.5408141", "0.5392551", "0.53712016", "0.5323929", "0.5288685", "0.5286743", "0.52866507", "0.5135332", "0.51346934", "0.51308876", "0.51164174", "0.510882", "0.51054704", "0.51004565", "0.5096431", "0.5088026", "0.50598115", "0.5042349", "0.5042349" ]
0.72450334
0
Adds an embedded campaign to the TLO.
def add_embedded_campaign(self, id, collection, campaign, confidence, analyst, date, description): if type(id) is not ObjectId: id = ObjectId(id) # TODO: Make sure the object does not already have the campaign # Return if it does. Add it if it doesn't obj = getattr(self.db, collection) result = obj.find({'_id': id, 'campaign.name': campaign}) if result.count() > 0: return else: log.debug('Adding campaign to set: {}'.format(campaign)) campaign_obj = { 'analyst': analyst, 'confidence': confidence, 'date': date, 'description': description, 'name': campaign } result = obj.update( {'_id': id}, {'$push': {'campaign': campaign_obj}} ) return result
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def add_campaign(self, campaign):\n self._campaigns += [campaign]", "def _set_campaign(self, campaign):\n if isinstance(campaign, str):\n campaign = TrackedCampaign.objects.create(name=campaign)\n\n campaign.save()\n\n self.campaign = campaign", "def campaign(self, campaign):\n\n self._campaign = campaign", "def create_campaign(client, customer_id, budget_resource_name):\n campaign_service = client.get_service(\"CampaignService\")\n campaign_operation = client.get_type(\"CampaignOperation\")\n campaign = campaign_operation.create\n campaign.name = f\"Interplanetary Cruise App #{uuid4()}\"\n campaign.campaign_budget = budget_resource_name\n # Recommendation: Set the campaign to PAUSED when creating it to\n # prevent the ads from immediately serving. Set to ENABLED once you've\n # added targeting and the ads are ready to serve.\n campaign.status = client.enums.CampaignStatusEnum.PAUSED\n # All App campaigns have an advertising_channel_type of\n # MULTI_CHANNEL to reflect the fact that ads from these campaigns are\n # eligible to appear on multiple channels.\n campaign.advertising_channel_type = (\n client.enums.AdvertisingChannelTypeEnum.MULTI_CHANNEL\n )\n campaign.advertising_channel_sub_type = (\n client.enums.AdvertisingChannelSubTypeEnum.APP_CAMPAIGN\n )\n # Sets the target CPA to $1 / app install.\n #\n # campaign_bidding_strategy is a 'oneof' message so setting target_cpa\n # is mutually exclusive with other bidding strategies such as\n # manual_cpc, commission, maximize_conversions, etc.\n # See https://developers.google.com/google-ads/api/reference/rpc\n # under current version / resources / Campaign\n campaign.target_cpa.target_cpa_micros = 1000000\n # Sets the App Campaign Settings.\n campaign.app_campaign_setting.app_id = \"com.google.android.apps.adwords\"\n campaign.app_campaign_setting.app_store = (\n client.enums.AppCampaignAppStoreEnum.GOOGLE_APP_STORE\n )\n # Optimize this campaign for getting new users for your app.\n campaign.app_campaign_setting.bidding_strategy_goal_type = (\n client.enums.AppCampaignBiddingStrategyGoalTypeEnum.OPTIMIZE_INSTALLS_TARGET_INSTALL_COST\n )\n # Optional fields\n campaign.start_date = (datetime.now() + timedelta(1)).strftime(\"%Y%m%d\")\n campaign.end_date = (datetime.now() + timedelta(365)).strftime(\"%Y%m%d\")\n # Optional: If you select the\n # OPTIMIZE_IN_APP_CONVERSIONS_TARGET_INSTALL_COST goal type, then also\n # specify your in-app conversion types so the Google Ads API can focus\n # your campaign on people who are most likely to complete the\n # corresponding in-app actions.\n #\n # campaign.selective_optimization.conversion_actions.extend(\n # [\"INSERT_CONVERSION_ACTION_RESOURCE_NAME_HERE\"]\n # )\n\n # Submits the campaign operation and print the results.\n campaign_response = campaign_service.mutate_campaigns(\n customer_id=customer_id, operations=[campaign_operation]\n )\n resource_name = campaign_response.results[0].resource_name\n print(f'Created App campaign with resource name: \"{resource_name}\".')\n return resource_name", "def _extend_record(self, campaign, fields, pull_ads):\n campaign_out = campaign.api_get(fields=fields).export_all_data()\n if pull_ads:\n campaign_out[\"ads\"] = {\"data\": []}\n ids = [ad[\"id\"] for ad in campaign.get_ads()]\n for ad_id in ids:\n campaign_out[\"ads\"][\"data\"].append({\"id\": ad_id})\n return campaign_out", "def trackEmbedding(self, tag):\n self.embeddedTags.append(tag)", "def test_admin_sms_campaign_view_add(self):\n response = self.client.get('/admin/sms_module/smscampaign/add/')\n self.failUnlessEqual(response.status_code, 200)", "def write_campaign(campaign_data):\n\n campaign = Campaign(**campaign_data)\n campaign.save()\n authorization.make_campaign_public(campaign)\n\n return campaign.id", "def embedded(self, embedded):\n self._embedded = embedded", "def create_campaign(account, row, name, acc_type):\n country = None\n if acc_type == Account.COUNTRY:\n country_name = row['LOCATION']\n country = Country.objects.filter(name__iexact=country_name).first()\n if not country:\n logging.getLogger('peacecorps.sync_accounting').warning(\n \"%s: Country does not exist: %s\",\n row['PROJ_NO'], row['LOCATION'])\n return\n\n account.save()\n summary = clean_description(row['SUMMARY'])\n campaign = Campaign.objects.create(\n name=name, account=account, campaigntype=acc_type,\n description=json.dumps({\"data\": [{\"type\": \"text\",\n \"data\": {\"text\": summary}}]}),\n country=country)\n if acc_type == Account.SECTOR:\n # Make sure we remember the sector this is marked as\n SectorMapping.objects.create(pk=row['SECTOR'], campaign=campaign)", "def ExecuteEmbed(self):\r\n \r\n Embed = DiscordEmbed(title=\"Test Title 123\", \r\n description=\"Test Description 321\",\r\n color=\"eb5e34\") \r\n Embed.set_timestamp()\r\n \r\n self.WEBHOOK.add_embed(Embed)\r\n Execute = self.WEBHOOK.execute()", "def campaign_id(self, campaign_id):\n\n self._campaign_id = campaign_id", "def create_custom_audience_from_website(\n self, account_id, name, domain, description=None,\n retention_days=30, prefill=True, batch=False):\n path = \"act_%s/customaudiences\" % account_id\n args = {\n 'name': name,\n 'subtype': \"WEBSITE\"\n }\n rule = {'url': {\n 'i_contains': domain,\n }}\n if rule:\n args['rule'] = json.dumps(rule)\n if retention_days:\n args['retention_days'] = retention_days\n if prefill:\n args['prefill'] = prefill\n return self.make_request(path, 'POST', args, batch=batch)", "def post(self):\n json_data = request.get_json()\n json_data[\"sender_id\"] = current_user.id\n try:\n new_campaign = self.schema.load(json_data)\n except ValidationError as err:\n return {\"message\": err.messages}, HTTPStatus.BAD_REQUEST\n if Campaign.query.filter_by(mailchimp_id=new_campaign.mailchimp_id).first() is not None:\n return {\"message\": \"Campaign already exists.\"}, HTTPStatus.CONFLICT\n db.session.add(new_campaign)\n db.session.commit()\n return self.schema.dump(new_campaign), HTTPStatus.CREATED", "def add_to_campaign(self, campaign_id):\n campaign = Campaign.objects.get(pk=campaign_id)\n if not ContactCampaignStatus.objects.filter(\n contact=self, campaign=campaign\n ).exists():\n # We first create the big object that will hold the status for the campaign\n ContactCampaignStatus.objects.create(contact=self, campaign=campaign)\n return _(\"Contact %s (ID: %s) added to campaign\") % (self.name, self.id)\n else:\n raise Exception(\n _(\"Contact %s (ID: %s) already in campaign\") % (self.name, self.id)\n )", "def embed():", "def embed(self, data, mime_type=\"text/plain\", encode_data_to_base64=True):\n if encode_data_to_base64:\n data = base64.standard_b64encode(data.encode()).decode()\n self.embeddings.append({\"data\": data, \"mime_type\": mime_type})", "def rpc_campaign_new(self, name):\n\t\tsession = db_manager.Session()\n\t\tcampaign = db_models.Campaign(name=name, user_id=self.basic_auth_user)\n\t\tsession.add(campaign)\n\t\tsession.commit()\n\t\treturn campaign.id", "def addAssociation(self, *args):\n return _libsbml.FbcAnd_addAssociation(self, *args)", "def create_custom_audience(self, account_id, name, subtype=None,\n description=None, rule=None, opt_out_link=None,\n retention_days=30, batch=False):\n path = \"act_%s/customaudiences\" % account_id\n args = {\n 'name': name,\n }\n if subtype:\n args['subtype'] = subtype\n if description:\n args['description'] = description\n if rule:\n args['rule'] = json.dumps(rule)\n if opt_out_link:\n args['opt_out_link'] = opt_out_link\n if retention_days:\n args['retention_days'] = retention_days\n return self.make_request(path, 'POST', args, batch=batch)", "def test_create_campaign(self):\n campaign = self.campaign\n\n self.assertTrue(isinstance(campaign, Campaign))\n self.assertEqual(campaign.name, \"Test Campaign\")", "async def _create_embed(self, event, info):\n\n e = discord.Embed(url=info.get(\"url\"))\n e.title = \"%s %s!\" % (info.get(\"streamer\"), info.get(\"live_status\"))\n e.add_field(name=\"Stream title\", value=info.get(\"title\"), inline=False)\n e.add_field(name=\"Begin:\", value=event.begin.format(\"HH:mm:ss ZZZ\") + \" (\" + event.begin.humanize() + \")\", inline=False)\n e.add_field(name=\"Duration: \", value=str(event.duration), inline=False)\n #e.add_field(name=\"Link\", value=info.get(\"url\"), inline=False)\n e.set_image(url=info.get(\"thumbnail\") or e.Empty)\n return e", "async def add(self, ctx, *, channel: discord.VoiceChannel):\n auto_channels = await self.config.guild(ctx.guild).auto_channels()\n if auto_channels is None:\n auto_channels = []\n auto_channels.append(channel.id)\n await self.config.guild(ctx.guild).auto_channels.set(auto_channels)\n await ctx.send(\n _(\"Startchannel used for automatic voicechannels added: {channel}\").format(\n channel=channel.name\n )\n )", "def add_embed_itmes(data):\n for k, v in data.items() :\n embed.add_embed_field(name=k, value=v)", "def testAddLandingPageToCampaign(self):\n if self.__class__.campaign1 is None:\n self.testSaveCampaign()\n landing_page = {\n 'id': '-1',\n 'url': 'http://www.example.com',\n 'name': 'Landing page #%s' % Utils.GetUniqueName()\n }\n self.assert_(isinstance(self.__class__.service.AddLandingPageToCampaign(\n self.__class__.campaign1['id'], [landing_page]), tuple))", "def get_adcampaign(self, campaign_id, fields, batch=False):\n path = '%s' % campaign_id\n args = {'fields': fields}\n return self.make_request(path, 'GET', args, batch=batch)", "def addAttachment(self,data,url,name,description):\n if self.PDFreactorConfiguration.in1[\"attachments\"] == None:\n self.PDFreactorConfiguration.in1[\"attachments\"] = []\n \n stylesArray = {'data':data, 'url':url, 'name':name, 'description':description}\n \n self.PDFreactorConfiguration.in1[\"attachments\"].append(stylesArray)", "def create_node_campaign(self, node_oid):\n node = Node.objects(id=node_oid)[0]\n list_id = DripCampaign.objects(id=node[\"drip_campaign_id\"])[0][\"list_id\"]\n segment = Segment.objects(id=node[\"segment_oid\"])[0]\n if segment[\"members_euid\"]:\n campaign_id = self.mw.create_campaign(\n list_id=list_id,\n segment_id=segment[\"segment_id\"],\n template_id=node[\"content\"][\"template_id\"],\n subject=node[\"content\"][\"subject\"],\n from_email=node[\"content\"][\"from_email\"],\n from_name=node[\"content\"][\"from_name\"],\n folder_id=self.folder_id,\n )\n node.update(set__campaign_id=campaign_id, set__updated_at=datetime.utcnow())\n return campaign_id\n else:\n return None", "def add(self, PlugLead):\n\n self.check_conflicts(PlugLead)\n self.plugleads.append(PlugLead)", "def add_podcast(_name_of_the_podcast, _duration_in_number_of_seconds,\r\n _host, _participants):\r\n # creating an instance of our Podcast constructor\r\n new_podcast = Podcast(name_of_the_podcast=_name_of_the_podcast,\r\n duration_in_number_of_seconds=_duration_in_number_of_seconds,\r\n host=_host, participants=_participants)\r\n db.session.add(new_podcast) # add new Podcast to database session\r\n db.session.commit() # commit changes to session\r" ]
[ "0.7053185", "0.55294055", "0.5502619", "0.53983206", "0.53974485", "0.53938943", "0.5390333", "0.5346445", "0.51280814", "0.51132816", "0.5100033", "0.5023265", "0.4961767", "0.49532866", "0.49409008", "0.49400547", "0.49252018", "0.49178767", "0.48796228", "0.48726276", "0.48115474", "0.481078", "0.47973284", "0.4793301", "0.47923282", "0.47914565", "0.4771861", "0.4746295", "0.47435823", "0.47363722" ]
0.7287195
0
Removes an item from the bucket list
def remove_bucket_list_item(self, id, collection, item): if type(id) is not ObjectId: id = ObjectId(id) obj = getattr(self.db, collection) result = obj.update( {'_id': id}, {'$pull': {'bucket_list': item}} ) return result
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def remove(self, item):\n try:\n entry = self.set.pop(item)\n entry[-1] = self.REMOVED\n except KeyError:\n print(\"Can't remove a non-existing item\")", "def remove(self, item):\n del self._dict[item]", "def _bucket_delitem(self, j, k):\n pass", "def __delitem__(self, key):\n\n bucket_key = self.key_for_bucket(key)\n del self.buckets[bucket_key][key]\n\n if not self.buckets[bucket_key]:\n del self.buckets[bucket_key]", "def remove(self, item):\n\n if item in self:\n item_index = self._index_map[item]\n last_item = self._list[-1]\n\n # Swap in the item from the end of the list\n self._list[item_index] = last_item\n self._list.pop()\n\n self._index_map[last_item] = item_index", "def remove(self, item):\n # type: (Any) -> None\n return list.remove(self, self.ref(item))", "def remove_item(self, idx_of_item):\n del self.items[idx_of_item]", "def remove(self, item: T) -> None:\n index = self.index(item)\n self.delete_at_index(index)", "def remove(self, item: T) -> None:\n index = self.index(item)\n self.delete_at_index(index)", "def removeItem(*args):", "def removeItem(*args):", "def __delitem__(self, key):\n bucket = self._buckets[self._index(key)]\n for node in bucket.linked_list:\n bucket_object_key, bucket_object_value = node.value\n if bucket_object_key.load_value() == key:\n # remove objects from object -> list_node dict\n key_list_node = self._object_to_list_node.pop(bucket_object_key)\n value_list_node = self._object_to_list_node.pop(bucket_object_value)\n # remove list_node from in_memory and disk objects\n self._in_memory_objects.remove(key_list_node)\n self._in_memory_objects.remove(value_list_node)\n self._disk_objects.remove(key_list_node)\n self._disk_objects.remove(value_list_node)\n # remove node from bucket linked list\n assert bucket.linked_list.remove(node) == True\n self._balance()\n return\n raise KeyError(\"Key `{}` is not exists\".format(key))", "def remove(self, item) -> None:\n entry = self.entry_finder.pop(item)\n entry[-1][0] = None", "def remove(self, key):\r\n\r\n\t\t# if the key doesn't exist, exit the function\r\n\t\tif not self.contains_key(key):\r\n\t\t\treturn\r\n\t\telse:\r\n\t\t\tindex = self.get_index(key) # get the index of the key\r\n\t\t\tlinked_list = self._buckets[index] # now get the entire linked list\r\n\t\t\tlinked_list.remove(key) # call the remove function from the linked list\r\n\t\t\tself.size -= 1 # subtract 1\r", "def item_remove(self, item):\n\t\treturn self._modify_object(item=item, new_item=\"\")", "def __delitem__(self, key):\n del self.list[key]", "def remove(self, key):\n index = key % self.size\n prev = cur = self.bucket[index]\n\n if not cur: return\n if cur.key == key:\n self.bucket[index] = cur.next\n return\n\n cur = cur.next\n while cur:\n if cur.key == key:\n prev.next = cur.next\n break\n else:\n cur = cur.next\n prev = prev.next", "def remove(self, item: Item) -> None:\n raise NotImplementedError(\"remove\")", "def remove (self, item):\n pass", "def __delitem__(self, k):\n j = self._hash_function(k)\n self._bucket_delitem(j, k)\n self._n -= 1", "def deleteItem(list,item):\n print \"I deleted this item:\", item\n list.remove(item)", "def remove(self, item):\n # self._probeCount = 0\n self._homeIndex = abs(self._hash(item)) % len(self._table)\n distance = 1\n index = self._homeIndex\n\n while (self._table[index] != HashTable.EMPTY or \\\n self._table[index] == HashTable.DELETED) and \\\n self._table[index] != item:\n\n if self._liner:\n increment = index + 1\n else:\n increment = self._homeIndex + distance ** 2\n distance += 1\n index = increment % len(self._table)\n self._probeCount += 1\n\n if self._table[index] == item:\n self._table[index] = HashTable.DELETED\n self._actualIndex = index\n self._size -= 1\n return index\n else:\n self._actualIndex = -1\n return -1", "def _bucket_delitem(self, j, k):\n bucket = self._table[j]\n if bucket is None: # no match found\n raise KeyError(\"Key Error: \" + repr(k))\n del bucket[k]", "def take(self, item): \n self.contents.remove(item)", "def remove_item(self, item_id):\n self.items.pop(item_id)", "def discard(self, item):\n try:\n self._del(item)\n except KeyError:\n pass", "def remove(self, key):\r\n for i in range(len(self.lis)):\r\n if self.lis[i][0] == key:\r\n self.lis.pop(i)\r\n break", "def __delitem__(self, key):\n self.f_remove(key)", "def remove(self, key: int) -> None:\n hashKey = key % 1000\n prev = node = self.bucket[hashKey]\n if not node: return\n if node.pair[0] == key:\n self.bucket[hashKey] = node.next\n else:\n node = node.next\n while node:\n if node.pair[0] == key:\n prev.next = node.next\n break\n else:\n prev, node = prev.next, node.next", "def delete(self, item):\r\n self.fetch()\r\n t = self.make_item_tuple(item)\r\n changed = False\r\n while t in self.data:\r\n self.data.remove(t)\r\n changed = True\r\n \r\n if changed:\r\n query_cache.set(self.iden, self.data)" ]
[ "0.791302", "0.7706514", "0.7652778", "0.75784016", "0.7561372", "0.7506943", "0.74443907", "0.7405826", "0.7405826", "0.73899317", "0.73899317", "0.73878163", "0.73668784", "0.7335001", "0.7286636", "0.7273813", "0.72560424", "0.72402847", "0.72392786", "0.72225547", "0.721951", "0.7215872", "0.72138774", "0.72032166", "0.7127759", "0.7120555", "0.7094752", "0.70845085", "0.70182854", "0.70019066" ]
0.8045885
0
Adds an item to the bucket list
def add_bucket_list_item(self, id, collection, item): if type(id) is not ObjectId: id = ObjectId(id) obj = getattr(self.db, collection) result = obj.update( {'_id': id}, {'$addToSet': {'bucket_list': item}} ) return result
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def add(self, item):", "def add_item(self, item):\n self.items.append(item)", "def add_to_bag(self, item):\n self._bag.append(item)", "def add(self, item):\n\n if item not in self:\n self._index_map[item] = len(self._list)\n self._list.append(item)", "def add(self, item):\n self._dict[item] = item", "def bfAdd(self, key, item):\n params = [key, item]\n \n return self.execute_command(self.BF_ADD, *params)", "def enqueue(self, item):\n self.list.append(item)", "def push(self, item):\n\t\tself.items.append(item)", "def add_item(self, item):\n self.items.append(item)\n self.length += 1", "def add_bucketlist_item(self, email, password, buckelist_id, item_name):\r\n test_date = str(date(2020, 9, 22))\r\n headers = self.authentication_headers(email=email, password=password)\r\n return self.client.post(\r\n '/api/v1/bucketlist/{}/items/'.format(buckelist_id),\r\n data=json.dumps({\"name\": item_name, \"finished_by\": test_date}),\r\n content_type=\"application/json\",\r\n headers=headers,\r\n follow_redirects=True\r\n )", "def push(self, item) -> None:\n self.items.append(item)", "def addItem(self, item, rank):\n with self.lock:\n if self.ItemHashList.get(item, -1) == -1:\n self.ItemHashList[item] = None\n if rank < 0:\n rank = 0\n heapq.heappush(self.ItemList, (rank, item))", "def push(self, item):\n self.items.append(item)", "def push(self, item):\n self.items.append(item)", "def push(self, item):\n self.items.append(item)", "def push(self, item):\n self.items.append(item)", "def push(self, item):\n self.items.append(item)", "def push(self,item):\n self.items.append(item)", "def add_item (self, item):\n new_item = CacheItem (item)\n cached = self.cache.get(hash(item))\n if cached is None:\n self.evict_or_add (new_item)\n cached.hits += 1", "def enqueue(self, item):\n self.items.insert(0, item)", "def enqueue(self, item):\n self.items.insert(0, item)", "def enqueue(self, item):\n\t\tself.items.insert(0, item)", "def push(self, item) -> None:\n self._items.append(item)", "def append(self, item):\n self.items.append(item)", "def enqueue(self, item):\n\n self.__items__.append(item)", "def add_item(self, item: Item):\n self.__items_list.append(item)", "def _add_item_by_item(self, item):\n self.item_list[item.call_number] = item", "def add_item(self, item):\r\n bag_res = consts.BAG_PUT_FAILED\r\n for i in range(len(self._items)):\r\n res = self.put_item_at(i, item, allow_switch=False)\r\n if res == consts.PUT_FORBIDDEN:\r\n return consts.BAG_PUT_FAILED\r\n if res == consts.PUT_SWITCH or \\\r\n res == consts.PUT_INTO_EMPTY or \\\r\n res == consts.PUT_MERGE_TOTALLY:\r\n return consts.BAG_PUT_TOTALLY\r\n if res == consts.PUT_MERGE_PARTIALLY:\r\n bag_res = consts.BAG_PUT_PARTIALLY\r\n continue\r\n if res == consts.PUT_MERGE_FAILED or \\\r\n res == consts.PUT_SWITCH_FORBIDDEN:\r\n continue\r\n return bag_res", "def put(self, item): \n self.__db.rpush(self.key, item)", "def push(self, item):\n\n self._list.append(item)" ]
[ "0.7640922", "0.75756705", "0.751956", "0.74900144", "0.7369312", "0.7361309", "0.7332356", "0.7283963", "0.7280235", "0.72764546", "0.72698563", "0.7263061", "0.72285056", "0.72285056", "0.72285056", "0.72285056", "0.72285056", "0.71988684", "0.7163348", "0.7156221", "0.7156221", "0.71561056", "0.7153414", "0.71492743", "0.71387446", "0.7135556", "0.713479", "0.71344966", "0.7122523", "0.71106565" ]
0.77858466
0
Returns a list of all valid campaign names
def get_campaign_name_list(self): campaigns = self.find('campaigns', {}) campaign_names = [] for campaign in campaigns: if 'name' in campaign: campaign_names.append(campaign['name']) return campaign_names
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_get_existent_campaigns_returns_campaigns_list(self):\n test_campaign = return_canned_campaign()\n test_campaign.create()\n response = self.client.get(self.endpoint_url)\n response_body = response.get_json()\n self.assertEqual(response.status_code, 200)\n self.assertEqual(\n response_body, {\"campaigns\": [{\"id\": 1, \"name\": \"Test Campaign\"}]}\n )", "def _list_of_availability_strings():\n names = [availability.name for availability in Availability]\n return names", "def _get_campaigns(self, params):\n return self._api.account.get_campaigns(params={**params, **self._state_filter()}, fields=[self.state_pk])", "def names(self) -> list[str]:", "def getNames(self) -> List[unicode]:\n ...", "def get_list_of_campaigns(self, limit=0, offset=0):\n logger.info(\"Function call: get_list_of_campaigns\")\n return self.__handle_result(self.__send_request('campaigns', 'GET', {'limit': limit or 0, 'offset': offset or 0}))", "def allowed(cls):\n # type: () -> List[Str]\n names = cls.names()\n allowed = names\n allowed.extend([name.lower() for name in names])\n return allowed", "def AllNames():\n return CheckForError(get_string_array(lib.Generators_Get_AllNames))", "def names(cls) -> List[str]:", "def get_blocked_usernames_list():\n return []", "def get_short_names(self) -> List[str]:\n result = []\n for elements in self._get_results_list():\n result.append(elements[0])\n return result", "def campaign_name(self):\n\n return self._campaign_name", "def campaign_name(self):\n\n return self._campaign_name", "def campaign_name(self):\n\n return self._campaign_name", "def get_valid_domains():\n msg = ''\n import glob\n validDomains = []\n for f in glob.glob('{}/*'.format(OPT_MANAGER_RESOURCES_PGAAS)):\n try:\n with open(f, \"r\") as fp:\n try:\n tmpdata = json.load(fp)\n if 'pubkey' in tmpdata:\n validDomains.append(os.path.basename(f))\n except: # pylint: disable=bare-except\n pass\n except: # pylint: disable=bare-except\n pass\n if len(validDomains) == 0:\n msg += '\\nNo valid PostgreSQL cluster information was found'\n else:\n msg += '\\nThese are the valid PostgreSQL cluster domains found on this manager:'\n for v in validDomains:\n msg += '\\n\\t\"{}\"'.format(v)\n return msg", "def test_get_non_existent_campaigns_returns_empty_list(self):\n response = self.client.get(self.endpoint_url)\n response_body = response.get_json()\n self.assertEqual(response.status_code, 200)\n self.assertEqual(response_body, {\"campaigns\": []})", "def list_campaigns(self, **kwargs) -> ApiResponse:\n return self._request(kwargs.pop('path'), params=kwargs)", "def list_domain_names():\n pass", "def embargoed_countries_list(self):\r\n if self.embargoed_countries == '':\r\n return []\r\n return [country.strip().upper() for country in self.embargoed_countries.split(',')] # pylint: disable=no-member\r", "def getAllCampaigns(service):\n # Using AWQL to retrieve campaigns.\n query = (adwords.ServiceQueryBuilder()\n .Select('Id', 'Name', 'Status', 'StartDate', 'EndDate',\n 'BudgetId', 'BudgetStatus', 'BudgetName', 'Amount',\n 'BudgetReferenceCount', 'IsBudgetExplicitlyShared')\n .Limit(0, pageSize)\n .Build())\n campaigns = []\n for page in query.Pager(service):\n if page['entries']:\n for campaign in page['entries']:\n campaigns.append(campaign)\n else:\n pass\n return campaigns", "def get_all_names(cls, exclude_values: Iterator['CommonBusinessAdvertisingType'] = None) -> Tuple[str]:\n name_list: Tuple[str] = tuple([value.name for value in cls.get_all(exclude_values=exclude_values)])\n return name_list", "def get_camp_ids_names_containing_str(marketer_id, string):\n all_campaigns = outb.get_campaigns_per_marketer(marketer_id).get(marketer_id[0])\n return [{\"id\": x.get(\"id\"), \"name\": x.get(\"name\")} for x in all_campaigns if string in x[\"name\"]]", "def get_conversations(filename):\n with open(filename, 'r') as fp:\n data = json.load(fp)\n convos = data['conversation_state']\n all_names = []\n for conv in convos:\n conv = conv['conversation_state']['conversation']\n # does if have a name?\n if 'name' in conv:\n name = conv['name']\n else:\n # get all the people in the conv\n people_names = [person['fallback_name']\n for person in conv['participant_data']]\n name = ','.join(people_names)\n all_names.append(name)\n return all_names", "def get_names_short(self):\r\n return [p.get_name() for p in self.people]", "def clean_embargoed_countries(self):\r\n embargoed_countries = self.cleaned_data[\"embargoed_countries\"]\r\n if not embargoed_countries:\r\n return ''\r\n\r\n error_countries = []\r\n\r\n for country in embargoed_countries.split(','):\r\n country = country.strip().upper()\r\n if not self._is_valid_code(country):\r\n error_countries.append(country)\r\n\r\n if error_countries:\r\n msg = 'COULD NOT PARSE COUNTRY CODE(S) FOR: {0}'.format(error_countries)\r\n msg += ' Please check the list of country codes and verify your entries.'\r\n raise forms.ValidationError(msg)\r\n\r\n return embargoed_countries", "def get_domain_whitelist(self):\n whitelist = self.charm_config[\"federation-domain-whitelist\"]\n return list(filter(None, whitelist.split(\",\")))", "def available_characterizations(self):\n return self.characterization_results.keys()", "def get_all_breaches_names_for_account(self, email):\n resp = self.get_all_breaches_for_account(email)\n return [breach[\"Name\"] for breach in resp]", "def testCMSNametoList(self):\n result = self.mySiteDB.cmsNametoList(\"T1_US*\", \"SE\")\n self.assertItemsEqual(result, [u'cmsdcadisk01.fnal.gov'])", "def _get_all_champs(self):\n return [self.user_values['Champion Name'][row].get().lower()\n for row, obj in enumerate(self.user_values['Champion Name'])]" ]
[ "0.5988995", "0.5805326", "0.5684534", "0.5676261", "0.5601181", "0.55940896", "0.5530269", "0.5515312", "0.5460801", "0.5450627", "0.5449519", "0.54434097", "0.54434097", "0.54434097", "0.542792", "0.5424966", "0.54231083", "0.54181385", "0.53910905", "0.5375468", "0.5351693", "0.53187054", "0.5313526", "0.5304703", "0.528853", "0.52840716", "0.5281339", "0.52515966", "0.5233569", "0.5220399" ]
0.7925846
0