code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
from flask import Flask, flash def display_error(error,preface="",postface=""): flash(f"{preface} {error} {postface}")
[ "flask.flash" ]
[((85, 123), 'flask.flash', 'flash', (['f"""{preface} {error} {postface}"""'], {}), "(f'{preface} {error} {postface}')\n", (90, 123), False, 'from flask import Flask, flash\n')]
import sys import typing from collections import deque from typing import Callable, Optional import mypy.errorcodes import mypy.errors import mypy.nodes import mypy.options import mypy.plugin import mypy.types if sys.version_info >= (3, 10): # pragma: no cover from typing import TypeGuard else: # pragma: no cov...
[ "collections.deque" ]
[((3972, 3986), 'collections.deque', 'deque', (['[type_]'], {}), '([type_])\n', (3977, 3986), False, 'from collections import deque\n')]
# -*- coding: utf-8 -*- # Generated by Django 1.10.7 on 2017-07-18 04:43 from __future__ import unicode_literals from django.db import migrations, models def update_names(apps, schema_editor): for x in apps.get_model('institutions', 'regon').objects.exclude(data=None).iterator(): x.name = x.data.get('naz...
[ "django.db.migrations.RunPython", "django.db.models.CharField" ]
[((1079, 1113), 'django.db.migrations.RunPython', 'migrations.RunPython', (['update_names'], {}), '(update_names)\n', (1099, 1113), False, 'from django.db import migrations, models\n'), ((740, 805), 'django.db.models.CharField', 'models.CharField', ([], {'default': '""""""', 'max_length': '(200)', 'verbose_name': '"""N...
# coding: utf-8 # pylint: disable=missing-docstring, invalid-name import flask import auth import config from main import app import model.user as user #import User#, UserVdr github_config = dict( access_token_method='POST', access_token_url='https://github.com/login/oauth/access_token', authorize_url='ht...
[ "flask.flash", "auth.signin_oauth", "flask.session.get", "flask.url_for", "main.app.route", "auth.signin_via_social", "auth.create_oauth_app" ]
[((574, 620), 'auth.create_oauth_app', 'auth.create_oauth_app', (['github_config', '"""github"""'], {}), "(github_config, 'github')\n", (595, 620), False, 'import auth\n'), ((624, 674), 'main.app.route', 'app.route', (['"""/_s/callback/github/oauth-authorized/"""'], {}), "('/_s/callback/github/oauth-authorized/')\n", (...
from pages.service_page.models import ServicePage from pages.topic_page.factories import JanisBasePageWithTopicsFactory from pages.base_page.fixtures.helpers.streamfieldify import streamfieldify class ServicePageFactory(JanisBasePageWithTopicsFactory): @classmethod def create(cls, *args, **kwargs): if...
[ "pages.base_page.fixtures.helpers.streamfieldify.streamfieldify" ]
[((390, 431), 'pages.base_page.fixtures.helpers.streamfieldify.streamfieldify', 'streamfieldify', (["kwargs['dynamic_content']"], {}), "(kwargs['dynamic_content'])\n", (404, 431), False, 'from pages.base_page.fixtures.helpers.streamfieldify import streamfieldify\n'), ((600, 636), 'pages.base_page.fixtures.helpers.strea...
# (c) Copyright [2017] Hewlett Packard Enterprise Development LP # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appli...
[ "os.remove", "os.walk", "json.dumps", "os.path.isfile", "multiprocessing.Queue", "os.path.join", "logging.error", "json.loads", "os.path.dirname", "os.path.exists", "json.dump", "copy.deepcopy", "subprocess.Popen", "importlib.import_module", "os.listdir", "re.compile", "json.load", ...
[((2094, 2120), 'os.path.dirname', 'os.path.dirname', (['file_name'], {}), '(file_name)\n', (2109, 2120), False, 'import os\n'), ((12416, 12435), 're.compile', 're.compile', (['pattern'], {}), '(pattern)\n', (12426, 12435), False, 'import re\n'), ((25162, 25183), 'copy.deepcopy', 'copy.deepcopy', (['config'], {}), '(co...
import json import os import requests from client import Submission from settings import API_BASE_URL, ACCESS_TOKEN def get_task_url(task_id: int): return API_BASE_URL + f"/tasks/{task_id}/download_grader/" def get_agent_url(submission_id: int): return API_BASE_URL + f"/submissions/{submission_id}/downloa...
[ "os.getenv", "json.loads", "requests.get" ]
[((510, 683), 'requests.get', 'requests.get', (["(API_BASE_URL + f'/jobs/{job_id}/start_job/')"], {'headers': "{'Authorization': f'Token {ACCESS_TOKEN}'}", 'data': "{'worker_name': worker_name, 'task_id': task_id}"}), "(API_BASE_URL + f'/jobs/{job_id}/start_job/', headers={\n 'Authorization': f'Token {ACCESS_TOKEN}'...
import time import logging from extended_networkx_tools import Analytics, AnalyticsGraph from timeit import default_timer as timer from utils import Solvers from utils.GraphUtils import GraphUtils from utils.ServerUtil import ServerUtil from datetime import datetime class GraphThread: @staticmethod def sta...
[ "utils.Solvers.Diff.solve", "utils.GraphUtils.GraphUtils.get_results", "logging.exception", "utils.Solvers.Random.solve", "timeit.default_timer", "utils.Solvers.Field.solve", "time.sleep", "utils.ServerUtil.ServerUtil", "utils.Solvers.Spec.solve", "datetime.datetime.now" ]
[((1097, 1117), 'utils.ServerUtil.ServerUtil', 'ServerUtil', (['base_url'], {}), '(base_url)\n', (1107, 1117), False, 'from utils.ServerUtil import ServerUtil\n'), ((1424, 1431), 'timeit.default_timer', 'timer', ([], {}), '()\n', (1429, 1431), True, 'from timeit import default_timer as timer\n'), ((1512, 1519), 'timeit...
from django import forms SACCO_DRIVER_STATUS_OPTIONS = [ ('Approved', ('Approved to operate')), ('Suspended', ('Suspended for the time being')), ('Blacklisted', ('Blacklisted from operating')) ] class VehicleForm(forms.Form): # sacco = forms.CharField(label="Sacco", max_length=100) regno = forms....
[ "django.forms.CharField", "django.forms.Select" ]
[((314, 372), 'django.forms.CharField', 'forms.CharField', ([], {'label': '"""Registration Number"""', 'max_length': '(7)'}), "(label='Registration Number', max_length=7)\n", (329, 372), False, 'from django import forms\n'), ((658, 687), 'django.forms.CharField', 'forms.CharField', ([], {'max_length': '(8)'}), '(max_le...
from typing import Optional from talon import Context from user.emacs.utils.voicemacs import rpc_call from user.utils.formatting import SurroundingText context = Context() context.matches = r""" tag: user.emacs """ @context.action_class("self") class UserActions: def surrounding_text() -> Optional[Surroundin...
[ "talon.Context", "user.emacs.utils.voicemacs.rpc_call", "user.utils.formatting.SurroundingText" ]
[((166, 175), 'talon.Context', 'Context', ([], {}), '()\n', (173, 175), False, 'from talon import Context\n'), ((416, 519), 'user.emacs.utils.voicemacs.rpc_call', 'rpc_call', (['"""voicemacs-surrounding-text"""', "[':chars-before', 30000, ':chars-after', 30000]"], {'timeout': '(10)'}), "('voicemacs-surrounding-text', [...
#!/usr/bin/env python # -*- coding: utf-8 -*- import sys import zipfile def make_nvz_main(output_file, nvm_file, target_file, pitch_file=None): if pitch_file is not None: files = [nvm_file, target_file, pitch_file] arc_names = ['target.nvm', 'target.pb', 'pitch.pb'] else: ...
[ "zipfile.ZipFile" ]
[((417, 484), 'zipfile.ZipFile', 'zipfile.ZipFile', (['output_file', '"""w"""'], {'compression': 'zipfile.ZIP_DEFLATED'}), "(output_file, 'w', compression=zipfile.ZIP_DEFLATED)\n", (432, 484), False, 'import zipfile\n')]
import click from mutacc.mutaccDB.remove_case import remove_case_from_db @click.command('remove') @click.argument('case_id') @click.pass_context def remove_command(context, case_id): """ Deletes case from mutacc DB """ adapter = context.obj['adapter'] remove_case_from_db(adapter, case_id)
[ "mutacc.mutaccDB.remove_case.remove_case_from_db", "click.argument", "click.command" ]
[((76, 99), 'click.command', 'click.command', (['"""remove"""'], {}), "('remove')\n", (89, 99), False, 'import click\n'), ((101, 126), 'click.argument', 'click.argument', (['"""case_id"""'], {}), "('case_id')\n", (115, 126), False, 'import click\n'), ((280, 317), 'mutacc.mutaccDB.remove_case.remove_case_from_db', 'remo...
# encoding: utf-8 # http://www.hexblog.com/?p=120 # Default IDA Pro Paths: # MAC /Applications/IDA\ Pro\ X/idaq.app/Contents/MacOS/plugins/ # Windows C:\Program Files (x86)\IDA X\plugins # to make it autoexec on openfile # add this to plugins.cfg # ; Other plugins #FullColor FullColor.py 0...
[ "idaapi.msg", "idc.set_color", "idc.get_screen_ea", "idc.print_operand", "idc.print_insn_mnem" ]
[((952, 995), 'idaapi.msg', 'idaapi.msg', (["('run() called with %d!\\n' % arg)"], {}), "('run() called with %d!\\n' % arg)\n", (962, 995), False, 'import idaapi\n'), ((2142, 2172), 'idaapi.msg', 'idaapi.msg', (['"""term() called!\n"""'], {}), "('term() called!\\n')\n", (2152, 2172), False, 'import idaapi\n'), ((1766, ...
import os import random import subprocess MYPATH = './out-of-dropbox-2020-08to12-' FILES = os.listdir(MYPATH) INP = '' while INP != 'q': INP = input('q to quit, enter anything else to continue') file_choice = random.choice(FILES) pathname_choice = MYPATH + '/' + file_choice subprocess.run(["open", path...
[ "subprocess.run", "random.choice", "os.listdir" ]
[((92, 110), 'os.listdir', 'os.listdir', (['MYPATH'], {}), '(MYPATH)\n', (102, 110), False, 'import os\n'), ((218, 238), 'random.choice', 'random.choice', (['FILES'], {}), '(FILES)\n', (231, 238), False, 'import random\n'), ((292, 333), 'subprocess.run', 'subprocess.run', (["['open', pathname_choice]"], {}), "(['open',...
import argparse import json from multiprocessing import Pool from puyotable.canonization import canonize_deals def all_deals(num_deals, num_colors): if not num_deals: return [[]] result = [] for c0 in range(num_colors): for c1 in range(num_colors): for deals in all_deals(num_d...
[ "puyotable.canonization.canonize_deals", "argparse.ArgumentParser", "multiprocessing.Pool" ]
[((1216, 1304), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Tabulate all opening sequences in Puyo Puyo."""'}), "(description=\n 'Tabulate all opening sequences in Puyo Puyo.')\n", (1239, 1304), False, 'import argparse\n'), ((1988, 1994), 'multiprocessing.Pool', 'Pool', ([], {}), '...
import matplotlib.pyplot as plt import numpy as np import numpy.polynomial.polynomial as nppol class Metawalk: def __init__(self, time_intervals=None, nodes=None, ): """ A basic constructor for a ``Metwalks`` object :param times : A list o...
[ "matplotlib.pyplot.tight_layout", "matplotlib.pyplot.plot", "numpy.polynomial.polynomial.Polynomial", "matplotlib.pyplot.vlines", "numpy.around", "numpy.math.factorial", "matplotlib.pyplot.gca", "matplotlib.pyplot.hlines", "matplotlib.pyplot.subplots" ]
[((4201, 4222), 'numpy.polynomial.polynomial.Polynomial', 'nppol.Polynomial', (['res'], {}), '(res)\n', (4217, 4222), True, 'import numpy.polynomial.polynomial as nppol\n'), ((7457, 7567), 'matplotlib.pyplot.plot', 'plt.plot', (['[self.time_intervals[0]]', '[id_source]'], {'color': 'color', 'marker': '"""o"""', 'alpha'...
from django.db import models from django.core.validators import MaxValueValidator, MinValueValidator from django.db.models.signals import post_delete from django.dispatch import receiver from myauth import models as myauth_models from products.models import Book class CommentProducts(models.Model): profile = mod...
[ "django.db.models.TextField", "django.db.models.ForeignKey", "django.core.validators.MinValueValidator", "django.dispatch.receiver", "django.db.models.DateTimeField", "django.core.validators.MaxValueValidator" ]
[((1334, 1379), 'django.dispatch.receiver', 'receiver', (['post_delete'], {'sender': 'CommentProducts'}), '(post_delete, sender=CommentProducts)\n', (1342, 1379), False, 'from django.dispatch import receiver\n'), ((317, 412), 'django.db.models.ForeignKey', 'models.ForeignKey', (['myauth_models.Profile'], {'on_delete': ...
from urllib.parse import urlparse, urlunparse from functools import wraps from flask import abort, request, current_app from lib.twilio import TwilioClient def validate_twilio_request(f): """Validates that incoming requests genuinely originated from Twilio""" # Adapted from https://www.twilio.com/docs/usage/...
[ "flask.request.headers.get", "urllib.parse.urlunparse", "flask.abort", "functools.wraps", "lib.twilio.TwilioClient", "urllib.parse.urlparse" ]
[((524, 532), 'functools.wraps', 'wraps', (['f'], {}), '(f)\n', (529, 532), False, 'from functools import wraps\n'), ((602, 718), 'lib.twilio.TwilioClient', 'TwilioClient', (["current_app.config['SECRETS'].TWILIO_ACCOUNT_SID", "current_app.config['SECRETS'].TWILIO_AUTH_TOKEN"], {}), "(current_app.config['SECRETS'].TWIL...
#!/usr/bin/env python3 import os from typing import Union from pathlib import Path import requests from rpicam.utils.logging_utils import get_logger class TelegramPoster: """ Bare-bones class to post videos to a Telegram chat. Uses per default credentials stored in environment. """ API_URL = '...
[ "requests.post", "rpicam.utils.logging_utils.get_logger", "os.getenv" ]
[((806, 852), 'rpicam.utils.logging_utils.get_logger', 'get_logger', (['self.__class__.__name__'], {'verb': '(True)'}), '(self.__class__.__name__, verb=True)\n', (816, 852), False, 'from rpicam.utils.logging_utils import get_logger\n'), ((1429, 1460), 'requests.post', 'requests.post', (['url'], {'files': 'files'}), '(u...
""" This module provides a class for interfacing with the Sense HAT add-on board for Raspberry Pi. """ import os from multiprocessing.managers import RemoteError from myDevices.utils.logger import error, exception, info from sensehat.manager import connect_client class SenseHAT(): """Class for interacting with a...
[ "myDevices.utils.logger.info", "myDevices.utils.logger.error", "os.path.join", "sensehat.manager.connect_client" ]
[((759, 836), 'os.path.join', 'os.path.join', (['"""/etc/myDevices/plugins/cayenne-plugin-sensehat/data/image.png"""'], {}), "('/etc/myDevices/plugins/cayenne-plugin-sensehat/data/image.png')\n", (771, 836), False, 'import os\n'), ((1086, 1102), 'sensehat.manager.connect_client', 'connect_client', ([], {}), '()\n', (11...
#!/usr/bin/env python from isrutils.looper import simpleloop # %% users param P = { "path": "~/data/2015-10-07/isr", "beamid": 64157, "acf": True, "vlimacf": (18, 45), "zlim_pl": [None, None], "vlim_pl": [72, 90], "flim_pl": [3.5, 5.5], "odir": "out/2015-10-07", "vlim": [25, 55], ...
[ "isrutils.looper.simpleloop" ]
[((383, 403), 'isrutils.looper.simpleloop', 'simpleloop', (['flist', 'P'], {}), '(flist, P)\n', (393, 403), False, 'from isrutils.looper import simpleloop\n')]
"""turbotutorial URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/4.0/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: path('', views.home, name='home') Class...
[ "django.views.generic.TemplateView.as_view", "django.urls.path" ]
[((789, 820), 'django.urls.path', 'path', (['"""admin/"""', 'admin.site.urls'], {}), "('admin/', admin.site.urls)\n", (793, 820), False, 'from django.urls import path\n'), ((1061, 1121), 'django.views.generic.TemplateView.as_view', 'TemplateView.as_view', ([], {'template_name': '"""broadcast_example.html"""'}), "(templ...
import os # ETL ETL_MODE = os.environ.get('ETL_MODE') ETL_CHUNK_SIZE = int(os.environ.get('ETL_CHUNK_SIZE')) ETL_SYNC_DELAY = int(os.environ.get('ETL_SYNC_DELAY')) ETL_FILE_STATE = os.environ.get('ETL_FILE_STATE') ETL_DEFAULT_DATE = os.environ.get('ETL_DEFAULT_DATE') # Postgres POSTGRES_NAME = os.environ.get('POSTGRE...
[ "os.environ.get" ]
[((28, 54), 'os.environ.get', 'os.environ.get', (['"""ETL_MODE"""'], {}), "('ETL_MODE')\n", (42, 54), False, 'import os\n'), ((182, 214), 'os.environ.get', 'os.environ.get', (['"""ETL_FILE_STATE"""'], {}), "('ETL_FILE_STATE')\n", (196, 214), False, 'import os\n'), ((234, 268), 'os.environ.get', 'os.environ.get', (['"""...
#!/usr/bin/env python from __future__ import absolute_import from goonpug import app def main(): app.run(debug=True) if __name__ == '__main__': main()
[ "goonpug.app.run" ]
[((105, 124), 'goonpug.app.run', 'app.run', ([], {'debug': '(True)'}), '(debug=True)\n', (112, 124), False, 'from goonpug import app\n')]
from setuptools import setup, find_packages from Cython.Distutils.extension import Extension from Cython.Build import cythonize, build_ext import numpy import os from glob import glob """ ext_modules = [Extension("traj_dist.cydist.basic_geographical", ["traj_dist/cydist/basic_geographical.pyx"]), Extens...
[ "numpy.get_include", "setuptools.find_packages", "glob.glob" ]
[((1294, 1324), 'glob.glob', 'glob', (['"""traj_dist/cydist/*.pyx"""'], {}), "('traj_dist/cydist/*.pyx')\n", (1298, 1324), False, 'from glob import glob\n'), ((1950, 1965), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (1963, 1965), False, 'from setuptools import setup, find_packages\n'), ((1700, 1719)...
import pyeccodes.accessors as _ def load(h): h.add(_.Unsigned('numberOfCategories', 1)) with h.list('categories'): for i in range(0, h.get_l('numberOfCategories')): h.add(_.Codetable('categoryType', 1, "4.91.table", _.Get('masterDir'), _.Get('localDir'))) h.add(_.Unsigned('co...
[ "pyeccodes.accessors.Unsigned", "pyeccodes.accessors.Get" ]
[((58, 93), 'pyeccodes.accessors.Unsigned', '_.Unsigned', (['"""numberOfCategories"""', '(1)'], {}), "('numberOfCategories', 1)\n", (68, 93), True, 'import pyeccodes.accessors as _\n'), ((306, 333), 'pyeccodes.accessors.Unsigned', '_.Unsigned', (['"""codeFigure"""', '(1)'], {}), "('codeFigure', 1)\n", (316, 333), True,...
import numpy as np from numba import guvectorize from pygama.dsp.errors import DSPFatal @guvectorize(["void(float32[:], float32, float32, float32, float32[:])", "void(float64[:], float64, float64, float64, float64[:])"], "(n),(),(),()->()", nopython=True, cache=True) def time_point_thresh(w_...
[ "numpy.floor", "pygama.dsp.errors.DSPFatal", "numba.guvectorize", "numpy.isnan" ]
[((90, 276), 'numba.guvectorize', 'guvectorize', (["['void(float32[:], float32, float32, float32, float32[:])',\n 'void(float64[:], float64, float64, float64, float64[:])']", '"""(n),(),(),()->()"""'], {'nopython': '(True)', 'cache': '(True)'}), "(['void(float32[:], float32, float32, float32, float32[:])',\n 'voi...
import tornado import json import uuid import pandas as pd from handler.mlsklearn.util import regqeust_arg_to_sklearn_arg from sklearn.model_selection import train_test_split from data.persistence import * from data.data_source import DataSource from data.data_storage import DataStorage class TrainTestSplitHandler(t...
[ "pandas.DataFrame", "json.loads", "sklearn.model_selection.train_test_split", "json.dumps", "data.data_storage.DataStorage.create_data_obj_by_pandas_data", "handler.mlsklearn.util.regqeust_arg_to_sklearn_arg", "data.data_storage.DataStorage.get_data_obj_by_data_id" ]
[((387, 416), 'json.loads', 'json.loads', (['self.request.body'], {}), '(self.request.body)\n', (397, 416), False, 'import json\n'), ((890, 934), 'data.data_storage.DataStorage.get_data_obj_by_data_id', 'DataStorage.get_data_obj_by_data_id', (['data_id'], {}), '(data_id)\n', (925, 934), False, 'from data.data_storage i...
import numpy as np import pandas as pd import yfinance as yf import matplotlib.pyplot as plt import datetime from yahoo_fin import stock_info as si plt.rcParams['figure.figsize'] = (15, 10) tickers = si.tickers_dow() individual_stock = input(f"Which of the following stocks would you like to backtest \n{tickers}\n:") ...
[ "matplotlib.pyplot.title", "matplotlib.pyplot.show", "yfinance.download", "matplotlib.pyplot.legend", "datetime.date.today", "yahoo_fin.stock_info.tickers_dow", "pandas.Series", "matplotlib.pyplot.subplots" ]
[((202, 218), 'yahoo_fin.stock_info.tickers_dow', 'si.tickers_dow', ([], {}), '()\n', (216, 218), True, 'from yahoo_fin import stock_info as si\n'), ((434, 467), 'yfinance.download', 'yf.download', (['tickers'], {'start': 'start'}), '(tickers, start=start)\n', (445, 467), True, 'import yfinance as yf\n'), ((705, 723), ...
# -*- coding: utf-8 -*- """ Created on Tue May 22 14:07:42 2018 @author: HORSE """ import logging import logging.handlers import os def ARLogger(log_filename = 'log.txt'): # if not os.path.exists('logs'): # os.makedirs('logs') fmt = '%(asctime)s %(levelname)s %(message)s' ...
[ "logging.Formatter", "logging.handlers.RotatingFileHandler", "logging.getLogger" ]
[((372, 401), 'logging.getLogger', 'logging.getLogger', (['"""ARLogger"""'], {}), "('ARLogger')\n", (389, 401), False, 'import logging\n'), ((454, 543), 'logging.handlers.RotatingFileHandler', 'logging.handlers.RotatingFileHandler', (['log_filename'], {'maxBytes': '(10000 * 4)', 'backupCount': '(5)'}), '(log_filename, ...
import unittest from desky.rect import Rect from desky.panel import Panel from enum import Enum from functools import reduce, partial from toolz.dicttoolz import valfilter # | Type of sizing | Maximum extra width allocation # -------------------------------------------------------------- # | Fixed (200 p...
[ "unittest.main", "functools.partial", "desky.panel.Panel.Rect", "toolz.dicttoolz.valfilter", "desky.rect.Rect", "desky.gui.Gui" ]
[((29067, 29082), 'unittest.main', 'unittest.main', ([], {}), '()\n', (29080, 29082), False, 'import unittest\n'), ((1911, 1955), 'toolz.dicttoolz.valfilter', 'valfilter', (['(lambda p: p != panel)', 'self.panels'], {}), '(lambda p: p != panel, self.panels)\n', (1920, 1955), False, 'from toolz.dicttoolz import valfilte...
from distutils.core import setup, Extension def main(): setup(name="seqrepc", version="beta1.0", description="SeqrepC is a module for fundamental operations related to numerical representations of genomic sequences.", author="<NAME>", author_email="<EMAIL>", url="...
[ "distutils.core.Extension" ]
[((429, 470), 'distutils.core.Extension', 'Extension', (['"""seqrepc"""', "['./src/seqrepc.c']"], {}), "('seqrepc', ['./src/seqrepc.c'])\n", (438, 470), False, 'from distutils.core import setup, Extension\n')]
# pylint: disable=no-member,invalid-name,line-too-long,trailing-whitespace """Add IsEnabled column to EventHandlerBASE Revision ID: <KEY> Revises: 6b5369ab5224 Create Date: 2021-02-17 20:15:42.776190 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = '<KEY>' down_...
[ "alembic.op.drop_column", "sqlalchemy.Column" ]
[((548, 595), 'alembic.op.drop_column', 'op.drop_column', (['"""EventHandlerBASE"""', '"""IsEnabled"""'], {}), "('EventHandlerBASE', 'IsEnabled')\n", (562, 595), False, 'from alembic import op\n'), ((458, 523), 'sqlalchemy.Column', 'sa.Column', (['"""IsEnabled"""', 'sa.Boolean'], {'nullable': '(False)', 'default': '(Fa...
from resource_management.libraries.script.script import Script from resource_management.core.resources.packaging import Package class Client(Script): def install(self, env): packages = ['percona-server-client'] Package(packages) self.configure(env) def configure(self, env): im...
[ "resource_management.core.resources.packaging.Package" ]
[((233, 250), 'resource_management.core.resources.packaging.Package', 'Package', (['packages'], {}), '(packages)\n', (240, 250), False, 'from resource_management.core.resources.packaging import Package\n')]
# Generated by Django 2.2.1 on 2019-07-19 12:36 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('website', '0035_auto_20190625_0900'), ] operations = [ migrations.RenameField( model_name='verenigingen', old_name='ontgroen...
[ "django.db.migrations.RenameField" ]
[((227, 332), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""verenigingen"""', 'old_name': '"""ontgroening"""', 'new_name': '"""introductietijd"""'}), "(model_name='verenigingen', old_name='ontgroening',\n new_name='introductietijd')\n", (249, 332), False, 'from django.db impor...
#!/usr/bin/python # Flask is used to create a somewhat lightweight listening server from flask import Flask from requests import get def spawn_proxy(): myproxy = Flask('__name__') # Quick health check override @myproxy.route('/healthcheck', methods=['GET']) def health(): return "OK" # ...
[ "flask.Flask", "requests.get" ]
[((170, 187), 'flask.Flask', 'Flask', (['"""__name__"""'], {}), "('__name__')\n", (175, 187), False, 'from flask import Flask\n'), ((774, 805), 'requests.get', 'get', (['f"""{target}/search?q={req}"""'], {}), "(f'{target}/search?q={req}')\n", (777, 805), False, 'from requests import get\n')]
#! /usr/bin/env python import pefile import datetime import os import re from pecli.plugins.base import Plugin from pecli.lib.utils import cli_out ASCII_BYTE = b" !\"#\$%&\'\(\)\*\+,-\./0123456789:;<=>\?@ABCDEFGHIJKLMNOPQRSTUVWXYZ\[\]\^_`abcdefghijklmnopqrstuvwxyz\{\|\}\\\~\t" class PluginStrings(Plugin): name...
[ "pecli.lib.utils.cli_out", "re.compile" ]
[((1130, 1180), 're.compile', 're.compile', (["(b'([%s]{%d,})' % (ASCII_BYTE, min_len))"], {}), "(b'([%s]{%d,})' % (ASCII_BYTE, min_len))\n", (1140, 1180), False, 'import re\n'), ((1199, 1257), 're.compile', 're.compile', (["(b'((?:[%s]\\x00){%d,})' % (ASCII_BYTE, min_len))"], {}), "(b'((?:[%s]\\x00){%d,})' % (ASCII_BY...
from django.core.management.base import BaseCommand from api.models import Country from django.db import transaction from django.db.models import Q from api.logger import logger class Command(BaseCommand): help = 'Update Countries initially to set/revoke their in_search field (probably one-time run only)' @t...
[ "api.logger.logger.info", "django.db.models.Q", "api.models.Country.objects.filter" ]
[((847, 902), 'api.logger.logger.info', 'logger.info', (['"""Successfully set in_search for Countries"""'], {}), "('Successfully set in_search for Countries')\n", (858, 902), False, 'from api.logger import logger\n'), ((472, 548), 'api.models.Country.objects.filter', 'Country.objects.filter', ([], {'independent': '(Tru...
from django import template from tos.models import CGUItem register = template.Library() @register.simple_tag def get_cgu_items(): return CGUItem.objects.filter(deleted_at__isnull=True)
[ "django.template.Library", "tos.models.CGUItem.objects.filter" ]
[((72, 90), 'django.template.Library', 'template.Library', ([], {}), '()\n', (88, 90), False, 'from django import template\n'), ((145, 192), 'tos.models.CGUItem.objects.filter', 'CGUItem.objects.filter', ([], {'deleted_at__isnull': '(True)'}), '(deleted_at__isnull=True)\n', (167, 192), False, 'from tos.models import CG...
# Copyright 2016 The Chromium OS Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """This is a factory test to check the brightness of LCD backlight or LEDs.""" from cros.factory.device import device_utils from cros.factory.test.i18n i...
[ "cros.factory.test.i18n.arg_utils.I18nArg", "cros.factory.utils.arg_utils.Arg", "cros.factory.device.device_utils.CreateDUTInterface" ]
[((538, 583), 'cros.factory.test.i18n.arg_utils.I18nArg', 'i18n_arg_utils.I18nArg', (['"""msg"""', '"""Message HTML"""'], {}), "('msg', 'Message HTML')\n", (560, 583), True, 'from cros.factory.test.i18n import arg_utils as i18n_arg_utils\n'), ((591, 669), 'cros.factory.utils.arg_utils.Arg', 'Arg', (['"""timeout_secs"""...
import pandas as pd import os import dotenv from dotenv import load_dotenv import datetime import plotly import plotly.graph_objects as go from plotly.subplots import make_subplots from app.other_data_pull import spy_pull, fred_pull from app.port_data_pull import port_data_pull from app.portfolio_import import portfol...
[ "app.port_data_pull.port_data_pull", "os.path.abspath", "app.other_data_pull.fred_pull", "app.other_data_pull.spy_pull", "dotenv.load_dotenv", "os.environ.get", "app.portfolio_import.portfolio_import", "plotly.subplots.make_subplots" ]
[((7076, 7089), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (7087, 7089), False, 'from dotenv import load_dotenv\n'), ((7111, 7148), 'os.environ.get', 'os.environ.get', (['"""PORTFOLIO_FILE_NAME"""'], {}), "('PORTFOLIO_FILE_NAME')\n", (7125, 7148), False, 'import os\n'), ((7166, 7204), 'os.environ.get', 'os....
from keras_tweaks import idseqs_to_mask import tensorflow as tf class AllTests(tf.test.TestCase): def test1(self): idseqs = [[1, 1, 0, 0, 2, 2, 3], [1, 3, 2, 1, 0, 0, 2]] target = tf.sparse.SparseTensor( indices=( [0, 0, 1], [0, 1, 1], ...
[ "tensorflow.test.main", "keras_tweaks.idseqs_to_mask", "tensorflow.sparse.to_dense" ]
[((4219, 4233), 'tensorflow.test.main', 'tf.test.main', ([], {}), '()\n', (4231, 4233), True, 'import tensorflow as tf\n'), ((646, 719), 'keras_tweaks.idseqs_to_mask', 'idseqs_to_mask', (['idseqs'], {'n_seqlen': '(6)', 'n_vocab_sz': '(3)', 'ignore': '[3]', 'dense': '(False)'}), '(idseqs, n_seqlen=6, n_vocab_sz=3, ignor...
# apis_v1/test_views_voter_email_address_save.py # Brought to you by We Vote. Be good. # -*- coding: UTF-8 -*- from django.urls import reverse from django.test import TestCase from email_outbound.models import EmailAddress, EmailManager import json class WeVoteAPIsV1TestsVoterEmailAddressRetrieve(TestCase): datab...
[ "django.urls.reverse" ]
[((417, 456), 'django.urls.reverse', 'reverse', (['"""apis_v1:deviceIdGenerateView"""'], {}), "('apis_v1:deviceIdGenerateView')\n", (424, 456), False, 'from django.urls import reverse\n'), ((489, 523), 'django.urls.reverse', 'reverse', (['"""apis_v1:voterCreateView"""'], {}), "('apis_v1:voterCreateView')\n", (496, 523)...
#!/usr/bin/python # -*- coding: utf-8 -*- """ Redundant misc. functions to be eventually removed from AC_tools. """ import os import numpy as np from matplotlib.backends.backend_pdf import PdfPages import matplotlib.pyplot as plt from pandas import DataFrame # time import time import datetime as datetime # math from m...
[ "pandas.DataFrame", "numpy.linspace" ]
[((4158, 4204), 'pandas.DataFrame', 'DataFrame', (['data'], {'index': 'dates', 'columns': "['data']"}), "(data, index=dates, columns=['data'])\n", (4167, 4204), False, 'from pandas import DataFrame\n'), ((2254, 2314), 'numpy.linspace', 'np.linspace', (['(0)', '(1)', 'extra_points_point_on_edge'], {'endpoint': '(True)'}...
from nanome._internal._util._serializers import _ArraySerializer, _StringSerializer, _ColorSerializer from . import _AtomSerializerID from . import _BondSerializer from .. import _Residue from nanome.util import Logs from nanome._internal._util._serializers import _TypeSerializer class _ResidueSerializer(_TypeSeriali...
[ "nanome._internal._util._serializers._StringSerializer", "nanome._internal._util._serializers._ArraySerializer", "nanome._internal._util._serializers._ColorSerializer" ]
[((419, 437), 'nanome._internal._util._serializers._ArraySerializer', '_ArraySerializer', ([], {}), '()\n', (435, 437), False, 'from nanome._internal._util._serializers import _ArraySerializer, _StringSerializer, _ColorSerializer\n'), ((537, 555), 'nanome._internal._util._serializers._ColorSerializer', '_ColorSerialize...
#!/usr/bin/env python3 import logging import subprocess from typing import Dict class Agent: name: str image: str environment: Dict[str, str] def __init__(self, name: str, image: str, environment: Dict[str, str]) -> None: self.name = name self.image = image self.environment =...
[ "logging.info", "subprocess.run" ]
[((369, 447), 'logging.info', 'logging.info', (['"""Starting agent \'%s\' based on image \'%s\'"""', 'self.name', 'self.image'], {}), '("Starting agent \'%s\' based on image \'%s\'", self.name, self.image)\n', (381, 447), False, 'import logging\n'), ((456, 573), 'subprocess.run', 'subprocess.run', (["['docker', 'run', ...
# -*- coding: utf-8 -*- ################################################################################################# import logging import urllib import requests from resources.lib.util import JSONRPC ################################################################################################## log = log...
[ "resources.lib.util.JSONRPC", "requests.head", "logging.getLogger" ]
[((317, 355), 'logging.getLogger', 'logging.getLogger', (["('DINGS.' + __name__)"], {}), "('DINGS.' + __name__)\n", (334, 355), False, 'import logging\n'), ((1228, 1263), 'resources.lib.util.JSONRPC', 'JSONRPC', (['"""Settings.GetSettingValue"""'], {}), "('Settings.GetSettingValue')\n", (1235, 1263), False, 'from resou...
#!/usr/bin/env python # Copyright (C) 2004 British Broadcasting Corporation and Kamaelia Contributors(1) # All Rights Reserved. # # You may only modify and redistribute this under the terms of any of the # following licenses(2): Mozilla Public License, V1.1, GNU General # Public License, V2.0, GNU Lesser Ge...
[ "SpatialIndexer.SpatialIndexer" ]
[((4558, 4596), 'SpatialIndexer.SpatialIndexer', 'SpatialIndexer', (['laws.maxInteractRadius'], {}), '(laws.maxInteractRadius)\n', (4572, 4596), False, 'from SpatialIndexer import SpatialIndexer\n')]
from rest_framework import parsers, renderers from rest_framework.authtoken.models import Token from rest_framework.authtoken.serializers import AuthTokenSerializer from rest_framework.response import Response from rest_framework.views import APIView from .serializers import ClientTokenSerializer from .models import C...
[ "rest_framework.response.Response" ]
[((1018, 1048), 'rest_framework.response.Response', 'Response', (["{'token': token.key}"], {}), "({'token': token.key})\n", (1026, 1048), False, 'from rest_framework.response import Response\n')]
# !/usr/bin/python from itertools import groupby def compress(data): return ((len(list(group)), name) for name, group in groupby(data)) def decompress(data): return (car * size for size, car in data) my_data = 'get uuuuuuuuuuuuuuuup' print(list(my_data)) compressed = compress(my_data) print(''.join(deco...
[ "itertools.groupby" ]
[((128, 141), 'itertools.groupby', 'groupby', (['data'], {}), '(data)\n', (135, 141), False, 'from itertools import groupby\n')]
import numpy as np import magpie # check cartesian def test_get_xedges(): xedges = magpie.grids.get_xedges(1., 2) xedges = np.round(xedges, decimals=2) assert len(xedges) == 3, "Length of xedges is incorrect." assert xedges[-1] - xedges[0] == 1., "xedges range is incorrect." xedges = magpie.grids...
[ "magpie.grids.polargrid", "numpy.sum", "magpie.grids.grid1d", "magpie.grids.grid3d", "magpie.grids.xmid2edges", "magpie.grids.get_xedges", "magpie.grids.polarEA_grid", "numpy.shape", "magpie.grids.polarEA_npix", "magpie.grids.polarEA_area", "magpie.grids.xedges2mid", "numpy.round", "magpie.g...
[((90, 121), 'magpie.grids.get_xedges', 'magpie.grids.get_xedges', (['(1.0)', '(2)'], {}), '(1.0, 2)\n', (113, 121), False, 'import magpie\n'), ((134, 162), 'numpy.round', 'np.round', (['xedges'], {'decimals': '(2)'}), '(xedges, decimals=2)\n', (142, 162), True, 'import numpy as np\n'), ((308, 350), 'magpie.grids.get_x...
from communication import * import socket s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) s.connect((socket.gethostname(), 1123)) while True: m = receive_message(s) if m: print(m, "\n") ping(s) print(s.getsockname()) print(socket.gethostbyname(socket.gethostname())) print(socket.get)
[ "socket.gethostname", "socket.socket" ]
[((47, 96), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (60, 96), False, 'import socket\n'), ((108, 128), 'socket.gethostname', 'socket.gethostname', ([], {}), '()\n', (126, 128), False, 'import socket\n'), ((262, 282), 'socket.gethostname'...
""" dmsetup commands - Command ``dmsetup`` ====================================== Parsers for parsing and extracting data from output of commands related to ``dmsetup``. Parsers contained in this module are: DmsetupInfo - command ``dmsetup info -C`` ----------------------------------------- """ from insights import...
[ "insights.parsers.parse_delimited_table", "insights.parser" ]
[((430, 456), 'insights.parser', 'parser', (['Specs.dmsetup_info'], {}), '(Specs.dmsetup_info)\n', (436, 456), False, 'from insights import parser, CommandParser\n'), ((2195, 2225), 'insights.parsers.parse_delimited_table', 'parse_delimited_table', (['content'], {}), '(content)\n', (2216, 2225), False, 'from insights.p...
import sys, os, random import numpy as np import matplotlib.pyplot as plt from matplotlib.colors import Normalize from matplotlib.colors import ListedColormap from sklearn.model_selection import train_test_split, GridSearchCV from sklearn.model_selection import StratifiedShuffleSplit from sklearn.preprocessing i...
[ "matplotlib.pyplot.title", "matplotlib.pyplot.figaspect", "numpy.load", "sklearn.preprocessing.StandardScaler", "random.shuffle", "matplotlib.pyplot.style.use", "matplotlib.pyplot.figure", "numpy.mean", "sklearn.neural_network.MLPClassifier", "sklearn.svm.SVC", "numpy.interp", "os.path.join", ...
[((1137, 1191), 'os.path.join', 'os.path.join', (['dirpath', '"""../datasets/breast-cancer.npz"""'], {}), "(dirpath, '../datasets/breast-cancer.npz')\n", (1149, 1191), False, 'import sys, os, random\n'), ((1202, 1251), 'os.path.join', 'os.path.join', (['dirpath', '"""../datasets/diabetes.npz"""'], {}), "(dirpath, '../d...
import unittest import maxixe from maxixe.tests import decorators from maxixe.tests import loader from maxixe.tests import parser from maxixe.tests import utils suite = unittest.TestSuite() suite.addTests(unittest.TestLoader().loadTestsFromModule(decorators)) suite.addTests(unittest.TestLoader().loadTestsFromModule(l...
[ "unittest.TestLoader", "unittest.TestSuite" ]
[((171, 191), 'unittest.TestSuite', 'unittest.TestSuite', ([], {}), '()\n', (189, 191), False, 'import unittest\n'), ((207, 228), 'unittest.TestLoader', 'unittest.TestLoader', ([], {}), '()\n', (226, 228), False, 'import unittest\n'), ((277, 298), 'unittest.TestLoader', 'unittest.TestLoader', ([], {}), '()\n', (296, 29...
from django.db import models class Blog(models.Model): title = models.CharField("标题", unique=True, max_length=200) class Meta: db_table = 'blog' verbose_name = '文章'
[ "django.db.models.CharField" ]
[((71, 122), 'django.db.models.CharField', 'models.CharField', (['"""标题"""'], {'unique': '(True)', 'max_length': '(200)'}), "('标题', unique=True, max_length=200)\n", (87, 122), False, 'from django.db import models\n')]
from wormer.tools import manager, downloader from wormer.data import strategy import re class Graber: synopsis_pattern = '''(?=lemma-summary")(.*?)(?<=config) ''' text_pattern = '>\s*?([^\&\b\n\[\]]*?)<' href_pattern = '<a target=_blank href="(/item/[\w\d%]*?)">' def __init__(self): self.urlM...
[ "wormer.tools.manager.ThreadManager", "wormer.tools.manager.LogManager", "wormer.tools.manager.TextManager", "wormer.tools.manager.UrlsManager", "wormer.tools.downloader.DownLoader" ]
[((329, 350), 'wormer.tools.manager.UrlsManager', 'manager.UrlsManager', ([], {}), '()\n', (348, 350), False, 'from wormer.tools import manager, downloader\n'), ((377, 400), 'wormer.tools.downloader.DownLoader', 'downloader.DownLoader', ([], {}), '()\n', (398, 400), False, 'from wormer.tools import manager, downloader\...
from telegram.ext import ConversationHandler from telegram.ext import MessageHandler from telegram.ext import Filters from telegram.ext import CallbackQueryHandler from Model.share import Share import Controllers.global_states as states from Utils.logging import get_logger as log import pandas as pd import datetime G...
[ "Model.share.Share", "Utils.logging.get_logger", "telegram.ext.MessageHandler", "pandas.to_datetime", "datetime.datetime.now" ]
[((1496, 1509), 'Model.share.Share', 'Share', (['ticker'], {}), '(ticker)\n', (1501, 1509), False, 'from Model.share import Share\n'), ((994, 999), 'Utils.logging.get_logger', 'log', ([], {}), '()\n', (997, 999), True, 'from Utils.logging import get_logger as log\n'), ((1388, 1393), 'Utils.logging.get_logger', 'log', (...
import gzip import sys import argparse import re import logging import numpy as np import pandas as p from itertools import product, tee from collections import Counter, OrderedDict from Bio import SeqIO def generate_feature_mapping(kmer_len): BASE_COMPLEMENT = {"A":"T","T":"A","G":"C","C":"G"} kmer_hash = ...
[ "gzip.open", "argparse.ArgumentParser", "Bio.SeqIO.parse", "numpy.zeros", "numpy.array", "itertools.product", "itertools.tee" ]
[((355, 387), 'itertools.product', 'product', (['"""ATGC"""'], {'repeat': 'kmer_len'}), "('ATGC', repeat=kmer_len)\n", (362, 387), False, 'from itertools import product, tee\n'), ((664, 675), 'itertools.tee', 'tee', (['seq', 'n'], {}), '(seq, n)\n', (667, 675), False, 'from itertools import product, tee\n'), ((971, 100...
# -*- coding: utf-8 -*- ''' <NAME> 1. a. Frequentist confidence intervals do not respect the physical limitations imposed on a system, ie non-negativity of a mass. b. Typically, that the probability to be found outside the interval on both sides of the distribution is 16% (or (100-CL)/2 %). Of...
[ "matplotlib.pyplot.title", "numpy.sum", "matplotlib.pyplot.clf", "numpy.polyfit", "matplotlib.pyplot.bar", "numpy.histogram", "numpy.exp", "numpy.random.normal", "numpy.linspace", "matplotlib.pyplot.errorbar", "math.isnan", "scipy.stats.chi2", "matplotlib.pyplot.ylabel", "numpy.log", "ma...
[((8154, 8178), 'numpy.linspace', 'np.linspace', (['(0)', '(5)', '(10000)'], {}), '(0, 5, 10000)\n', (8165, 8178), True, 'import numpy as np\n'), ((8204, 8236), 'numpy.exp', 'np.exp', (['(-theExponential.lookup_x)'], {}), '(-theExponential.lookup_x)\n', (8210, 8236), True, 'import numpy as np\n'), ((8448, 8457), 'matpl...
from fontTools.misc.fixedTools import floatToFixed from fontTools.ttLib import TTFont, newTable, registerCustomTableClass from fontTools.varLib.models import VariationModel, allEqual from fontTools.varLib.varStore import OnlineVarStoreBuilder from rcjktools.varco import VarCoFont from rcjktools.table_VarC import ( ...
[ "fontTools.varLib.models.VariationModel", "fontTools.ttLib.TTFont", "argparse.ArgumentParser", "fontTools.varLib.models.allEqual", "fontTools.ttLib.newTable", "fontTools.varLib.varStore.OnlineVarStoreBuilder", "pathlib.Path", "rcjktools.table_VarC.getToFixedConverterForNumIntBitsForScale", "fontTool...
[((615, 653), 'fontTools.varLib.models.VariationModel', 'VariationModel', (['allLocations', 'axisTags'], {}), '(allLocations, axisTags)\n', (629, 653), False, 'from fontTools.varLib.models import VariationModel, allEqual\n'), ((673, 704), 'fontTools.varLib.varStore.OnlineVarStoreBuilder', 'OnlineVarStoreBuilder', (['ax...
from unifi.objects.base import UnifiBaseObject from unifi.helper import find_by_attr, json_print class UnifiDeviceObject(UnifiBaseObject): def get_port_profile(self, **filter_kwargs): port = find_by_attr(self.port_table, **filter_kwargs) port_override = find_by_attr(self.port_overrides, port_idx=...
[ "unifi.helper.find_by_attr" ]
[((206, 252), 'unifi.helper.find_by_attr', 'find_by_attr', (['self.port_table'], {}), '(self.port_table, **filter_kwargs)\n', (218, 252), False, 'from unifi.helper import find_by_attr, json_print\n'), ((277, 337), 'unifi.helper.find_by_attr', 'find_by_attr', (['self.port_overrides'], {'port_idx': "port['port_idx']"}), ...
# Copyright 2019 The DMLab2D Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in...
[ "dmlab2d.runfiles_helper.find", "absl.testing.absltest.main", "numpy.dtype", "numpy.testing.assert_array_equal" ]
[((8160, 8175), 'absl.testing.absltest.main', 'absltest.main', ([], {}), '()\n', (8173, 8175), False, 'from absl.testing import absltest\n'), ((3413, 3470), 'numpy.testing.assert_array_equal', 'np.testing.assert_array_equal', (['observations[0]', '[1, 2, 3]'], {}), '(observations[0], [1, 2, 3])\n', (3442, 3470), True, ...
#!/usr/bin/env python # -*- coding: utf-8 -*- import os from json import dumps, loads # for python 2 # from httplib import HTTPConnection # for python 3 from http.client import HTTPConnection # connect with REST server connection = HTTPConnection('127.0.0.1', 80) connection.connect() data = {"ip": "192...
[ "http.client.HTTPConnection", "json.dumps" ]
[((245, 276), 'http.client.HTTPConnection', 'HTTPConnection', (['"""127.0.0.1"""', '(80)'], {}), "('127.0.0.1', 80)\n", (259, 276), False, 'from http.client import HTTPConnection\n'), ((489, 500), 'json.dumps', 'dumps', (['data'], {}), '(data)\n', (494, 500), False, 'from json import dumps, loads\n')]
# coding: utf-8 # Copyright (c) 2016, 2019, Oracle and/or its affiliates. All rights reserved. """ NOTE: This class should always comply to the API definition of NfsDatasetClient present in services/dts/src/oci_cli_dts/physical_appliance_control_plane/client/nfs_dataset_client.py """ from oci_cli import cli_util fr...
[ "oci_cli.cli_util.build_config", "services.dts.src.oci_cli_dts.appliance_config_manager.ApplianceConfigManager", "services.dts.src.oci_cli_dts.physical_appliance_control_plane.client.nfs_dataset_client.NfsDatasetClient" ]
[((769, 819), 'services.dts.src.oci_cli_dts.appliance_config_manager.ApplianceConfigManager', 'ApplianceConfigManager', (['APPLIANCE_CONFIGS_BASE_DIR'], {}), '(APPLIANCE_CONFIGS_BASE_DIR)\n', (791, 819), False, 'from services.dts.src.oci_cli_dts.appliance_config_manager import ApplianceConfigManager\n'), ((1077, 1107),...
from spreaduler import ParamsSheet from train_attention import train from options import get_parser class YourParamsSheet(ParamsSheet): """ Your model Params Sheet class """ params_sheet_id = '...' client_credentials = { "type": "service_account", "project_id": "....", "pr...
[ "options.get_parser" ]
[((1351, 1363), 'options.get_parser', 'get_parser', ([], {}), '()\n', (1361, 1363), False, 'from options import get_parser\n')]
import os import sys import time import torch import utils import logging import argparse import torch.nn as nn import torch.utils from adaptive_augmentor import AdaAug from networks import get_model from networks.projection import Projection from dataset import get_num_class, get_dataloaders, get_label_name, get_data...
[ "argparse.ArgumentParser", "utils.create_exp_dir", "dataset.get_dataloaders", "dataset.get_dataset_dimension", "time.strftime", "logging.Formatter", "dataset.get_num_class", "utils.AvgrageMeter", "torch.no_grad", "os.path.join", "utils.reproducibility", "config.get_warmup_config", "torch.cud...
[((433, 467), 'argparse.ArgumentParser', 'argparse.ArgumentParser', (['"""ada_aug"""'], {}), "('ada_aug')\n", (456, 467), False, 'import argparse\n'), ((3752, 3783), 'utils.create_exp_dir', 'utils.create_exp_dir', (['args.save'], {}), '(args.save)\n', (3772, 3783), False, 'import utils\n'), ((3823, 3934), 'logging.basi...
import tensorflow as tf sess = tf.Session() #在名字为foo的命名空间内创建名字为v的变量 with tf.variable_scope("foo"): #创建一个常量为1的v v= tf.get_variable('v1',[1],initializer = tf.constant_initializer(1.0)) #因为在foo空间已经创建v的变量,所以下面的代码会报错 #with tf.variable_scope("foo"): # v= tf.get_variable('v',[1]) #在生成上下文管理器时,将参数reuse设置为True。这样tf.g...
[ "tensorflow.constant_initializer", "tensorflow.Session", "tensorflow.variable_scope", "tensorflow.initialize_all_variables" ]
[((33, 45), 'tensorflow.Session', 'tf.Session', ([], {}), '()\n', (43, 45), True, 'import tensorflow as tf\n'), ((75, 99), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""foo"""'], {}), "('foo')\n", (92, 99), True, 'import tensorflow as tf\n'), ((437, 482), 'tensorflow.variable_scope', 'tf.variable_scope', (['"...
from collections import namedtuple from typing import Dict, List, Callable Node = namedtuple('Node', 'name parent children data') def make_tree_from_adj_list(adj_list): root = 'COM' nodes: Dict['str', Node] = {root: Node(root, None, [], {})} for parent, child in adj_list: node = Node(child, pare...
[ "collections.namedtuple" ]
[((83, 130), 'collections.namedtuple', 'namedtuple', (['"""Node"""', '"""name parent children data"""'], {}), "('Node', 'name parent children data')\n", (93, 130), False, 'from collections import namedtuple\n')]
from iotbx import mtz mtz_obj = mtz.object(file_name="3nd4.mtz") # Only works with mtz.object. # Does not work if mtz is read in with iotbx.file_reader. miller_arrays_dict = mtz_obj.as_miller_arrays_dict()
[ "iotbx.mtz.object" ]
[((32, 64), 'iotbx.mtz.object', 'mtz.object', ([], {'file_name': '"""3nd4.mtz"""'}), "(file_name='3nd4.mtz')\n", (42, 64), False, 'from iotbx import mtz\n')]
""" Copyright (c) 2019 Intel Corporation Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writin...
[ "torch.ones", "functools.partial", "torch.IntTensor", "collections.namedtuple", "torch.zeros", "torch.distributed.broadcast", "logging.getLogger" ]
[((1000, 1027), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1017, 1027), False, 'import logging\n'), ((1314, 1409), 'collections.namedtuple', 'namedtuple', (['"""QuantizationParams"""', "['bits', 'mode', 'signed', 'signed_scope', 'per_channel']"], {}), "('QuantizationParams', ['bits',...
import curses, sys, os #Servo controller connected to IC2 import Adafruit_PCA9685 pwm = Adafruit_PCA9685.PCA9685() pwm.set_pwm_freq(60) from time import sleep #ESC Brushles motor states: direction, on/off toggleState = 400 throttle = 450 delta = 20 print("toggleState1") pwm.set_pwm(2,0,toggleState) sleep(0.2) for i...
[ "Adafruit_PCA9685.PCA9685", "time.sleep" ]
[((89, 115), 'Adafruit_PCA9685.PCA9685', 'Adafruit_PCA9685.PCA9685', ([], {}), '()\n', (113, 115), False, 'import Adafruit_PCA9685\n'), ((304, 314), 'time.sleep', 'sleep', (['(0.2)'], {}), '(0.2)\n', (309, 314), False, 'from time import sleep\n'), ((494, 504), 'time.sleep', 'sleep', (['(0.4)'], {}), '(0.4)\n', (499, 50...
import pytropos.internals.values as pv from pytropos.internals.values.builtin_values import * from pytropos.internals.values.python_values.builtin_mutvalues import * from pytropos.internals.values.python_values.wrappers import * from pytropos.internals.values.python_values.python_values import PythonValue, PT exitcode...
[ "pytropos.internals.values.int", "pytropos.internals.values.python_values.python_values.PythonValue" ]
[((380, 399), 'pytropos.internals.values.python_values.python_values.PythonValue', 'PythonValue', (['PT.Top'], {}), '(PT.Top)\n', (391, 399), False, 'from pytropos.internals.values.python_values.python_values import PythonValue, PT\n'), ((440, 454), 'pytropos.internals.values.python_values.python_values.PythonValue', '...
import re import setuptools def find_version(fname): """Attempts to find the version number in the file names fname. Raises RuntimeError if not found. """ version = '' with open(fname, 'r') as fp: reg = re.compile(r'__version__ = [\'"]([^\'"]*)[\'"]') for line in fp: m ...
[ "setuptools.find_packages", "re.compile" ]
[((233, 286), 're.compile', 're.compile', (['"""__version__ = [\\\\\'"]([^\\\\\'"]*)[\\\\\'"]"""'], {}), '(\'__version__ = [\\\\\\\'"]([^\\\\\\\'"]*)[\\\\\\\'"]\')\n', (243, 286), False, 'import re\n'), ((990, 1016), 'setuptools.find_packages', 'setuptools.find_packages', ([], {}), '()\n', (1014, 1016), False, 'import ...
#!/usr/bin/python3 import os.path import openpyxl import requests import json import argparse BASE_URL_XIV_API_CHARACTER: str = "https://xivapi.com/character/" GERMAN_TO_ENGLISH_CLASS_DICT: dict = {} SUB_30_MAPPING_DICT: dict = {} CONFIG_LOCATION = os.getcwd() DEBUG_ENABLED = False def main(filepath): """main ...
[ "argparse.ArgumentParser", "json.load", "openpyxl.load_workbook", "requests.get" ]
[((417, 449), 'openpyxl.load_workbook', 'openpyxl.load_workbook', (['filepath'], {}), '(filepath)\n', (439, 449), False, 'import openpyxl\n'), ((3683, 3708), 'requests.get', 'requests.get', (['request_url'], {}), '(request_url)\n', (3695, 3708), False, 'import requests\n'), ((5160, 5232), 'argparse.ArgumentParser', 'ar...
import datetime from functools import partial from typing import List, Optional, Union from quickbuild.helpers import ContentType, response2py class Configurations: def __init__(self, quickbuild): self.quickbuild = quickbuild def _get(self, params: dict) -> List[dict]: return self.quickbui...
[ "functools.partial" ]
[((2333, 2380), 'functools.partial', 'partial', (['response2py'], {'content_type': 'content_type'}), '(response2py, content_type=content_type)\n', (2340, 2380), False, 'from functools import partial\n')]
from os import path from pathlib import Path def curr_file_path() -> Path: """Get cuurent file path.""" return Path(__file__).absolute() def out_folder_path() -> Path: """Get output folder path.""" return curr_file_path().parents[3].joinpath("out").absolute() def out_geom_path() -> Path: """Ge...
[ "pathlib.Path" ]
[((121, 135), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (125, 135), False, 'from pathlib import Path\n')]
#!/usr/bin/env python # -*- coding: utf-8 -*- import requests, json, base64 def post_cli(auth, command): url_cli = "http://" + auth.ipaddr + "/rest/" + auth.version + "/cli" command_dict = {"cmd": command} try: post_command = requests.post(url_cli, headers=auth.cookie, data=json.dumps(command_dict...
[ "base64.b64decode", "json.dumps" ]
[((297, 321), 'json.dumps', 'json.dumps', (['command_dict'], {}), '(command_dict)\n', (307, 321), False, 'import requests, json, base64\n'), ((418, 448), 'base64.b64decode', 'base64.b64decode', (['cli_response'], {}), '(cli_response)\n', (434, 448), False, 'import requests, json, base64\n')]
from model.group import Group class GroupHelper: def __init__(self, app): self.app = app def create(self, group): wd = self.app.wd self.go_to_group_page() wd.find_element_by_name("new").click() self.fill_form_group(group) # Submit group creation wd.fin...
[ "model.group.Group" ]
[((3260, 3289), 'model.group.Group', 'Group', ([], {'name': 'text', 'id': 'group_id'}), '(name=text, id=group_id)\n', (3265, 3289), False, 'from model.group import Group\n')]
import math def sieve(n): primes = list(range(2, n+1)) i = 0 while i < len(primes): no = primes[i] m = 2 while (no * m) <= max(primes): if primes.count(no * m) > 0: primes.remove(no * m) m+=1 i+=1 return primes def maxPower(n, ...
[ "math.pow" ]
[((348, 366), 'math.pow', 'math.pow', (['n', '(i + 1)'], {}), '(n, i + 1)\n', (356, 366), False, 'import math\n')]
from os import system def ler_qtd(n, msg): n = int(input(msg)) while (n < 1) or (n > 10000): n = int(input(f' - Entrada invalida!{msg}')) return n def preencher_set_cartas(cartas, qtd, p): """ set de cartas, qtd de cartas, p de pessoa """ from time import sleep print()...
[ "os.system", "time.sleep" ]
[((1764, 1777), 'os.system', 'system', (['"""cls"""'], {}), "('cls')\n", (1770, 1777), False, 'from os import system\n'), ((641, 649), 'time.sleep', 'sleep', (['(1)'], {}), '(1)\n', (646, 649), False, 'from time import sleep\n')]
import numpy as np import random import numbers import cv2 from PIL import Image import wpcv from wpcv.utils.ops import pil_ops, polygon_ops from wpcv.utils.data_aug.base import Compose, Zip from wpcv.utils.data_aug import img_aug class ToPILImage(object): def __init__(self): self.to = img_aug.ToPILImage(...
[ "wpcv.utils.ops.pil_ops.vflip", "wpcv.utils.ops.polygon_ops.get_translate_range", "wpcv.utils.ops.pil_ops.resize_keep_ratio", "wpcv.utils.ops.polygon_ops.scale", "random.randint", "wpcv.utils.ops.pil_ops.scale", "wpcv.utils.ops.polygon_ops.translate", "wpcv.utils.data_aug.img_aug.ToPILImage", "wpcv....
[((301, 321), 'wpcv.utils.data_aug.img_aug.ToPILImage', 'img_aug.ToPILImage', ([], {}), '()\n', (319, 321), False, 'from wpcv.utils.data_aug import img_aug\n'), ((1855, 1891), 'wpcv.utils.ops.pil_ops.scale', 'pil_ops.scale', (['img', '(scaleX, scaleY)'], {}), '(img, (scaleX, scaleY))\n', (1868, 1891), False, 'from wpcv...
# coding: utf-8 """ IbIocProfile.py The Clear BSD License Copyright (c) – 2016, NetApp, Inc. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted (subject to the limitations in the disclaimer below) provided that the following conditions are met: *...
[ "six.iteritems" ]
[((18511, 18540), 'six.iteritems', 'iteritems', (['self.swagger_types'], {}), '(self.swagger_types)\n', (18520, 18540), False, 'from six import iteritems\n')]
from csv import reader from sklearn import preprocessing from plotly import graph_objects def import_data(path): return [[float(f) for f in r] for r in reader(open(path, "r"))] def normalize_data(dataset): scaler = preprocessing.MinMaxScaler(feature_range=(0,1)) normalized = scaler.fit_transform(dataset...
[ "plotly.graph_objects.Figure", "sklearn.preprocessing.MinMaxScaler" ]
[((4307, 4341), 'plotly.graph_objects.Figure', 'graph_objects.Figure', ([], {'data': '[table]'}), '(data=[table])\n', (4327, 4341), False, 'from plotly import graph_objects\n'), ((227, 275), 'sklearn.preprocessing.MinMaxScaler', 'preprocessing.MinMaxScaler', ([], {'feature_range': '(0, 1)'}), '(feature_range=(0, 1))\n'...
#!/bin/python3 from selenium import webdriver from selenium.webdriver.support.ui import WebDriverWait from selenium.webdriver.common.by import By from selenium.webdriver.support import expected_conditions as EC from selenium.common.exceptions import NoSuchElementException import sys,os options = webdriver.ChromeOptio...
[ "selenium.webdriver.support.expected_conditions.presence_of_element_located", "selenium.webdriver.ChromeOptions", "selenium.webdriver.Chrome", "selenium.webdriver.support.ui.WebDriverWait", "sys.exit" ]
[((299, 324), 'selenium.webdriver.ChromeOptions', 'webdriver.ChromeOptions', ([], {}), '()\n', (322, 324), False, 'from selenium import webdriver\n'), ((499, 532), 'selenium.webdriver.Chrome', 'webdriver.Chrome', ([], {'options': 'options'}), '(options=options)\n', (515, 532), False, 'from selenium import webdriver\n')...
# Generated by Django 2.2.4 on 2019-09-11 14:12 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('users', '0004_auto_20190907_1334'), ] operations = [ migrations.AddField( model_name='pic', name='classification152'...
[ "django.db.models.ImageField", "django.db.models.TextField" ]
[((340, 368), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)'}), '(blank=True)\n', (356, 368), False, 'from django.db import migrations, models\n'), ((495, 523), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)'}), '(blank=True)\n', (511, 523), False, 'from django.db im...
# Copyright 2019 DeepMind Technologies Ltd. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
[ "open_spiel.python.mfg.algorithms.distribution.DistributionPolicy", "open_spiel.python.mfg.algorithms.EGTA.init_oracle.init_br_oracle", "open_spiel.python.mfg.algorithms.EGTA.inner_loop.InnerLoop" ]
[((2652, 2700), 'open_spiel.python.mfg.algorithms.EGTA.inner_loop.InnerLoop', 'inner_loop.InnerLoop', (['self._meta_strategy_method'], {}), '(self._meta_strategy_method)\n', (2672, 2700), False, 'from open_spiel.python.mfg.algorithms.EGTA import inner_loop\n'), ((4084, 4132), 'open_spiel.python.mfg.algorithms.EGTA.inne...
# encoding:utf-8 import os, sys basepath = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(os.path.join(basepath, 'FaceDetector')) import requests import base64 import cv2 import numpy as np import urllib.request import base64 def fetchImageFromHttp(image_url, timeout_s=1): # 该函数是读取url...
[ "os.path.abspath", "cv2.imwrite", "cv2.imdecode", "cv2.imread", "requests.get", "cv2.rectangle", "requests.post", "os.path.join" ]
[((119, 157), 'os.path.join', 'os.path.join', (['basepath', '"""FaceDetector"""'], {}), "(basepath, 'FaceDetector')\n", (131, 157), False, 'import os, sys\n'), ((886, 904), 'requests.get', 'requests.get', (['host'], {}), '(host)\n', (898, 904), False, 'import requests\n'), ((1739, 1795), 'requests.post', 'requests.post...
# # INF 552 Homework 3 # Part 2: Fast Map # Group Members: <NAME> (zhan198), <NAME> (minyihua), <NAME> (jeffyjac) # Date: 2/27/2018 # Programming Language: Python 3.6 # import numpy as np import matplotlib.pyplot as plt DIMENSION = 2 DATA_SIZE = 10 # WORDS = ["acting", "activist", "compute", "coward","forward","in...
[ "matplotlib.pyplot.title", "matplotlib.pyplot.show", "matplotlib.pyplot.scatter", "numpy.power", "numpy.zeros", "numpy.amax", "numpy.where", "matplotlib.pyplot.subplots" ]
[((470, 508), 'numpy.zeros', 'np.zeros', ([], {'shape': '(DATA_SIZE, DATA_SIZE)'}), '(shape=(DATA_SIZE, DATA_SIZE))\n', (478, 508), True, 'import numpy as np\n'), ((516, 554), 'numpy.zeros', 'np.zeros', ([], {'shape': '(DATA_SIZE, DIMENSION)'}), '(shape=(DATA_SIZE, DIMENSION))\n', (524, 554), True, 'import numpy as np\...
#!/usr/bin/python # -*- coding: utf-8 -*- __author__ = 'ar' import os import glob from app.backend.core.utils import getDirectorySizeInBytes, humanReadableSize if __name__ == '__main__': path='../../../data/datasets' for ii,pp in enumerate(glob.glob('%s/*' % path)): tbn=os.path.basename(pp) ts...
[ "app.backend.core.utils.getDirectorySizeInBytes", "os.path.basename", "app.backend.core.utils.humanReadableSize", "glob.glob" ]
[((250, 274), 'glob.glob', 'glob.glob', (["('%s/*' % path)"], {}), "('%s/*' % path)\n", (259, 274), False, 'import glob\n'), ((289, 309), 'os.path.basename', 'os.path.basename', (['pp'], {}), '(pp)\n', (305, 309), False, 'import os\n'), ((326, 353), 'app.backend.core.utils.getDirectorySizeInBytes', 'getDirectorySizeInB...
# # Copyright (c) 2019 UAVCAN Development Team # This software is distributed under the terms of the MIT License. # Author: <NAME> <<EMAIL>> # import pytest import subprocess from ._subprocess import run_cli_tool def _unittest_trivial() -> None: run_cli_tool('show-transport', timeout=2.0) with pytest.raises...
[ "pytest.raises" ]
[((307, 351), 'pytest.raises', 'pytest.raises', (['subprocess.CalledProcessError'], {}), '(subprocess.CalledProcessError)\n', (320, 351), False, 'import pytest\n'), ((397, 441), 'pytest.raises', 'pytest.raises', (['subprocess.CalledProcessError'], {}), '(subprocess.CalledProcessError)\n', (410, 441), False, 'import pyt...
import time import numpy as np import torch import torch.nn as nn import torch.optim as optim import queue from insomnia.utils import empty_torch_queue from insomnia.explores.gaussian_noise import GaussianActionNoise from insomnia.numeric_models import d4pg from insomnia.numeric_models.misc import l2_projection cla...
[ "torch.from_numpy", "numpy.abs", "torch.nn.BCELoss", "insomnia.utils.empty_torch_queue", "insomnia.numeric_models.d4pg.CriticNetwork", "numpy.asarray", "numpy.zeros", "time.time", "torch.cuda.is_available", "torch.sum", "torch.tensor", "insomnia.numeric_models.misc.l2_projection._l2_project" ]
[((1153, 1270), 'insomnia.numeric_models.d4pg.CriticNetwork', 'd4pg.CriticNetwork', (['beta', 'input_dims', 'fc1_dims', 'fc2_dims', 'n_actions', 'name', 'self.v_min', 'self.v_max', 'self.num_atoms'], {}), '(beta, input_dims, fc1_dims, fc2_dims, n_actions, name,\n self.v_min, self.v_max, self.num_atoms)\n', (1171, 12...
import logging from wrapper import * logger = logging.getLogger(__name__) # noinspection PyUnusedLocal def get_user(url='', key='', timeout=60, **kwargs): return get(url + '/user', headers={'Authorization': "Bearer " + key}, timeout=timeout).json() # noinspection PyUnusedLocal def get_user_tokens(url='', key='...
[ "logging.getLogger" ]
[((47, 74), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (64, 74), False, 'import logging\n')]
from __future__ import absolute_import # Need to import lwr_client absolutely. from ..objectstore import ObjectStore try: from galaxy.jobs.runners.lwr_client.manager import ObjectStoreClientManager except ImportError: from lwr.lwr_client.manager import ObjectStoreClientManager class LwrObjectStore(ObjectStor...
[ "lwr.lwr_client.manager.ObjectStoreClientManager" ]
[((2814, 2857), 'lwr.lwr_client.manager.ObjectStoreClientManager', 'ObjectStoreClientManager', ([], {}), '(**manager_options)\n', (2838, 2857), False, 'from lwr.lwr_client.manager import ObjectStoreClientManager\n')]
""" Demo of json_required decorator for API input validation/error handling """ import inspect import functools import json from traceback import format_exception from flask import jsonify, request import sys from flask.exceptions import JSONBadRequest from flask import Flask import re app = Flask(__name__) def...
[ "functools.partial", "traceback.format_exception", "json.loads", "flask.Flask", "json.dumps", "inspect.getargspec", "functools.wraps", "sys.exc_info", "flask.request.json.get", "re.compile" ]
[((300, 315), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (305, 315), False, 'from flask import Flask\n'), ((3436, 3619), 're.compile', 're.compile', (['"""[A-Za-z0-9!#$%&\'*+/=?^_`{|}~-]+(?:\\\\.[A-Za-z0-9!#$%&\'*+/=?^_`{|}~-]+)*@(?:[A-Za-z0-9](?:[A-Za-z0-9-]*[A-Za-z0-9])?\\\\.)+[A-Za-z0-9](?:[A-Za-z0-...
import torch from ..math.cross import * from ..math.normvec import * class CameraExtrinsic(object): """ A class representing the camera extrinsic properties Attributes ---------- position : Tensor the camera position target : Tensor the camera target up_vector : Tensor ...
[ "torch.tensor" ]
[((1340, 1396), 'torch.tensor', 'torch.tensor', (['position'], {'dtype': 'torch.float', 'device': 'device'}), '(position, dtype=torch.float, device=device)\n', (1352, 1396), False, 'import torch\n'), ((1423, 1477), 'torch.tensor', 'torch.tensor', (['target'], {'dtype': 'torch.float', 'device': 'device'}), '(target, dty...
import uuid from textwrap import dedent from IPython.core.display import display, HTML from string import Template import numpy as np # function to initialize a scatter plot def init_chart(data,features): chart_id = 'mychart-' + str(uuid.uuid4()) feature_types = {} # map each feature to type num_feature_ra...
[ "IPython.core.display.HTML", "uuid.uuid4", "textwrap.dedent" ]
[((782, 853), 'IPython.core.display.HTML', 'HTML', (['"""<script src="/static/components/requirejs/require.js"></script>"""'], {}), '(\'<script src="/static/components/requirejs/require.js"></script>\')\n', (786, 853), False, 'from IPython.core.display import display, HTML\n'), ((238, 250), 'uuid.uuid4', 'uuid.uuid4', ...
############################################################################## # Copyright (c) 2017 Huawei Technologies Co.,Ltd and others. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, ...
[ "yardstick.common.openstack_utils.create_image", "yardstick.common.openstack_utils.get_shade_client", "logging.getLogger" ]
[((619, 646), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (636, 646), False, 'import logging\n'), ((1721, 1755), 'yardstick.common.openstack_utils.get_shade_client', 'openstack_utils.get_shade_client', ([], {}), '()\n', (1753, 1755), False, 'from yardstick.common import openstack_utils...
#!/usr/bin/env python # # Copyright © 2022 Github Lzhiyong # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
[ "subprocess.run", "pathlib.Path", "argparse.ArgumentParser", "time.time" ]
[((1459, 1482), 'subprocess.run', 'subprocess.run', (['command'], {}), '(command)\n', (1473, 1482), False, 'import subprocess\n'), ((1495, 1506), 'time.time', 'time.time', ([], {}), '()\n', (1504, 1506), False, 'import time\n'), ((1965, 1979), 'pathlib.Path', 'Path', (['args.ndk'], {}), '(args.ndk)\n', (1969, 1979), Fa...
import pathlib from datetime import timedelta from airflow import DAG from airflow.operators.python import PythonOperator import pandas as pd from sklearn.compose import ColumnTransformer from sklearn.pipeline import Pipeline from sklearn.preprocessing import OneHotEncoder from sklearn.preprocessing import Stan...
[ "json.dump", "pickle.dump", "sklearn.preprocessing.StandardScaler", "airflow.operators.python.PythonOperator", "pandas.read_csv", "sklearn.model_selection.train_test_split", "sklearn.metrics.accuracy_score", "sklearn.preprocessing.OneHotEncoder", "sklearn.metrics.roc_auc_score", "sklearn.linear_mo...
[((621, 675), 'pandas.read_csv', 'pd.read_csv', (['"""/opt/airflow/data/raw/{{ ds }}/data.csv"""'], {}), "('/opt/airflow/data/raw/{{ ds }}/data.csv')\n", (632, 675), True, 'import pandas as pd\n'), ((693, 749), 'pandas.read_csv', 'pd.read_csv', (['"""/opt/airflow/data/raw/{{ ds }}/target.csv"""'], {}), "('/opt/airflow/...