code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
from flask import Flask, flash
def display_error(error,preface="",postface=""):
flash(f"{preface} {error} {postface}") | [
"flask.flash"
] | [((85, 123), 'flask.flash', 'flash', (['f"""{preface} {error} {postface}"""'], {}), "(f'{preface} {error} {postface}')\n", (90, 123), False, 'from flask import Flask, flash\n')] |
import sys
import typing
from collections import deque
from typing import Callable, Optional
import mypy.errorcodes
import mypy.errors
import mypy.nodes
import mypy.options
import mypy.plugin
import mypy.types
if sys.version_info >= (3, 10): # pragma: no cover
from typing import TypeGuard
else: # pragma: no cov... | [
"collections.deque"
] | [((3972, 3986), 'collections.deque', 'deque', (['[type_]'], {}), '([type_])\n', (3977, 3986), False, 'from collections import deque\n')] |
# -*- coding: utf-8 -*-
# Generated by Django 1.10.7 on 2017-07-18 04:43
from __future__ import unicode_literals
from django.db import migrations, models
def update_names(apps, schema_editor):
for x in apps.get_model('institutions', 'regon').objects.exclude(data=None).iterator():
x.name = x.data.get('naz... | [
"django.db.migrations.RunPython",
"django.db.models.CharField"
] | [((1079, 1113), 'django.db.migrations.RunPython', 'migrations.RunPython', (['update_names'], {}), '(update_names)\n', (1099, 1113), False, 'from django.db import migrations, models\n'), ((740, 805), 'django.db.models.CharField', 'models.CharField', ([], {'default': '""""""', 'max_length': '(200)', 'verbose_name': '"""N... |
# coding: utf-8
# pylint: disable=missing-docstring, invalid-name
import flask
import auth
import config
from main import app
import model.user as user #import User#, UserVdr
github_config = dict(
access_token_method='POST',
access_token_url='https://github.com/login/oauth/access_token',
authorize_url='ht... | [
"flask.flash",
"auth.signin_oauth",
"flask.session.get",
"flask.url_for",
"main.app.route",
"auth.signin_via_social",
"auth.create_oauth_app"
] | [((574, 620), 'auth.create_oauth_app', 'auth.create_oauth_app', (['github_config', '"""github"""'], {}), "(github_config, 'github')\n", (595, 620), False, 'import auth\n'), ((624, 674), 'main.app.route', 'app.route', (['"""/_s/callback/github/oauth-authorized/"""'], {}), "('/_s/callback/github/oauth-authorized/')\n", (... |
from pages.service_page.models import ServicePage
from pages.topic_page.factories import JanisBasePageWithTopicsFactory
from pages.base_page.fixtures.helpers.streamfieldify import streamfieldify
class ServicePageFactory(JanisBasePageWithTopicsFactory):
@classmethod
def create(cls, *args, **kwargs):
if... | [
"pages.base_page.fixtures.helpers.streamfieldify.streamfieldify"
] | [((390, 431), 'pages.base_page.fixtures.helpers.streamfieldify.streamfieldify', 'streamfieldify', (["kwargs['dynamic_content']"], {}), "(kwargs['dynamic_content'])\n", (404, 431), False, 'from pages.base_page.fixtures.helpers.streamfieldify import streamfieldify\n'), ((600, 636), 'pages.base_page.fixtures.helpers.strea... |
# (c) Copyright [2017] Hewlett Packard Enterprise Development LP
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appli... | [
"os.remove",
"os.walk",
"json.dumps",
"os.path.isfile",
"multiprocessing.Queue",
"os.path.join",
"logging.error",
"json.loads",
"os.path.dirname",
"os.path.exists",
"json.dump",
"copy.deepcopy",
"subprocess.Popen",
"importlib.import_module",
"os.listdir",
"re.compile",
"json.load",
... | [((2094, 2120), 'os.path.dirname', 'os.path.dirname', (['file_name'], {}), '(file_name)\n', (2109, 2120), False, 'import os\n'), ((12416, 12435), 're.compile', 're.compile', (['pattern'], {}), '(pattern)\n', (12426, 12435), False, 'import re\n'), ((25162, 25183), 'copy.deepcopy', 'copy.deepcopy', (['config'], {}), '(co... |
import json
import os
import requests
from client import Submission
from settings import API_BASE_URL, ACCESS_TOKEN
def get_task_url(task_id: int):
return API_BASE_URL + f"/tasks/{task_id}/download_grader/"
def get_agent_url(submission_id: int):
return API_BASE_URL + f"/submissions/{submission_id}/downloa... | [
"os.getenv",
"json.loads",
"requests.get"
] | [((510, 683), 'requests.get', 'requests.get', (["(API_BASE_URL + f'/jobs/{job_id}/start_job/')"], {'headers': "{'Authorization': f'Token {ACCESS_TOKEN}'}", 'data': "{'worker_name': worker_name, 'task_id': task_id}"}), "(API_BASE_URL + f'/jobs/{job_id}/start_job/', headers={\n 'Authorization': f'Token {ACCESS_TOKEN}'... |
import time
import logging
from extended_networkx_tools import Analytics, AnalyticsGraph
from timeit import default_timer as timer
from utils import Solvers
from utils.GraphUtils import GraphUtils
from utils.ServerUtil import ServerUtil
from datetime import datetime
class GraphThread:
@staticmethod
def sta... | [
"utils.Solvers.Diff.solve",
"utils.GraphUtils.GraphUtils.get_results",
"logging.exception",
"utils.Solvers.Random.solve",
"timeit.default_timer",
"utils.Solvers.Field.solve",
"time.sleep",
"utils.ServerUtil.ServerUtil",
"utils.Solvers.Spec.solve",
"datetime.datetime.now"
] | [((1097, 1117), 'utils.ServerUtil.ServerUtil', 'ServerUtil', (['base_url'], {}), '(base_url)\n', (1107, 1117), False, 'from utils.ServerUtil import ServerUtil\n'), ((1424, 1431), 'timeit.default_timer', 'timer', ([], {}), '()\n', (1429, 1431), True, 'from timeit import default_timer as timer\n'), ((1512, 1519), 'timeit... |
from django import forms
SACCO_DRIVER_STATUS_OPTIONS = [
('Approved', ('Approved to operate')),
('Suspended', ('Suspended for the time being')),
('Blacklisted', ('Blacklisted from operating'))
]
class VehicleForm(forms.Form):
# sacco = forms.CharField(label="Sacco", max_length=100)
regno = forms.... | [
"django.forms.CharField",
"django.forms.Select"
] | [((314, 372), 'django.forms.CharField', 'forms.CharField', ([], {'label': '"""Registration Number"""', 'max_length': '(7)'}), "(label='Registration Number', max_length=7)\n", (329, 372), False, 'from django import forms\n'), ((658, 687), 'django.forms.CharField', 'forms.CharField', ([], {'max_length': '(8)'}), '(max_le... |
from typing import Optional
from talon import Context
from user.emacs.utils.voicemacs import rpc_call
from user.utils.formatting import SurroundingText
context = Context()
context.matches = r"""
tag: user.emacs
"""
@context.action_class("self")
class UserActions:
def surrounding_text() -> Optional[Surroundin... | [
"talon.Context",
"user.emacs.utils.voicemacs.rpc_call",
"user.utils.formatting.SurroundingText"
] | [((166, 175), 'talon.Context', 'Context', ([], {}), '()\n', (173, 175), False, 'from talon import Context\n'), ((416, 519), 'user.emacs.utils.voicemacs.rpc_call', 'rpc_call', (['"""voicemacs-surrounding-text"""', "[':chars-before', 30000, ':chars-after', 30000]"], {'timeout': '(10)'}), "('voicemacs-surrounding-text', [... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import sys
import zipfile
def make_nvz_main(output_file, nvm_file, target_file, pitch_file=None):
if pitch_file is not None:
files = [nvm_file, target_file, pitch_file]
arc_names = ['target.nvm', 'target.pb', 'pitch.pb']
else:
... | [
"zipfile.ZipFile"
] | [((417, 484), 'zipfile.ZipFile', 'zipfile.ZipFile', (['output_file', '"""w"""'], {'compression': 'zipfile.ZIP_DEFLATED'}), "(output_file, 'w', compression=zipfile.ZIP_DEFLATED)\n", (432, 484), False, 'import zipfile\n')] |
import click
from mutacc.mutaccDB.remove_case import remove_case_from_db
@click.command('remove')
@click.argument('case_id')
@click.pass_context
def remove_command(context, case_id):
"""
Deletes case from mutacc DB
"""
adapter = context.obj['adapter']
remove_case_from_db(adapter, case_id)
| [
"mutacc.mutaccDB.remove_case.remove_case_from_db",
"click.argument",
"click.command"
] | [((76, 99), 'click.command', 'click.command', (['"""remove"""'], {}), "('remove')\n", (89, 99), False, 'import click\n'), ((101, 126), 'click.argument', 'click.argument', (['"""case_id"""'], {}), "('case_id')\n", (115, 126), False, 'import click\n'), ((280, 317), 'mutacc.mutaccDB.remove_case.remove_case_from_db', 'remo... |
# encoding: utf-8
# http://www.hexblog.com/?p=120
# Default IDA Pro Paths:
# MAC /Applications/IDA\ Pro\ X/idaq.app/Contents/MacOS/plugins/
# Windows C:\Program Files (x86)\IDA X\plugins
# to make it autoexec on openfile
# add this to plugins.cfg
# ; Other plugins
#FullColor FullColor.py 0... | [
"idaapi.msg",
"idc.set_color",
"idc.get_screen_ea",
"idc.print_operand",
"idc.print_insn_mnem"
] | [((952, 995), 'idaapi.msg', 'idaapi.msg', (["('run() called with %d!\\n' % arg)"], {}), "('run() called with %d!\\n' % arg)\n", (962, 995), False, 'import idaapi\n'), ((2142, 2172), 'idaapi.msg', 'idaapi.msg', (['"""term() called!\n"""'], {}), "('term() called!\\n')\n", (2152, 2172), False, 'import idaapi\n'), ((1766, ... |
import os
import random
import subprocess
MYPATH = './out-of-dropbox-2020-08to12-'
FILES = os.listdir(MYPATH)
INP = ''
while INP != 'q':
INP = input('q to quit, enter anything else to continue')
file_choice = random.choice(FILES)
pathname_choice = MYPATH + '/' + file_choice
subprocess.run(["open", path... | [
"subprocess.run",
"random.choice",
"os.listdir"
] | [((92, 110), 'os.listdir', 'os.listdir', (['MYPATH'], {}), '(MYPATH)\n', (102, 110), False, 'import os\n'), ((218, 238), 'random.choice', 'random.choice', (['FILES'], {}), '(FILES)\n', (231, 238), False, 'import random\n'), ((292, 333), 'subprocess.run', 'subprocess.run', (["['open', pathname_choice]"], {}), "(['open',... |
import argparse
import json
from multiprocessing import Pool
from puyotable.canonization import canonize_deals
def all_deals(num_deals, num_colors):
if not num_deals:
return [[]]
result = []
for c0 in range(num_colors):
for c1 in range(num_colors):
for deals in all_deals(num_d... | [
"puyotable.canonization.canonize_deals",
"argparse.ArgumentParser",
"multiprocessing.Pool"
] | [((1216, 1304), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Tabulate all opening sequences in Puyo Puyo."""'}), "(description=\n 'Tabulate all opening sequences in Puyo Puyo.')\n", (1239, 1304), False, 'import argparse\n'), ((1988, 1994), 'multiprocessing.Pool', 'Pool', ([], {}), '... |
import matplotlib.pyplot as plt
import numpy as np
import numpy.polynomial.polynomial as nppol
class Metawalk:
def __init__(self,
time_intervals=None,
nodes=None,
):
"""
A basic constructor for a ``Metwalks`` object
:param times : A list o... | [
"matplotlib.pyplot.tight_layout",
"matplotlib.pyplot.plot",
"numpy.polynomial.polynomial.Polynomial",
"matplotlib.pyplot.vlines",
"numpy.around",
"numpy.math.factorial",
"matplotlib.pyplot.gca",
"matplotlib.pyplot.hlines",
"matplotlib.pyplot.subplots"
] | [((4201, 4222), 'numpy.polynomial.polynomial.Polynomial', 'nppol.Polynomial', (['res'], {}), '(res)\n', (4217, 4222), True, 'import numpy.polynomial.polynomial as nppol\n'), ((7457, 7567), 'matplotlib.pyplot.plot', 'plt.plot', (['[self.time_intervals[0]]', '[id_source]'], {'color': 'color', 'marker': '"""o"""', 'alpha'... |
from django.db import models
from django.core.validators import MaxValueValidator, MinValueValidator
from django.db.models.signals import post_delete
from django.dispatch import receiver
from myauth import models as myauth_models
from products.models import Book
class CommentProducts(models.Model):
profile = mod... | [
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.core.validators.MinValueValidator",
"django.dispatch.receiver",
"django.db.models.DateTimeField",
"django.core.validators.MaxValueValidator"
] | [((1334, 1379), 'django.dispatch.receiver', 'receiver', (['post_delete'], {'sender': 'CommentProducts'}), '(post_delete, sender=CommentProducts)\n', (1342, 1379), False, 'from django.dispatch import receiver\n'), ((317, 412), 'django.db.models.ForeignKey', 'models.ForeignKey', (['myauth_models.Profile'], {'on_delete': ... |
from urllib.parse import urlparse, urlunparse
from functools import wraps
from flask import abort, request, current_app
from lib.twilio import TwilioClient
def validate_twilio_request(f):
"""Validates that incoming requests genuinely originated from Twilio"""
# Adapted from https://www.twilio.com/docs/usage/... | [
"flask.request.headers.get",
"urllib.parse.urlunparse",
"flask.abort",
"functools.wraps",
"lib.twilio.TwilioClient",
"urllib.parse.urlparse"
] | [((524, 532), 'functools.wraps', 'wraps', (['f'], {}), '(f)\n', (529, 532), False, 'from functools import wraps\n'), ((602, 718), 'lib.twilio.TwilioClient', 'TwilioClient', (["current_app.config['SECRETS'].TWILIO_ACCOUNT_SID", "current_app.config['SECRETS'].TWILIO_AUTH_TOKEN"], {}), "(current_app.config['SECRETS'].TWIL... |
#!/usr/bin/env python3
import os
from typing import Union
from pathlib import Path
import requests
from rpicam.utils.logging_utils import get_logger
class TelegramPoster:
"""
Bare-bones class to post videos to a Telegram chat.
Uses per default credentials stored in environment.
"""
API_URL = '... | [
"requests.post",
"rpicam.utils.logging_utils.get_logger",
"os.getenv"
] | [((806, 852), 'rpicam.utils.logging_utils.get_logger', 'get_logger', (['self.__class__.__name__'], {'verb': '(True)'}), '(self.__class__.__name__, verb=True)\n', (816, 852), False, 'from rpicam.utils.logging_utils import get_logger\n'), ((1429, 1460), 'requests.post', 'requests.post', (['url'], {'files': 'files'}), '(u... |
"""
This module provides a class for interfacing with the Sense HAT add-on board for Raspberry Pi.
"""
import os
from multiprocessing.managers import RemoteError
from myDevices.utils.logger import error, exception, info
from sensehat.manager import connect_client
class SenseHAT():
"""Class for interacting with a... | [
"myDevices.utils.logger.info",
"myDevices.utils.logger.error",
"os.path.join",
"sensehat.manager.connect_client"
] | [((759, 836), 'os.path.join', 'os.path.join', (['"""/etc/myDevices/plugins/cayenne-plugin-sensehat/data/image.png"""'], {}), "('/etc/myDevices/plugins/cayenne-plugin-sensehat/data/image.png')\n", (771, 836), False, 'import os\n'), ((1086, 1102), 'sensehat.manager.connect_client', 'connect_client', ([], {}), '()\n', (11... |
#!/usr/bin/env python
from isrutils.looper import simpleloop
# %% users param
P = {
"path": "~/data/2015-10-07/isr",
"beamid": 64157,
"acf": True,
"vlimacf": (18, 45),
"zlim_pl": [None, None],
"vlim_pl": [72, 90],
"flim_pl": [3.5, 5.5],
"odir": "out/2015-10-07",
"vlim": [25, 55],
... | [
"isrutils.looper.simpleloop"
] | [((383, 403), 'isrutils.looper.simpleloop', 'simpleloop', (['flist', 'P'], {}), '(flist, P)\n', (393, 403), False, 'from isrutils.looper import simpleloop\n')] |
"""turbotutorial URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/4.0/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class... | [
"django.views.generic.TemplateView.as_view",
"django.urls.path"
] | [((789, 820), 'django.urls.path', 'path', (['"""admin/"""', 'admin.site.urls'], {}), "('admin/', admin.site.urls)\n", (793, 820), False, 'from django.urls import path\n'), ((1061, 1121), 'django.views.generic.TemplateView.as_view', 'TemplateView.as_view', ([], {'template_name': '"""broadcast_example.html"""'}), "(templ... |
import os
# ETL
ETL_MODE = os.environ.get('ETL_MODE')
ETL_CHUNK_SIZE = int(os.environ.get('ETL_CHUNK_SIZE'))
ETL_SYNC_DELAY = int(os.environ.get('ETL_SYNC_DELAY'))
ETL_FILE_STATE = os.environ.get('ETL_FILE_STATE')
ETL_DEFAULT_DATE = os.environ.get('ETL_DEFAULT_DATE')
# Postgres
POSTGRES_NAME = os.environ.get('POSTGRE... | [
"os.environ.get"
] | [((28, 54), 'os.environ.get', 'os.environ.get', (['"""ETL_MODE"""'], {}), "('ETL_MODE')\n", (42, 54), False, 'import os\n'), ((182, 214), 'os.environ.get', 'os.environ.get', (['"""ETL_FILE_STATE"""'], {}), "('ETL_FILE_STATE')\n", (196, 214), False, 'import os\n'), ((234, 268), 'os.environ.get', 'os.environ.get', (['"""... |
#!/usr/bin/env python
from __future__ import absolute_import
from goonpug import app
def main():
app.run(debug=True)
if __name__ == '__main__':
main()
| [
"goonpug.app.run"
] | [((105, 124), 'goonpug.app.run', 'app.run', ([], {'debug': '(True)'}), '(debug=True)\n', (112, 124), False, 'from goonpug import app\n')] |
from setuptools import setup, find_packages
from Cython.Distutils.extension import Extension
from Cython.Build import cythonize, build_ext
import numpy
import os
from glob import glob
"""
ext_modules = [Extension("traj_dist.cydist.basic_geographical", ["traj_dist/cydist/basic_geographical.pyx"]),
Extens... | [
"numpy.get_include",
"setuptools.find_packages",
"glob.glob"
] | [((1294, 1324), 'glob.glob', 'glob', (['"""traj_dist/cydist/*.pyx"""'], {}), "('traj_dist/cydist/*.pyx')\n", (1298, 1324), False, 'from glob import glob\n'), ((1950, 1965), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (1963, 1965), False, 'from setuptools import setup, find_packages\n'), ((1700, 1719)... |
import pyeccodes.accessors as _
def load(h):
h.add(_.Unsigned('numberOfCategories', 1))
with h.list('categories'):
for i in range(0, h.get_l('numberOfCategories')):
h.add(_.Codetable('categoryType', 1, "4.91.table", _.Get('masterDir'), _.Get('localDir')))
h.add(_.Unsigned('co... | [
"pyeccodes.accessors.Unsigned",
"pyeccodes.accessors.Get"
] | [((58, 93), 'pyeccodes.accessors.Unsigned', '_.Unsigned', (['"""numberOfCategories"""', '(1)'], {}), "('numberOfCategories', 1)\n", (68, 93), True, 'import pyeccodes.accessors as _\n'), ((306, 333), 'pyeccodes.accessors.Unsigned', '_.Unsigned', (['"""codeFigure"""', '(1)'], {}), "('codeFigure', 1)\n", (316, 333), True,... |
import numpy as np
from numba import guvectorize
from pygama.dsp.errors import DSPFatal
@guvectorize(["void(float32[:], float32, float32, float32, float32[:])",
"void(float64[:], float64, float64, float64, float64[:])"],
"(n),(),(),()->()", nopython=True, cache=True)
def time_point_thresh(w_... | [
"numpy.floor",
"pygama.dsp.errors.DSPFatal",
"numba.guvectorize",
"numpy.isnan"
] | [((90, 276), 'numba.guvectorize', 'guvectorize', (["['void(float32[:], float32, float32, float32, float32[:])',\n 'void(float64[:], float64, float64, float64, float64[:])']", '"""(n),(),(),()->()"""'], {'nopython': '(True)', 'cache': '(True)'}), "(['void(float32[:], float32, float32, float32, float32[:])',\n 'voi... |
import tornado
import json
import uuid
import pandas as pd
from handler.mlsklearn.util import regqeust_arg_to_sklearn_arg
from sklearn.model_selection import train_test_split
from data.persistence import *
from data.data_source import DataSource
from data.data_storage import DataStorage
class TrainTestSplitHandler(t... | [
"pandas.DataFrame",
"json.loads",
"sklearn.model_selection.train_test_split",
"json.dumps",
"data.data_storage.DataStorage.create_data_obj_by_pandas_data",
"handler.mlsklearn.util.regqeust_arg_to_sklearn_arg",
"data.data_storage.DataStorage.get_data_obj_by_data_id"
] | [((387, 416), 'json.loads', 'json.loads', (['self.request.body'], {}), '(self.request.body)\n', (397, 416), False, 'import json\n'), ((890, 934), 'data.data_storage.DataStorage.get_data_obj_by_data_id', 'DataStorage.get_data_obj_by_data_id', (['data_id'], {}), '(data_id)\n', (925, 934), False, 'from data.data_storage i... |
import numpy as np
import pandas as pd
import yfinance as yf
import matplotlib.pyplot as plt
import datetime
from yahoo_fin import stock_info as si
plt.rcParams['figure.figsize'] = (15, 10)
tickers = si.tickers_dow()
individual_stock = input(f"Which of the following stocks would you like to backtest \n{tickers}\n:")
... | [
"matplotlib.pyplot.title",
"matplotlib.pyplot.show",
"yfinance.download",
"matplotlib.pyplot.legend",
"datetime.date.today",
"yahoo_fin.stock_info.tickers_dow",
"pandas.Series",
"matplotlib.pyplot.subplots"
] | [((202, 218), 'yahoo_fin.stock_info.tickers_dow', 'si.tickers_dow', ([], {}), '()\n', (216, 218), True, 'from yahoo_fin import stock_info as si\n'), ((434, 467), 'yfinance.download', 'yf.download', (['tickers'], {'start': 'start'}), '(tickers, start=start)\n', (445, 467), True, 'import yfinance as yf\n'), ((705, 723), ... |
# -*- coding: utf-8 -*-
"""
Created on Tue May 22 14:07:42 2018
@author: HORSE
"""
import logging
import logging.handlers
import os
def ARLogger(log_filename = 'log.txt'):
# if not os.path.exists('logs'):
# os.makedirs('logs')
fmt = '%(asctime)s %(levelname)s %(message)s'
... | [
"logging.Formatter",
"logging.handlers.RotatingFileHandler",
"logging.getLogger"
] | [((372, 401), 'logging.getLogger', 'logging.getLogger', (['"""ARLogger"""'], {}), "('ARLogger')\n", (389, 401), False, 'import logging\n'), ((454, 543), 'logging.handlers.RotatingFileHandler', 'logging.handlers.RotatingFileHandler', (['log_filename'], {'maxBytes': '(10000 * 4)', 'backupCount': '(5)'}), '(log_filename, ... |
import unittest
from desky.rect import Rect
from desky.panel import Panel
from enum import Enum
from functools import reduce, partial
from toolz.dicttoolz import valfilter
# | Type of sizing | Maximum extra width allocation
# --------------------------------------------------------------
# | Fixed (200 p... | [
"unittest.main",
"functools.partial",
"desky.panel.Panel.Rect",
"toolz.dicttoolz.valfilter",
"desky.rect.Rect",
"desky.gui.Gui"
] | [((29067, 29082), 'unittest.main', 'unittest.main', ([], {}), '()\n', (29080, 29082), False, 'import unittest\n'), ((1911, 1955), 'toolz.dicttoolz.valfilter', 'valfilter', (['(lambda p: p != panel)', 'self.panels'], {}), '(lambda p: p != panel, self.panels)\n', (1920, 1955), False, 'from toolz.dicttoolz import valfilte... |
from distutils.core import setup, Extension
def main():
setup(name="seqrepc",
version="beta1.0",
description="SeqrepC is a module for fundamental operations related to numerical representations of genomic sequences.",
author="<NAME>",
author_email="<EMAIL>",
url="... | [
"distutils.core.Extension"
] | [((429, 470), 'distutils.core.Extension', 'Extension', (['"""seqrepc"""', "['./src/seqrepc.c']"], {}), "('seqrepc', ['./src/seqrepc.c'])\n", (438, 470), False, 'from distutils.core import setup, Extension\n')] |
# pylint: disable=no-member,invalid-name,line-too-long,trailing-whitespace
"""Add IsEnabled column to EventHandlerBASE
Revision ID: <KEY>
Revises: 6b5369ab5224
Create Date: 2021-02-17 20:15:42.776190
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = '<KEY>'
down_... | [
"alembic.op.drop_column",
"sqlalchemy.Column"
] | [((548, 595), 'alembic.op.drop_column', 'op.drop_column', (['"""EventHandlerBASE"""', '"""IsEnabled"""'], {}), "('EventHandlerBASE', 'IsEnabled')\n", (562, 595), False, 'from alembic import op\n'), ((458, 523), 'sqlalchemy.Column', 'sa.Column', (['"""IsEnabled"""', 'sa.Boolean'], {'nullable': '(False)', 'default': '(Fa... |
from resource_management.libraries.script.script import Script
from resource_management.core.resources.packaging import Package
class Client(Script):
def install(self, env):
packages = ['percona-server-client']
Package(packages)
self.configure(env)
def configure(self, env):
im... | [
"resource_management.core.resources.packaging.Package"
] | [((233, 250), 'resource_management.core.resources.packaging.Package', 'Package', (['packages'], {}), '(packages)\n', (240, 250), False, 'from resource_management.core.resources.packaging import Package\n')] |
# Generated by Django 2.2.1 on 2019-07-19 12:36
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('website', '0035_auto_20190625_0900'),
]
operations = [
migrations.RenameField(
model_name='verenigingen',
old_name='ontgroen... | [
"django.db.migrations.RenameField"
] | [((227, 332), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""verenigingen"""', 'old_name': '"""ontgroening"""', 'new_name': '"""introductietijd"""'}), "(model_name='verenigingen', old_name='ontgroening',\n new_name='introductietijd')\n", (249, 332), False, 'from django.db impor... |
#!/usr/bin/python
# Flask is used to create a somewhat lightweight listening server
from flask import Flask
from requests import get
def spawn_proxy():
myproxy = Flask('__name__')
# Quick health check override
@myproxy.route('/healthcheck', methods=['GET'])
def health():
return "OK"
# ... | [
"flask.Flask",
"requests.get"
] | [((170, 187), 'flask.Flask', 'Flask', (['"""__name__"""'], {}), "('__name__')\n", (175, 187), False, 'from flask import Flask\n'), ((774, 805), 'requests.get', 'get', (['f"""{target}/search?q={req}"""'], {}), "(f'{target}/search?q={req}')\n", (777, 805), False, 'from requests import get\n')] |
#! /usr/bin/env python
import pefile
import datetime
import os
import re
from pecli.plugins.base import Plugin
from pecli.lib.utils import cli_out
ASCII_BYTE = b" !\"#\$%&\'\(\)\*\+,-\./0123456789:;<=>\?@ABCDEFGHIJKLMNOPQRSTUVWXYZ\[\]\^_`abcdefghijklmnopqrstuvwxyz\{\|\}\\\~\t"
class PluginStrings(Plugin):
name... | [
"pecli.lib.utils.cli_out",
"re.compile"
] | [((1130, 1180), 're.compile', 're.compile', (["(b'([%s]{%d,})' % (ASCII_BYTE, min_len))"], {}), "(b'([%s]{%d,})' % (ASCII_BYTE, min_len))\n", (1140, 1180), False, 'import re\n'), ((1199, 1257), 're.compile', 're.compile', (["(b'((?:[%s]\\x00){%d,})' % (ASCII_BYTE, min_len))"], {}), "(b'((?:[%s]\\x00){%d,})' % (ASCII_BY... |
from django.core.management.base import BaseCommand
from api.models import Country
from django.db import transaction
from django.db.models import Q
from api.logger import logger
class Command(BaseCommand):
help = 'Update Countries initially to set/revoke their in_search field (probably one-time run only)'
@t... | [
"api.logger.logger.info",
"django.db.models.Q",
"api.models.Country.objects.filter"
] | [((847, 902), 'api.logger.logger.info', 'logger.info', (['"""Successfully set in_search for Countries"""'], {}), "('Successfully set in_search for Countries')\n", (858, 902), False, 'from api.logger import logger\n'), ((472, 548), 'api.models.Country.objects.filter', 'Country.objects.filter', ([], {'independent': '(Tru... |
from django import template
from tos.models import CGUItem
register = template.Library()
@register.simple_tag
def get_cgu_items():
return CGUItem.objects.filter(deleted_at__isnull=True)
| [
"django.template.Library",
"tos.models.CGUItem.objects.filter"
] | [((72, 90), 'django.template.Library', 'template.Library', ([], {}), '()\n', (88, 90), False, 'from django import template\n'), ((145, 192), 'tos.models.CGUItem.objects.filter', 'CGUItem.objects.filter', ([], {'deleted_at__isnull': '(True)'}), '(deleted_at__isnull=True)\n', (167, 192), False, 'from tos.models import CG... |
# Copyright 2016 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""This is a factory test to check the brightness of LCD backlight or LEDs."""
from cros.factory.device import device_utils
from cros.factory.test.i18n i... | [
"cros.factory.test.i18n.arg_utils.I18nArg",
"cros.factory.utils.arg_utils.Arg",
"cros.factory.device.device_utils.CreateDUTInterface"
] | [((538, 583), 'cros.factory.test.i18n.arg_utils.I18nArg', 'i18n_arg_utils.I18nArg', (['"""msg"""', '"""Message HTML"""'], {}), "('msg', 'Message HTML')\n", (560, 583), True, 'from cros.factory.test.i18n import arg_utils as i18n_arg_utils\n'), ((591, 669), 'cros.factory.utils.arg_utils.Arg', 'Arg', (['"""timeout_secs"""... |
import pandas as pd
import os
import dotenv
from dotenv import load_dotenv
import datetime
import plotly
import plotly.graph_objects as go
from plotly.subplots import make_subplots
from app.other_data_pull import spy_pull, fred_pull
from app.port_data_pull import port_data_pull
from app.portfolio_import import portfol... | [
"app.port_data_pull.port_data_pull",
"os.path.abspath",
"app.other_data_pull.fred_pull",
"app.other_data_pull.spy_pull",
"dotenv.load_dotenv",
"os.environ.get",
"app.portfolio_import.portfolio_import",
"plotly.subplots.make_subplots"
] | [((7076, 7089), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (7087, 7089), False, 'from dotenv import load_dotenv\n'), ((7111, 7148), 'os.environ.get', 'os.environ.get', (['"""PORTFOLIO_FILE_NAME"""'], {}), "('PORTFOLIO_FILE_NAME')\n", (7125, 7148), False, 'import os\n'), ((7166, 7204), 'os.environ.get', 'os.... |
from keras_tweaks import idseqs_to_mask
import tensorflow as tf
class AllTests(tf.test.TestCase):
def test1(self):
idseqs = [[1, 1, 0, 0, 2, 2, 3], [1, 3, 2, 1, 0, 0, 2]]
target = tf.sparse.SparseTensor(
indices=(
[0, 0, 1],
[0, 1, 1],
... | [
"tensorflow.test.main",
"keras_tweaks.idseqs_to_mask",
"tensorflow.sparse.to_dense"
] | [((4219, 4233), 'tensorflow.test.main', 'tf.test.main', ([], {}), '()\n', (4231, 4233), True, 'import tensorflow as tf\n'), ((646, 719), 'keras_tweaks.idseqs_to_mask', 'idseqs_to_mask', (['idseqs'], {'n_seqlen': '(6)', 'n_vocab_sz': '(3)', 'ignore': '[3]', 'dense': '(False)'}), '(idseqs, n_seqlen=6, n_vocab_sz=3, ignor... |
# apis_v1/test_views_voter_email_address_save.py
# Brought to you by We Vote. Be good.
# -*- coding: UTF-8 -*-
from django.urls import reverse
from django.test import TestCase
from email_outbound.models import EmailAddress, EmailManager
import json
class WeVoteAPIsV1TestsVoterEmailAddressRetrieve(TestCase):
datab... | [
"django.urls.reverse"
] | [((417, 456), 'django.urls.reverse', 'reverse', (['"""apis_v1:deviceIdGenerateView"""'], {}), "('apis_v1:deviceIdGenerateView')\n", (424, 456), False, 'from django.urls import reverse\n'), ((489, 523), 'django.urls.reverse', 'reverse', (['"""apis_v1:voterCreateView"""'], {}), "('apis_v1:voterCreateView')\n", (496, 523)... |
#!/usr/bin/python
# -*- coding: utf-8 -*-
"""
Redundant misc. functions to be eventually removed from AC_tools.
"""
import os
import numpy as np
from matplotlib.backends.backend_pdf import PdfPages
import matplotlib.pyplot as plt
from pandas import DataFrame
# time
import time
import datetime as datetime
# math
from m... | [
"pandas.DataFrame",
"numpy.linspace"
] | [((4158, 4204), 'pandas.DataFrame', 'DataFrame', (['data'], {'index': 'dates', 'columns': "['data']"}), "(data, index=dates, columns=['data'])\n", (4167, 4204), False, 'from pandas import DataFrame\n'), ((2254, 2314), 'numpy.linspace', 'np.linspace', (['(0)', '(1)', 'extra_points_point_on_edge'], {'endpoint': '(True)'}... |
from nanome._internal._util._serializers import _ArraySerializer, _StringSerializer, _ColorSerializer
from . import _AtomSerializerID
from . import _BondSerializer
from .. import _Residue
from nanome.util import Logs
from nanome._internal._util._serializers import _TypeSerializer
class _ResidueSerializer(_TypeSeriali... | [
"nanome._internal._util._serializers._StringSerializer",
"nanome._internal._util._serializers._ArraySerializer",
"nanome._internal._util._serializers._ColorSerializer"
] | [((419, 437), 'nanome._internal._util._serializers._ArraySerializer', '_ArraySerializer', ([], {}), '()\n', (435, 437), False, 'from nanome._internal._util._serializers import _ArraySerializer, _StringSerializer, _ColorSerializer\n'), ((537, 555), 'nanome._internal._util._serializers._ColorSerializer', '_ColorSerialize... |
#!/usr/bin/env python3
import logging
import subprocess
from typing import Dict
class Agent:
name: str
image: str
environment: Dict[str, str]
def __init__(self, name: str, image: str, environment: Dict[str, str]) -> None:
self.name = name
self.image = image
self.environment =... | [
"logging.info",
"subprocess.run"
] | [((369, 447), 'logging.info', 'logging.info', (['"""Starting agent \'%s\' based on image \'%s\'"""', 'self.name', 'self.image'], {}), '("Starting agent \'%s\' based on image \'%s\'", self.name, self.image)\n', (381, 447), False, 'import logging\n'), ((456, 573), 'subprocess.run', 'subprocess.run', (["['docker', 'run', ... |
# -*- coding: utf-8 -*-
#################################################################################################
import logging
import urllib
import requests
from resources.lib.util import JSONRPC
##################################################################################################
log = log... | [
"resources.lib.util.JSONRPC",
"requests.head",
"logging.getLogger"
] | [((317, 355), 'logging.getLogger', 'logging.getLogger', (["('DINGS.' + __name__)"], {}), "('DINGS.' + __name__)\n", (334, 355), False, 'import logging\n'), ((1228, 1263), 'resources.lib.util.JSONRPC', 'JSONRPC', (['"""Settings.GetSettingValue"""'], {}), "('Settings.GetSettingValue')\n", (1235, 1263), False, 'from resou... |
#!/usr/bin/env python
# Copyright (C) 2004 British Broadcasting Corporation and Kamaelia Contributors(1)
# All Rights Reserved.
#
# You may only modify and redistribute this under the terms of any of the
# following licenses(2): Mozilla Public License, V1.1, GNU General
# Public License, V2.0, GNU Lesser Ge... | [
"SpatialIndexer.SpatialIndexer"
] | [((4558, 4596), 'SpatialIndexer.SpatialIndexer', 'SpatialIndexer', (['laws.maxInteractRadius'], {}), '(laws.maxInteractRadius)\n', (4572, 4596), False, 'from SpatialIndexer import SpatialIndexer\n')] |
from rest_framework import parsers, renderers
from rest_framework.authtoken.models import Token
from rest_framework.authtoken.serializers import AuthTokenSerializer
from rest_framework.response import Response
from rest_framework.views import APIView
from .serializers import ClientTokenSerializer
from .models import C... | [
"rest_framework.response.Response"
] | [((1018, 1048), 'rest_framework.response.Response', 'Response', (["{'token': token.key}"], {}), "({'token': token.key})\n", (1026, 1048), False, 'from rest_framework.response import Response\n')] |
# !/usr/bin/python
from itertools import groupby
def compress(data):
return ((len(list(group)), name) for name, group in groupby(data))
def decompress(data):
return (car * size for size, car in data)
my_data = 'get uuuuuuuuuuuuuuuup'
print(list(my_data))
compressed = compress(my_data)
print(''.join(deco... | [
"itertools.groupby"
] | [((128, 141), 'itertools.groupby', 'groupby', (['data'], {}), '(data)\n', (135, 141), False, 'from itertools import groupby\n')] |
import numpy as np
import magpie
# check cartesian
def test_get_xedges():
xedges = magpie.grids.get_xedges(1., 2)
xedges = np.round(xedges, decimals=2)
assert len(xedges) == 3, "Length of xedges is incorrect."
assert xedges[-1] - xedges[0] == 1., "xedges range is incorrect."
xedges = magpie.grids... | [
"magpie.grids.polargrid",
"numpy.sum",
"magpie.grids.grid1d",
"magpie.grids.grid3d",
"magpie.grids.xmid2edges",
"magpie.grids.get_xedges",
"magpie.grids.polarEA_grid",
"numpy.shape",
"magpie.grids.polarEA_npix",
"magpie.grids.polarEA_area",
"magpie.grids.xedges2mid",
"numpy.round",
"magpie.g... | [((90, 121), 'magpie.grids.get_xedges', 'magpie.grids.get_xedges', (['(1.0)', '(2)'], {}), '(1.0, 2)\n', (113, 121), False, 'import magpie\n'), ((134, 162), 'numpy.round', 'np.round', (['xedges'], {'decimals': '(2)'}), '(xedges, decimals=2)\n', (142, 162), True, 'import numpy as np\n'), ((308, 350), 'magpie.grids.get_x... |
from communication import *
import socket
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
s.connect((socket.gethostname(), 1123))
while True:
m = receive_message(s)
if m:
print(m, "\n")
ping(s)
print(s.getsockname())
print(socket.gethostbyname(socket.gethostname()))
print(socket.get) | [
"socket.gethostname",
"socket.socket"
] | [((47, 96), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (60, 96), False, 'import socket\n'), ((108, 128), 'socket.gethostname', 'socket.gethostname', ([], {}), '()\n', (126, 128), False, 'import socket\n'), ((262, 282), 'socket.gethostname'... |
"""
dmsetup commands - Command ``dmsetup``
======================================
Parsers for parsing and extracting data from output of commands related to
``dmsetup``.
Parsers contained in this module are:
DmsetupInfo - command ``dmsetup info -C``
-----------------------------------------
"""
from insights import... | [
"insights.parsers.parse_delimited_table",
"insights.parser"
] | [((430, 456), 'insights.parser', 'parser', (['Specs.dmsetup_info'], {}), '(Specs.dmsetup_info)\n', (436, 456), False, 'from insights import parser, CommandParser\n'), ((2195, 2225), 'insights.parsers.parse_delimited_table', 'parse_delimited_table', (['content'], {}), '(content)\n', (2216, 2225), False, 'from insights.p... |
import sys, os, random
import numpy as np
import matplotlib.pyplot as plt
from matplotlib.colors import Normalize
from matplotlib.colors import ListedColormap
from sklearn.model_selection import train_test_split, GridSearchCV
from sklearn.model_selection import StratifiedShuffleSplit
from sklearn.preprocessing i... | [
"matplotlib.pyplot.title",
"matplotlib.pyplot.figaspect",
"numpy.load",
"sklearn.preprocessing.StandardScaler",
"random.shuffle",
"matplotlib.pyplot.style.use",
"matplotlib.pyplot.figure",
"numpy.mean",
"sklearn.neural_network.MLPClassifier",
"sklearn.svm.SVC",
"numpy.interp",
"os.path.join",
... | [((1137, 1191), 'os.path.join', 'os.path.join', (['dirpath', '"""../datasets/breast-cancer.npz"""'], {}), "(dirpath, '../datasets/breast-cancer.npz')\n", (1149, 1191), False, 'import sys, os, random\n'), ((1202, 1251), 'os.path.join', 'os.path.join', (['dirpath', '"""../datasets/diabetes.npz"""'], {}), "(dirpath, '../d... |
import unittest
import maxixe
from maxixe.tests import decorators
from maxixe.tests import loader
from maxixe.tests import parser
from maxixe.tests import utils
suite = unittest.TestSuite()
suite.addTests(unittest.TestLoader().loadTestsFromModule(decorators))
suite.addTests(unittest.TestLoader().loadTestsFromModule(l... | [
"unittest.TestLoader",
"unittest.TestSuite"
] | [((171, 191), 'unittest.TestSuite', 'unittest.TestSuite', ([], {}), '()\n', (189, 191), False, 'import unittest\n'), ((207, 228), 'unittest.TestLoader', 'unittest.TestLoader', ([], {}), '()\n', (226, 228), False, 'import unittest\n'), ((277, 298), 'unittest.TestLoader', 'unittest.TestLoader', ([], {}), '()\n', (296, 29... |
from django.db import models
class Blog(models.Model):
title = models.CharField("标题", unique=True, max_length=200)
class Meta:
db_table = 'blog'
verbose_name = '文章' | [
"django.db.models.CharField"
] | [((71, 122), 'django.db.models.CharField', 'models.CharField', (['"""标题"""'], {'unique': '(True)', 'max_length': '(200)'}), "('标题', unique=True, max_length=200)\n", (87, 122), False, 'from django.db import models\n')] |
from wormer.tools import manager, downloader
from wormer.data import strategy
import re
class Graber:
synopsis_pattern = '''(?=lemma-summary")(.*?)(?<=config) '''
text_pattern = '>\s*?([^\&\b\n\[\]]*?)<'
href_pattern = '<a target=_blank href="(/item/[\w\d%]*?)">'
def __init__(self):
self.urlM... | [
"wormer.tools.manager.ThreadManager",
"wormer.tools.manager.LogManager",
"wormer.tools.manager.TextManager",
"wormer.tools.manager.UrlsManager",
"wormer.tools.downloader.DownLoader"
] | [((329, 350), 'wormer.tools.manager.UrlsManager', 'manager.UrlsManager', ([], {}), '()\n', (348, 350), False, 'from wormer.tools import manager, downloader\n'), ((377, 400), 'wormer.tools.downloader.DownLoader', 'downloader.DownLoader', ([], {}), '()\n', (398, 400), False, 'from wormer.tools import manager, downloader\... |
from telegram.ext import ConversationHandler
from telegram.ext import MessageHandler
from telegram.ext import Filters
from telegram.ext import CallbackQueryHandler
from Model.share import Share
import Controllers.global_states as states
from Utils.logging import get_logger as log
import pandas as pd
import datetime
G... | [
"Model.share.Share",
"Utils.logging.get_logger",
"telegram.ext.MessageHandler",
"pandas.to_datetime",
"datetime.datetime.now"
] | [((1496, 1509), 'Model.share.Share', 'Share', (['ticker'], {}), '(ticker)\n', (1501, 1509), False, 'from Model.share import Share\n'), ((994, 999), 'Utils.logging.get_logger', 'log', ([], {}), '()\n', (997, 999), True, 'from Utils.logging import get_logger as log\n'), ((1388, 1393), 'Utils.logging.get_logger', 'log', (... |
import gzip
import sys
import argparse
import re
import logging
import numpy as np
import pandas as p
from itertools import product, tee
from collections import Counter, OrderedDict
from Bio import SeqIO
def generate_feature_mapping(kmer_len):
BASE_COMPLEMENT = {"A":"T","T":"A","G":"C","C":"G"}
kmer_hash = ... | [
"gzip.open",
"argparse.ArgumentParser",
"Bio.SeqIO.parse",
"numpy.zeros",
"numpy.array",
"itertools.product",
"itertools.tee"
] | [((355, 387), 'itertools.product', 'product', (['"""ATGC"""'], {'repeat': 'kmer_len'}), "('ATGC', repeat=kmer_len)\n", (362, 387), False, 'from itertools import product, tee\n'), ((664, 675), 'itertools.tee', 'tee', (['seq', 'n'], {}), '(seq, n)\n', (667, 675), False, 'from itertools import product, tee\n'), ((971, 100... |
# -*- coding: utf-8 -*-
'''
<NAME>
1. a. Frequentist confidence intervals do not respect the physical limitations imposed on a system, ie non-negativity of a mass.
b. Typically, that the probability to be found outside the interval on both sides of the distribution is 16% (or (100-CL)/2 %).
Of... | [
"matplotlib.pyplot.title",
"numpy.sum",
"matplotlib.pyplot.clf",
"numpy.polyfit",
"matplotlib.pyplot.bar",
"numpy.histogram",
"numpy.exp",
"numpy.random.normal",
"numpy.linspace",
"matplotlib.pyplot.errorbar",
"math.isnan",
"scipy.stats.chi2",
"matplotlib.pyplot.ylabel",
"numpy.log",
"ma... | [((8154, 8178), 'numpy.linspace', 'np.linspace', (['(0)', '(5)', '(10000)'], {}), '(0, 5, 10000)\n', (8165, 8178), True, 'import numpy as np\n'), ((8204, 8236), 'numpy.exp', 'np.exp', (['(-theExponential.lookup_x)'], {}), '(-theExponential.lookup_x)\n', (8210, 8236), True, 'import numpy as np\n'), ((8448, 8457), 'matpl... |
from fontTools.misc.fixedTools import floatToFixed
from fontTools.ttLib import TTFont, newTable, registerCustomTableClass
from fontTools.varLib.models import VariationModel, allEqual
from fontTools.varLib.varStore import OnlineVarStoreBuilder
from rcjktools.varco import VarCoFont
from rcjktools.table_VarC import (
... | [
"fontTools.varLib.models.VariationModel",
"fontTools.ttLib.TTFont",
"argparse.ArgumentParser",
"fontTools.varLib.models.allEqual",
"fontTools.ttLib.newTable",
"fontTools.varLib.varStore.OnlineVarStoreBuilder",
"pathlib.Path",
"rcjktools.table_VarC.getToFixedConverterForNumIntBitsForScale",
"fontTool... | [((615, 653), 'fontTools.varLib.models.VariationModel', 'VariationModel', (['allLocations', 'axisTags'], {}), '(allLocations, axisTags)\n', (629, 653), False, 'from fontTools.varLib.models import VariationModel, allEqual\n'), ((673, 704), 'fontTools.varLib.varStore.OnlineVarStoreBuilder', 'OnlineVarStoreBuilder', (['ax... |
from unifi.objects.base import UnifiBaseObject
from unifi.helper import find_by_attr, json_print
class UnifiDeviceObject(UnifiBaseObject):
def get_port_profile(self, **filter_kwargs):
port = find_by_attr(self.port_table, **filter_kwargs)
port_override = find_by_attr(self.port_overrides, port_idx=... | [
"unifi.helper.find_by_attr"
] | [((206, 252), 'unifi.helper.find_by_attr', 'find_by_attr', (['self.port_table'], {}), '(self.port_table, **filter_kwargs)\n', (218, 252), False, 'from unifi.helper import find_by_attr, json_print\n'), ((277, 337), 'unifi.helper.find_by_attr', 'find_by_attr', (['self.port_overrides'], {'port_idx': "port['port_idx']"}), ... |
# Copyright 2019 The DMLab2D Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in... | [
"dmlab2d.runfiles_helper.find",
"absl.testing.absltest.main",
"numpy.dtype",
"numpy.testing.assert_array_equal"
] | [((8160, 8175), 'absl.testing.absltest.main', 'absltest.main', ([], {}), '()\n', (8173, 8175), False, 'from absl.testing import absltest\n'), ((3413, 3470), 'numpy.testing.assert_array_equal', 'np.testing.assert_array_equal', (['observations[0]', '[1, 2, 3]'], {}), '(observations[0], [1, 2, 3])\n', (3442, 3470), True, ... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
from json import dumps, loads
# for python 2
# from httplib import HTTPConnection
# for python 3
from http.client import HTTPConnection
# connect with REST server
connection = HTTPConnection('127.0.0.1', 80)
connection.connect()
data = {"ip": "192... | [
"http.client.HTTPConnection",
"json.dumps"
] | [((245, 276), 'http.client.HTTPConnection', 'HTTPConnection', (['"""127.0.0.1"""', '(80)'], {}), "('127.0.0.1', 80)\n", (259, 276), False, 'from http.client import HTTPConnection\n'), ((489, 500), 'json.dumps', 'dumps', (['data'], {}), '(data)\n', (494, 500), False, 'from json import dumps, loads\n')] |
# coding: utf-8
# Copyright (c) 2016, 2019, Oracle and/or its affiliates. All rights reserved.
"""
NOTE: This class should always comply to the API definition of NfsDatasetClient present in
services/dts/src/oci_cli_dts/physical_appliance_control_plane/client/nfs_dataset_client.py
"""
from oci_cli import cli_util
fr... | [
"oci_cli.cli_util.build_config",
"services.dts.src.oci_cli_dts.appliance_config_manager.ApplianceConfigManager",
"services.dts.src.oci_cli_dts.physical_appliance_control_plane.client.nfs_dataset_client.NfsDatasetClient"
] | [((769, 819), 'services.dts.src.oci_cli_dts.appliance_config_manager.ApplianceConfigManager', 'ApplianceConfigManager', (['APPLIANCE_CONFIGS_BASE_DIR'], {}), '(APPLIANCE_CONFIGS_BASE_DIR)\n', (791, 819), False, 'from services.dts.src.oci_cli_dts.appliance_config_manager import ApplianceConfigManager\n'), ((1077, 1107),... |
from spreaduler import ParamsSheet
from train_attention import train
from options import get_parser
class YourParamsSheet(ParamsSheet):
"""
Your model Params Sheet class
"""
params_sheet_id = '...'
client_credentials = {
"type": "service_account",
"project_id": "....",
"pr... | [
"options.get_parser"
] | [((1351, 1363), 'options.get_parser', 'get_parser', ([], {}), '()\n', (1361, 1363), False, 'from options import get_parser\n')] |
import os
import sys
import time
import torch
import utils
import logging
import argparse
import torch.nn as nn
import torch.utils
from adaptive_augmentor import AdaAug
from networks import get_model
from networks.projection import Projection
from dataset import get_num_class, get_dataloaders, get_label_name, get_data... | [
"argparse.ArgumentParser",
"utils.create_exp_dir",
"dataset.get_dataloaders",
"dataset.get_dataset_dimension",
"time.strftime",
"logging.Formatter",
"dataset.get_num_class",
"utils.AvgrageMeter",
"torch.no_grad",
"os.path.join",
"utils.reproducibility",
"config.get_warmup_config",
"torch.cud... | [((433, 467), 'argparse.ArgumentParser', 'argparse.ArgumentParser', (['"""ada_aug"""'], {}), "('ada_aug')\n", (456, 467), False, 'import argparse\n'), ((3752, 3783), 'utils.create_exp_dir', 'utils.create_exp_dir', (['args.save'], {}), '(args.save)\n', (3772, 3783), False, 'import utils\n'), ((3823, 3934), 'logging.basi... |
import tensorflow as tf
sess = tf.Session()
#在名字为foo的命名空间内创建名字为v的变量
with tf.variable_scope("foo"):
#创建一个常量为1的v
v= tf.get_variable('v1',[1],initializer = tf.constant_initializer(1.0))
#因为在foo空间已经创建v的变量,所以下面的代码会报错
#with tf.variable_scope("foo"):
# v= tf.get_variable('v',[1])
#在生成上下文管理器时,将参数reuse设置为True。这样tf.g... | [
"tensorflow.constant_initializer",
"tensorflow.Session",
"tensorflow.variable_scope",
"tensorflow.initialize_all_variables"
] | [((33, 45), 'tensorflow.Session', 'tf.Session', ([], {}), '()\n', (43, 45), True, 'import tensorflow as tf\n'), ((75, 99), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""foo"""'], {}), "('foo')\n", (92, 99), True, 'import tensorflow as tf\n'), ((437, 482), 'tensorflow.variable_scope', 'tf.variable_scope', (['"... |
from collections import namedtuple
from typing import Dict, List, Callable
Node = namedtuple('Node', 'name parent children data')
def make_tree_from_adj_list(adj_list):
root = 'COM'
nodes: Dict['str', Node] = {root: Node(root, None, [], {})}
for parent, child in adj_list:
node = Node(child, pare... | [
"collections.namedtuple"
] | [((83, 130), 'collections.namedtuple', 'namedtuple', (['"""Node"""', '"""name parent children data"""'], {}), "('Node', 'name parent children data')\n", (93, 130), False, 'from collections import namedtuple\n')] |
from iotbx import mtz
mtz_obj = mtz.object(file_name="3nd4.mtz")
# Only works with mtz.object.
# Does not work if mtz is read in with iotbx.file_reader.
miller_arrays_dict = mtz_obj.as_miller_arrays_dict()
| [
"iotbx.mtz.object"
] | [((32, 64), 'iotbx.mtz.object', 'mtz.object', ([], {'file_name': '"""3nd4.mtz"""'}), "(file_name='3nd4.mtz')\n", (42, 64), False, 'from iotbx import mtz\n')] |
"""
Copyright (c) 2019 Intel Corporation
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writin... | [
"torch.ones",
"functools.partial",
"torch.IntTensor",
"collections.namedtuple",
"torch.zeros",
"torch.distributed.broadcast",
"logging.getLogger"
] | [((1000, 1027), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1017, 1027), False, 'import logging\n'), ((1314, 1409), 'collections.namedtuple', 'namedtuple', (['"""QuantizationParams"""', "['bits', 'mode', 'signed', 'signed_scope', 'per_channel']"], {}), "('QuantizationParams', ['bits',... |
import curses, sys, os
#Servo controller connected to IC2
import Adafruit_PCA9685
pwm = Adafruit_PCA9685.PCA9685()
pwm.set_pwm_freq(60)
from time import sleep
#ESC Brushles motor states: direction, on/off
toggleState = 400
throttle = 450
delta = 20
print("toggleState1")
pwm.set_pwm(2,0,toggleState)
sleep(0.2)
for i... | [
"Adafruit_PCA9685.PCA9685",
"time.sleep"
] | [((89, 115), 'Adafruit_PCA9685.PCA9685', 'Adafruit_PCA9685.PCA9685', ([], {}), '()\n', (113, 115), False, 'import Adafruit_PCA9685\n'), ((304, 314), 'time.sleep', 'sleep', (['(0.2)'], {}), '(0.2)\n', (309, 314), False, 'from time import sleep\n'), ((494, 504), 'time.sleep', 'sleep', (['(0.4)'], {}), '(0.4)\n', (499, 50... |
import pytropos.internals.values as pv
from pytropos.internals.values.builtin_values import *
from pytropos.internals.values.python_values.builtin_mutvalues import *
from pytropos.internals.values.python_values.wrappers import *
from pytropos.internals.values.python_values.python_values import PythonValue, PT
exitcode... | [
"pytropos.internals.values.int",
"pytropos.internals.values.python_values.python_values.PythonValue"
] | [((380, 399), 'pytropos.internals.values.python_values.python_values.PythonValue', 'PythonValue', (['PT.Top'], {}), '(PT.Top)\n', (391, 399), False, 'from pytropos.internals.values.python_values.python_values import PythonValue, PT\n'), ((440, 454), 'pytropos.internals.values.python_values.python_values.PythonValue', '... |
import re
import setuptools
def find_version(fname):
"""Attempts to find the version number in the file names fname.
Raises RuntimeError if not found.
"""
version = ''
with open(fname, 'r') as fp:
reg = re.compile(r'__version__ = [\'"]([^\'"]*)[\'"]')
for line in fp:
m ... | [
"setuptools.find_packages",
"re.compile"
] | [((233, 286), 're.compile', 're.compile', (['"""__version__ = [\\\\\'"]([^\\\\\'"]*)[\\\\\'"]"""'], {}), '(\'__version__ = [\\\\\\\'"]([^\\\\\\\'"]*)[\\\\\\\'"]\')\n', (243, 286), False, 'import re\n'), ((990, 1016), 'setuptools.find_packages', 'setuptools.find_packages', ([], {}), '()\n', (1014, 1016), False, 'import ... |
#!/usr/bin/python3
import os.path
import openpyxl
import requests
import json
import argparse
BASE_URL_XIV_API_CHARACTER: str = "https://xivapi.com/character/"
GERMAN_TO_ENGLISH_CLASS_DICT: dict = {}
SUB_30_MAPPING_DICT: dict = {}
CONFIG_LOCATION = os.getcwd()
DEBUG_ENABLED = False
def main(filepath):
"""main ... | [
"argparse.ArgumentParser",
"json.load",
"openpyxl.load_workbook",
"requests.get"
] | [((417, 449), 'openpyxl.load_workbook', 'openpyxl.load_workbook', (['filepath'], {}), '(filepath)\n', (439, 449), False, 'import openpyxl\n'), ((3683, 3708), 'requests.get', 'requests.get', (['request_url'], {}), '(request_url)\n', (3695, 3708), False, 'import requests\n'), ((5160, 5232), 'argparse.ArgumentParser', 'ar... |
import datetime
from functools import partial
from typing import List, Optional, Union
from quickbuild.helpers import ContentType, response2py
class Configurations:
def __init__(self, quickbuild):
self.quickbuild = quickbuild
def _get(self, params: dict) -> List[dict]:
return self.quickbui... | [
"functools.partial"
] | [((2333, 2380), 'functools.partial', 'partial', (['response2py'], {'content_type': 'content_type'}), '(response2py, content_type=content_type)\n', (2340, 2380), False, 'from functools import partial\n')] |
from os import path
from pathlib import Path
def curr_file_path() -> Path:
"""Get cuurent file path."""
return Path(__file__).absolute()
def out_folder_path() -> Path:
"""Get output folder path."""
return curr_file_path().parents[3].joinpath("out").absolute()
def out_geom_path() -> Path:
"""Ge... | [
"pathlib.Path"
] | [((121, 135), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (125, 135), False, 'from pathlib import Path\n')] |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import requests, json, base64
def post_cli(auth, command):
url_cli = "http://" + auth.ipaddr + "/rest/" + auth.version + "/cli"
command_dict = {"cmd": command}
try:
post_command = requests.post(url_cli, headers=auth.cookie, data=json.dumps(command_dict... | [
"base64.b64decode",
"json.dumps"
] | [((297, 321), 'json.dumps', 'json.dumps', (['command_dict'], {}), '(command_dict)\n', (307, 321), False, 'import requests, json, base64\n'), ((418, 448), 'base64.b64decode', 'base64.b64decode', (['cli_response'], {}), '(cli_response)\n', (434, 448), False, 'import requests, json, base64\n')] |
from model.group import Group
class GroupHelper:
def __init__(self, app):
self.app = app
def create(self, group):
wd = self.app.wd
self.go_to_group_page()
wd.find_element_by_name("new").click()
self.fill_form_group(group)
# Submit group creation
wd.fin... | [
"model.group.Group"
] | [((3260, 3289), 'model.group.Group', 'Group', ([], {'name': 'text', 'id': 'group_id'}), '(name=text, id=group_id)\n', (3265, 3289), False, 'from model.group import Group\n')] |
import math
def sieve(n):
primes = list(range(2, n+1))
i = 0
while i < len(primes):
no = primes[i]
m = 2
while (no * m) <= max(primes):
if primes.count(no * m) > 0:
primes.remove(no * m)
m+=1
i+=1
return primes
def maxPower(n, ... | [
"math.pow"
] | [((348, 366), 'math.pow', 'math.pow', (['n', '(i + 1)'], {}), '(n, i + 1)\n', (356, 366), False, 'import math\n')] |
from os import system
def ler_qtd(n, msg):
n = int(input(msg))
while (n < 1) or (n > 10000):
n = int(input(f' - Entrada invalida!{msg}'))
return n
def preencher_set_cartas(cartas, qtd, p):
""" set de cartas, qtd de cartas, p de pessoa """
from time import sleep
print()... | [
"os.system",
"time.sleep"
] | [((1764, 1777), 'os.system', 'system', (['"""cls"""'], {}), "('cls')\n", (1770, 1777), False, 'from os import system\n'), ((641, 649), 'time.sleep', 'sleep', (['(1)'], {}), '(1)\n', (646, 649), False, 'from time import sleep\n')] |
import numpy as np
import random
import numbers
import cv2
from PIL import Image
import wpcv
from wpcv.utils.ops import pil_ops, polygon_ops
from wpcv.utils.data_aug.base import Compose, Zip
from wpcv.utils.data_aug import img_aug
class ToPILImage(object):
def __init__(self):
self.to = img_aug.ToPILImage(... | [
"wpcv.utils.ops.pil_ops.vflip",
"wpcv.utils.ops.polygon_ops.get_translate_range",
"wpcv.utils.ops.pil_ops.resize_keep_ratio",
"wpcv.utils.ops.polygon_ops.scale",
"random.randint",
"wpcv.utils.ops.pil_ops.scale",
"wpcv.utils.ops.polygon_ops.translate",
"wpcv.utils.data_aug.img_aug.ToPILImage",
"wpcv.... | [((301, 321), 'wpcv.utils.data_aug.img_aug.ToPILImage', 'img_aug.ToPILImage', ([], {}), '()\n', (319, 321), False, 'from wpcv.utils.data_aug import img_aug\n'), ((1855, 1891), 'wpcv.utils.ops.pil_ops.scale', 'pil_ops.scale', (['img', '(scaleX, scaleY)'], {}), '(img, (scaleX, scaleY))\n', (1868, 1891), False, 'from wpcv... |
# coding: utf-8
"""
IbIocProfile.py
The Clear BSD License
Copyright (c) – 2016, NetApp, Inc. All rights reserved.
Redistribution and use in source and binary forms, with or without modification, are permitted (subject to the limitations in the disclaimer below) provided that the following conditions are met:
*... | [
"six.iteritems"
] | [((18511, 18540), 'six.iteritems', 'iteritems', (['self.swagger_types'], {}), '(self.swagger_types)\n', (18520, 18540), False, 'from six import iteritems\n')] |
from csv import reader
from sklearn import preprocessing
from plotly import graph_objects
def import_data(path):
return [[float(f) for f in r] for r in reader(open(path, "r"))]
def normalize_data(dataset):
scaler = preprocessing.MinMaxScaler(feature_range=(0,1))
normalized = scaler.fit_transform(dataset... | [
"plotly.graph_objects.Figure",
"sklearn.preprocessing.MinMaxScaler"
] | [((4307, 4341), 'plotly.graph_objects.Figure', 'graph_objects.Figure', ([], {'data': '[table]'}), '(data=[table])\n', (4327, 4341), False, 'from plotly import graph_objects\n'), ((227, 275), 'sklearn.preprocessing.MinMaxScaler', 'preprocessing.MinMaxScaler', ([], {'feature_range': '(0, 1)'}), '(feature_range=(0, 1))\n'... |
#!/bin/python3
from selenium import webdriver
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.common.by import By
from selenium.webdriver.support import expected_conditions as EC
from selenium.common.exceptions import NoSuchElementException
import sys,os
options = webdriver.ChromeOptio... | [
"selenium.webdriver.support.expected_conditions.presence_of_element_located",
"selenium.webdriver.ChromeOptions",
"selenium.webdriver.Chrome",
"selenium.webdriver.support.ui.WebDriverWait",
"sys.exit"
] | [((299, 324), 'selenium.webdriver.ChromeOptions', 'webdriver.ChromeOptions', ([], {}), '()\n', (322, 324), False, 'from selenium import webdriver\n'), ((499, 532), 'selenium.webdriver.Chrome', 'webdriver.Chrome', ([], {'options': 'options'}), '(options=options)\n', (515, 532), False, 'from selenium import webdriver\n')... |
# Generated by Django 2.2.4 on 2019-09-11 14:12
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('users', '0004_auto_20190907_1334'),
]
operations = [
migrations.AddField(
model_name='pic',
name='classification152'... | [
"django.db.models.ImageField",
"django.db.models.TextField"
] | [((340, 368), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)'}), '(blank=True)\n', (356, 368), False, 'from django.db import migrations, models\n'), ((495, 523), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)'}), '(blank=True)\n', (511, 523), False, 'from django.db im... |
# Copyright 2019 DeepMind Technologies Ltd. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | [
"open_spiel.python.mfg.algorithms.distribution.DistributionPolicy",
"open_spiel.python.mfg.algorithms.EGTA.init_oracle.init_br_oracle",
"open_spiel.python.mfg.algorithms.EGTA.inner_loop.InnerLoop"
] | [((2652, 2700), 'open_spiel.python.mfg.algorithms.EGTA.inner_loop.InnerLoop', 'inner_loop.InnerLoop', (['self._meta_strategy_method'], {}), '(self._meta_strategy_method)\n', (2672, 2700), False, 'from open_spiel.python.mfg.algorithms.EGTA import inner_loop\n'), ((4084, 4132), 'open_spiel.python.mfg.algorithms.EGTA.inne... |
# encoding:utf-8
import os, sys
basepath = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
sys.path.append(os.path.join(basepath, 'FaceDetector'))
import requests
import base64
import cv2
import numpy as np
import urllib.request
import base64
def fetchImageFromHttp(image_url, timeout_s=1):
# 该函数是读取url... | [
"os.path.abspath",
"cv2.imwrite",
"cv2.imdecode",
"cv2.imread",
"requests.get",
"cv2.rectangle",
"requests.post",
"os.path.join"
] | [((119, 157), 'os.path.join', 'os.path.join', (['basepath', '"""FaceDetector"""'], {}), "(basepath, 'FaceDetector')\n", (131, 157), False, 'import os, sys\n'), ((886, 904), 'requests.get', 'requests.get', (['host'], {}), '(host)\n', (898, 904), False, 'import requests\n'), ((1739, 1795), 'requests.post', 'requests.post... |
#
# INF 552 Homework 3
# Part 2: Fast Map
# Group Members: <NAME> (zhan198), <NAME> (minyihua), <NAME> (jeffyjac)
# Date: 2/27/2018
# Programming Language: Python 3.6
#
import numpy as np
import matplotlib.pyplot as plt
DIMENSION = 2
DATA_SIZE = 10
# WORDS = ["acting", "activist", "compute", "coward","forward","in... | [
"matplotlib.pyplot.title",
"matplotlib.pyplot.show",
"matplotlib.pyplot.scatter",
"numpy.power",
"numpy.zeros",
"numpy.amax",
"numpy.where",
"matplotlib.pyplot.subplots"
] | [((470, 508), 'numpy.zeros', 'np.zeros', ([], {'shape': '(DATA_SIZE, DATA_SIZE)'}), '(shape=(DATA_SIZE, DATA_SIZE))\n', (478, 508), True, 'import numpy as np\n'), ((516, 554), 'numpy.zeros', 'np.zeros', ([], {'shape': '(DATA_SIZE, DIMENSION)'}), '(shape=(DATA_SIZE, DIMENSION))\n', (524, 554), True, 'import numpy as np\... |
#!/usr/bin/python
# -*- coding: utf-8 -*-
__author__ = 'ar'
import os
import glob
from app.backend.core.utils import getDirectorySizeInBytes, humanReadableSize
if __name__ == '__main__':
path='../../../data/datasets'
for ii,pp in enumerate(glob.glob('%s/*' % path)):
tbn=os.path.basename(pp)
ts... | [
"app.backend.core.utils.getDirectorySizeInBytes",
"os.path.basename",
"app.backend.core.utils.humanReadableSize",
"glob.glob"
] | [((250, 274), 'glob.glob', 'glob.glob', (["('%s/*' % path)"], {}), "('%s/*' % path)\n", (259, 274), False, 'import glob\n'), ((289, 309), 'os.path.basename', 'os.path.basename', (['pp'], {}), '(pp)\n', (305, 309), False, 'import os\n'), ((326, 353), 'app.backend.core.utils.getDirectorySizeInBytes', 'getDirectorySizeInB... |
#
# Copyright (c) 2019 UAVCAN Development Team
# This software is distributed under the terms of the MIT License.
# Author: <NAME> <<EMAIL>>
#
import pytest
import subprocess
from ._subprocess import run_cli_tool
def _unittest_trivial() -> None:
run_cli_tool('show-transport', timeout=2.0)
with pytest.raises... | [
"pytest.raises"
] | [((307, 351), 'pytest.raises', 'pytest.raises', (['subprocess.CalledProcessError'], {}), '(subprocess.CalledProcessError)\n', (320, 351), False, 'import pytest\n'), ((397, 441), 'pytest.raises', 'pytest.raises', (['subprocess.CalledProcessError'], {}), '(subprocess.CalledProcessError)\n', (410, 441), False, 'import pyt... |
import time
import numpy as np
import torch
import torch.nn as nn
import torch.optim as optim
import queue
from insomnia.utils import empty_torch_queue
from insomnia.explores.gaussian_noise import GaussianActionNoise
from insomnia.numeric_models import d4pg
from insomnia.numeric_models.misc import l2_projection
cla... | [
"torch.from_numpy",
"numpy.abs",
"torch.nn.BCELoss",
"insomnia.utils.empty_torch_queue",
"insomnia.numeric_models.d4pg.CriticNetwork",
"numpy.asarray",
"numpy.zeros",
"time.time",
"torch.cuda.is_available",
"torch.sum",
"torch.tensor",
"insomnia.numeric_models.misc.l2_projection._l2_project"
] | [((1153, 1270), 'insomnia.numeric_models.d4pg.CriticNetwork', 'd4pg.CriticNetwork', (['beta', 'input_dims', 'fc1_dims', 'fc2_dims', 'n_actions', 'name', 'self.v_min', 'self.v_max', 'self.num_atoms'], {}), '(beta, input_dims, fc1_dims, fc2_dims, n_actions, name,\n self.v_min, self.v_max, self.num_atoms)\n', (1171, 12... |
import logging
from wrapper import *
logger = logging.getLogger(__name__)
# noinspection PyUnusedLocal
def get_user(url='', key='', timeout=60, **kwargs):
return get(url + '/user', headers={'Authorization': "Bearer " + key}, timeout=timeout).json()
# noinspection PyUnusedLocal
def get_user_tokens(url='', key='... | [
"logging.getLogger"
] | [((47, 74), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (64, 74), False, 'import logging\n')] |
from __future__ import absolute_import # Need to import lwr_client absolutely.
from ..objectstore import ObjectStore
try:
from galaxy.jobs.runners.lwr_client.manager import ObjectStoreClientManager
except ImportError:
from lwr.lwr_client.manager import ObjectStoreClientManager
class LwrObjectStore(ObjectStor... | [
"lwr.lwr_client.manager.ObjectStoreClientManager"
] | [((2814, 2857), 'lwr.lwr_client.manager.ObjectStoreClientManager', 'ObjectStoreClientManager', ([], {}), '(**manager_options)\n', (2838, 2857), False, 'from lwr.lwr_client.manager import ObjectStoreClientManager\n')] |
"""
Demo of json_required decorator for API input validation/error handling
"""
import inspect
import functools
import json
from traceback import format_exception
from flask import jsonify, request
import sys
from flask.exceptions import JSONBadRequest
from flask import Flask
import re
app = Flask(__name__)
def... | [
"functools.partial",
"traceback.format_exception",
"json.loads",
"flask.Flask",
"json.dumps",
"inspect.getargspec",
"functools.wraps",
"sys.exc_info",
"flask.request.json.get",
"re.compile"
] | [((300, 315), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (305, 315), False, 'from flask import Flask\n'), ((3436, 3619), 're.compile', 're.compile', (['"""[A-Za-z0-9!#$%&\'*+/=?^_`{|}~-]+(?:\\\\.[A-Za-z0-9!#$%&\'*+/=?^_`{|}~-]+)*@(?:[A-Za-z0-9](?:[A-Za-z0-9-]*[A-Za-z0-9])?\\\\.)+[A-Za-z0-9](?:[A-Za-z0-... |
import torch
from ..math.cross import *
from ..math.normvec import *
class CameraExtrinsic(object):
"""
A class representing the camera extrinsic properties
Attributes
----------
position : Tensor
the camera position
target : Tensor
the camera target
up_vector : Tensor
... | [
"torch.tensor"
] | [((1340, 1396), 'torch.tensor', 'torch.tensor', (['position'], {'dtype': 'torch.float', 'device': 'device'}), '(position, dtype=torch.float, device=device)\n', (1352, 1396), False, 'import torch\n'), ((1423, 1477), 'torch.tensor', 'torch.tensor', (['target'], {'dtype': 'torch.float', 'device': 'device'}), '(target, dty... |
import uuid
from textwrap import dedent
from IPython.core.display import display, HTML
from string import Template
import numpy as np
# function to initialize a scatter plot
def init_chart(data,features):
chart_id = 'mychart-' + str(uuid.uuid4())
feature_types = {} # map each feature to type
num_feature_ra... | [
"IPython.core.display.HTML",
"uuid.uuid4",
"textwrap.dedent"
] | [((782, 853), 'IPython.core.display.HTML', 'HTML', (['"""<script src="/static/components/requirejs/require.js"></script>"""'], {}), '(\'<script src="/static/components/requirejs/require.js"></script>\')\n', (786, 853), False, 'from IPython.core.display import display, HTML\n'), ((238, 250), 'uuid.uuid4', 'uuid.uuid4', ... |
##############################################################################
# Copyright (c) 2017 Huawei Technologies Co.,Ltd and others.
#
# All rights reserved. This program and the accompanying materials
# are made available under the terms of the Apache License, Version 2.0
# which accompanies this distribution, ... | [
"yardstick.common.openstack_utils.create_image",
"yardstick.common.openstack_utils.get_shade_client",
"logging.getLogger"
] | [((619, 646), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (636, 646), False, 'import logging\n'), ((1721, 1755), 'yardstick.common.openstack_utils.get_shade_client', 'openstack_utils.get_shade_client', ([], {}), '()\n', (1753, 1755), False, 'from yardstick.common import openstack_utils... |
#!/usr/bin/env python
#
# Copyright © 2022 Github Lzhiyong
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"subprocess.run",
"pathlib.Path",
"argparse.ArgumentParser",
"time.time"
] | [((1459, 1482), 'subprocess.run', 'subprocess.run', (['command'], {}), '(command)\n', (1473, 1482), False, 'import subprocess\n'), ((1495, 1506), 'time.time', 'time.time', ([], {}), '()\n', (1504, 1506), False, 'import time\n'), ((1965, 1979), 'pathlib.Path', 'Path', (['args.ndk'], {}), '(args.ndk)\n', (1969, 1979), Fa... |
import pathlib
from datetime import timedelta
from airflow import DAG
from airflow.operators.python import PythonOperator
import pandas as pd
from sklearn.compose import ColumnTransformer
from sklearn.pipeline import Pipeline
from sklearn.preprocessing import OneHotEncoder
from sklearn.preprocessing import Stan... | [
"json.dump",
"pickle.dump",
"sklearn.preprocessing.StandardScaler",
"airflow.operators.python.PythonOperator",
"pandas.read_csv",
"sklearn.model_selection.train_test_split",
"sklearn.metrics.accuracy_score",
"sklearn.preprocessing.OneHotEncoder",
"sklearn.metrics.roc_auc_score",
"sklearn.linear_mo... | [((621, 675), 'pandas.read_csv', 'pd.read_csv', (['"""/opt/airflow/data/raw/{{ ds }}/data.csv"""'], {}), "('/opt/airflow/data/raw/{{ ds }}/data.csv')\n", (632, 675), True, 'import pandas as pd\n'), ((693, 749), 'pandas.read_csv', 'pd.read_csv', (['"""/opt/airflow/data/raw/{{ ds }}/target.csv"""'], {}), "('/opt/airflow/... |